Full-Time
Confirmed live in the last 24 hours
Platform for improving machine learning models
$148k - $160kAnnually
Senior
Greenville, NC, USA
This is a hybrid position, providing care in person and remotely.
You match the following Galileo's candidate preferences
Employers are more likely to interview you if you match these preferences:
Galileo offers a platform that helps machine learning teams enhance their models and lower annotation costs by using data-centric algorithms for Natural Language Processing. It allows teams to quickly identify and fix data issues that affect model performance and provides a collaborative space to manage models from raw data to production. Unlike competitors, Galileo integrates easily with existing tools and focuses on actionability, security, and privacy, while also streamlining the data labeling process. The company's goal is to equip machine learning teams with efficient tools to improve their models and reduce costs.
Company Size
51-200
Company Stage
Series B
Total Funding
$66.2M
Headquarters
San Francisco, California
Founded
2021
Help us improve and share your feedback! Did you find this helpful?
Health Insurance
Dental Insurance
Vision Insurance
Disability Insurance
Parental Leave
Flexible Work Hours
401(k) Retirement Plan
401(k) Company Match
When deploying generative artificial intelligence (AI), one of the most fundamental decisions businesses face is whether to choose open-source or proprietary AI models — or aim for a hybrid of the two. “This basic choice between the open source ecosystem and a proprietary setting impacts countless business and technical decision, making it ‘the AI developer’s dilemma,’” according to an Intel Labs blog post. This choice is critical because it affects a company’s AI development, accessibility, security and innovation. Businesses must navigate these options carefully to maximize benefits while mitigating risks
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More. Galileo, a San Francisco-based startup, is betting that the future of artificial intelligence depends on trust. Today, the company launched a new product, Agentic Evaluations, to address a growing challenge in the world of AI: making sure the increasingly complex systems known as AI agents actually work as intended.AI agents—autonomous systems that perform multi-step tasks like generating reports or analyzing customer data—are gaining traction across industries. But their rapid adoption raises a crucial question: How can companies verify these systems remain reliable after deployment? Galileo’s CEO, Vikram Chatterji, believes his company has found the answer.“Over the last six to eight months, we started to see some of our customers trying to adopt agentic systems,” said Chatterji in an interview. “Now LLMs can be used as a smart router to pick and choose the right API calls towards actually completing a task
Galileo Technologies Inc., a provider of enterprise AI observability and assessment platforms, today announced that it has raised $45 million in new funding.
/PRNewswire/ -- Galileo, a leader in generative AI evaluation and observability for enterprises, today announced it raised $45M in Series B funding led by...
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More. Researchers at Apple have introduced ToolSandbox, a novel benchmark designed to assess the real-world capabilities of AI assistants more comprehensively than ever before. The research, published on arXiv, addresses crucial gaps in existing evaluation methods for large language models (LLMs) that use external tools to complete tasks.ToolSandbox incorporates three key elements often missing from other benchmarks: stateful interactions, conversational abilities, and dynamic evaluation. Lead author Jiarui Lu explains, “ToolSandbox includes stateful tool execution, implicit state dependencies between tools, a built-in user simulator supporting on-policy conversational evaluation and a dynamic evaluation strategy.”This new benchmark aims to mirror real-world scenarios more closely. For instance, it can test whether an AI assistant understands that it needs to enable a device’s cellular service before sending a text message — a task that requires reasoning about the current state of the system and making appropriate changes.Proprietary models outshine open-source, but challenges remainThe researchers tested a range of AI models using ToolSandbox, revealing a significant performance gap between proprietary and open-source models.This finding challenges recent reports suggesting that open-source AI is rapidly catching up to proprietary systems