Full-Time

Software Engineer

Model Evaluation Systems

Confirmed live in the last 24 hours

Groq

Groq

201-500 employees

AI inference hardware for cloud and on-premises

Compensation Overview

$154.4k - $209k/yr

Mid, Senior

Palo Alto, CA, USA

Category
Backend Engineering
Software Engineering
Required Skills
Python
Go
Data Analysis
Requirements
  • Proven experience building or working with evaluation systems or working on novel applied AI projects.
  • Proficiency in Python and Golang, with strong software engineering fundamentals and best practices.
  • Self-Driven: Ability to see projects through end-to-end independently, taking initiative to optimize systems proactively.
  • Deep problem-solving skills — candidates should be able to tackle non-trivial challenges with minimal guidance.
  • Ability to work across unconventional AI problems, iterate quickly, and ship impactful solutions.
Responsibilities
  • Write simple, concise, high-performance code in Python and Golang to support evaluation pipelines and tooling.
  • Develop and implement benchmarking tools to assess the performance of models hosted on Groq, ensuring Groq’s AI infrastructure remains at the cutting edge of efficiency and performance.
  • Build automated evaluation systems and testing infrastructure to streamline model deployment and validation processes.
  • Work closely with cross-functional teams to integrate models into Groq’s infrastructure, ensuring seamless operation and accurate evaluation across various system configurations.
  • Collect and analyze performance data, providing actionable insights and recommendations for improving model performance, scalability, and reliability.
Desired Qualifications
  • Contributions to open-source projects or prior startup experience is a big plus.

Groq specializes in AI inference technology, providing the Groq LPU™, which is known for its high compute speed, quality, and energy efficiency. The Groq LPU™ is designed to handle AI processing tasks quickly and effectively, making it suitable for both cloud and on-premises applications. Unlike many competitors, Groq ensures that all its products are designed, fabricated, and assembled in North America, which helps maintain high quality and performance standards. The company targets a wide range of clients who need fast and efficient AI processing capabilities. Groq's goal is to deliver scalable AI inference solutions that meet the demands of industries requiring rapid data processing.

Company Size

201-500

Company Stage

Growth Equity (Non-Venture Capital)

Total Funding

$2.8B

Headquarters

Mountain View, California

Founded

2016

Simplify Jobs

Simplify's Take

What believers are saying

  • Growing demand for energy-efficient AI hardware benefits Groq's innovative technology.
  • Groq's $1.5 billion investment from Saudi Arabia boosts its Middle East market presence.
  • Partnership with PlayAI enhances Groq's capabilities in voice AI applications.

What critics are saying

  • Competition from NVIDIA and AMD threatens Groq's market share.
  • Dependence on Saudi Arabia for funding poses financial risks amid geopolitical changes.
  • Rapid AI model development by tech giants may outpace Groq's integration capabilities.

What makes Groq unique

  • Groq's LPU offers unmatched compute speed and energy efficiency for AI inference.
  • GroqCloud provides exclusive access to Meta's Llama 4 models in the Middle East.
  • Groq's vertically integrated stack ensures no delays or bottlenecks in AI model deployment.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Remote Work Options

Company Equity

Growth & Insights and Company News

Headcount

6 month growth

-4%

1 year growth

0%

2 year growth

-6%
PR Newswire
Apr 8th, 2025
Groq Delivers Exclusive Access To Llama 4 In Saudi Arabia

RIYADH, Saudi Arabia, April 8, 2025 /PRNewswire/ -- Groq announced today the exclusive launch of Meta's Llama 4 Scout and Maverick models in the Middle East. Available only on GroqCloud™, developers now have day-zero access to Meta's most advanced openly-available models.This launch marks a significant milestone in positioning the Middle East as a hub for cutting-edge AI infrastructure, following the activation of the largest inference cluster in the region, located in Dammam. The data center, which has been live since February, is now serving Llama 4 globally."The integration of Llama 4 with Groq technology marks a major step forward in the Kingdom of Saudi Arabia's journey toward technological leadership," said Tareq Amin."We built Groq to drive the cost of compute to zero," said Jonathan Ross, CEO and Founder of Groq. "Together with our partners, we're delivering Llama 4 to the region with high-performance inference that runs faster, costs less, and doesn't compromise."Llama 4 Now Available — Only on GroqCloudPowered by the custom-built Groq LPU, GroqCloud gives developers instant access to Llama 4 with no tuning, no cold starts, and no trade-offs.Llama 4 Scout: $0.11 / M input tokens and $0.34 / M output tokens, at a blended rate of $0.13/ M input tokens and / M output tokens, at a blended rate of Llama 4 Maverick: $0.50 / M input tokens and $0.77 / M output tokens, at a blended rate of $0.53Learn more about Groq pricing here.About the ModelsLlama 4 is Meta's latest openly-available model family, featuring Mixture of Experts (MoE) architecture and native multimodality.Llama 4 Scout (17Bx16E): A strong general-purpose model, ideal for summarization, reasoning, and code. Runs at over 625 tokens per second on Groq.Llama 4 Maverick (17Bx128E): A larger, more capable model optimized for multilingual and multimodal tasks—great for assistants, chat, and creative applications. Supports 12 languages, including Arabic.Start Building TodayAccess Llama 4 via:GroqChatGroqCloud ConsoleGroq API (model IDs available in-console)Start free at console.groq.com .Upgrade for worry-free rate limits and higher throughput.About GroqGroq is the AI inference platform redefining price and performance

PR Newswire
Apr 5th, 2025
Llama 4 Live Day-Zero On Groq At Lowest Cost

MOUNTAIN VIEW, Calif., April 5, 2025 /PRNewswire/ -- Groq, the pioneer in AI inference, has launched Meta's Llama 4 Scout and Maverick models, now live on GroqCloud™. Developers and enterprises get day-zero access to the most advanced open-source AI models available.That speed is possible because Groq controls the full stack—from our custom-built LPU to our vertically integrated cloud. The result: models go live with no delay, no tuning, and no bottlenecks—and run at the lowest cost per token in the industry, with full performance

VentureBeat
Mar 26th, 2025
Groq And Playai Just Made Voice Ai Sound Way More Human — Here’S How

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More. Groq and PlayAI announced a partnership today to bring Dialog, an advanced text-to-speech model, to market through Groq’s high-speed inference platform.The partnership combines PlayAI’s expertise in voice AI with Groq’s specialized processing infrastructure, creating what the companies claim is one of the most natural-sounding and responsive text-to-speech systems available.“Groq provides a complete, low latency system for automatic speech recognition (ASR), GenAI, and text-to-speech, all in one place,” said Ian Andrews, Chief Revenue Officer at Groq, in an exclusive interview with VentureBeat. “With Dialog now running on GroqCloud, this means customers won’t have to use multiple providers for a single use case — Groq is a one stop solution.”. VIDEO

Weebseat
Mar 26th, 2025
Groq and PlayAI Revolutionize Human-like Voice AI

In a groundbreaking development in the field of Artificial Intelligence, Groq has partnered with PlayAI to introduce an advanced text-to-speech model named Dialog.

CISCO Investments
Feb 26th, 2025
Startup Snapshot - Catching Up with our Portfolio Companies in February 2025

Groq unveiled a Developer Tier on GroqCloud to broaden access to its high-performance compute platform for AI and machine learning.