Full-Time
Confirmed live in the last 24 hours
AI inference hardware for cloud and on-premises
$154.4k - $209k/yr
Mid, Senior
Palo Alto, CA, USA
Groq specializes in AI inference technology, providing the Groq LPU™, which is known for its high compute speed, quality, and energy efficiency. The Groq LPU™ is designed to handle AI processing tasks quickly and effectively, making it suitable for both cloud and on-premises applications. Unlike many competitors, Groq ensures that all its products are designed, fabricated, and assembled in North America, which helps maintain high quality and performance standards. The company targets a wide range of clients who need fast and efficient AI processing capabilities. Groq's goal is to deliver scalable AI inference solutions that meet the demands of industries requiring rapid data processing.
Company Size
201-500
Company Stage
Growth Equity (Non-Venture Capital)
Total Funding
$2.8B
Headquarters
Mountain View, California
Founded
2016
Help us improve and share your feedback! Did you find this helpful?
Remote Work Options
Company Equity
RIYADH, Saudi Arabia, April 8, 2025 /PRNewswire/ -- Groq announced today the exclusive launch of Meta's Llama 4 Scout and Maverick models in the Middle East. Available only on GroqCloud™, developers now have day-zero access to Meta's most advanced openly-available models.This launch marks a significant milestone in positioning the Middle East as a hub for cutting-edge AI infrastructure, following the activation of the largest inference cluster in the region, located in Dammam. The data center, which has been live since February, is now serving Llama 4 globally."The integration of Llama 4 with Groq technology marks a major step forward in the Kingdom of Saudi Arabia's journey toward technological leadership," said Tareq Amin."We built Groq to drive the cost of compute to zero," said Jonathan Ross, CEO and Founder of Groq. "Together with our partners, we're delivering Llama 4 to the region with high-performance inference that runs faster, costs less, and doesn't compromise."Llama 4 Now Available — Only on GroqCloudPowered by the custom-built Groq LPU, GroqCloud gives developers instant access to Llama 4 with no tuning, no cold starts, and no trade-offs.Llama 4 Scout: $0.11 / M input tokens and $0.34 / M output tokens, at a blended rate of $0.13/ M input tokens and / M output tokens, at a blended rate of Llama 4 Maverick: $0.50 / M input tokens and $0.77 / M output tokens, at a blended rate of $0.53Learn more about Groq pricing here.About the ModelsLlama 4 is Meta's latest openly-available model family, featuring Mixture of Experts (MoE) architecture and native multimodality.Llama 4 Scout (17Bx16E): A strong general-purpose model, ideal for summarization, reasoning, and code. Runs at over 625 tokens per second on Groq.Llama 4 Maverick (17Bx128E): A larger, more capable model optimized for multilingual and multimodal tasks—great for assistants, chat, and creative applications. Supports 12 languages, including Arabic.Start Building TodayAccess Llama 4 via:GroqChatGroqCloud ConsoleGroq API (model IDs available in-console)Start free at console.groq.com .Upgrade for worry-free rate limits and higher throughput.About GroqGroq is the AI inference platform redefining price and performance
MOUNTAIN VIEW, Calif., April 5, 2025 /PRNewswire/ -- Groq, the pioneer in AI inference, has launched Meta's Llama 4 Scout and Maverick models, now live on GroqCloud™. Developers and enterprises get day-zero access to the most advanced open-source AI models available.That speed is possible because Groq controls the full stack—from our custom-built LPU to our vertically integrated cloud. The result: models go live with no delay, no tuning, and no bottlenecks—and run at the lowest cost per token in the industry, with full performance
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More. Groq and PlayAI announced a partnership today to bring Dialog, an advanced text-to-speech model, to market through Groq’s high-speed inference platform.The partnership combines PlayAI’s expertise in voice AI with Groq’s specialized processing infrastructure, creating what the companies claim is one of the most natural-sounding and responsive text-to-speech systems available.“Groq provides a complete, low latency system for automatic speech recognition (ASR), GenAI, and text-to-speech, all in one place,” said Ian Andrews, Chief Revenue Officer at Groq, in an exclusive interview with VentureBeat. “With Dialog now running on GroqCloud, this means customers won’t have to use multiple providers for a single use case — Groq is a one stop solution.”. VIDEO
In a groundbreaking development in the field of Artificial Intelligence, Groq has partnered with PlayAI to introduce an advanced text-to-speech model named Dialog.
Groq unveiled a Developer Tier on GroqCloud to broaden access to its high-performance compute platform for AI and machine learning.