Full-Time

Principal Performance Modeling Engineer

Posted on 8/8/2024

Groq

Groq

201-500 employees

AI inference technology for scalable solutions

AI & Machine Learning

Compensation Overview

$240.2k - $420.4kAnnually

Senior, Expert

Remote in USA

Category
Applied Machine Learning
AI Research
AI & Machine Learning
Required Skills
Python
C/C++
Requirements
  • Computer science, mathematics, ECE or equivalent background and/or experience in this domain
  • Strong fundamentals in computer architecture, with deep knowledge and experience of working on domain specific AI architectures, is highly preferred
  • In-depth understanding of latest AI/ML algorithms and their hardware implications
  • Ability to analyze and simplify complex hardware designs into simple abstracted timing models
  • Past experience on modeling AI/ML workloads, and creating necessary tools for performance optimization. Experience with modeling LLM performance is beneficial, but not required
  • Proficient in programming languages such as C/C++ and Python
  • Experience with cycle-accurate simulators for benchmarking analysis
  • Experience with developing ASIC microarchitecture design is a plus
  • Experience with understanding and simulating RTL (systemVerilog) designs is a plus
Responsibilities
  • Develop and maintain performance models for multiple generations of Groq hardware on the latest AI/ML workloads (LLMs, CNNs, LSTMs, etc.)
  • Analyze AI/ML algorithms to understand their compute, networking and memory requirements, and map them effectively onto the underlying hardware architecture
  • Lead a matrixed team to enable SW/HW co-optimization across chip, system and software teams
  • Identify performance bottlenecks and help drive next generation chip architecture through a solid understanding of Groq's software and hardware
  • Work with silicon and system integration engineers to evaluate the costs & benefits of new technologies on Groq systems
  • Provide what-if scenarios / continuous guidance directly to CEO & senior leadership
  • Develop the Design Space Exploration (DSE) tool for performance analysis and exploration of both chip and system across various workloads
  • Define custom hardware solutions for high profile customers

Groq specializes in AI inference technology, providing the Groq LPU™, which is known for its high compute speed, quality, and energy efficiency. The Groq LPU™ is designed to handle AI processing tasks quickly and effectively, making it suitable for both cloud and on-premises applications. Unlike many competitors, Groq's products are designed, fabricated, and assembled in North America, which helps maintain high standards of quality and performance. The company targets a variety of clients across different industries that require fast and efficient AI processing capabilities. Groq's goal is to deliver scalable AI inference solutions that meet the growing demands for rapid data processing in the AI and machine learning market.

Company Stage

Series D

Total Funding

$1.3B

Headquarters

Mountain View, California

Founded

2016

Growth & Insights
Headcount

6 month growth

8%

1 year growth

-1%

2 year growth

-4%
Simplify Jobs

Simplify's Take

What believers are saying

  • Groq secured $640M in Series D funding, boosting its expansion capabilities.
  • Partnership with Aramco Digital aims to build the world's largest inferencing data center.
  • Integration with Touchcast's Cognitive Caching enhances Groq's hardware for hyper-speed inference.

What critics are saying

  • Increased competition from SambaNova Systems and Gradio in high-speed AI inference.
  • Geopolitical risks in the MENA region may affect the Saudi Arabia data center project.
  • Rapid expansion could strain Groq's operational capabilities and supply chain.

What makes Groq unique

  • Groq's LPU offers exceptional compute speed and energy efficiency for AI inference.
  • The company's products are designed and assembled in North America, ensuring high quality.
  • Groq emphasizes deterministic performance, providing predictable outcomes in AI computations.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Remote Work Options

Company Equity

INACTIVE