Full-Time

Director – Business Operations

Cerebras

Cerebras

501-1,000 employees

AI accelerator hardware replacing GPUs

No salary listed

Sunnyvale, CA, USA

In Person

Category
Business & Strategy (1)
Required Skills
Python
SQL
Data Analysis
Requirements
  • 6 to 10 years of total experience is a reasonable guide.
  • 2+ years at a top-tier strategy consulting firm (McKinsey, BCG, Bain, or similar), with readiness to turn generalist problem-solving into company operating impact.
  • Experience driving operational change inside a scaling company. This can come from operations, strategy, analytics, or FP&A, as long as you have owned real outcomes.
  • High learning velocity. You want an apprenticeship, not just a title, and you learn quickly from direct feedback while owning real outcomes.
  • Strong analytical skills and comfort with imperfect data. You can go from ambiguity to a clear framework, then execute.
  • High agency. You do not wait for perfect inputs or perfect direction. You move, communicate, and close loops.
  • Systems thinker. You build infrastructure that continues working when the pace increases and the data gets messy.
  • Technical fluency beyond spreadsheets (preferred). Comfortable with SQL, Python, or adjacent tooling to pull and shape data and automate recurring work.
  • Clear executive communication. You distill complexity into concise narratives that drive decisions.
  • AI affinity. You proactively apply modern tools to speed up workflows, improve quality, and reduce manual work.
Responsibilities
  • Partner with 5 to 10 operational leaders across supply chain, manufacturing, and data center deployment to drive insight and action.
  • Own and deliver the right information in the right way at the right time. Build the context that allows the organization to know what happened, the implications, and what to do next.
  • Drive closed-loop operational change. Diagnose bottlenecks, redesign processes, and follow through until adoption and measurable improvement.
  • 50/50 analytics and execution. You build the assets (metrics, dashboards, operating packets) and you drive the behaviors (cadence, accountability, decisions).
  • Apprenticeship into operating leadership. We mean it. The hiring manager has used this model repeatedly over the years and can provide references from alumni who have enjoyed meaningful career acceleration.
  • Elegant entry point in the cutting edge of AI. If your pace, horsepower, agency, and ambition are elite, this role gives you room to run.. If you make an impact you can chart your own path.
  • Small, elite, high-standards team. You are a hands-on leader who learns fast, raises the pace, and may selectively add exceptional talent over time to amplify leverage. This will be a small and mighty team.
  • Operational analytics infrastructure required to scale supply chain, manufacturing, inventory management, and data center operations with uncompromising quality and speed.
  • A decision-quality KPI and reporting architecture: leading indicators, dashboards, recurring reviews, and crisp narratives that operators trust.
  • Closed-loop mechanisms that turn operational complexity into repeatable processes: metric definitions, data ownership, reconciliation paths, and accountability loops.
  • System integration between operational data sources and finance systems so that decisions are grounded in consistent, auditable definitions.
  • Automation and tooling that compounds output, including structured data pulls, workflow automation, and pragmatic use of AI tools to reduce manual work.
  • Public-company-ready operating rhythms: clean close-to-insight timelines, documented definitions, and durable operating packets that scale with scrutiny.
  • Operations partnering: Work shoulder-to-shoulder with operations leadership to translate operational reality into clear priorities, tradeoffs, and actions.
  • KPI standards and insight assets: Define the metrics that matter, build the dashboards and operating packets, and keep them accurate as systems and processes evolve.
  • Operating cadence: Design and run weekly and monthly performance reviews, ensure decisions are made, and close the loop on follow-through.
  • Process change: Identify the highest-leverage process gaps, drive redesign and adoption, and measure impact in throughput, cycle time, quality, and predictability.
  • Data quality and reconciliation: Build trust in the numbers by instituting clear definitions, checks, and ownership across operational and finance systems.
  • Executive communication: Deliver concise narratives that clearly separate signal from noise and drive action.
  • You earn trust. Operators proactively pull you into decisions because your work improves outcomes, not just visibility.
  • You grow as an operator. You turn consulting-grade problem solving into operating judgment, cross-functional credibility, broader ownership and ultimately IMPACT.
  • Durable insight assets are live (dashboards, weekly and monthly operating packets, KPI definitions) with a clear cadence and single-source-of-truth inputs.
  • A small number of critical operational processes are redesigned and adopted, with measurable improvements in speed, predictability, and execution quality.
  • Cross-functional decisions move faster because the organization shares consistent definitions and a clear view of tradeoffs and constraints.
  • Manual reporting load declines materially as automation and self-serve assets replace ad hoc requests and one-off analyses.
Desired Qualifications
  • Experience in hardware, semiconductor, manufacturing, supply chain, or data center operations.
  • Comfort with Python or similar tooling as evidence of your ability to be a leading creator and user of agents
  • A track record of building KPI architectures, operating cadences, and repeatable mechanisms from scratch.
  • Experience partnering directly with COO, C-suite, or VP-level operational leaders.
  • Demonstrated pattern of high agency: you see problems early and fix them before being asked.
  • Technical fluency beyond spreadsheets (preferred).
  • AI affinity. You proactively apply modern tools to speed up workflows, improve quality, and reduce manual work.

Cerebras Systems creates AI acceleration hardware and software. Its CS-2 system is designed to replace traditional GPU clusters for AI workloads, speeding up training and inference while simplifying the setup by eliminating the need for parallel programming, distributed training, and cluster management. The product works as a single, large processor-based accelerator with accompanying software and cloud services to run AI models efficiently, reducing latency and time to results. Compared with competitors, Cerebras differentiates itself with the largest processor in the industry and an integrated hardware-software stack that aims to streamline AI workflows rather than relying on multi-GPU clusters. The company’s goal is to help research labs, healthcare, finance, and other industries achieve faster, more cost-effective AI development and deployment by offering a turnkey high-performance AI compute solution.

Company Size

501-1,000

Company Stage

Debt Financing

Total Funding

$3.7B

Headquarters

Sunnyvale, California

Founded

2016

Simplify Jobs

Simplify's Take

What believers are saying

  • OpenAI committed $20B+ over three years for Cerebras-powered servers, validating demand.
  • AWS partnership deploys CS-3 on Amazon Bedrock for generative AI inference H2 2026.
  • Cerebras trains Llama2-70B in one day, earning TIME Magazine Best Invention 2024.

What critics are saying

  • TSMC 3nm process enables competitors' custom chips with 10x density within 18-24 months.
  • AWS Trainium3 hybrid integration exposes prefill-decode bottlenecks, risking partnership termination.
  • IPO valuation collapse mirrors 2025 GPU declines, forcing 50%+ market cap dilution.

What makes Cerebras unique

  • WSE-3 integrates 900,000 cores on single wafer, eliminating multi-GPU communication bottlenecks.
  • 44GB on-chip SRAM delivers 21 petabytes/sec memory bandwidth, 7,000x exceeding H100.
  • Pipeline execution maps model layers to wafer regions, enabling fraction-of-chip inference.

Help us improve and share your feedback! Did you find this helpful?

Your Connections

People at Cerebras who can refer or advise you

Benefits

Professional Development Budget

Flexible Work Hours

Remote Work Options

401(k) Company Match

401(k) Retirement Plan

Mental Health Support

Wellness Program

Paid Sick Leave

Paid Holidays

Paid Vacation

Parental Leave

Family Planning Benefits

Fertility Treatment Support

Adoption Assistance

Childcare Support

Elder Care Support

Pet Insurance

Bereavement Leave

Employee Discounts

Company Social Events

Growth & Insights and Company News

Headcount

6 month growth

0%

1 year growth

2%

2 year growth

0%
Yahoo Finance
Mar 13th, 2026
Cerebras and Amazon partner to combine AI chips on AWS cloud

Amazon and Cerebras Systems have struck a deal to combine their computing chips in a new service aimed at accelerating AI applications like chatbots and coding tools. Cerebras chips will be installed in Amazon Web Services data centres and linked to Amazon's Trainium3 AI chips using custom networking technology. The partnership will tackle AI inference by splitting tasks between the two chip types: Amazon's Trainium3 will handle "prefill", transforming user requests into AI-readable tokens, whilst Cerebras chips will manage the "decode" stage, generating answers. Both companies declined to disclose the deal's financial terms. Valued at $23.1 billion, Cerebras recently signed a $10 billion deal with OpenAI. Amazon expects its service to launch in the second half of this year and believes it will lead in price-performance versus traditional GPUs.

The Register
Feb 20th, 2026
Cerebras to build 8 exaFLOPS AI supercomputer in India backed by UAE's G42

Cerebras Systems will power an 8 exaFLOPS AI supercomputer in India through a collaboration between UAE's Mohamed Bin Zayed University of AI and India's Center for Development of Advanced Computing. The system will be deployed by UAE technology company G42, one of Cerebras' largest backers. The supercomputer will feature approximately 64 WSE-3 wafer-scale accelerators, each delivering 125 petaFLOPS of performance. Unlike traditional GPUs using high-bandwidth memory, Cerebras chips use on-chip SRAM offering 21 petabytes per second of memory bandwidth — roughly 1,000 times faster than Nvidia's HBM4. G42 will deploy the system under India-defined governance frameworks, with all data remaining within Indian borders. The supercomputer will serve Indian universities, startups and SMEs whilst maintaining data sovereignty.

The Tech Buzz
Feb 20th, 2026
G42 and Cerebras deploy 8 exaflops of AI compute in India

G42 and Cerebras deploy 8 exaflops of AI compute in India. UAE's G42 partners with Cerebras to build massive AI infrastructure in India PUBLISHED: Fri, Feb 20, 2026, 12:12 PM UTC | UPDATED: Sun, Feb 22, 2026, 4:53 PM UTC 4 mins read * | G42 and Cerebras announce deployment of 8 exaflops of AI compute infrastructure in India, unveiled at the India AI Impact Summit 2026 * | The partnership represents one of the largest enterprise AI infrastructure investments in Asia, addressing the region's growing compute shortage * | Cerebras brings its wafer-scale engine technology, which offers significant advantages over traditional GPU clusters for training large language models * | The move positions India as a critical AI infrastructure hub and highlights the growing geopolitical importance of compute capacity G42, the Abu Dhabi-based AI powerhouse, just announced a massive infrastructure partnership with chipmaker Cerebras to deploy eight exaflops of compute capacity in India. The deal, revealed at the India AI Impact Summit 2026, represents one of the largest AI infrastructure investments in Asia and signals a major shift in the global AI compute landscape. It's a bold move that positions India as a critical node in the emerging AI supply chain while cementing G42's role as a bridge between Middle Eastern capital and Asian tech ambitions. G42 is making a massive bet on India's AI future. The Abu Dhabi tech giant just unveiled plans to deploy eight exaflops of compute capacity across India in partnership with Cerebras, the U.S. chipmaker known for its wafer-scale engine technology. The announcement, made at the India AI Impact Summit 2026, marks one of the largest AI infrastructure commitments in Asia and comes as the region faces a critical shortage of compute resources. The timing couldn't be more strategic. India's been scrambling to build out AI infrastructure as demand from startups and enterprises skyrockets. According to industry estimates, the country currently has less than 2% of global AI compute capacity despite having one of the world's largest developer populations. This deal could change that calculus overnight. Cerebras brings something different to the table than the usual Nvidia GPU clusters dominating the market. The company's CS-3 systems use wafer-scale engines - essentially chips the size of dinner plates - that can train large language models faster and more efficiently than traditional setups. For G42, which has been aggressively expanding its AI infrastructure footprint across the Middle East and Asia, Cerebras offers a way to differentiate from competitors betting entirely on Nvidia's ecosystem. The eight exaflops figure is eye-popping. To put it in perspective, that's roughly equivalent to eight quintillion floating-point operations per second, enough computational power to train multiple frontier AI models simultaneously. It's the kind of capacity typically reserved for national supercomputing initiatives or hyperscale cloud providers. G42's been on a tear lately. The company, backed by Abu Dhabi's Royal Group and with strategic investments from Microsoft, has positioned itself as a neutral AI infrastructure provider at a time when U.S.-China tech tensions are reshaping global supply chains. The India deployment follows similar announcements in the UAE, Saudi Arabia, and several African nations, suggesting a coordinated strategy to build compute capacity in regions underserved by Western cloud giants. For Cerebras, this represents validation of its alternative approach to AI infrastructure. The company's been competing against Nvidia's CUDA ecosystem dominance by focusing on customers who need to train models quickly rather than just run inference at scale. Landing a deal of this magnitude with G42 gives Cerebras a major reference customer and manufacturing scale it hasn't previously enjoyed. The geopolitical implications are significant. India's government has been pushing for AI sovereignty, wanting to ensure the country isn't dependent on foreign cloud providers for critical infrastructure. A partnership between a UAE-based company and a U.S. chipmaker, deploying hardware on Indian soil, threads a delicate needle - bringing capital and technology without the baggage of direct Chinese or pure Western control. Neither company disclosed the financial terms, but industry sources estimate infrastructure deployments of this scale typically run into the billions of dollars. The systems are expected to come online in phases starting later this year, with full deployment targeted for 2027. The announcement also raises questions about power and cooling infrastructure. Eight exaflops of compute requires massive amounts of electricity and sophisticated cooling systems. India's data center infrastructure has been growing rapidly, but projects of this scale will need dedicated facilities with direct access to power substations and water resources for cooling. What remains unclear is who'll be the primary customers for this capacity. G42 operates both as a cloud provider and as an infrastructure partner for governments and large enterprises. The India deployment could serve regional startups building AI applications, support government AI initiatives, or potentially lease capacity to other cloud providers looking to expand in Asia without building their own infrastructure. This partnership represents more than just another infrastructure deal - it's a glimpse at how AI compute is becoming a geopolitical chess piece. As countries race to build AI capabilities, access to cutting-edge compute matters as much as access to talent or data. G42's ability to move capital and technology across borders while Cerebras provides an alternative to Nvidia's ecosystem creates a playbook other regions will watch closely. For India, it's a chance to leapfrog into the top tier of AI infrastructure nations. The real test will be whether the country can build the ecosystem of customers, applications, and governance frameworks to make full use of all those exaflops. More Topics:

Bloomberg L.P.
Feb 12th, 2026
OpenAI debuts first AI model using Cerebras chips, challenging Nvidia dominance

OpenAI is releasing its first AI model running on chips from Cerebras Systems, marking a move to diversify beyond Nvidia. The model, GPT-5.3-Codex-Spark, launches Thursday as a faster but less powerful version of its Codex software for automating coding. The new model enables software engineers to quickly complete tasks like editing code segments and running tests. Users can interrupt the model or redirect it to different coding tasks without waiting for lengthy computing processes to finish. The release represents OpenAI's effort to broaden its chipmaker partnerships beyond its primary reliance on Nvidia hardware.

TechCrunch
Feb 12th, 2026
OpenAI launches GPT-5.3-Codex-Spark powered by Cerebras' $10B chip partnership

OpenAI has launched GPT-5.3-Codex-Spark, a lightweight version of its coding tool designed for faster inference, powered by Cerebras' Wafer Scale Engine 3 chip. The model represents the "first milestone" in OpenAI's multi-year, $10 billion partnership with Cerebras announced last month. Spark is optimised for rapid prototyping and real-time collaboration, targeting daily productivity tasks rather than the heavier workloads handled by the standard 5.3 model. It's currently available in research preview for ChatGPT Pro users in the Codex app. The integration marks a new level of hardware partnership for OpenAI. Cerebras' WSE-3 chip contains 4 trillion transistors and excels at low-latency workflows. Cerebras recently raised $1 billion at a $23 billion valuation.