Full-Time

Principal Systems Performance Engineer

Posted on 11/23/2025

Micron Technology

Micron Technology

10,001+ employees

Memory and storage semiconductor manufacturer

No salary listed

Hyderabad, Telangana, India

In Person

Category
AI & Machine Learning (2)
,
Requirements
  • Strong programming skills in Python and familiarity with ML frameworks such as TensorFlow, PyTorch, or scikit-learn.
  • Experience in data preparation: cleaning, splitting, and transforming data for training, validation, and testing.
  • Proficiency in model training and development: creating and training machine learning models.
  • Expertise in model evaluation: testing models to assess their performance.
  • Skills in model deployment: launching server, live inference, batched inference
  • Experience with AI inference and distributed training techniques.
  • Strong foundation in GPU and CPU processor architecture
  • Familiarity with and knowledge of server system memory (DRAM)
  • Strong experience with benchmarking and performance analysis
  • Strong software development skills using leading scripting, programming languages and technologies (Python, CUDA, C, C++)
  • Familiarity with PCIe and NVLINK connectivity
Responsibilities
  • Design, implement, and maintain scalable & reliable ML infrastructure and pipelines.
  • Collaborate with data scientists and ML engineers to deploy machine learning models into production environments.
  • Automate and optimize ML workflows, including data preprocessing, model training, evaluation, and deployment.
  • Monitor and manage the performance, reliability, and scalability of ML systems.
  • Troubleshoot and resolve issues related to ML infrastructure and deployments.
  • Implement and manage distributed training and inference solutions to enhance model performance and scalability.
  • Utilize DeepSpeed, TensorRT, vLLM for optimizing and accelerating AI inference and training processes.
  • Understand key care abouts when it comes to ML models such as: transformer architectures, precision, quantization, distillation, attention span & KV cache, MoE, etc.
  • Build workload memory access traces from AI models.
  • Study system balance ratios for DRAM to HBM in terms of capacity and bandwidth to understand and model TCO.
  • Study data movement between CPU, GPU and the associated memory subsystems (DDR, HBM) in heterogeneous system architectures via connectivity such as PCIe/NVLINK/Infinity Fabric to understand the bottlenecks in data movement for different workloads.
  • Develop an automated testing framework through scripting.
  • Customer engagements and conference presentations to showcase findings and develop whitepapers.
Desired Qualifications
  • Experience in quickly building AI workflows: building pipelines and model workflows to design, deploy, and manage consistent model delivery.
  • Ability to easily deploy models anywhere: using managed endpoints to deploy models and workflows across accessible CPU and GPU machines.
  • Understanding of MLOps: the overarching concept covering the core tools, processes, and best practices for end-to-end machine learning system development and operations in production.
  • Knowledge of GenAIOps: extending MLOps to develop and operationalize generative AI solutions, including the management of and interaction with a foundation model.
  • Familiarity with LLMOps: focused specifically on developing and productionizing LLM-based solutions.
  • Experience with RAGOps: focusing on the delivery and operation of RAGs, considered the ultimate reference architecture for generative AI and LLMs.
  • Data management: collect, ingest, store, process, and label data for training and evaluation. Configure role-based access control; dataset search, browsing, and exploration; data provenance tracking, data logging, dataset versioning, metadata indexing, data quality validation, dataset cards, and dashboards for data visualization.
  • Workflow and pipeline management: work with cloud resources or a local workstation; connect data preparation, model training, model evaluation, model optimization, and model deployment steps into an end-to-end automated and scalable workflow combining data and compute.
  • Model management: train, evaluate, and optimize models for production; store and version models along with their model cards in a centralized model registry; assess model risks, and ensure compliance with standards.
  • Experiment management and observability: track and compare different machine learning model experiments, including changes in training data, models, and hyperparameters. Automatically search the space of possible model architectures and hyperparameters for a given model architecture; analyze model performance during inference, monitor model inputs and outputs for concept drift.
  • Synthetic data management: extend data management with a new native generative AI capability. Generate synthetic training data through domain randomization to increase transfer learning capabilities. Declaratively define and generate edge cases to evaluate, validate, and certify model accuracy and robustness.
  • Embedding management: represent data samples of any modality as dense multi-dimensional embedding vectors; generate, store, and version embeddings in a vector database. Visualize embeddings for improvised exploration. Find relevant contextual information through vector similarity search for RAGs.

Micron Technology designs and manufactures memory and storage products, including DRAM, NAND, and NOR flash. These products power devices across computing, networking, automotive, industrial, and mobile markets. The company sells to OEMs, distributors, and end users worldwide and funds ongoing research and development to meet evolving needs. Micron aims to provide scalable memory solutions and maintain an inclusive, growth‑oriented workplace for its employees.

Company Size

10,001+

Company Stage

IPO

Headquarters

Boise, Idaho

Founded

1978

Simplify Jobs

Simplify's Take

What believers are saying

  • Entire HBM4 supply for 2026 sold out under three-to-five-year contracts.
  • Q2 FY2026 revenue hits $23.9B, up 196% year-over-year on AI demand.
  • Mizuho raises price target to $740 on 6 May 2026 citing SSD shipments.

What critics are saying

  • Samsung and SK Hynix ramp HBM4 production in 2H 2026 eroding pricing power.
  • U.S. export restrictions to China eliminate 15-20% of Micron's market by 2027.
  • Hyperscaler capex pullback in 2027 strands Micron's $25B capex investments.

What makes Micron Technology unique

  • Micron ships 245TB Micron 6600 ION SSD using G9 QLC NAND for AI workloads.
  • Micron volume ships 36GB 12-high HBM4 for NVIDIA Vera Rubin systems.
  • Micron owns fabs in U.S., Taiwan, Japan, Singapore for vertical integration.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Health Insurance

Dental Insurance

Vision Insurance

Disability Insurance

Parental Leave

Unlimited Paid Time Off

Paid Holidays

Flexible Work Hours

Company News

Yahoo Finance
Apr 10th, 2026
Micron stock surges 230% since 2025 on AI demand and record $23.86B quarterly revenue

Micron Technology, a global leader in memory and storage solutions, saw its stock rise following ceasefire news, though shares remain 13% below their mid-March peak of $471. The company has surged over 230% since late 2025, significantly outperforming the S&P 500 Information Technology Index. Micron recently announced its exit from the consumer-facing Crucial brand to focus on high-growth AI infrastructure, including High Bandwidth Memory and low-power server modules for large language models. The company reported record second-quarter revenue of $23.86 billion, up 196% year-over-year, beating analyst expectations of $19.19 billion. Non-GAAP diluted earnings per share reached $12.20, surpassing the $8.79 consensus. The semiconductor company's performance is driven by rising memory prices and strong demand from the AI sector.

Gulf & Main Magazine
Apr 8th, 2026
SiMa.ai secures strategic investment from Micron to scale power-efficient physical AI solutions

SiMa.ai, a physical AI company, has secured a strategic investment from Micron Technology to scale production of high-performance, power-efficient AI solutions for edge applications. The investment strengthens collaboration between the companies on integrated compute and memory architectures for robotics, autonomous systems and industrial automation. SiMa.ai's Modalix MLSoC platform integrates Micron's LPDDR5X memory to deliver superior bandwidth efficiency and power optimisation. The company offers system-on-modules featuring Micron memory that fit into existing platforms for robotics, industrial automation and autonomous vehicles. The investment expands SiMa.ai's technology partnership ecosystem, which includes Arm, Synopsys, TSMC and Wind River. San Jose-based SiMa.ai focuses on scaling physical AI across multiple sectors including automotive, aerospace and healthcare.

Yahoo Finance
Apr 8th, 2026
Wall Street sees Palantir and Micron AI stocks soaring 35% and 50% despite recent dips

Palantir Technologies and Micron Technology have seen significant upward revisions to earnings estimates, creating a buying opportunity according to Wall Street analysts. Palantir's median target price of $200 per share implies 35% upside, whilst Micron's target of $550 suggests 50% potential gains. Palantir develops data integration and analytics platforms, particularly for government agencies, with its AI platform AIP distinguishing itself through a decision-making framework called an ontology. The company reported fourth-quarter revenue growth of 70% to $1.4 billion and achieved a record Rule of 40 score of 127%. Despite trading at 197 times adjusted earnings, analysts have raised current-year earnings forecasts by 30% to $1.31 per diluted share, implying 75% growth. The stock is down 28% from its high.

Yahoo Finance
Apr 8th, 2026
Micron stock drop overdone: Nvidia's AI chips still require HBM despite Google's TurboQuant

Micron Technology shares plunged in late March after Alphabet unveiled TurboQuant, software that compresses memory footprints in large language models during inference. The sell-off stemmed from concerns about Micron's relationship with Nvidia, as its high-bandwidth memory solutions power Nvidia's graphics processing units. However, the panic-selling appears premature. TurboQuant reduces working memory requirements during operation but doesn't shrink AI models themselves. Nvidia's GPUs remain dependent on external memory systems, including Micron's stacked dynamic random-access memory layers, for seamless data transfer between parameters and compute networks. Whilst TurboQuant minimises storage space whilst preserving model accuracy, it doesn't eliminate the need for rapid data transfer capabilities that Micron's chips provide. Nvidia's AI ecosystem continues to require massive amounts of specialised memory to process the terabytes needed for today's models.

Yahoo Finance
Apr 6th, 2026
Micron's revenue triples to $23.9B on mid-110% DRAM price surge, but growth faces sustainability concerns

Micron Technology's shares have more than quadrupled over the past year, but the growth story relies heavily on soaring prices rather than volume. In its latest quarter ending 26 February, revenue reached $23.9 billion, triple the prior year's $8.1 billion, whilst net income of $13.8 billion was nearly nine times higher. However, the revenue surge was primarily price-driven. DRAM revenue rose 207%, with average selling prices up mid-110% whilst shipments increased only mid-40%. NAND products showed similar patterns, with sales up 169% as prices more than doubled. The sustainability of this growth faces challenges as Micron will soon compare against these inflated figures. If supply catches up with demand and prices decline, growth could turn negative. Despite trading at only six times estimated future earnings, the stock has pulled back over 20% from its 52-week high.

INACTIVE