Internship

Machine Learning Software Engineering Intern

Confirmed live in the last 24 hours

d-Matrix

d-Matrix

201-500 employees

AI compute platform for datacenters

Enterprise Software
AI & Machine Learning

Santa Clara, CA, USA

Hybrid position requiring onsite work at Santa Clara headquarters 3 days per week.

Category
Applied Machine Learning
AI Research
AI & Machine Learning
Software Engineering
Required Skills
Python
Machine Learning
C/C++
Requirements
  • Enrolled in a Bachelor's degree in Computer Science, Electrical and Computer Engineering, or a related scientific discipline.
  • A problem solver, be able to break-down and simplify complex problems to come up with elegant and efficient solutions
  • Proficient in programming with either Python/C/C++ programming languages.
Responsibilities
  • Develop performant implementations of SOTA ML models such as LLaMA, GPT, BERT, DLRM, etc.
  • Develop and maintain tools for performance simulation, analysis, debugging, profiling.
  • Develop AI infra software such as kernel compiler, inference engine, model factory, etc.
  • Develop QA systems/automation software.
  • Engage and collaborate with the rest of the SW team to meet development milestones.
  • Contribute to publication of papers and intellectual properties as applicable.
Desired Qualifications
  • Enrolled in either a MS or PhD in Computer Science, Electrical and Computer Engineering, or a related scientific discipline.
  • Understanding of CPU / GPU architectures and their memory systems.
  • Experience with specialized HW accelerators for deep neural networks.
  • Experience developing high performance kernels, simulators, debuggers, etc. targeting GPUs/Other accelerators.
  • Experience using Machine Learning frameworks, like PyTorch (preferrable), Tensorflow, etc.
  • Experience with Machine Learning compilers, like MLIR (preferrable), TVM etc.
  • Experience deploying inference pipelines.
  • Experience using or developing inference engines such as vLLM, TensorRT-LLM.
  • Passionate about AI and thriving in a fast-paced and dynamic startup culture.

d-Matrix focuses on improving the efficiency of AI computing for large datacenter customers. Its main product is the digital in-memory compute (DIMC) engine, which combines computing capabilities directly with programmable memory. This design helps reduce power consumption and enhances data processing speed while ensuring accuracy. d-Matrix differentiates itself from competitors by offering a modular and scalable approach, utilizing low-power chiplets that can be tailored for different applications. The company's goal is to provide high-performance, energy-efficient AI inference solutions to large-scale datacenter operators.

Company Stage

Series B

Total Funding

$149.8M

Headquarters

Santa Clara, California

Founded

2019

Growth & Insights
Headcount

6 month growth

2%

1 year growth

0%

2 year growth

9%
Simplify Jobs

Simplify's Take

What believers are saying

  • Growing demand for energy-efficient AI solutions boosts d-Matrix's low-power chiplets appeal.
  • Partnerships with companies like Microsoft could lead to strategic alliances.
  • Increasing adoption of modular AI hardware in data centers benefits d-Matrix's offerings.

What critics are saying

  • Competition from Nvidia, AMD, and Intel may pressure d-Matrix's market share.
  • Complex AI chip design could lead to delays or increased production costs.
  • Rapid AI innovation may render d-Matrix's technology obsolete if not updated.

What makes d-Matrix unique

  • d-Matrix's DIMC engine integrates compute into memory, enhancing efficiency and accuracy.
  • The company offers scalable AI solutions through modular, low-power chiplets.
  • d-Matrix focuses on brain-inspired AI compute engines for diverse inferencing workloads.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Hybrid Work Options