Internship

Machine Learning Intern

Confirmed live in the last 24 hours

d-Matrix

d-Matrix

51-200 employees

AI compute platform for datacenters

Enterprise Software
AI & Machine Learning

Santa Clara, CA, USA

Hybrid position requiring onsite presence in Santa Clara, CA for 3 days per week.

Category
Applied Machine Learning
AI & Machine Learning
Required Skills
Python
CUDA
Data Structures & Algorithms
Pytorch
Requirements
  • Currently pursuing a degree in Computer Science, Electrical Engineering, Machine Learning, or a related field.
  • Familiarity with PyTorch and deep learning concepts, particularly regarding model optimization and memory management.
  • Understanding of CUDA programming and hardware-accelerated computation (experience with CUDA is a plus).
  • Strong programming skills in Python, with experience in PyTorch.
  • Analytical mindset with the ability to approach problems creatively.
  • Experience with deep learning model inference optimization.
  • Knowledge of data structures used in machine learning for memory and compute efficiency.
  • Experience with hardware-specific optimization, especially on custom hardware like D-Matrix, is an advantage.
Responsibilities
  • Research and analyze existing KV-Cache implementations used in LLM inference, particularly those utilizing lists of past-key-values PyTorch tensors.
  • Investigate 'Paged Attention' mechanisms that leverage dedicated CUDA data structures to optimize memory for variable sequence lengths.
  • Design and implement a torch-native dynamic KV-Cache model that can be integrated seamlessly within PyTorch.
  • Model KV-Cache behavior within the PyTorch compute graph to improve compatibility with torch.compile and facilitate the export of the compute graph.
  • Conduct experiments to evaluate memory utilization and inference efficiency on D-Matrix hardware.

d-Matrix focuses on improving the efficiency of AI computing for large datacenter customers. Its main product is the digital in-memory compute (DIMC) engine, which combines computing capabilities directly within programmable memory. This design helps reduce power consumption and enhances data processing speed while ensuring accuracy. Unlike many competitors, d-Matrix offers a modular and scalable approach, utilizing low-power chiplets that can be tailored for different applications. The company's goal is to provide high-performance, energy-efficient AI inference solutions to large-scale datacenter operators.

Company Stage

Series B

Total Funding

$149.8M

Headquarters

Santa Clara, California

Founded

2019

Growth & Insights
Headcount

6 month growth

11%

1 year growth

-2%

2 year growth

219%
Simplify Jobs

Simplify's Take

What believers are saying

  • d-Matrix raised $110 million in Series B funding, showing strong investor confidence.
  • The launch of the Corsair AI processor positions d-Matrix as a competitor to Nvidia.
  • Jayhawk II silicon advances low-latency AI inference for large language models.

What critics are saying

  • Competition from Nvidia, AMD, and Intel could pressure d-Matrix's market share.
  • Rapid AI innovation may lead to obsolescence if d-Matrix doesn't continuously innovate.
  • Potential regulatory changes in AI technology could impose new compliance costs.

What makes d-Matrix unique

  • d-Matrix's DIMC engine integrates compute into memory, enhancing efficiency and accuracy.
  • The company's chiplet-based modular design allows for scalable and customizable AI solutions.
  • d-Matrix focuses on power efficiency, addressing critical issues in AI datacenter workloads.

Help us improve and share your feedback! Did you find this helpful?