Full-Time

AI / ML System Software Engineer

Staff

Posted on 5/7/2024

d-Matrix

d-Matrix

51-200 employees

AI compute platform for datacenters

Enterprise Software
AI & Machine Learning

Senior

Santa Clara, CA, USA

Hybrid, working onsite at our Santa Clara, CA headquarters 3 days per week.

Category
Backend Engineering
FinTech Engineering
Software Engineering
Required Skills
Python
Tensorflow
Data Structures & Algorithms
Pytorch
C/C++
Linux/Unix
Requirements
  • BS Computer Science, Engineering, Math, Physics or related degree with 5+ years of industry software development experience / MS Computer Science, Engineering, Math, Physics or related degree preferred with 2+ years of industry software development experience
  • Strong grasp of system software, data structures, parallel programming, computer architecture basics, and machine learning fundamentals.
  • Proficient in C/C++/Python development in Linux environment and using standard development tools.
  • Experience with distributed, high performance software design and implementation.
  • Self-motivated team player with a strong sense of ownership and leadership.
  • MS or PhD in Computer Science, Electrical Engineering, or related fields (preferred).
  • Prior startup, small team or incubation experience (preferred).
  • Work experience at a cloud provider or AI compute / sub-system company (preferred).
  • Experience with deep learning frameworks (such as PyTorch, Tensorflow) (preferred).
  • Experience with deep learning runtimes (such as ONNX Runtime, TensorRT,…) (preferred).
  • Experience with inference servers/model serving frameworks (such as Triton, TFServ, KubeFlow,…) (preferred).
  • Experience with distributed systems collectives such as NCCL, OpenMPI (preferred).
  • Experience deploying ML workloads on distributed systems (preferred).
Responsibilities
  • Be part of the team that helps productize the SW stack for our AI compute engine.
  • Responsible for the development, enhancement, and maintenance of the next-generation AI Deployment software.
  • Work across all aspects of the full stack tool chain and understand the nuances of what it takes to optimize and trade-off various aspects of hardware-software co-design.
  • Build and scale software deliverables in a tight development window.
  • Work with a team of compiler experts to build out the compiler infrastructure working closely with other software (ML, Systems) and hardware (mixed signal, DSP, CPU) experts in the company.

d-Matrix focuses on improving the efficiency of AI computing for large datacenter customers. Its main product is the digital in-memory compute (DIMC) engine, which combines computing capabilities directly within programmable memory. This design helps reduce power consumption and enhances data processing speed while ensuring accuracy. d-Matrix differentiates itself from competitors by offering a modular and scalable approach, utilizing low-power chiplets that can be tailored for different applications. The company's goal is to provide high-performance, energy-efficient AI inference solutions to large-scale datacenter operators.

Company Stage

Series B

Total Funding

$149.8M

Headquarters

Santa Clara, California

Founded

2019

Growth & Insights
Headcount

6 month growth

11%

1 year growth

-2%

2 year growth

219%
Simplify Jobs

Simplify's Take

What believers are saying

  • d-Matrix's AI chip, Corsair, challenges Nvidia's dominance in AI hardware.
  • The demand for energy-efficient AI inference solutions is increasing in data centers.
  • d-Matrix raised $110 million in Series B funding, backed by Microsoft.

What critics are saying

  • Increased competition from Nvidia, AMD, and Intel could pressure d-Matrix's market share.
  • Regulatory challenges from the EU's AI Act and US guidelines may affect product strategies.
  • Reliance on large-scale datacenter operators could impact revenue if purchasing strategies shift.

What makes d-Matrix unique

  • d-Matrix integrates compute into programmable memory, enhancing efficiency and accuracy.
  • The DIMC engine offers a unique digital in-memory compute solution for AI inference.
  • d-Matrix's modular chiplet design allows for customizable and scalable AI compute solutions.

Help us improve and share your feedback! Did you find this helpful?

INACTIVE