Simplify Logo

Full-Time

ML Compiler Engineer

Staff

Confirmed live in the last 24 hours

d-Matrix

d-Matrix

51-200 employees

AI compute platform using in-memory computing

AI & Machine Learning

Senior, Expert

Toronto, ON, Canada

Hybrid position requiring onsite presence in Toronto for 3 days per week.

Category
Backend Engineering
Software QA & Testing
Software Engineering
Required Skills
Tensorflow
Requirements
  • Bachelor's degree in Computer Science with 7+ Yrs of relevant industry experience, MSCS Preferred with 5+ yrs of relevant industry experience.
  • Ability to deliver production quality code in modern C++.
  • Experience in modern compiler infrastructures, for example: LLVM, MLIR.
  • Experience in machine learning frameworks and interfaces, for example: ONNX, TensorFlow and PyTorch.
  • Experience in production compiler development.
Responsibilities
  • The d-Matrix compiler team is looking for exceptional candidates to help develop the compiler backend - specifically the problem of assigning hardware resources in a spatial architecture to execute low level instructions.
  • The successful candidate will be motivated, capable of solving algorithmic compiler problems and interested in learning intricate details of the underlining hardware and software architectures.
  • The successful candidate will join a team of experienced compiler developers, which will be guiding the candidate for a quick ramp up in the compiler infrastructure, in order to attack the important problem of mapping low level instructions to hardware resources.
  • We have opportunities specifically in the following areas: Model partitioning (pipelined, tensor, model and data parallelism), tiling, resource allocation, memory management, scheduling and optimization (for latency, bandwidth and throughput).

d-Matrix is developing a unique AI compute platform using in-memory computing (IMC) techniques with chiplet level scale-out interconnects, revolutionizing datacenter AI inferencing. Their innovative circuit techniques, ML tools, software, and algorithms have successfully addressed the memory-compute integration problem, enhancing AI compute efficiency.

Company Stage

Series B

Total Funding

$161.5M

Headquarters

Santa Clara, California

Founded

2019

Growth & Insights
Headcount

6 month growth

-18%

1 year growth

82%

2 year growth

219%
Simplify Jobs

Simplify's Take

What believers are saying

  • Securing $110 million in Series B funding positions d-Matrix for rapid growth and technological advancements.
  • Their Jayhawk II silicon aims to solve critical issues in AI inference, such as cost, latency, and throughput, making generative AI more commercially viable.
  • The company's focus on efficient AI inference could attract significant interest from data centers and enterprises looking to deploy large language models.

What critics are saying

  • Competing against industry giants like Nvidia poses a significant challenge in terms of market penetration and customer acquisition.
  • The high dependency on continuous innovation and technological advancements could strain resources and lead to potential setbacks.

What makes d-Matrix unique

  • d-Matrix focuses on developing AI hardware specifically optimized for Transformer models, unlike general-purpose AI chip providers like Nvidia.
  • Their digital in-memory compute (DIMC) architecture with chiplet interconnect is a first-of-its-kind innovation, setting them apart in the AI hardware market.
  • Backed by major investors like Microsoft, d-Matrix has the financial support to challenge established players like Nvidia.