Simplify Logo

Full-Time

High Performance Computing

Triton + MPI, Engineer

Posted on 3/22/2024

Lamini AI

Lamini AI

11-50 employees

Customizes private language models for developers

Enterprise Software
AI & Machine Learning

Junior, Mid, Senior

Palo Alto, CA, USA

Category
Network Administration
System Administration
IT & Security
Requirements
  • High performance computing experience
  • Experience with developing and optimizing high performance collective and kernel libraries
  • Knowledge of technologies including MPI, ROCe, UCX, collectives, OpenAI Triton, etc.
Responsibilities
  • Developing and optimizing high performance collective and kernel libraries for running LLMs on AMD GPUs
  • Using technologies including MPI, ROCe, UCX, collectives, OpenAI Triton, etc.

Lamini's LLM platform allows developers to create customized, private language models using technologies such as fine-tuning, RLHF, retrieval-augmented training, data augmentation, and GPU optimization, providing data privacy, ownership, flexibility, and cost control. The platform leverages the latest generation of models, including those from HuggingFace and OpenAI, to specialize language models based on specific use cases and data constraints.

Company Stage

Series A

Total Funding

$25M

Headquarters

Menlo Park, California

Founded

2022

Growth & Insights
Headcount

6 month growth

20%

1 year growth

71%

2 year growth

100%
INACTIVE