Full-Time

Machine Learning Engineer

Inference

Confirmed live in the last 24 hours

Together AI

Together AI

51-200 employees

Decentralized cloud services for AI development

Enterprise Software
AI & Machine Learning

Compensation Overview

$160k - $230kAnnually

+ Equity + Benefits

Mid, Senior

San Francisco, CA, USA

Category
Applied Machine Learning
Deep Learning
AI & Machine Learning
Required Skills
Rust
Python
CUDA
Pytorch
Requirements
  • 3+ years of experience writing high-performance, well-tested, production-quality code.
  • Proficiency with Python and PyTorch.
  • Demonstrated experience in building high performance libraries and tooling.
  • Excellent understanding of low-level operating systems concepts including multi-threading, memory management, networking, storage, performance, and scale.
  • Preferred: Knowledge of existing AI inference systems such as TGI, vLLM, TensorRT-LLM, Optimum
  • Preferred: Knowledge of AI inference techniques such as speculative decoding.
  • Preferred: Knowledge of CUDA/Triton programming.
  • Nice to have: Knowledge of Rust, Cython and compilers.
Responsibilities
  • Design and build the production systems that power the Together AI inference engine, enabling reliability and performance at scale.
  • Develop and optimize runtime inference services for large-scale AI applications.
  • Collaborate with researchers, engineers, product managers, and designers to bring new features and research capabilities to the world.
  • Conduct design and code reviews to ensure high standards of quality.
  • Create services, tools, and developer documentation to support the inference engine.
  • Implement robust and fault-tolerant systems for data ingestion and processing.

Together AI focuses on enhancing artificial intelligence through open-source contributions. The company offers decentralized cloud services that allow developers and researchers from various organizations to train, fine-tune, and deploy generative AI models. Their services cater to a wide range of clients, including small startups, large enterprises, and academic institutions. Together AI's business model is based on providing cloud-based solutions that support the development and deployment of AI models, generating revenue through service subscriptions and usage fees. The company stands out from competitors by emphasizing open and transparent AI systems, which fosters innovation and aims to achieve beneficial outcomes for society.

Company Stage

Series A

Total Funding

$222.3M

Headquarters

Menlo Park, California

Founded

2022

Growth & Insights
Headcount

6 month growth

69%

1 year growth

134%

2 year growth

617%
Simplify Jobs

Simplify's Take

What believers are saying

  • The rise of AI model interpretability tools aligns with Together AI's open-source approach.
  • Increasing adoption of AI in edge computing creates opportunities for lightweight model development.
  • Together AI's partnership with Meta for Llama 3.2 Vision model enhances its service offerings.

What critics are saying

  • Advanced AI models like Flux 1.1 Pro could overshadow Together AI's offerings.
  • Meta's free Llama 3.2 Vision model may attract developers away from Together AI.
  • Stricter AI regulations could impose additional compliance costs for Together AI.

What makes Together AI unique

  • Together AI focuses on open-source contributions, unlike many proprietary AI companies.
  • The company offers decentralized cloud services for AI model training and deployment.
  • Together AI's acquisition of CodeSandbox enhances its AI development platform with a code interpreter.

Help us improve and share your feedback! Did you find this helpful?