Internship

Summer 2025 Engineering Intern

Posted on 2/1/2025

Together AI

Together AI

51-200 employees

Decentralized cloud services for AI development

No salary listed

San Francisco, CA, USA

Internship duration is approximately 12 weeks during Summer 2025.

Category
Backend Engineering
Software Engineering
Required Skills
Kubernetes
Rust
Python
JavaScript
Data Structures & Algorithms
Java
Go
C/C++
Requirements
  • Pursuing a degree in Computer Science, Computer Engineering or related field.
  • Proficient coding in one of these languages: Go, Rust, Python, C/C++, Java, JavaScript
  • Strong understanding of data structures, algorithms, and system design
  • Thrives in a fast paced environment
Responsibilities
  • Work on scalable, high-performance systems that power critical business processes.
  • Design and implement systems, services and APIs powering an AI-focused cloud platform
  • Dive into real-world challenges, including: building Kubernetes operators, building services and UX to manage cloud resources, efficiently scaling inference, real time monitoring and fault-tolerant AI systems
  • Collaborate with cross-functional teams to ensure seamless integration of backend systems with front-end applications.
  • Communicate the plans, progress, and results of projects to the broader team

Together AI focuses on enhancing artificial intelligence through open-source contributions. The company offers decentralized cloud services that allow developers and researchers to train, fine-tune, and deploy generative AI models. Their platform supports a wide range of clients, from small startups to large enterprises and academic institutions, by providing cloud-based solutions that simplify the development and deployment of AI models. Unlike many competitors, Together AI emphasizes open and transparent AI systems, which fosters innovation and aims to achieve beneficial outcomes for society. The company's goal is to advance the field of AI while ensuring accessibility and collaboration among users.

Company Size

51-200

Company Stage

Series B

Total Funding

$533.5M

Headquarters

Menlo Park, California

Founded

2022

Simplify Jobs

Simplify's Take

What believers are saying

  • Together AI secured $305M to expand its AI Acceleration Cloud with NVIDIA GPUs.
  • The open-source release of DeepCoder-14B enhances Together AI's reputation in the AI community.
  • Growing interest in open-source AI models boosts demand for Together AI's services.

What critics are saying

  • Emerging competitors like Deep Cogito may challenge Together AI's market position.
  • Snowflake's AI hub could divert talent and resources from Together AI.
  • Increasing infrastructure demand may strain Together AI's resources and service delivery.

What makes Together AI unique

  • Together AI focuses on open-source contributions, unlike many proprietary AI companies.
  • The company offers decentralized cloud services for scalable AI model deployment.
  • Together AI's DeepCoder-14B model is smaller yet efficient, with only 14 billion parameters.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Health Insurance

Company Equity

Growth & Insights and Company News

Headcount

6 month growth

-7%

1 year growth

8%

2 year growth

-3%
VentureBeat
Apr 10th, 2025
Deepcoder Delivers Top Coding Performance In Efficient 14B Open Model

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More. Researchers at Together AI and Agentica have released DeepCoder-14B, a new coding model that delivers impressive performance comparable to leading proprietary models like OpenAI’s o3-mini. Built on top of DeepSeek-R1, this model gives more flexibility to integrate high-performance code generation and reasoning capabilities into real-world applications. Importantly, the teams have fully open-sourced the model, its training data, code, logs and system optimizations, which can help researchers improve their work and accelerate progress.Competitive coding capabilities in a smaller packageThe research team’s experiments show that DeepCoder-14B performs strongly across several challenging coding benchmarks, including LiveCodeBench (LCB), Codeforces and HumanEval+.“Our model demonstrates strong performance across all coding benchmarks… comparable to the performance of o3-mini (low) and o1,” the researchers write in a blog post that describes the model.Interestingly, despite being trained primarily on coding tasks, the model shows improved mathematical reasoning, scoring 73.8% on the AIME 2024 benchmark, a 4.1% improvement over its base model (DeepSeek-R1-Distill-Qwen-14B). This suggests that the reasoning skills developed through RL on code can be generalized effectively to other domains.Credit: Together AIThe most striking aspect is achieving this level of performance with only 14 billion parameters. This makes DeepCoder significantly smaller and potentially more efficient to run than many frontier models.Innovations driving DeepCoder’s performanceWhile developing the model, the researchers solved some of the key challenges in training coding models using reinforcement learning (RL).The first challenge was curating the training data

VentureBeat
Apr 8th, 2025
New Open Source Ai Company Deep Cogito Releases First Models And They’Re Already Topping The Charts

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn MoreDeep Cogito, a new AI research startup based in San Francisco, officially emerged from stealth today with Cogito v1, a new line of open source large language models (LLMs) fine-tuned from Meta’s Llama 3.2 and equipped with hybrid reasoning capabilities — the ability to answer quickly and immediately, or “self-reflect” like OpenAI’s “o” series and DeepSeek R1.The company aims to push the boundaries of AI beyond current human-overseer limitations by enabling models to iteratively refine and internalize their own improved reasoning strategies. It’s ultimately on a quest toward developing superintelligence — AI smarter than all humans in all domains — yet the company says that “All models we create will be open sourced.”Deep Cogito’s CEO and co-founder Drishan Arora — a former Senior Software Engineer at Google who says he led the large language model (LLM) modeling for Google’s generative search product —also said in a post on X they are “the strongest open models at their scale – including those from LLaMA, DeepSeek, and Qwen.”The initial model lineup includes five base sizes: 3 billion, 8 billion, 14 billion, 32 billion, and 70 billion parameters, available now on AI code sharing community Hugging Face, Ollama and through application programming interfaces (API) on Fireworks and Together AI.They’re available under the Llama licensing terms which allows for commercial usage — so third-party enterprises could put them to work in paid products — up to 700 million monthly users, at which point they need to obtain a paid license from Meta.The company plans to release even larger models — up to 671 billion parameters — in the coming months.Arora describes the company’s training approach, iterated distillation and amplification (IDA), as a novel alternative to traditional reinforcement learning from human feedback (RLHF) or teacher-model distillation.The core idea behind IDA is to allocate more compute for a model to generate improved solutions, then distill the improved reasoning process into the model’s own parameters — effectively creating a feedback loop for capability growth. Arora likens this approach to Google AlphaGo’s self-play strategy, applied to natural language.The Cogito models are open-source and available for download via Hugging Face and Ollama, or through APIs provided by Fireworks AI and Together AI. Each model supports both a standard mode for direct answers and a reasoning mode, where the model reflects internally before responding.Benchmarks and evaluationsThe company shared a broad set of evaluation results comparing Cogito models to open-source peers across general knowledge, mathematical reasoning, and multilingual tasks. Highlights include:Cogito 3B (Standard) outperforms LLaMA 3.2 3B on MMLU by 6.7 percentage points (65.4% vs

Startup by Doc
Apr 2nd, 2025
Composio Secures $24 Million in Series A Funding: A Leap Forward for Agentic AI - Startup By DOC

In a significant boost to the rapidly evolving field of artificial intelligence, Composio, an emerging leader in Agentic AI, has successfully raised $24

PR Newswire
Mar 18th, 2025
Weka Expands Nvidia Integrations And Certifications, Unveils Augmented Memory Grid At Gtc 2025

Accelerates AI Reasoning With the NVIDIA AI Data Platform and Achieves New Storage Certifications for NVIDIA Cloud Partners and Enterprise DeploymentsSAN JOSE, Calif. and CAMPBELL, Calif., March 18, 2025 /PRNewswire/ -- From GTC 2025: WEKA, the AI-native data platform company, announced it is integrating with the NVIDIA AI Data Platform reference design and has achieved NVIDIA storage certifications to provide optimized AI infrastructure for the future of agentic AI and reasoning models. Additionally, the company announced new certifications for the NVIDIA Cloud Partner (NCP) Reference Architecture with NVIDIA GB200 NVL72 and the NVIDIA-Certified Systems™ Storage designation for enterprise AI factory deployments with NVIDIA Enterprise Reference Architectures

Pulse 2.0
Mar 4th, 2025
Together AI Secures $305M for Expansion

Together AI secured a $305 million Series B funding round, led by General Catalyst and Prosperity7, valuing the company at $3.3 billion. The investment will support the expansion of its AI Acceleration Cloud, including large-scale deployment of NVIDIA Blackwell GPUs. Together AI aims to enhance its platform for open source AI models, offering faster and more efficient solutions for AI developers and enterprises. The company has grown rapidly, now serving over 450,000 AI developers.

INACTIVE