Internship

Research Scientist Intern

Posted on 7/19/2023

Mosaic ML

Mosaic ML

51-200 employees

Training and deploying generative AI models

No salary listed

San Francisco, CA, USA + 1 more

More locations: New York, NY, USA

Category
AI & Machine Learning
Lab & Research
Required Skills
Pytorch
Connection
Connection
Connection
logo

Get referrals →

You have ways to get a Mosaic ML referral from your network.

💡

Applications through a referral are 3x more likely to get an interview!

Requirements
  • Required:
  • Pursuing an undergraduate or graduate degree in computer science or related fields (electrical engineering, neuroscience, physics, math, etc.)
  • Some proficiency with the fundamentals of deep learning
  • Proficient software engineering skills, including with PyTorch
  • Nice to have:
  • Knowledge of the systems aspects of how neural networks train and the resources utilized in the process of doing so
  • Research experience in deep learning
  • US work authorization required
  • Hourly Rate: $58.00
Responsibilities
  • Adapting, improving, and evaluating a method from the literature
  • Designing an entirely new method
  • Composing together multiple methods to create new recipes for efficient training
  • Scientifically investigating how neural networks learn in practice
  • Exploring new approaches for training neural networks

MosaicML focuses on training and deploying generative AI models for businesses that need AI solutions for tasks like code generation and data analysis. Their platform allows clients to integrate Large Language Models (LLMs) into their applications easily and efficiently. It is user-friendly and designed to provide significant cost savings, up to 15 times less than traditional methods, while ensuring clients maintain full control of their data. MosaicML also offers services to train and serve large AI models at scale, managing complex aspects like orchestration and infrastructure. A key feature of their platform is its ability to integrate with existing data pipelines and tools, and it is cloud-agnostic, meaning it can work in any cloud environment. This flexibility and focus on efficiency set MosaicML apart from competitors. The company's goal is to empower businesses to leverage AI technologies effectively while ensuring data security and cost-effectiveness.

Company Size

51-200

Company Stage

Acquired

Total Funding

$1.4B

Headquarters

San Francisco, California

Founded

2021

Simplify Jobs

Simplify's Take

What believers are saying

  • Partnership with Weights & Biases increases visibility and attracts more platform users.
  • Integration with Oracle Cloud enhances cloud-agnostic capabilities, offering clients more flexibility.
  • Acquisition by Databricks provides additional resources and market reach for scaling.

What critics are saying

  • Increased competition from open-source models like OLMo challenges MosaicML's market position.
  • Potential AI bubble burst could reduce investment and interest in AI startups.
  • Reliance on Oracle Cloud poses risks if service terms or pricing change.

What makes Mosaic ML unique

  • MosaicML offers a user-friendly, open-source platform for integrating large language models.
  • The platform promises up to 15x cost savings, enhancing its appeal to businesses.
  • MosaicML's cloud-agnostic design allows seamless integration with existing data pipelines.

Help us improve and share your feedback! Did you find this helpful?

Growth & Insights and Company News

Headcount

6 month growth

-1%

1 year growth

5%

2 year growth

6%
Plants Need CO2
Mar 27th, 2025
The TAO of data: How Databricks is optimizing AI LLM fine-tuning without data labels

The Mosaic research team at Databricks developed the new TAO method.

Forbes
Feb 5th, 2024
How Olmo From Ai2 Redefines Llm Innovation

OLMo LLM AI2The Allen Institute for AI created the Open Language Model, or OLMo, which is an open-source large language model with the aim of advancing the science of language models through open research. It marks a major milestone in the evolution of large language models.Unlike current open large language models like Llama and Mistral, which might limit access to their training data, architectures, or evaluation methodologies, OLMo stands out by providing full access to its pre-training data, training code, model weights, and evaluation suite. This openness is aimed at empowering academics and researchers to collectively study and advance the field of language modeling.OLMo represents a collaborative effort to advance the science of language models. The developers behind the LLM have a mission to empower academics and researchers by providing access to training code, models, and evaluation code necessary for open research.OLMo's architecture is built on AI2’s Dolma dataset, which features a three trillion-token open corpus. It includes full model weights for four model variants at the 7B scale, each trained to at least 2T tokens. OLMo's innovative aspects include its training approaches, size, and the diversity of data it was trained on

Analytics India Magazine
Jan 2nd, 2024
MosaicML Announces Beyond Chinchilla-Optimal for LLM Scaling Laws in Inference

MosaicML has unveiled its latest research, titled "Beyond Chinchilla-Optimal: Accounting for Inference in Language Model Scaling Laws."

LangLabs
Nov 6th, 2023
Exciting Updates from the AI Industry: British Startups, Accessible AI, and Cutting-Edge Innovations

Weights & Biases has partnered with @MosaicML to provide a FREE course covering LLM Evaluation, Dataset Curation, Distributed Training, and Practical Tips from industry experts.

Blocks and Files
Oct 25th, 2023
Storage news ticker - October 25

Arcion's software will be integrated into MosaicML's GenAI software.

INACTIVE