Full-Time

Developer Relations Lead

Multiple Teams

Posted on 7/25/2025

LMArena

LMArena

11-50 employees

Crowdsourced LLM evaluation and leaderboard

No salary listed

San Francisco, CA, USA

Hybrid

Hybrid position, requiring some in-office presence.

Category
Developer Relations
Required Skills
Machine Learning
Requirements
  • 5+ years of experience in Developer Relations, Developer Advocacy, or technical community engagement roles
  • Strong technical fluency with APIs, data infrastructure, or machine learning tools—you can read and write code, even if you're not a full-time engineer
  • Proven track record of creating technical content such as tutorials, documentation, demos, or blog posts for developer audiences
  • Experience engaging with technical users (e.g., AI researchers, ML engineers, open-source contributors), especially in community or support contexts
  • Demonstrated ability to gather developer feedback and translate it into actionable product or platform improvements
  • Excellent communication skills: able to explain complex technical systems clearly across mediums (written, verbal, live demos)
  • Comfortable working cross-functionally with engineering, product, and communications teams in a fast-paced, early-stage environment
  • A proactive, self-directed work style with the ability to operate independently and prioritize impact
Responsibilities
  • Driving developer adoption by making it easier for researchers and engineers to use our evaluation tools, APIs, and datasets.
  • Creating technical content such as tutorials, demo apps, notebooks, deep dives, and walkthroughs that explain LMArena’s infrastructure and methodologies.
  • Building bridges with the AI community through direct support, workshops, conferences, hackathons, and Discord.
  • Collecting and surfacing developer feedback to inform product, API, and tooling improvements.
  • Partnering with model providers and contributors to help them integrate into the platform and make the most of evaluation.
  • Championing community-built tools, libraries, and insights, helping amplify external contributions and shape standards for transparent evaluation.
  • Collaborating cross-functionally with Product, Engineering, Comms, and Research to help define how we explain and scale the platform.
Desired Qualifications
  • Experience with AI evaluation, LLM fine-tuning, or benchmarks.
  • Familiarity with open-source development, especially in ML ecosystems.
  • Experience hosting workshops, tutorials, or online dev-focused content.

LMArena is a crowdsourced platform for evaluating and ranking large language models using side-by-side prompts and votes, publishing a public Elo-based leaderboard. It evolved from Chatbot Arena and operates under Arena Intelligence Inc., with the FastChat framework powering its UI and backend, which remains open source. Users submit prompts, compare two anonymous model responses, and reveal identities after voting, creating a real-world, data-driven performance dataset. The company monetizes insights from model performance and user preferences to guide researchers, developers, and enterprises in model development and selection.

Company Size

11-50

Company Stage

Series A

Total Funding

$250M

Headquarters

San Francisco, California

Founded

2025

Simplify Jobs

Simplify's Take

What believers are saying

  • Annualized consumption run rate hit $30 million four months after September 2025 AI Evaluations launch.
  • Raised $150 million Series A at $1.7 billion valuation in January 2026 from Felicis and UC Investments.
  • Partners with OpenAI, Google, xAI for enterprise evaluations in law, medicine, and software engineering.

What critics are saying

  • April 2025 competitor paper proves OpenAI, Google partnerships enable Elo gaming, eroding neutrality.
  • OpenAI synthetic preference data replaces human voting, obsoleting leaderboards within 6-12 months.
  • Hugging Face leaderboards capture 40% open model evaluations, fragmenting LMArena market share.

What makes LMArena unique

  • LMArena uses crowdsourced Elo ratings from 60 million monthly conversations across 150 countries.
  • Platform evaluates models in text, vision, web development, and multimodal tasks beyond static benchmarks.
  • Open-source FastChat backend enables transparent, real-world human preference data collection.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Health Insurance

Dental Insurance

Vision Insurance

Company Equity

Growth & Insights and Company News

Headcount

6 month growth

0%

1 year growth

13%

2 year growth

13%
Quanmeishe
Jan 7th, 2026
LMArena achieves $1.7B valuation four months after launching its product

LMArena achieves $1.7B valuation four months after launching its product. LMArena, a startup that originally launched as a UC Berkeley research project in 2023, announced on Tuesday that it raised a $150 million Series A at a post-money valuation of $1.7 billion. The round was led by Felicis and the university's fund, UC Investments. The startup bolted out of the gate as a commercial venture with a $100 million seed round in May at a $600 million valuation. This new round means it raised $250 million in about seven months. LMArena is best known for its crowdsourced AI model performance leaderboards. Its consumer website lets a user type a prompt that it sends to two models, with the user then choosing which model did a better job. Those results, which now span more than 5 million monthly users across 150 countries and 60 million conversations a month, the company says, fuel the leaderboards. It ranks various models on a variety of tasks including text, web development, vision, text-to-image, and other criteria. The models it tests include various flavors of OpenAI GPT, Google Gemini, Anthropic Claude, and Grok, as well as ones that are geared toward specialties like image generation, text to image, or reasoning. The company began as Chatbot Arena, an open research project built by UC Berkeley researchers Anastasios Angelopoulos and Wei-Lin Chiang, and was originally funded through grants and donations. LMArena's leaderboards became something of an obsession among model makers. When LMArena started pursuing revenue, it partnered with select model companies such as OpenAI, Google, and Anthropic to make their flagship models available for its community to evaluate. In April, a group of competitors published a paper alleging that this helped those model makers game the startup's benchmarks, an allegation LMArena has vehemently denied. In September, it publicly launched a commercial service, AI Evaluations, in which enterprises, model labs, and developers can hire the company to perform model evaluations through its community. This gave LMArena an annualized "consumption rate" - as the company describes its annual recurring revenue (ARR) - of $30 million as of December, less than four months after launch. Join the disrupt 2026 waitlist. Add yourself to the disrupt 2026 waitlist to be first in line when early bird tickets drop. Past disrupts have brought Google cloud, netflix, microsoft, box, phia, a16z, elevenlabs, wayve, hugging face, elad gil, and vinod khosla to the stages - part of 250+ industry leaders driving 200+ sessions built to fuel your growth and sharpen your edge. Plus, meet the hundreds of startups innovating across every sector.

Dealroom.co
Jan 6th, 2026
LMArena company information, funding & investors

LMArena, open community platform to benchmark and compare ai models through side-by-side evaluations and user voting. Here you'll find information about their funding, investors and team.

Arena
Jan 6th, 2026
Fueling the World’s Most Trusted AI Evaluation Platform

Today, we’re excited to share a major milestone in LMArena’s journey. We’ve raised $150M of Series A led by Felicis and UC Investments (University of California), with participation from Andreessen Horowitz, The House Fund, LDVP, Kleiner Perkins, Lightspeed Venture Partners and Laude Ventures.

TipRanks
Jan 6th, 2026
LMArena raises $150M at $1.7B valuation to expand AI model ranking platform

LMArena, a startup operating a widely cited ranking system for AI models, has raised $150 million at a $1.7 billion valuation, according to The Information. The funding round was co-led by existing investors Felicis and the University of California's investment arm. The company evaluates AI model performance for clients including OpenAI, Google, xAI and Microsoft. The new capital will fund computing power needed to run the AI models it assesses and support hiring of technical staff. LMArena's ranking system has become an influential benchmark in the AI industry for comparing model capabilities across different providers.

intelligence360
May 30th, 2025
LMArena Secures $100M for AI Evaluation

LMArena, an AI evaluation platform, has secured $100 million in seed funding led by a16z and UC Investments, with participation from others. The funding supports the relaunch of LMArena, featuring a rebuilt UI, mobile-first design, and new features. LMArena aims to provide rigorous, community-driven AI evaluation, focusing on real-world performance. The platform is open and transparent, helping shape AI reliability. Learn more at lmarena.ai.

INACTIVE