Full-Time

Research Scientist

Scientific Discovery

Posted on 11/19/2025

The Allen Institute for AI

The Allen Institute for AI

201-500 employees

Non-profit AI research institute and tools

Compensation Overview

$167k - $260.6k/yr

+ Bonus plans

Seattle, WA, USA

Hybrid

On-site work from Seattle office; on-site requirements vary by position/team.

Category
AI & Machine Learning (2)
,
Requirements
  • PhD (or nearing completion) with research experience in Computer Science or related field
Responsibilities
  • Collaborate with AI researchers and software engineers to advance ambitious, team-based research in intelligent agents and scientific discovery
Desired Qualifications
  • Strong publication record in AI-related areas with venues such as ACL, ICLR, EMNLP, NeurIPS, ICML, NAACL
  • Contributions to the research community (e.g., workshop organization, tutorials) are a plus
  • Strong software engineering skills
  • Ability to work on-site in Seattle
The Allen Institute for AI

The Allen Institute for AI

View

AI2 advances artificial intelligence through nonprofit research and open engineering that benefits society. It builds open projects—such as AllenNLP, Aristo, Semantic Scholar and others—that researchers use to do language understanding, reasoning, science Q&A, and literature search. Unlike many AI firms that sell products, AI2 is funded by grants and donations and releases its tools and datasets openly to the public. Its goal is to improve AI reasoning and language understanding and apply these advances to real-world problems in education, science, and policy for the public good.

Company Size

201-500

Company Stage

N/A

Total Funding

N/A

Headquarters

Seattle, Washington

Founded

2014

Simplify Jobs

Simplify's Take

What believers are saying

  • MolmoBot achieves zero-shot sim-to-real transfer using 230,000 diverse simulation scenes.
  • SERA-32B coding agent matches larger models at 54.2% SWE-bench rate using 26x less compute.
  • Molmo 2 video models process arbitrary-length videos with full training transparency.

What critics are saying

  • Microsoft hired CEO Ali Farhadi, Hanna Hajishirzi, and Ranjay Krishna in March 2026.
  • FFST shifts to proposal-based funding prioritizing applied AI over frontier models.
  • Key departures jeopardize $152M NSF-Nvidia OMAI grant within 12-18 months.

What makes The Allen Institute for AI unique

  • AI2 releases fully open models like MolmoWeb and Olmo 3 with weights, data, and code.
  • AI2 pioneered byte-level Bolmo models bypassing subword tokenization for better multilingual fairness.
  • AI2's AutoDiscovery autonomously generates hypotheses from datasets in oncology and ecology.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Medical

401(k)

Visa Sponsorship

Time off

Catered meals & free snacks

Training & development

Tuition reimbursement

Remote & hybrid work options

Company News

SiliconANGLE Media
Mar 24th, 2026
Ai2 releases open-source visual AI agent that can take control of web browsers.

Ai2 releases open-source visual AI agent that can take control of web browsers. Allen Institute for AI, a prominent Seattle-based nonprofit research organization working on advancing artificial intelligence models and systems, today launched a new open-source AI agent that can take control of web browsers on a user's behalf and automate tasks. Web agents represent the next step of what is called vision-language models, which move large language models from understanding images and text through captions and answering questions to taking actions. Today, the company announced MolmoWeb, built on the Molmo 2 multimodal model family, available in two sizes: 4 billion and 8 billion parameters. It will be available for free, along with the weights, training data and code (coming soon), as well as the evaluation tools used to build it. It's designed to be self-hosted locally or in the cloud. To take actions, AI agents must interpret instructions from humans and what can be seen. That includes a set of tasks written in conversational language and a live web page. The AI model observes the web page through a series of screenshots and then interacts directly with it via the interface by predicting what will happen when it takes actions such as clicking, typing characters into text fields, or scrolling up and down. The company said that, unlike other open-weight web agents, MolmoWeb was trained without compressing a proprietary vision-based agent. The data comes from synthetically generated text-only accessibility agents and human usage of actual web browsing activities. The agent interface supports navigating URLs, clicking on screen coordinates, typing text into fields, scrolling through pages, opening and switching browser tabs and sending a message back to the user. All of these actions work directly within the browser, with click locations represented as coordinates in pixels when executed. Ai2 said the agent was designed this way so that it won't break if the underlying webpage code or HTML changes on the fly. For example, some web pages obfuscate, or hide, how they operate under the hood in order to protect themselves. Some of them use specialized JavaScript engines in order to detect bots, stop ad blockers, display animations, track users and more. Using the underlying code can also consume tens of thousands of tokens, the essential currency of AI operations. Visual interfaces also behave much more closely to how humans interact with web interfaces: What a person sees is how they will approach the page. It means it's easier to debug why the model did what it did. In spite of the compact size, Ai2 said MolmoWeb achieves state-of-the-art results among open-weight web agents. When tested on popular evaluation suites, the 8B model scored 78.2% on WebVoyager, 42.3% on DeepShop, and 49.5% on TailBench. It outperformed leading open-weight models such as Fara-7B across all four benchmarks. The company said that MolmoWeb can also outperform agents built on GPT-4 that rely on annotated and structured page data. Ai2 said that's a particularly important result given that those models can "see" deeply into the very code of the webpage and also have substantially larger parameter sizes - by colossal orders of magnitude. like comparing a mouse to an elephant. More access to open-weight browser AI agents will also help researchers and hobbyists develop their own web-using automations. Closed-source large language model providers have already dipped their toes into the market with agentic web browsers capable of automating web tasks, including OpenAI Group PBC and Perplexity AI Inc., with ChatGPT Atlas and Perplexity Comet, respectively. Image: allen Institute for AI. A message from John Furrier, co-founder of SiliconANGLE: Support its mission to keep content open and free by engaging with theCUBE community. Join theCUBE's Alumni Trust Network, where technology leaders connect, share intelligence and create opportunities. * 15M+ viewers of theCUBE videos, powering conversations across AI, cloud, cybersecurity and more * 11.4k+ theCUBE alumni - Connect with more than 11,400 tech and business leaders shaping the future through a unique trusted-based network. About SiliconANGLE Media SiliconANGLE Media is a recognized leader in digital media innovation, uniting breakthrough technology, strategic insights and real-time audience engagement. As the parent company of SiliconANGLE, theCUBE Network, theCUBE Research, CUBE365, theCUBE AI and theCUBE SuperStudios - with flagship locations in Silicon Valley and the New York Stock Exchange - SiliconANGLE Media operates at the intersection of media, technology and AI. Founded by tech visionaries John Furrier and Dave Vellante, SiliconANGLE Media has built a dynamic ecosystem of industry-leading digital media brands that reach 15+ million elite tech professionals. Its new proprietary theCUBE AI Video Cloud is breaking ground in audience interaction, leveraging theCUBEai.com neural network to help technology companies make data-driven decisions and stay at the forefront of industry conversations.

NJUDEV
Mar 24th, 2026
AI2's VLA harness: robot tests get standardized.

AI2's VLA harness: robot tests get standardized. AI2 released an open-source framework for evaluating visual-language-action (VLA) models on robot simulations. This provides a unified way to compare model performance across different benchmarks. What it means for you. This tool simplifies testing AI models for robotics, potentially allowing you to integrate more advanced automation into your services and gain a competitive edge.

NJUDEV
Mar 24th, 2026
AI2's MolmoPoint: sharper vision-language understanding.

AI2's MolmoPoint: sharper vision-language understanding. AI2 presents MolmoPoint, an advancement improving how vision-language models process information. This enhancement leads to more accurate and nuanced understanding in AI systems that combine images and text. What it means for you. You'll likely see AI tools that can better understand images and text together, leading to more accurate content analysis and creation.

GeekWire
Mar 23rd, 2026
Microsoft hires former Ai2 CEO Ali Farhadi and key researchers for Suleyman's AI team.

Microsoft hires former Ai2 CEO Ali Farhadi and key researchers for Suleyman's AI team. by Todd Bishop on Mar 23, 2026 at 3:32 pm Microsoft is hiring a group of top AI researchers from the Seattle-based Allen Institute for AI and the University of Washington, including former Ai2 CEO Ali Farhadi, GeekWire has learned. Farhadi, Hanna Hajishirzi, and Ranjay Krishna are expected to join Mustafa Suleyman's organization at Microsoft while retaining their faculty positions at the UW's Allen School of Computer Science and Engineering. Also joining is Sophie Lebrecht, the former Ai2 chief operating officer. The move follows Farhadi's departure from Ai2, announced March 12. Farhadi had led the Seattle-based nonprofit research institute for more than two and a half years. In a post on LinkedIn on Tuesday, Farhadi wrote that he was excited to join Microsoft. "I believe this is an opportunity to work on something that goes beyond what a frontier lab can do alone: exploring what becomes possible when AI is built from within a frontier ecosystem," he said. Suleyman, the CEO of Microsoft AI, narrowed his focus last week from overseeing consumer-oriented Copilot products to leading Microsoft's Superintelligence team. The hires come as Microsoft works to reduce its dependence on OpenAI for frontier AI models, competing against Amazon, Google, and others. Suleyman's Superintelligence team, formed in November, is part of a broader push to further develop advanced foundation models. Microsoft has already hired researchers from Google DeepMind, Meta, OpenAI, and Anthropic, and the addition of the Ai2 and UW group would bring deep expertise in open-source model development and training efficiency - where Ai2 has punched well above its weight. Backing from NSF and Nvidia. The exits represent a notable collective loss for Ai2, which was founded in 2014 by the late Microsoft co-founder Paul Allen. Hajishirzi is a co-lead of the OLMo open-source language model project and a co-principal investigator on a $152 million, five-year initiative backed by the National Science Foundation and Nvidia to build open AI models for scientific research. She represented Ai2 in multiple sessions last week at Nvidia's GTC conference in San Jose, including a panel on the future of open models alongside Nvidia CEO Jensen Huang. Krishna has led the development of Ai2's Molmo multimodal models, among other projects. He also presented at the Nvidia conference last week on behalf of the institute. Farhadi, a computer vision specialist, co-founded Ai2 spinout Xnor.ai, which Apple acquired in 2020 for an estimated $200 million. He led machine learning efforts at Apple before returning to lead Ai2 as CEO in July 2023. Ai2 interim CEO Peter Clark acknowledged the departures in a statement, saying the institute remains committed to its mission and its partnerships with the NSF and Nvidia, including the OMAI initiative. "These initiatives are backed by a broad, experienced team with the expertise and continuity needed to carry this work forward," Clark said. "We're confident in our ability to build on the strong foundation already in place and to expand the impact of these efforts in the months ahead." He added that the institute is "grateful for the leadership and contributions of Ali, Hanna, Ranjay, and others" in advancing Ai2's work, and wished them well. In a post about the hires on LinkedIn, Suleyman praised Farhadi for leading Ai2 in releasing more than 100 models in a single year and called Hajishirzi "one of the most cited researchers of natural language processing in the world, full stop." Suleyman described Lebrecht as having scaled Ai2's operations and open-source efforts, noting that she also co-founded the AI company Neon Labs and holds a PhD in cognitive neuroscience from Brown University. He said they will help pursue Microsoft's mission of "humanist superintelligence: safer, controllable, more capable AI systems in service of humanity and our toughest problems." When news broke earlier this month that Farhadi was leaving, Ai2 board chair Bill Hilf told GeekWire that Farhadi wanted to pursue research at the extreme frontier of AI, where for-profit companies are spending billions on training the most advanced models. At the time, Hilf said the board had to weigh whether a nonprofit's philanthropic dollars were best spent trying to keep pace, acknowledging that competing against tech giants at the largest scale of model development had become extraordinarily difficult. Changes in Ai2's funding realities. Behind the scenes, the changing nature of Ai2's funding environment has also been playing a role in the exits, according to people with knowledge of the situation. Ai2 was originally funded by Allen's Vulcan Inc. and later by his estate. Its primary backer is now the Fund for Science and Technology, a $3.1 billion foundation created under Allen's instructions and publicly launched in August, with a focus on applying science and technology to problems in areas aligned with Allen's passions, including AI, bioscience, and the environment. FFST, led by CEO Dr. Lynda Stuart, a physician-scientist who previously led the Institute for Protein Design at the UW, favors applied uses of AI over the costly work of frontier models. In addition, while all Ai2 programs for 2026 are fully funded, these people said, FFST is moving from providing Ai2 with overall annual funding to a proposal-based process, with future support expected to favor real-world applications of AI over building open-source foundation models. The shift helps explain the departures of researchers focused on model development. A spokesperson for the Fund for Science and Technology said Ai2's "work and mission remain the same" and that FFST's broader program strategies are still under development. Farhadi, Hajishirzi, and Krishna are researchers whose work centers on building and advancing AI models. Microsoft's Superintelligence team, backed by billions in compute investment, offers the resources and mandate to pursue that work at a much larger scale.

HPCwire
Mar 13th, 2026
New Ai2 Robotics Models Aim to Bridge the Sim-to-Real Gap

New Ai2 robotics models aim to bridge the sim-to-real gap. Press play to listen to this content Ai2 is making a bold claim in robotics: It's possible for a robot to learn useful manipulation skills entirely in simulation, then carry those skills into the real world without ever being trained on real-world robot data. That is the core idea behind MolmoSpaces and MolmoBot, two new open releases from the Allen Institute for AI (Ai2). MolmoSpaces is a large simulation environment for embodied AI research, built from more than 230,000 indoor scenes, over 130,000 object models, and over 42 million robotic grasp annotations. MolmoBot is the manipulation model trained on top of it. Together, these models are designed to address the sim-to-real gap, a foundational robotics challenge of ensuring that behaviors learned in simulation still work when deployed on physical robots. In evaluations, MolmoBot successfully performed several core manipulation tasks. It was able to perform pick-and-place operations, manipulate articulated objects such as drawers and cabinets, and open doors on two robot platforms, including the Franka FR3 arm and the RB-Y1 mobile manipulator, without real-world fine-tuning. This kind of result is known as zero-shot sim-to-real transfer, and it challenges the common view in robotics that synthetic training can only go so far before human-guided demonstrations become necessary. This claim of zero-shot sim-to-real transfer is what makes this release notable. Much of today's robotics work depends on large volumes of teleoperated data collected by people guiding robots through tasks in the physical world. Those datasets are expensive to produce, usually closed, and difficult to reproduce. To learn more about the problem, AIwire interviewed Ranjay Krishna, a researcher who leads Ai2's PRIOR team, which works on multimodal and embodied AI. The group develops models that combine vision, language, and audio, and has released open vision-language systems alongside projects in areas like satellite imagery analysis. More recently, the team has expanded that work into robotics, where those models are used to guide physical actions. Krishna said his group wanted to approach the sim-to-real problem differently by asking what simulation itself was missing, rather than assuming real-world data was the only solution. His answer was diversity. "Our overall goal is to deliver breakthrough AI models that can tackle really hard challenges. Our approach has been to turn to simulation and figure out what it is about simulation that makes it less appealing for people, versus going out and collecting large, expensive data sets," Krishna said. "Our hypothesis going into this set of projects a couple of years ago was that simulation data is just not diverse enough for us to generate enough useful data to train these models." Krishna explained how the issue is not that simulation is inherently too artificial, but that it often lacks the range of conditions robots encounter outside the lab. Ai2 addressed that by drastically expanding the variety of environments and situations available during training. Robots trained in MolmoSpaces encounter different layouts, lighting conditions, object placements, camera viewpoints, and physical interactions, all generated inside the simulator. The idea is that if a model sees enough variation during training, the real world begins to look like just another environment. MolmoSpaces provides an extensive simulation environment used to generate that diversity. The system includes more than 230,000 indoor scenes ranging from homes and offices to hospitals and museums, each populated with objects whose physical properties, like weight, material, and articulation, are modeled for robotic interaction. Because everything is simulated, researchers can generate large volumes of robot trajectories and test how well models generalize across many environments rather than evaluating them in a single fixed setup. Krishna gave an especially vivid example. In many robot systems, even a slight camera shift can cause performance to collapse because the model has become overly dependent on a fixed visual viewpoint. Ai2 wanted to break that brittleness by introducing those kinds of disruptions directly into simulation. As a result, Krishna said, MolmoBot could continue performing tasks even when the camera was moved around substantially, including by hand. As it typically does with its research releases, Ai2 is releasing MolmoSpaces and MolmoBot as open infrastructure. The organization is publishing not just the models, but the whole stack: data, generation pipelines, assets, benchmarks, and tools. MolmoSpaces is also designed to work across common simulators, including MuJoCo, ManiSkill, and Nvidia Isaac Lab and Sim, making it more accessible as shared research infrastructure. "A big principle that we've adhered to here at the Allen Institute is that everything that we do is completely reproducible, completely open source," Krishna said. "You can take all of our data, all of our environments, and build from scratch everything that we build in-house." Krishna acknowledged that the sim-to-real gap has become a central debate in robotics research: "It's been such a hot topic. There are blog posts written by some of these larger corporations where they talk about how real data is what we need to make these robots work in the real world, and anything you do in simulation just isn't going to transfer to the real world," he said. "We're empirically showing that it actually does work. The thing that was missing is a diversity of environments in our simulation engines, a high-quality set of assets, a high-quality set of grasps, and a large amount of data that we can generate." None of this means the sim-to-real gap is closed, Krishna was careful to note. Ai2 has demonstrated transfer for a small set of manipulation tasks in environments that are still relatively controlled, not the full complexity of real-world robotics. Tasks involving dynamic motion, such as catching or throwing objects, remain largely unexplored, as do more dexterous manipulations that require rotating or repositioning items with greater precision. Robots also still struggle with reasoning tasks such as searching through cluttered spaces for a specific object. Krishna said the team is already working on a follow-up to MolmoBot aimed at tackling more complex, multi-step tasks. Instead of simple manipulation, the goal is to handle multi-step instructions, such as finding an item in a home and retrieving it or cleaning up a room by breaking the request into smaller actions. That requires models that can map their surroundings, remember where they have already looked, and plan a series of steps toward a goal. Krishna added that simulation alone will not solve the problem. While large simulated environments can accelerate training, robots will still need to learn continuously from real-world experience and demonstrations. "Simulation is definitely part of the answer," he said, but developing robots that can keep learning after deployment remains a major challenge. For Krishna, the biggest takeaway from this project is that simulation can play a much larger role in robotics than many researchers once believed: "The main message for us is that sim-to-real is possible and you can reproduce these results, and we're hoping that it's going to lead to better, more equitable models that anyone can build."

INACTIVE