model release

AI2 releases robotics models trained entirely in simulation, achieving zero-shot real-world transfer

TL;DR

AI2 has released MolmoSpaces and MolmoBot, robotics models trained exclusively in simulation that transfer directly to real robots without manual real-world data collection or fine-tuning. The approach eliminates months of teleoperated demonstrations typically required for simulation-trained robots. Both systems are open-source.

2 min read
0

AI2 Releases Robotics Models Trained Entirely in Simulation, Achieving Zero-Shot Real-World Transfer

AI research institute AI2 has released two open-source robotics models—MolmoSpaces and MolmoBot—trained exclusively in simulation that achieve zero-shot transfer to real robots without any manually collected real-world data or fine-tuning.

What's New

The models represent a significant shift in robotics training methodology. Conventional approaches require researchers to spend months collecting teleoperated real-world demonstrations before simulation-trained robots become reliable. AI2's approach eliminates this bottleneck entirely.

MolmoSpaces, the foundation dataset, contains:

  • 230,000+ indoor scenes
  • 130,000+ curated objects
  • 42 million physics-based robotic grasping annotations

MolmoBot, built on MolmoSpaces, demonstrates capabilities including:

  • Object picking and placement
  • Opening drawers
  • Operating doors

All tasks execute without training data from real-world demonstrations.

The Technical Approach

According to Ranjay Krishna, director of the PRIOR team at AI2, the key insight is straightforward: the simulation-to-reality gap shrinks dramatically when researchers increase the variety of simulated environments, objects, and camera conditions. Rather than improving physics simulation fidelity, the models benefit from diversity in training conditions.

This aligns with recent findings in robotics research showing that breadth of training distribution often matters more than pixel-perfect simulation accuracy. By exposing models to hundreds of thousands of variations in scene configuration, object types, and viewpoints, the models learn generalizable behaviors that transfer directly to physical systems.

Open-Source Release

Both models and supporting tools are available publicly. Technical details are available in the accompanying research paper. The open-source approach allows other research groups and robotics companies to build on the foundation rather than starting from zero with their own simulation data collection.

What This Means

This work addresses one of robotics' most significant friction points: the cost and time required to train deployable systems. If the zero-shot transfer results hold up in broader testing, the implications are substantial. Companies and research labs could dramatically reduce development timelines—from months of manual demonstration collection to weeks of model training. This could accelerate deployment of manipulation tasks in warehousing, manufacturing, and service robotics.

The emphasis on simulation diversity over physics accuracy also reframes how the robotics community should approach simulation tools. Rather than competing on fidelity, platforms that generate high-variance synthetic training data may prove more valuable. This could shift investment and resource allocation within the robotics software ecosystem.

Related Articles

model release

Allen Institute releases EMO, 14B parameter MoE model with selective 12.5% expert use

Allen Institute for AI released EMO, a 1B-active, 14B-total-parameter mixture-of-experts model trained on 1 trillion tokens. The model uses 8 active experts per token from a pool of 128 total experts, and can maintain near full-model performance while using just 12.5% of its experts for specific tasks.

model release

IBM Releases Granite Embedding 311M R2 With 32K Context, 200+ Language Support

IBM released Granite Embedding 311M Multilingual R2, a 311-million parameter dense embedding model with 32,768-token context length and support for 200+ languages. The model scores 64.0 on Multilingual MTEB Retrieval (18 tasks), an 11.8-point improvement over its predecessor, and ships with ONNX and OpenVINO models for production deployment.

model release

IBM releases Apache 2.0 Granite 4.1 LLMs in 3B, 8B, and 30B sizes

IBM has released the Granite 4.1 family of language models under Apache 2.0 license. The models come in 3B, 8B, and 30B parameter sizes. Unsloth has released 21 GGUF quantized variants of the 3B model ranging from 1.2GB to 6.34GB.

model release

IBM Releases Granite 4.1 30B With 131K Context Window and Enhanced Tool-Calling

IBM released Granite 4.1 30B, a 30-billion parameter instruction-following model with a 131,072 token context window. The model scores 80.16 on MMLU 5-shot and 88.41 on HumanEval pass@1, with enhanced tool-calling capabilities following OpenAI's function definition schema.

Comments

Loading...