Multimodal AI

DreamDojo: A Generalist Robot World Model from Large-Scale Human Videos

SShenyuan GaoWWilliam LiangKKaiyuan ZhengAAyaan MalikSSeonghyeon YeSSihyun YuWWei-Cheng TsengYYuzhu DongKKaichun MoCChen-Hsuan LinQQianli MaSSeungjun NahLLoic MagneJJiannan XiangYYuqi XieRRuijie ZhengDDantong NiuYYou Liang TanKK. R. ZentnerGGeorge KurianSSuneel IndupuruPPooya JannatyJJinwei GuJJun ZhangJJitendra MalikPPieter AbbeelMMing-Yu LiuYYuke ZhuJJoel JangLLinxi "Jim" Fan
Published
February 6, 2026
Authors
30
Word Count
13,837

Pretraining robots using large-scale human videos.

Abstract

Being able to simulate the outcomes of actions in varied environments will revolutionize the development of generalist agents at scale. However, modeling these world dynamics, especially for dexterous robotics tasks, poses significant challenges due to limited data coverage and scarce action labels. As an endeavor towards this end, we introduce DreamDojo, a foundation world model that learns diverse interactions and dexterous controls from 44k hours of egocentric human videos. Our data mixture represents the largest video dataset to date for world model pretraining, spanning a wide range of daily scenarios with diverse objects and skills. To address the scarcity of action labels, we introduce continuous latent actions as unified proxy actions, enhancing interaction knowledge transfer from unlabeled videos. After post-training on small-scale target robot data, DreamDojo demonstrates a strong understanding of physics and precise action controllability. We also devise a distillation pipeline that accelerates DreamDojo to a real-time speed of 10.81 FPS and further improves context consistency. Our work enables several important applications based on generative world models, including live teleoperation, policy evaluation, and model-based planning. Systematic evaluation on multiple challenging out-of-distribution (OOD) benchmarks verifies the significance of our method for simulating open-world, contact-rich tasks, paving the way for general-purpose robot world models.

Key Takeaways

  • 1

    Utilizes large-scale human videos for robot pretraining.

  • 2

    Introduces continuous latent actions for effective transfer.

  • 3

    Achieves real-time prediction with high visual quality.

Limitations

  • Requires extensive human video data for pretraining.

  • Dependent on the quality and diversity of human videos.

Keywords

world modelegocentric videoscontinuous latent actionsaction labelsdistillation pipelinereal-time speedteleoperationpolicy evaluationmodel-based planningout-of-distribution benchmarks

More in Multimodal AI

View all
DreamDojo: A Generalist Robot World Model from Large-Scale Human Videos | Paperchime