
Nvidia unveils DreamDojo — a robot world model trained on 44,000 hours of human video
Read Our Expert Analysis
Create an account or login for free to unlock our expert analysis and key takeaways for this development.
By continuing, you agree to receive marketing communications and our weekly newsletter. You can opt-out at any time.
Recommended for you

ABB accelerates robot training with NVIDIA simulation libraries
ABB and NVIDIA are integrating high-fidelity simulation to tighten robot behavior between digital training and factory floors, with Foxconn piloting camera-guided assembly and a planned product launch in H2 2026. The move sits inside a broader industry shift — Alphabet’s Intrinsic is also piloting Foxconn collaborations but emphasizes continuous, field-driven adaptation — highlighting two competing strategies for production-ready robotics.

Alibaba, ByteDance and Kuaishou Unveil Next-Gen Robotics and Video AI
Chinese technology leaders released distinct AI models this week: Alibaba introduced a robotics-focused model for real-world object interaction, ByteDance launched an improved text-to-video generator, and Kuaishou rolled out a paywalled video model with longer outputs. These releases sharpen competition with Western labs on robotics, video synthesis, and agentic capabilities while raising consent and commercialisation questions.

Alibaba pushes robotics forward with open-source RynnBrain foundation model
Alibaba’s DAMO Academy released RynnBrain, an open-source foundation model that links spatial-temporal perception to task sequencing for embodied robots. The move aims to speed real-world deployments by lowering custom engineering needs, though success will hinge on compute costs, transferability across hardware and rigorous safety validation.

DeepMind opens Project Genie to U.S. Google AI Ultra users, seeks real-world feedback on interactive world models
DeepMind has opened a constrained preview of Project Genie to U.S. Google AI Ultra subscribers to collect hands-on feedback for its Genie 3-powered world model. The prototype generates short, explorable virtual environments from text or images but is limited by compute, safety guardrails, and nascent interactivity.
World Models: AMI Labs, World Labs, DeepMind Recast Physical AI
Two >$1B financings and a flurry of strategic partnerships have redirected venture capital toward physically grounded world models; AMI Labs (led scientifically by Yann LeCun) and World Labs (led by Fei‑Fei Li, with an Autodesk commitment) exemplify divergent go‑to‑market paths—industrial pilots versus media/design integrations—that together reprice risks and supplier leverage across robotics, autonomy and spatial computing.

Nvidia mobilizes $26B to launch open-weight model program
Nvidia plans a multi-year, $26 billion program to develop and publish open-weight models, and concurrently released Nemotron 3 Super , a 128‑billion‑parameter model. The move tightens hardware-model coupling, amplifies demand for Nvidia systems, and reshapes competitive dynamics between US cloud providers and open-weight ecosystems.

NVIDIA unveils Nemotron 3 Super for enterprise agents
NVIDIA released Nemotron 3 Super, a reasoning‑first model aimed at sustained, multi‑step enterprise agents and published with open weights, datasets and recipes to enable on‑prem deployment and fine‑tuning. Public reports differ on headline parameters (the company and some outlets cite ~120B while other engineering notes and press accounts describe ~128B), but all sources confirm a runtime sparsity mode (reported as ~12B active parameters) plus a wider program and hardware roadmap—NemoClaw, NVL72/Rubin racks and privileged partner access—that together reshape procurement and vendor leverage for enterprise agent stacks.
Nvidia Nemotron-Cascade 2: Post‑Training Playbook Upsets Size Orthodoxy
Nvidia’s Nemotron-Cascade 2 uses a sequential post-training recipe to deliver top-tier math and coding performance while activating only 3B parameters at inference. The Cascade RL pipeline plus MOPD token-level distillation signals a shift toward intelligence-density strategies that cut serving cost and raise the value of training orchestration. Public materials across the Nemotron family sometimes report divergent headline sizes, a difference that likely reflects measurement conventions rather than an architectural contradiction.