Odyssey-2 Pro
Odyssey-2 Pro is a frontier general-purpose world model that generates continuous, interactive simulations you can integrate into products via the Odyssey API, marking a pivotal moment for world models similar to GPT-2 in language. It’s trained on large amounts of video and interaction data to learn how the world evolves frame-by-frame and outputs minutes-long simulations that can be interacted with in real time, not fixed short clips. Odyssey-2 Pro delivers improved physics, richer dynamics, more authentic behaviors, and sharper visuals by streaming 720p video at up to ~22 FPS that responds instantly to prompts and actions, and it supports embedding interactive streams, viewable streams, and parameterized simulations into applications with simple SDKs in JavaScript and Python. Developers can integrate the model with under ten lines of code to create open-ended, interactive video experiences where users’ inputs shape evolving scenes.
Learn more
Mirage 2
Mirage 2 is an AI-driven Generative World Engine that lets anyone instantly transform images or descriptions into fully playable, interactive game environments directly in the browser. Upload sketches, concept art, photos, or prompts, like “Ghibli-style village” or “Paris street scene”, and Mirage 2 builds immersive worlds you can explore in real time. The experience isn’t pre-scripted: you can modify your world mid-play using natural-language chat, evolving settings dynamically, from a cyberpunk city to a rainforest or a mountaintop castle, all with minimal latency (around 200 ms) on a single consumer GPU. Mirage 2 supports smooth rendering, real-time prompt control, and extended gameplay stretches beyond ten minutes. It outpaces earlier world-model systems by offering true general-domain generation, no upper limit on styles or genres, as well as seamless world adaptation and sharing features.
Learn more
Odyssey
Odyssey is a frontier interactive video model that enables instant, real-time generation of video you can interact with. Just type a prompt, and the system begins streaming minutes of video that respond to your input. It shifts video from a static playback format to a dynamic, action-aware stream: the model is causal and autoregressive, generating each frame based solely on prior frames and your actions rather than a fixed timeline, enabling continuous adaptation of camera angles, scenery, characters, and events. The platform begins streaming video almost instantly, producing new frames every ~50 milliseconds (about 20 fps), so you don’t wait minutes for a clip, you engage in an evolving experience. Under the hood, the model is trained via a novel multi-stage pipeline to transition from fixed-clip generation to open-ended interactive video, allowing you to type or speak commands and explore an AI-imagined world that reacts in real time.
Learn more
NVIDIA Cosmos
NVIDIA Cosmos is a developer-first platform of state-of-the-art generative World Foundation Models (WFMs), advanced video tokenizers, guardrails, and an accelerated data processing and curation pipeline designed to supercharge physical AI development. It enables developers working on autonomous vehicles, robotics, and video analytics AI agents to generate photorealistic, physics-aware synthetic video data, trained on an immense dataset including 20 million hours of real-world and simulated video, to rapidly simulate future scenarios, train world models, and fine‑tune custom behaviors. It includes three core WFM types; Cosmos Predict, capable of generating up to 30 seconds of continuous video from multimodal inputs; Cosmos Transfer, which adapts simulations across environments and lighting for versatile domain augmentation; and Cosmos Reason, a vision-language model that applies structured reasoning to interpret spatial-temporal data for planning and decision-making.
Learn more