Marble
Marble is an experimental AI model internally tested by World Labs, a variant and extension of their Large World Model technology. It is a web service that turns a single 2D image into a navigable spatial environment. Marble offers two generation modes: a smaller, fast model for rough previews that’s quick to iterate on, and a larger, high-fidelity model that takes longer (around ten minutes in the example) but produces a significantly more convincing result. The value proposition is instant, photogrammetry-like image-to-world creation without a full capture rig, turning a single shot into an explorable space for memory capture, mood boards, archviz previews, or creative experiments.
Learn more
Odyssey-2 Pro
Odyssey-2 Pro is a frontier general-purpose world model that generates continuous, interactive simulations you can integrate into products via the Odyssey API, marking a pivotal moment for world models similar to GPT-2 in language. It’s trained on large amounts of video and interaction data to learn how the world evolves frame-by-frame and outputs minutes-long simulations that can be interacted with in real time, not fixed short clips. Odyssey-2 Pro delivers improved physics, richer dynamics, more authentic behaviors, and sharper visuals by streaming 720p video at up to ~22 FPS that responds instantly to prompts and actions, and it supports embedding interactive streams, viewable streams, and parameterized simulations into applications with simple SDKs in JavaScript and Python. Developers can integrate the model with under ten lines of code to create open-ended, interactive video experiences where users’ inputs shape evolving scenes.
Learn more
Mirage 2
Mirage 2 is an AI-driven Generative World Engine that lets anyone instantly transform images or descriptions into fully playable, interactive game environments directly in the browser. Upload sketches, concept art, photos, or prompts, like “Ghibli-style village” or “Paris street scene”, and Mirage 2 builds immersive worlds you can explore in real time. The experience isn’t pre-scripted: you can modify your world mid-play using natural-language chat, evolving settings dynamically, from a cyberpunk city to a rainforest or a mountaintop castle, all with minimal latency (around 200 ms) on a single consumer GPU. Mirage 2 supports smooth rendering, real-time prompt control, and extended gameplay stretches beyond ten minutes. It outpaces earlier world-model systems by offering true general-domain generation, no upper limit on styles or genres, as well as seamless world adaptation and sharing features.
Learn more
Genie 3
Genie 3 is DeepMind’s next-generation, general-purpose world model capable of generating richly interactive 3D environments in real time at 24 frames per second and 720p resolution that remain consistent for several minutes. Prompted by text input, the system constructs dynamic virtual worlds where users (or embodied agents) can navigate and interact with natural phenomena from multiple perspectives, like first-person or isometric. A standout feature is its emergent long-horizon visual memory: Genie 3 maintains environmental consistency over extended durations, preserving off-screen elements and spatial coherence across revisits. It also supports “promptable world events,” enabling users to modify scenes, such as changing weather or introducing new objects, on the fly. Designed to support embodied agent research, Genie 3 seamlessly integrates with agents like SIMA, facilitating goal-based navigation and complex task accomplishment.
Learn more