Runway announces GWM-1 family of world models

AI firm Runway has unveiled GWM-1, its initial set of world models designed to extend beyond video generation into areas like robotics and avatars. Built on the Gen-4.5 text-to-video model, these three specialized autoregression models enable real-time simulations, synthetic data creation, and natural human-like interactions. The launch highlights Runway's push into a competitive field dominated by tech giants.

Runway, known for its video generation tools, introduced GWM-1 as a trio of models post-trained on domain-specific data from its Gen-4.5 foundation. This move signals the company's expansion from creative industries into broader AI applications.

The first, GWM Worlds, provides an interface for exploring digital environments with real-time user inputs influencing frame generation. Users can specify world elements, appearances, physics rules, and actions such as camera movements or environmental changes, maintaining consistency over extended sequences. Potential uses include pre-visualization in game development, virtual reality setups, and educational simulations of historical sites. It also supports training AI agents, including those for robotics.

GWM Robotics focuses on producing synthetic training data to enhance robotics datasets, incorporating novel objects, task instructions, and environmental variations. This aids in simulating challenging real-world conditions like varying weather and allows safer, cost-effective policy testing in virtual settings before physical trials. Runway offers a Python SDK for its robotics API on a per-request basis.

GWM Avatars integrates video and speech generation to create avatars that move and emote naturally during speaking and listening, sustaining long conversations without quality loss. It will soon integrate into Runway's web app and API.

While aiming for more unified models across domains, Runway's current versions are distinct. CEO Cristóbal Valenzuela described GWM-1 on X as "a major step toward universal simulation." The company enters a crowded space with players like Google and Nvidia, targeting robotics, physics, and life sciences alongside film and games.

Additionally, Runway revealed Gen-4.5 updates with native audio, audio editing, and multi-shot video capabilities, plus a partnership with CoreWeave for Nvidia's GB300 NVL72 racks to support future AI training and inference.

Awọn iroyin ti o ni ibatan

Photo illustration of Google executives unveiling the Gemini 3 AI model and Antigravity IDE in a conference setting.
Àwòrán tí AI ṣe

Google unveils Gemini 3 AI model and Antigravity IDE

Ti AI ṣe iroyin Àwòrán tí AI ṣe

Google has released Gemini 3 Pro, its latest flagship AI model, emphasizing improved reasoning, visual outputs, and coding capabilities. The company also introduced Antigravity, an AI-first integrated development environment. Both are available in limited preview starting today.

Experts foresee 2026 as the pivotal year for world models, AI systems designed to comprehend the physical world more deeply than large language models. These models aim to ground AI in reality, enabling advancements in robotics and autonomous vehicles. Industry leaders like Yann LeCun and Fei-Fei Li highlight their potential to revolutionize spatial intelligence.

Ti AI ṣe iroyin

Google has announced that its experimental AI prototype, Genie 3, is now available to subscribers of its highest-tier AI plan. The tool allows users to generate and navigate interactive 3D worlds using simple text prompts. Previously limited to trusted testers, this expansion marks a step toward broader access for the 18-and-older audience.

Google has enhanced its AI-powered virtual try-on feature, allowing users to generate a full-body avatar from a simple selfie. The update eliminates the need for full-body photos and supports sizes from XS to 4XL+. This tool is now available in the US, building on the original version launched in May.

Ti AI ṣe iroyin

Google has released Gemini 3.1 Pro, an updated version of its flagship AI model, emphasizing improvements in problem-solving and reasoning. The model is available in preview for developers and consumers starting today. It builds on the Gemini 3 release from November.

Anthropic announced its new Claude Mythos Preview model and Project Glasswing, a consortium involving Apple, Google, and more than 45 other organizations. The initiative aims to test advancing AI cybersecurity capabilities amid growing concerns over powerful models. The formal reveal followed leaked details about the model at the end of March.

Ti AI ṣe iroyin

OpenAI has rolled out an updated image generation model for ChatGPT, making it four times faster and better at following user instructions. The upgrade includes improved editing capabilities and enhanced text rendering. This comes shortly after the release of GPT-5.2 and amid competition from Google's Gemini.

 

 

 

Ojú-ìwé yìí nlo kuki

A nlo kuki fun itupalẹ lati mu ilọsiwaju wa. Ka ìlànà àṣírí wa fun alaye siwaju sii.
Kọ