Runway claims its latest text-to-video model generates even more accurate visuals than its last. In a blog post on Monday, Runway says its Gen-4.5 model can produce “cinematic and highly realistic outputs,” potentially making it even more difficult to distinguish between what’s real and what’s AI.
“Gen-4.5 achieves unprecedented physical accuracy and visual precision,” Runway’s announcement says. It adds that the new AI model is better at adhering to prompts, allowing it to produce detailed scenes without compromising video quality. Runway says that AI-generated objects “move with realistic weight, momentum and force,” while liquids “flow with proper dynamics.”
The Gen-4.5 model is rolling out to all users gradually and will offer the same speed and efficiency as its predecessor, according to Runway. There are still some limitations, though, as the model may experience issues with object permanence and causal reasoning, meaning effects may happen before the cause, such as a door opening before someone uses a handle.
Along with Runway, OpenAI is ramping up efforts to make its AI-generated videos look more lifelike. OpenAI highlighted upgrades to physics with the release of its Sora 2 text-to-video model in September, with Sora head Bill Peebles saying, “You can accurately do backflips on top of a paddleboard on a body of water, and all of the fluid dynamics and buoyancy are accurately modeled.”
Runway says its Gen-4.5 model is better at handling different visual styles, too, allowing it to produce more consistent photorealistic, stylized, and cinematic visuals. The startup claims that photorealistic visuals created with Gen-4.5 can be “indistinguishable from real-world footage with lifelike detail and accuracy.”
