← Back to stories Close-up of a stylish mesh dress adorned with rhinestones, showcasing intricate detail and fashion elegance.
Photo by PREM GS on Pexels
凤凰科技 2026-04-11

谷歌AI凭什么赢了OpenAI?

Demo that stops the scroll

Ailing Zeng of the Anuttacon技术团队 (Anuttacon technical team) has posted a striking demo of LPM 1.0 on X, and the results underscore how fast real‑time video role generation is advancing. In seconds the model renders an on‑screen character cycling through a battery of micro‑expressions—hesitation, a tightened mouth, a sharp intake of breath—then reacts in real time when the human speaker on the other side talks. The footage is unnerving in its specificity: eye shifts, subtle eyebrow tension, altered mouth aperture on stressed syllables. Reportedly, the demo shows both listening behavior (silent, attentive reactions) and responsive speaking, making the interaction look much closer to a genuine video call than to a stitched‑together clip.

How it works

It has been reported that LPM 1.0 is trained on a person‑centric multimodal dataset augmented with performance understanding and identity awareness to preserve long‑term consistency. The team reportedly uses a 17‑billion‑parameter diffusion Transformer to model spatial–temporal relations—lip sync, breathing rhythm, and layered emotional expression—while a causal flow generator provides low‑latency, effectively unbounded real‑time streaming. The combination aims to let a character both "listen" (react while another party speaks) and then produce coherent, identity‑consistent speech videos even over long interactions—examples cited by the team include generated clips of 22 and 48 minutes with sustained consistency.

What this means for the AI race

Why mention this when headlines ask whether Google is ahead of OpenAI? Because breakthroughs like LPM 1.0 show that the competitive landscape is broader than two labs and that real‑time, multimodal interaction is now a critical battleground. Large‑scale architectures and clever streaming generators can produce user‑facing interactivity that changes product roadmaps: conversational agents that look like real people, live tutoring avatars, and new video‑first interfaces. Who will commercialize these capabilities most responsibly and at scale? That remains an open question.

Risks and governance

Realistic, low‑latency video agents raise immediate policy concerns. It has been reported that regulators and industry observers are increasingly focused on deepfake risks, misuse, and export‑control implications for advanced AI hardware—issues that shape which labs can scale these systems and how they are deployed across borders. The demos are technically impressive, but deployment will hinge as much on governance, compute access and trust frameworks as on raw model quality.

AI
View original source →