← Back to stories Researchers working with advanced robotics technology in a laboratory setting.
Photo by Pavel Danilyuk on Pexels
虎嗅 2026-03-19

Are robots abandoning language? GTC 2026 offers the real answer

The headline shift: language steps back, control moves forward

It has been reported that NVIDIA (英伟达) used GTC 2026 to showcase a suite of robot-focused updates — a refreshed Isaac Platform, multimodal foundation models and tighter sim-to-real training loops — illustrated on stage by Disney’s Olaf (雪宝) interacting live. The demos point to a clear engineering trend: robots are reducing reliance on explicit intermediate representations such as step‑by‑step language instructions or hand‑rolled “future imagination” rolls, and instead letting perception feed policy more directly. Language is not gone; it is being re‑positioned as a training-time supervisory or high‑level constraint rather than a per‑step control signal.

From VLA and world models to implicit policies

For years the VLA (vision‑language‑action) paradigm and world‑action models (WAMs) used explicit middle layers — semantic tokens, decomposition, or imagined future state sequences — to make decision chains human‑interpretable. Recent work, including a paper from Tsinghua University’s Institute for Interdisciplinary Information (清华大学交叉信息研究院) with Galaxea AI titled Fast‑WAM, questions whether those inference‑time steps are necessary. The paper reportedly cuts inference delay to about 190 ms and claims roughly fourfold speedups versus approaches that perform explicit future imagination, showing that when models learn rich dynamical representations during training, they can skip explicit intermediate generation at run time without losing performance.

Why this is happening now — data, simulation and compute

The shift is driven by three converging conditions: vastly improved simulation and data‑generation (NVIDIA’s Isaac Sim and large-scale behavioral datasets), architectures that handle temporal signals directly (temporal Transformers, diffusion‑style policies), and more available compute at both GPU and edge levels. These elements let a unified representation map raw perception to continuous control internally, avoiding noisy symbol conversions and high‑latency cross‑module interfaces. That said, geopolitics matters: export controls, trade policy and U.S.–China tech tensions shape which chips and software stacks are accessible in different markets, and thus influence how fast these end‑to‑end approaches can be fielded worldwide.

What it means for robotics — faster, less interpretable, still human‑guided

The practical payoff is faster, more responsive robots that better match the closed‑loop timing of physical control. The trade‑off is interpretability: systems become more opaque because they no longer “think aloud” in human‑readable language or explicit imagined futures. But language remains valuable — just in different roles, as training supervision or high‑level constraints rather than a step‑by‑step controller. So are robots abandoning language or simply moving it out of the tight control loop? The answer from GTC 2026 is: they’re doing the latter — keeping language for people, and letting implicit models handle the mechanics.

AI
View original source →