← Back to stories A black smart speaker resting on a light-colored wooden table in a cozy indoor setting.
Photo by Fabian Hurnaus on Pexels
虎嗅 2026-03-10

Crossing the "Valley of Horror" Trap: Making AI Voice Truly "Touch the Heart"

A shift in the race: from mimicry to experience

The AI voice race in China is no longer just about technical realism. Firms such as Baidu (百度) and iFlytek (科大讯飞), alongside consumer brands like NIO (蔚来), are discovering that closer imitation of human speech can backfire. But how realistic is too realistic? The key argument: chasing the last percentage points of human likeness risks triggering an auditory "uncanny valley" — described here as the "valley of horror" — where user affinity plunges even as development costs skyrocket.

Where synthetic voice matters

AI-generated speech is already penetrating customer service, in-car assistants, and content creation. Companies use bespoke tones as part of a "sonic identity" to convey brand personality across phones, apps and devices. In publishing and media, neural text-to-speech (TTS) has lowered production costs and enabled scalable audiobooks and voiceovers; it has been reported that platforms have even used AI to restore or "revive" actors' and performers' voices for films and archives, a practice that raises both creative opportunity and ethical questions.

The auditory uncanny valley and business risk

Research and platform data suggest a consistent pattern: as TTS moves from robotic to near-human, listener ratings rise — until a subtle threshold where likeness becomes discomforting and ratings collapse, only to recover if the voice becomes virtually indistinguishable from a human. Why? Two mechanisms: small deviations in near-human voices are amplified by listeners, and identity ambiguity creates cognitive dissonance during long listening sessions. The result for businesses can be heavy R&D spend with little or negative return, and noisy but negative engagement metrics that masquerade as popularity.

Strategy and geopolitics: design, not just fidelity

The remedy is strategic: shift from "technical mimicry" to "experience adaptation." Prioritise warmth, contextual fit and long-duration comfort over pixel-perfect realism; design voices to match user expectations and scenario (service bots vs. storytelling), and treat sonic identity as a cross-platform product asset. Broader constraints also matter. It has been reported that export controls on advanced chips and tightening data rules have raised the cost and complexity of training large generative speech models, a geopolitical factor that will shape which players can afford to pursue extreme realism. In short: success will favour teams that turn sound into emotion and identity, not simply pursuit of the last few decibels of fidelity.

AI
View original source →