Comment by palsecam

1 day ago

> My pet theory is the BigCo's are walking a tightrope of model safety and are intentionally incorporating some uncanny valley into their products, since if people really knew that AI could "talk like Pete" they would get uneasy. The cognitive dissonance doesn't kick in when a bot talks like a drone from HR instead of a real person.

FTR, Bruce Schneier (famed cryptologist) is advocating for such an approach:

We have a simple proposal: all talking AIs and robots should use a ring modulator. In the mid-twentieth century, before it was easy to create actual robotic-sounding speech synthetically, ring modulators were used to make actors’ voices sound robotic. Over the last few decades, we have become accustomed to robotic voices, simply because text-to-speech systems were good enough to produce intelligible speech that was not human-like in its sound. Now we can use that same technology to make robotic speech that is indistinguishable from human sound robotic again.https://www.schneier.com/blog/archives/2025/02/ais-and-robot...

Reminds me of the robot voice from The Incredibles[1]. It had an obviously-robotic cadence where it would pause between every word. Text-to-speech at the time already knew how to make words flow into each other, but I thought the voice from The Incredibles sounded much nicer than the contemporaneous text-to-speech bots, while also still sounding robotic.

[1] https://www.youtube.com/watch?v=_dxV4BvyV2w

That doesn't sound like ring modulation in a musical sense (IIRC it has a modulator above 30 Hz, or inverts the signal instead of attenuating?), so much as crackling, cutting in and out, or an overdone tremolo effect. I checked in Audacity and the signal only gets cut out, not inverted.