Most people know that robots no longer sound like tinny trash cans. They sound like Siri, Alexa, and Gemini. They sound like the voices in labyrinthine customer support phone trees. And even those robot voices are being made obsolete by new AI-generated voices that can mimic every vocal nuance and tic of human speech, down to specific regional accents. And with just a few seconds of audio, AI can now clone someone’s specific voice.

  • MHLoppy@fedia.io
    link
    fedilink
    arrow-up
    1
    ·
    5 days ago

    I can get behind the general idea, but in this implementation specifically it seems like the low modulation example isn’t distinct enough from simply lower-quality audio, but the higher modulation example (where the effect is more distinct as an intentional effect), is just not nice to listen to. Maybe there are other ways to distort the voice that don’t have as much of that downside?