Close Menu
    Trending
    • From Equal Weights to Smart Weights: OTPO’s Approach to Better LLM Alignment
    • The Future of AI Agent Communication with ACP
    • Vad världen har frågat ChatGPT under 2025
    • Google’s generative video model Veo 3 has a subtitles problem
    • MedGemma – Nya AI-modeller för hälso och sjukvård
    • AI text-to-speech programs could “unlearn” how to imitate certain people
    • AI’s giants want to take over the classroom
    • What Can the History of Data Tell Us About the Future of AI?
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » AI text-to-speech programs could “unlearn” how to imitate certain people
    AI Technology

    AI text-to-speech programs could “unlearn” how to imitate certain people

    ProfitlyAIBy ProfitlyAIJuly 15, 2025No Comments5 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    AI firms usually hold a good grip on their fashions to discourage misuse. For instance, if you happen to ask ChatGPT to offer you somebody’s cellphone quantity or directions for doing one thing unlawful, it would doubtless simply let you know it can’t assist. Nonetheless, as many examples over time have proven, intelligent immediate engineering or mannequin fine-tuning can generally get these fashions to say issues they in any other case wouldn’t. The undesirable data should be hiding someplace contained in the mannequin in order that it may be accessed with the precise methods. 

    At current, firms are inclined to cope with this problem by making use of guardrails; the thought is to examine whether or not the prompts or the AI’s responses include disallowed materials. Machine unlearning as an alternative asks whether or not an AI might be made to neglect a chunk of knowledge that the corporate doesn’t need it to know. The method takes a leaky mannequin and the particular coaching knowledge to be redacted and makes use of them to create a brand new mannequin—primarily, a model of the unique that by no means discovered that piece of knowledge. Whereas machine unlearning has ties to older methods in AI analysis, it’s solely previously couple of years that it’s been utilized to giant language fashions.

    Jinju Kim, a grasp’s pupil at Sungkyunkwan College who labored on the paper with Ko and others, sees guardrails as fences across the unhealthy knowledge put in place to maintain individuals away from it. “You’ll be able to’t get by means of the fence, however some individuals will nonetheless attempt to go underneath the fence or over the fence,” says Kim. However unlearning, she says, makes an attempt to take away the unhealthy knowledge altogether, so there may be nothing behind the fence in any respect. 

    The way in which present text-to-speech techniques are designed complicates this just a little extra, although. These so-called “zero-shot” fashions use examples of individuals’s speech to be taught to re-create any voice, together with these not within the coaching set—with sufficient knowledge, it may be a superb mimic when provided with even a small pattern of somebody’s voice. So “unlearning” means a mannequin not solely must “neglect” voices it was educated on but additionally has to be taught to not mimic particular voices it wasn’t educated on. All of the whereas, it nonetheless must carry out properly for different voices. 

    To reveal learn how to get these outcomes, Kim taught a recreation of VoiceBox, a speech era mannequin from Meta, that when it was prompted to provide a textual content pattern in one of many voices to be redacted, it ought to as an alternative reply with a random voice. To make these voices practical, the mannequin “teaches” itself utilizing random voices of its personal creation. 

    In accordance with the crew’s results, that are to be introduced this week on the Worldwide Convention on Machine Studying, prompting the mannequin to mimic a voice it has “unlearned” offers again a end result that—in line with state-of-the-art tools that measure voice similarity—mimics the forgotten voice greater than 75% much less successfully than the mannequin did earlier than. In observe, this makes the brand new voice unmistakably totally different. However the forgetfulness comes at a price: The mannequin is about 2.8% worse at mimicking permitted voices. Whereas these percentages are a bit arduous to interpret, the demo the researchers launched online presents very convincing outcomes, each for a way properly redacted audio system are forgotten and the way properly the remaining are remembered. A pattern from the demo is given beneath. 

    A voice pattern of a speaker to be forgotten by the mannequin.
    The generated text-to-speech audio from the unique mannequin utilizing the above as a immediate.
    The generated text-to-speech audio utilizing the identical immediate, however now from the mannequin the place the speaker was forgotten.

    Ko says the unlearning course of can take “a number of days,” relying on what number of audio system the researchers need the mannequin to neglect. Their methodology additionally requires an audio clip about 5 minutes lengthy for every speaker whose voice is to be forgotten.

    In machine unlearning, items of knowledge are sometimes changed with randomness in order that they’ll’t be reverse-engineered again to the unique. On this paper, the randomness for the forgotten audio system may be very excessive—an indication, the authors declare, that they’re actually forgotten by the mannequin. 

     “I’ve seen individuals optimizing for randomness in different contexts,” says Vaidehi Patil, a PhD pupil on the College of North Carolina at Chapel Hill who researches machine unlearning. “This is among the first works I’ve seen for speech.” Patil is organizing a machine unlearning workshop affiliated with the convention, and the voice unlearning analysis will even be introduced there. 



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleAI’s giants want to take over the classroom
    Next Article MedGemma – Nya AI-modeller för hälso och sjukvård
    ProfitlyAI
    • Website

    Related Posts

    AI Technology

    Google’s generative video model Veo 3 has a subtitles problem

    July 15, 2025
    AI Technology

    AI’s giants want to take over the classroom

    July 15, 2025
    AI Technology

    The three-layer AI strategy for supply chains

    July 10, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Baidu släpper ERNIE 4.5 som öppen källkod

    June 30, 2025

    Creating a common language | MIT News

    April 5, 2025

    The Complete Guide to NetSuite SuiteScript

    April 4, 2025

    10,000x Faster Bayesian Inference: Multi-GPU SVI vs. Traditional MCMC

    June 11, 2025

    Stop Building AI Platforms | Towards Data Science

    June 14, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    Explained: How Does L1 Regularization Perform Feature Selection?

    April 23, 2025

    Reducing Time to Value for Data Science Projects: Part 3

    July 10, 2025

    Chain-of-Thought Prompting: Everything You Need to Know About It

    April 5, 2025
    Our Picks

    From Equal Weights to Smart Weights: OTPO’s Approach to Better LLM Alignment

    July 15, 2025

    The Future of AI Agent Communication with ACP

    July 15, 2025

    Vad världen har frågat ChatGPT under 2025

    July 15, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.