Close Menu
    Trending
    • Generative AI improves a wireless vision system that sees through obstructions | MIT News
    • A better method for identifying overconfident large language models | MIT News
    • DataRobot + Nebius: An enterprise-ready AI Factory optimized for agents
    • Two-Stage Hurdle Models: Predicting Zero-Inflated Outcomes
    • The New Experience of Coding with AI
    • Why You Should Stop Worrying About AI Taking Data Science Jobs
    • Why Chatbots Are Coming for Your Medical Records
    • One Model to Rule Them All? SAP-RPT-1 and the Future of Tabular Foundation Models
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » A better method for identifying overconfident large language models | MIT News
    Artificial Intelligence

    A better method for identifying overconfident large language models | MIT News

    ProfitlyAIBy ProfitlyAIMarch 19, 2026No Comments5 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Giant language fashions (LLMs) can generate credible however inaccurate responses, so researchers have developed uncertainty quantification strategies to verify the reliability of predictions. One in style technique includes submitting the identical immediate a number of instances to see if the mannequin generates the identical reply.

    However this technique measures self-confidence, and even essentially the most spectacular LLM could be confidently mistaken. Overconfidence can mislead customers concerning the accuracy of a prediction, which could lead to devastating penalties in high-stakes settings like well being care or finance.   

    To deal with this shortcoming, MIT researchers launched a brand new technique for measuring a unique kind of uncertainty that extra reliably identifies assured however incorrect LLM responses.

    Their technique includes evaluating a goal mannequin’s response to responses from a bunch of comparable LLMs. They discovered that measuring cross-model disagreement extra precisely captures this sort of uncertainty than conventional approaches.

    They mixed their method with a measure of LLM self-consistency to create a complete uncertainty metric, and evaluated it on 10 lifelike duties, resembling question-answering and math reasoning. This complete uncertainty metric constantly outperformed different measures and was higher at figuring out unreliable predictions.

    “Self-consistency is being utilized in a variety of totally different approaches for uncertainty quantification, but when your estimate of uncertainty solely depends on a single mannequin’s consequence, it isn’t essentially trustable. We went again to the start to grasp the restrictions of present approaches and used these as a place to begin to design a complementary technique that may empirically enhance the outcomes,” says Kimia Hamidieh, {an electrical} engineering and pc science (EECS) graduate scholar at MIT and lead writer of a paper on this technique.

    She is joined on the paper by Veronika Thost, a analysis scientist on the MIT-IBM Watson AI Lab; Walter Gerych, a former MIT postdoc who’s now an assistant professor at Worcester Polytechnic Institute; Mikhail Yurochkin, a employees analysis scientist on the MIT-IBM Watson AI Lab; and senior writer Marzyeh Ghassemi, an affiliate professor in EECS and a member of the Institute of Medical Engineering Sciences and the Laboratory for Info and Resolution Methods.

    Understanding overconfidence

    Many in style strategies for uncertainty quantification contain asking a mannequin for a confidence rating or testing the consistency of its responses to the identical immediate. These strategies estimate aleatoric uncertainty, or how internally assured a mannequin is in its personal prediction.

    Nonetheless, LLMs will be assured when they’re utterly mistaken. Analysis has proven that epistemic uncertainty, or uncertainty about whether or not one is utilizing the best mannequin, could be a higher strategy to assess true uncertainty when a mannequin is overconfident.

    The MIT researchers estimate epistemic uncertainty by measuring disagreement throughout the same group of LLMs.    

    “If I ask ChatGPT the identical query a number of instances and it offers me the identical reply again and again, that doesn’t imply the reply is essentially right. If I change to Claude or Gemini and ask them the identical query, and I get a unique reply, that’s going to offer me a way of the epistemic uncertainty,” Hamidieh explains.

    Epistemic uncertainty makes an attempt to seize how far a goal mannequin diverges from the perfect mannequin for that job. However since it’s unattainable to construct an excellent mannequin, researchers use surrogates or approximations that usually depend on defective assumptions.

    To enhance uncertainty quantification, the MIT researchers wanted a extra correct strategy to estimate epistemic uncertainty.

    An ensemble method

    The tactic they developed includes measuring the divergence between the goal mannequin and a small ensemble of fashions with related measurement and structure. They discovered that evaluating semantic similarity, or how intently the meanings of the responses match, might present a greater estimate of epistemic uncertainty.

    To realize essentially the most correct estimate, the researchers wanted a set of LLMs that lined numerous responses, weren’t too just like the goal mannequin, and had been weighted primarily based on credibility.

    “We discovered that the best strategy to fulfill all these properties is to take fashions which are educated by totally different firms. We tried many various approaches that had been extra advanced, however this quite simple method ended up working greatest,” Hamidieh says.

    As soon as that they had developed this technique for estimating epistemic uncertainty, they mixed it with an ordinary method that measures aleatoric uncertainty. This complete uncertainty metric (TU) provided essentially the most correct reflection of whether or not a mannequin’s confidence stage is reliable.

    “Uncertainty is dependent upon the uncertainty of the given immediate in addition to how shut our mannequin is to the optimum mannequin. Because of this summing up these two uncertainty metrics goes to offer us the most effective estimate,” Hamidieh says.

    TU might extra successfully establish conditions the place an LLM is hallucinating, since epistemic uncertainty can flag confidently mistaken outputs that aleatoric uncertainty may miss. It might additionally allow researchers to bolster an LLM’s confidently right solutions throughout coaching, which can enhance efficiency.

    They examined TU utilizing a number of LLMs on 10 frequent duties, resembling question-answering, summarization, translation, and math reasoning. Their technique extra successfully recognized unreliable predictions than both measure by itself.

    Measuring complete uncertainty usually required fewer queries than calculating aleatoric uncertainty, which might cut back computational prices and save vitality.

    Their experiments additionally revealed that epistemic uncertainty is only on duties with a novel right reply, like factual question-answering, however might underperform on extra open-ended duties.

    Sooner or later, the researchers might adapt their approach to enhance its efficiency on open-ended queries. They could additionally construct on this work by exploring different types of aleatoric uncertainty.

    This work is funded, partially, by the MIT-IBM Watson AI Lab.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleDataRobot + Nebius: An enterprise-ready AI Factory optimized for agents
    Next Article Generative AI improves a wireless vision system that sees through obstructions | MIT News
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    Generative AI improves a wireless vision system that sees through obstructions | MIT News

    March 19, 2026
    Artificial Intelligence

    Two-Stage Hurdle Models: Predicting Zero-Inflated Outcomes

    March 18, 2026
    Artificial Intelligence

    The New Experience of Coding with AI

    March 18, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Let’s Analyze OpenAI’s Claims About ChatGPT Energy Use

    June 16, 2025

    [The AI Show Episode 156]: AI Answers

    June 26, 2025

    The brain power behind sustainable AI | MIT News

    October 24, 2025

    Optimizing PyTorch Model Inference on AWS Graviton

    December 10, 2025

    Gamers Nexus avslöjar omfattande GPU-smugglingsimperium från Kina

    August 19, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    Why AI is the New Social Media: A Shift from Connection to Personalization

    December 5, 2025

    Critical Mistakes Companies Make When Integrating AI/ML into Their Processes

    November 14, 2025

    Preventing Context Overload: Controlled Neo4j MCP Cypher Responses for LLMs

    September 7, 2025
    Our Picks

    Generative AI improves a wireless vision system that sees through obstructions | MIT News

    March 19, 2026

    A better method for identifying overconfident large language models | MIT News

    March 19, 2026

    DataRobot + Nebius: An enterprise-ready AI Factory optimized for agents

    March 18, 2026
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.