Close Menu
    Trending
    • From Transactions to Trends: Predict When a Customer Is About to Stop Buying
    • America’s coming war over AI regulation
    • “Dr. Google” had its issues. Can ChatGPT Health do better?
    • Evaluating Multi-Step LLM-Generated Content: Why Customer Journeys Require Structural Metrics
    • Why SaaS Product Management Is the Best Domain for Data-Driven Professionals in 2026
    • Stop Writing Messy Boolean Masks: 10 Elegant Ways to Filter Pandas DataFrames
    • What Other Industries Can Learn from Healthcare’s Knowledge Graphs
    • Everyone wants AI sovereignty. No one can truly have it.
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » Which Hallucinates Less And How To Fix Both » Ofemwire
    AI Tools & Technologies

    Which Hallucinates Less And How To Fix Both » Ofemwire

    ProfitlyAIBy ProfitlyAIJanuary 20, 2026No Comments10 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    You ask your AI a easy query. It provides you a assured reply with particular particulars.

    Then you definitely fact-check it. Every little thing’s improper.

    Welcome to AI hallucinations. The issue that makes you second-guess each response, even when the AI sounds utterly sure.

    GPT-5 and Claude 4 are essentially the most superior language fashions out there. However they nonetheless make issues up.

    The query isn’t in the event that they’ll hallucinate. It’s how typically, why, and what you are able to do about it.

    I’ve spent three months testing each fashions.

    Operating similar prompts. Reality-checking responses.

    Monitoring hallucination charges throughout completely different duties.

    Right here’s what I discovered, which mannequin is extra dependable, and the way to scale back hallucinations in each.

    The Hallucination Actuality Test

    First, let’s be clear about what we’re coping with.

    AI hallucinations aren’t bugs. They’re options of how these fashions work.

    They predict the subsequent phrase based mostly on patterns of their coaching knowledge.

    Generally these predictions sound correct however are utterly improper.

    Each GPT-5 and Claude 4 hallucinate. However they do it in a different way, in numerous conditions, at completely different charges.

    Understanding these variations helps you select the precise mannequin and use it appropriately.

    GPT-5 Hallucination Patterns: What I Discovered

    GPT-5 is quick, artistic, and assured. Generally too assured.

    The place GPT-5 hallucinates most:

    • ‍Particular information and figures: Ask for statistics, dates, or numbers and GPT-5 provides you with exact solutions. Typically improper.

    I requested it for the inhabitants of mid-sized cities. It gave me numbers that have been off by 20-30%. But it surely introduced them with full certainty.‍

    • ‍Latest occasions: Something after its data cutoff is a big gamble. GPT-5 will fill gaps with plausible-sounding info relatively than admitting uncertainty.

    Requested a couple of tech firm’s Q3 2024 earnings. It gave me income figures that sounded affordable. All fabricated.‍

    • ‍Citations and sources: Request sources and GPT-5 will present them. E-book titles. Article names. URLs. Many don’t exist.

    I examined this with tutorial analysis requests. About 40% of the citations it generated have been utterly made up. Actual-sounding titles. Pretend papers.

    • ‍Technical specs: Product specs, API particulars, model options. GPT-5 blends what it is aware of with what sounds proper.

    Requested about particular options in React 19. It listed capabilities that don’t exist but, combined with actual options.

    The place GPT-5 is dependable:

    Additionally Learn: chatgpt prompts for social media

    • ‍Common data: Widespread information, well-documented historical past, widely-known info. Right here, GPT-5 is stable.‍
    • Code patterns: Customary programming options and customary implementations. It’s seen tens of millions of examples.‍
    • Inventive work: When accuracy doesn’t matter, hallucinations don’t harm. Writing fiction? GPT-5 is okay.‍
    • Conceptual explanations: How issues work usually. Rules and ideas relatively than particular information.

    READ ALSO  How to Choose the Perfect AI Tool in 2024

    Hallucination price in my testing:

    Factual questions: 25-35% contained no less than one hallucination

    Technical particulars: 20-25% hallucination price

    Latest occasions: 40-50% hallucination price

    Common data: 10-15% hallucination price

    Claude 4 Hallucination Patterns: What I Discovered

    Claude 4 takes a unique method. It’s extra cautious, extra more likely to specific uncertainty, and usually extra correct on information.

    The place Claude 4 hallucinates most:

    • ‍Obscure info: When coping with area of interest matters or uncommon particulars, Claude typically fills gaps relatively than admitting it doesn’t know.

    I requested a couple of small regional pageant. Claude gave me dates and particulars that sounded particular. Couldn’t confirm any of it.

    • ‍Connecting unrelated information: Claude is sweet at reasoning, however typically makes logical leaps that aren’t supported.

    Requested about correlations in dataset. Claude confidently defined relationships that weren’t truly there.

    • ‍Finishing patterns: Whenever you give it partial info, Claude tries to finish it. Generally these completions are invented.

    Began describing a hypothetical product. Claude added options and specs I by no means talked about, treating them as actual.

    The place Claude 4 is dependable:

    • ‍Factual warning: Claude typically says “I’m not sure” or “Based mostly on my coaching knowledge” relatively than making issues up. That is large.‍
    • Reasoning by way of issues: When Claude reveals its considering course of (prolonged considering), hallucinations drop considerably.‍
    • Admitting limitations: Claude is extra more likely to say “I don’t have details about that” than to manufacture a solution.‍
    • Technical accuracy: For well-documented technical matters, Claude is persistently extra correct than GPT-5.

    Hallucination price in my testing:

    Factual questions: 15-20% contained no less than one hallucination

    Technical particulars: 10-15% hallucination price

    Latest occasions: 25-30% hallucination price

    Common data: 5-10% hallucination price

    Claude 4 hallucinates roughly 40% lower than GPT-5 throughout most classes.

    The Key Distinction: Confidence vs. Warning

    The most important distinction isn’t simply accuracy. It’s how every mannequin handles uncertainty.

    GPT-5 method: At all times provides you a solution. Even when it’s unsure. Confidence over accuracy.

    Claude 4 method: Extra more likely to specific uncertainty or admit gaps. Accuracy over confidence.

    This issues in apply.

    GPT-5 feels extra useful as a result of it by no means says “I don’t know.” However that helpfulness consists of making issues up.

    Claude 4 feels extra trustworthy as a result of it admits limitations.

    However typically you need a solution, even when it’s imperfect.

    Select based mostly in your use case. Want creativity and don’t care about excellent accuracy?

    GPT-5 works. Want information you’ll be able to belief? Claude 4 is safer.

    The way to Scale back GPT-5 Hallucinations

    Listed here are the methods that truly work for GPT-5:

    1. Request Sources and Citations

    Dangerous immediate: “What’s the common wage for knowledge scientists in 2024?”

    Higher immediate: “What’s the common wage for knowledge scientists in 2024? Please cite your sources and observe if you happen to’re unsure about any figures.”

    Whenever you ask for sources, GPT-5 is extra cautious. It gained’t eradicate hallucinations, however it reduces them by about 30% in my testing.

    READ ALSO  How to Create Effective and Affordable AI Content

    2. Use Step-by-Step Reasoning

    Dangerous immediate: “Is that this funding technique sound?”

    Higher immediate: “Analyze this funding technique step-by-step. First, determine the important thing assumptions. Then, consider every assumption. Lastly, give your evaluation.”

    Breaking down reasoning reduces logical leaps and makes hallucinations extra apparent.

    3. Set Conservative Parameters

    Use decrease temperature settings (0.3-0.5) for factual duties. Greater creativity means extra hallucinations.

    Use this within the API or ask ChatGPT to “be conservative and fact-focused” in your immediate.

    4. Confirm Latest Data

    Add to prompts: “If this info is after your data cutoff date, please say so explicitly.”

    This forces GPT-5 to acknowledge when it’s guessing about current occasions.

    5. Request Confidence Ranges

    Add to prompts: “Fee your confidence on this reply from 1-10 and clarify why.”

    GPT-5 will typically price itself decrease on info it’s much less sure about. Not excellent, however useful.

    6. Use Unfavorable Examples

    Add to prompts: “Don’t make up citations, dates, or statistics. In case you’re not sure a couple of particular element, say so.”

    Specific directions to keep away from hallucinations assist. Not utterly, however measurably.

    7. Double-Test Particular Claims

    By no means belief particular numbers, dates, citations, or current occasions with out verification. Interval.

    The way to Scale back Claude 4 Hallucinations

    Claude 4 wants completely different methods as a result of it hallucinates in a different way:

    1. Use Prolonged Pondering

    When out there, use Claude’s prolonged considering mode for advanced queries. The hallucination price drops by 50% when Claude reveals its reasoning.

    Customary immediate: “Clarify this technical idea.”

    Prolonged considering immediate: “Take time to assume by way of this technical idea fastidiously. Present your reasoning course of.”

    2. Ask for Uncertainty Markers

    Add to prompts: “Please mark any statements you’re unsure about with [uncertain] tags.”

    Claude is trustworthy about uncertainty whenever you ask. That is its energy.

    3. Request Reasoning Chains

    Higher immediate: “Clarify your reasoning for this conclusion. What proof helps it? What proof may contradict it?”

    Claude’s hallucinations typically occur in its conclusions, not its reasoning. Make it present each.

    4. Keep away from Main Questions

    Claude typically tries to agree together with your assumptions. Body questions neutrally.

    Dangerous: “This knowledge reveals X is true, proper?”

    Higher: “What does this knowledge truly present? Take into account various interpretations.”

    5. Use Structured Outputs

    Add to prompts: “Format your response as: Info (what you’re sure about), Inferences (what you’re reasoning towards), Uncertainties (what you don’t know).”

    Construction reduces the prospect of blending information with hypothesis.

    6. Leverage Citations Mode

    When Claude cites sources (in modes the place that is out there), hallucinations drop considerably. Request citations each time potential.

    7. Problem Assured Claims

    When Claude states one thing definitively, push again: “How sure are you about that? What would you should confirm it?”

    Claude will typically again down from overconfident claims when challenged.

    Methods That Work for Each Fashions

    Some methods scale back hallucinations in each GPT-5 and Claude 4:

    READ ALSO  When ChatGPT isn’t the answer: 17 disadvantages of ChatGPT content

    1. Present Context

    The extra context you give, the much less the AI must guess.

    Dangerous: “What’s the perfect framework?”

    Higher: “I’m constructing a real-time dashboard with 10K concurrent customers. Information updates each second. Workforce is aware of React. What framework ought to I exploit and why?”

    2. Break Advanced Questions Down

    Don’t ask one huge query. Break it into steps.

    As an alternative of: “Design an entire system structure for my app.”

    Do: “First, what are the important thing parts of this method? [Wait for response] Now, how ought to these parts talk? [Wait for response] What database is sensible given these necessities?”

    3. Confirm Essential Data

    For something necessary, use a number of approaches:

    • Ask the identical query in a different way and examine solutions
    • Request the AI to fact-check itself
    • Cross-reference with different sources
    • Use net search options when out there

    4. Use Particular Constraints

    Add to prompts: “Solely present info you could have excessive confidence in. For the rest, say ‘I’m not sure about this’ explicitly.”

    Each fashions reply higher to express guardrails.

    5. Check with Identified Solutions

    Earlier than trusting a mannequin on unknown info, take a look at it with questions you already know the reply to. See the way it handles uncertainty and accuracy.

    When Every Mannequin is the Higher Selection

    Select GPT-5 when:

    • Pace issues greater than excellent accuracy
    • You’re brainstorming or being artistic
    • You possibly can simply confirm the output
    • You want broad normal data
    • You need conversational, assured responses

    Select Claude 4 when:

    • Accuracy is vital
    • You’re working with technical particulars
    • You want clear reasoning
    • You worth honesty about limitations
    • You possibly can look forward to prolonged considering on advanced issues

    The Actuality: Good Accuracy Doesn’t Exist

    Right here’s the reality: You can not eradicate AI hallucinations utterly. Not with GPT-5. Not with Claude 4. Not with any language mannequin.

    These instruments predict textual content. They don’t confirm fact. That’s not how they work.

    The very best you are able to do is:

    • Perceive when every mannequin is more likely to hallucinate
    • Use prompting methods that scale back the speed
    • Confirm something necessary
    • Select the precise mannequin for every job
    • Set life like expectations

    In my testing, good prompting methods decreased hallucinations by 40-60%. That’s vital. But it surely’s not elimination.

    The winners aren’t individuals who eradicate hallucinations. They’re individuals who work with these limitations intelligently.

    The Backside Line

    Claude 4 hallucinates lower than GPT-5. About 40% much less in my testing. It’s extra cautious, extra trustworthy about uncertainty, and extra correct on information.

    However GPT-5 is quicker, extra assured, and typically that’s what you want.

    Each require the identical method: Sensible prompting. Wholesome skepticism.

    Verification of vital information.

    Use Claude 4 when accuracy issues. Use GPT-5 when velocity and creativity matter. Use verification for each.

    The way forward for AI isn’t hallucination-free responses.

    It’s customers who know the way to work with imperfect instruments to get dependable outcomes.

    That’s the talent that issues now.

    Commercials

    You’ve got been blocked from seeing adverts.

    ‘;
    endif;
    ?>





    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleWhy Package Installs Are Slow (And How to Fix It)
    Next Article You Probably Don’t Need a Vector Database for Your RAG — Yet
    ProfitlyAI
    • Website

    Related Posts

    AI Tools & Technologies

    Proofig or TruthScan? Which Should You Use?

    January 12, 2026
    AI Tools & Technologies

    TruthScan vs. BrandWell: Which Should Be Your AI Image Detector?

    December 19, 2025
    AI Tools & Technologies

    5 Business Ideas To Start With Chatgpt In 2026 » Ofemwire

    December 9, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    The cost of thinking | MIT News

    November 19, 2025

    Artificial intelligence enhances air mobility planning | MIT News

    April 25, 2025

    IT as the new HR: Managing your AI workforce

    November 7, 2025

    How to Consistently Extract Metadata from Complex Documents

    October 24, 2025

    YOLOv1 Paper Walkthrough: The Day YOLO First Saw the World

    December 5, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    Introducing ShaTS: A Shapley-Based Method for Time-Series Models

    November 17, 2025

    How to Automate Workflows with AI

    November 15, 2025

    Back office automation for insurance companies: A success story

    April 24, 2025
    Our Picks

    From Transactions to Trends: Predict When a Customer Is About to Stop Buying

    January 23, 2026

    America’s coming war over AI regulation

    January 23, 2026

    “Dr. Google” had its issues. Can ChatGPT Health do better?

    January 22, 2026
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.