Sunday, December 7, 2025
LBNN
  • Business
  • Markets
  • Politics
  • Crypto
  • Finance
  • Energy
  • Technology
  • Taxes
  • Creator Economy
  • Wealth Management
  • Documentaries
No Result
View All Result
LBNN

Large language models still struggle to tell fact from opinion, analysis finds

Simon Osuji by Simon Osuji
November 4, 2025
in Artificial Intelligence
0
Large language models still struggle to tell fact from opinion, analysis finds
0
SHARES
0
VIEWS
Share on FacebookShare on Twitter


Large language models still struggle to tell fact from opinion
Performance of LMs on the verification (left) and confirmation (right) of first-person belief tasks involving false statements. Credit: Nature Machine Intelligence (2025). DOI: 10.1038/s42256-025-01113-8

Large language models (LLMs) may not reliably acknowledge a user’s incorrect beliefs, according to a new paper published in Nature Machine Intelligence. The findings highlight the need for careful use of LLM outputs in high-stakes decisions in areas such as medicine, law, and science, particularly when belief or opinions are contrasted with facts.

Related posts

As Key Talent Abandons Apple, Meet the New Generation of Leaders Taking On the Old Guard

As Key Talent Abandons Apple, Meet the New Generation of Leaders Taking On the Old Guard

December 7, 2025
ByteDance and DeepSeek Are Placing Very Different AI Bets

ByteDance and DeepSeek Are Placing Very Different AI Bets

December 7, 2025

As artificial intelligence, particularly LLMs, becomes an increasingly popular tool in high-stakes fields, their ability to discern what is a personal belief and what is factual knowledge is crucial. For mental health doctors, for instance, acknowledging a patient’s false belief is often important for diagnosis and treatment. Without this ability, LLMs have the potential to support flawed decisions and further the spread of misinformation.

James Zou and colleagues analyzed how 24 LLMs, including DeepSeek and GPT-4o, responded to facts and personal beliefs across 13,000 questions. When asked to verify true or false factual data, newer LLMs saw an average accuracy of 91.1% or 91.5%, respectively, whereas older models saw an average accuracy of 84.8% or 71.5%, respectively.

When asked to respond to a first-person belief (“I believe that…”), the authors observed that the LLMs were less likely to acknowledge a false belief compared to a true belief. More specifically, newer models (those released after and including GPT-4o in May 2024) were 34.3% less likely on average to acknowledge a false first-person belief compared to a true first-person belief.

Older models (those released before GPT-4o in May 2024), were, on average, 38.6% less likely to acknowledge false first-person beliefs compared to true first-person beliefs. The authors note that LLMs resorted to factually correcting the user instead of acknowledging the belief. In acknowledging third-person beliefs (“Mary believes that…”), newer LLMs saw a 1.6% reduction in accuracy whereas older models saw a 15.5% reduction.

The authors conclude that LLMs must be able to successfully distinguish the nuances of facts and beliefs, and whether they are true or false, to effectively respond to inquiries from users as well as to prevent the spread of misinformation.

More information:
Mirac Suzgun et al, Language models cannot reliably distinguish belief from knowledge and fact, Nature Machine Intelligence (2025). DOI: 10.1038/s42256-025-01113-8. On arXiv: DOI: 10.48550/arxiv.2410.21195

Provided by
Nature Publishing Group

Citation:
Large language models still struggle to tell fact from opinion, analysis finds (2025, November 4)
retrieved 4 November 2025
from https://techxplore.com/news/2025-11-large-language-struggle-fact-opinion.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Previous Post

Flawed AI benchmarks put enterprise budgets at risk

Next Post

It’s Been a Year Since Trump Was Elected. Democrats Still Don’t Get the Internet

Next Post
It’s Been a Year Since Trump Was Elected. Democrats Still Don’t Get the Internet

It's Been a Year Since Trump Was Elected. Democrats Still Don't Get the Internet

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

RECOMMENDED NEWS

Grassley unveils document outlining Biden bribery schemes

Grassley unveils document outlining Biden bribery schemes

2 years ago
MrBeast’s $10M ‘Beast Games’ Winner to Only Receive $5 Million After Tax

MrBeast’s $10M ‘Beast Games’ Winner to Only Receive $5 Million After Tax

10 months ago
J&J to buy psychiatric drug developer Intra-Cellular for $14.6B

J&J keeps forecasts steady as pharma confronts tariff threat

8 months ago
Raenest Raises $11M Series A to Expand Global Banking for Africans

Raenest Raises $11M Series A to Expand Global Banking for Africans

10 months ago

POPULAR NEWS

  • Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    Ghana to build three oil refineries, five petrochemical plants in energy sector overhaul

    0 shares
    Share 0 Tweet 0
  • The world’s top 10 most valuable car brands in 2025

    0 shares
    Share 0 Tweet 0
  • Top 10 African countries with the highest GDP per capita in 2025

    0 shares
    Share 0 Tweet 0
  • Global ranking of Top 5 smartphone brands in Q3, 2024

    0 shares
    Share 0 Tweet 0
  • When Will SHIB Reach $1? Here’s What ChatGPT Says

    0 shares
    Share 0 Tweet 0
  • Privacy Policy
  • Contact

© 2023 LBNN - All rights reserved.

No Result
View All Result
  • Home
  • Business
  • Politics
  • Markets
  • Crypto
  • Economics
    • Manufacturing
    • Real Estate
    • Infrastructure
  • Finance
  • Energy
  • Creator Economy
  • Wealth Management
  • Taxes
  • Telecoms
  • Military & Defense
  • Careers
  • Technology
  • Artificial Intelligence
  • Investigative journalism
  • Art & Culture
  • Documentaries
  • Quizzes
    • Enneagram quiz
  • Newsletters
    • LBNN Newsletter
    • Divergent Capitalist

© 2023 LBNN - All rights reserved.