Citiverse

gert@poliversity.it (@gert@poliversity.it)

Psychologist | Licensed Professional Psychoterapist

𝙿𝚊𝚜𝚜𝚒𝚘𝚗𝚊𝚝𝚎 𝚌𝚕𝚒𝚖𝚋𝚎𝚛, 𝚌𝚘𝚍𝚎𝚛 𝚊𝚗𝚍 𝙻𝚒𝚗𝚞𝚡 𝚞𝚜𝚎𝚛

  • Language models cannot reliably distinguish belief from knowledge and fact

    Language models cannot reliably distinguish belief from knowledge and fact

    Abstract
    -----------
    «As language models (LMs) increasingly infiltrate into high-stakes domains such as law, medicine, journalism and science, their ability to distinguish belief from knowledge, and fact from fiction, becomes imperative. Failure to make such distinctions can mislead diagnoses, distort judicial judgments and amplify misinformation. Here we evaluate 24 cutting-edge LMs using a new KaBLE benchmark of 13,000 questions across 13 epistemic tasks. Our findings reveal crucial limitations. In particular, all models tested systematically fail to acknowledge first-person false beliefs, with GPT-4o dropping from 98.2% to 64.4% accuracy and DeepSeek R1 plummeting from over 90% to 14.4%. Further, models process third-person false beliefs with substantially higher accuracy (95% for newer models; 79% for older ones) than first-person false beliefs (62.6% for newer; 52.5% for older), revealing a troubling attribution bias. We also find that, while recent models show competence in recursive knowledge tasks, they still rely on inconsistent reasoning strategies, suggesting superficial pattern matching rather than robust epistemic understanding. Most models lack a robust understanding of the factive nature of knowledge, that knowledge inherently requires truth. These limitations necessitate urgent improvements before deploying LMs in high-stakes domains where epistemic distinctions are crucial.»

    https://www.nature.com/articles/s42256-025-01113-8

     Senza categoria llms epistemology knowledge