PulseAugur
LIVE 03:42:18
research · [53 sources] ·
0
research

New AI research tackles LLM hallucinations with novel detection and intervention methods

Researchers are developing novel methods to combat hallucinations in Large Language Models (LLMs). Several papers propose new frameworks and techniques, including LaaB, which bridges neural features and symbolic judgments, and CuraView, a multi-agent system for medical hallucination detection using GraphRAG. Other approaches focus on neuro-symbolic agents for hallucination-free requirements reuse, adaptive unlearning for surgical hallucination suppression in code generation, and harnessing reasoning trajectories via answer-agreement representation shaping. Additionally, new benchmarks like HalluScan are being created to systematically evaluate detection and mitigation strategies. AI

Summary written by gemini-2.5-flash-lite from 53 sources. How we write summaries →

IMPACT New research offers diverse strategies to improve LLM factual accuracy, crucial for reliable deployment in sensitive domains like healthcare and code generation.

RANK_REASON Multiple arXiv papers present novel research on detecting and mitigating LLM hallucinations.

Read on Hugging Face Daily Papers →

COVERAGE [53]

  1. arXiv cs.AI TIER_1 · Ali Baheri ·

    Where Does Reasoning Break? Step-Level Hallucination Detection via Hidden-State Transport Geometry

    Large language models hallucinate during multi-step reasoning, but most existing detectors operate at the trace level: they assign one confidence score to a full output, fail to localize the first error, and often require multiple sampled completions. We frame hallucination inste…

  2. arXiv cs.AI TIER_1 · Amine Trabelsi ·

    CAAFC: Chronological Actionable Automated Fact-Checker for misinformation / non-factual hallucination detection and correction

    With the vast amount of content uploaded every hour, along with the AI generated content that can include hallucinations, Automated Fact-Checking (AFC) has become increasingly vital, as it is infeasible for human fact-checkers to manually verify the sheer volume of information ge…

  3. arXiv cs.AI TIER_1 · Yi R. Fung ·

    Scalable Token-Level Hallucination Detection in Large Language Models

    Large language models (LLMs) have demonstrated remarkable capabilities, but they still frequently produce hallucinations. These hallucinations are difficult to detect in reasoning-intensive tasks, where the content appears coherent but contains errors like logical flaws and unrel…

  4. arXiv cs.LG TIER_1 · Ruixuan Wang ·

    Instruction Lens Score: Your Instruction Contributes a Powerful Object Hallucination Detector for Multimodal Large Language Models

    Multimodal large language models (MLLMs) have achieved remarkable progress, yet the object hallucination remains a critical challenge for reliable deployment. In this paper, we present an in-depth analysis of instruction token embeddings and reveal that they implicitly encode vis…

  5. arXiv cs.AI TIER_1 · Yian Yin ·

    LLM hallucinations in the wild: Large-scale evidence from non-existent citations

    Large language models (LLMs) are known to generate plausible but false information across a wide range of contexts, yet the real-world magnitude and consequences of this hallucination problem remain poorly understood. Here we leverage a uniquely verifiable object - scientific cit…

  6. arXiv cs.CL TIER_1 · Brandon C. Colelough, Davis Bartels, Dina Demner-Fushman ·

    Quantifying Hallucinations in Language Language Models on Medical Textbooks

    arXiv:2603.09986v2 Announce Type: replace Abstract: Hallucinations, the tendency for large language models to provide responses with factually incorrect and unsupported claims, is a serious problem within natural language processing for which we do not yet have an effective solut…

  7. arXiv cs.CL TIER_1 · Erik Nielsen, Elia Cunegatti, Marcus Vukojevic, Giovanni Iacca ·

    Hallucination as an Anomaly: Dynamic Intervention via Probabilistic Circuits

    arXiv:2605.05953v1 Announce Type: new Abstract: One of the most critical challenges in Large Language Models is their tendency to hallucinate, i.e., produce factually incorrect responses. Existing approaches show promising results in terms of hallucination correction, but still s…

  8. arXiv cs.CL TIER_1 · Giovanni Iacca ·

    Hallucination as an Anomaly: Dynamic Intervention via Probabilistic Circuits

    One of the most critical challenges in Large Language Models is their tendency to hallucinate, i.e., produce factually incorrect responses. Existing approaches show promising results in terms of hallucination correction, but still suffer from a main limitation: they apply correct…

  9. Hugging Face Daily Papers TIER_1 ·

    Hallucination as an Anomaly: Dynamic Intervention via Probabilistic Circuits

    One of the most critical challenges in Large Language Models is their tendency to hallucinate, i.e., produce factually incorrect responses. Existing approaches show promising results in terms of hallucination correction, but still suffer from a main limitation: they apply correct…

  10. arXiv cs.LG TIER_1 · Linggang Kong, Lei Wu, Yunlong Zhang, Xiaofeng Zhong, Zhen Wang, Yongjie Wang, Yao Pan ·

    CausalGaze: Unveiling Hallucinations via Counterfactual Graph Intervention in Large Language Models

    arXiv:2604.11087v2 Announce Type: replace Abstract: Despite the groundbreaking advancements made by large language models (LLMs), hallucination remains a critical bottleneck for their deployment in high-stakes domains. Existing classification-based methods mainly rely on static a…

  11. arXiv cs.CL TIER_1 · Mina Gabriel ·

    The First Token Knows: Single-Decode Confidence for Hallucination Detection

    arXiv:2605.05166v1 Announce Type: new Abstract: Self-consistency detects hallucinations by generating multiple sampled answers to a question and measuring agreement, but this requires repeated decoding and can be sensitive to lexical variation. Semantic self-consistency improves …

  12. arXiv cs.CL TIER_1 · Gijs van Dijk ·

    Detecting Hallucinations in Large Language Models via Internal Attention Divergence Signals

    arXiv:2605.05025v1 Announce Type: new Abstract: We propose a lightweight and single-pass uncertainty quantification method for detecting hallucinations in Large Language Models. The method uses attention matrices to estimate uncertainty without requiring repeated sampling or exte…

  13. arXiv cs.LG TIER_1 · Dan Wilson, Mohamed Akrout ·

    Low-Cost Black-Box Detection of LLM Hallucinations via Dynamical System Prediction

    arXiv:2605.05134v1 Announce Type: new Abstract: Large Language Models (LLMs) frequently generate plausible but non-factual content, a phenomenon known as hallucination. While existing detection methods typically rely on computationally expensive sampling-based consistency checks …

  14. arXiv cs.CL TIER_1 · Philip Wootaek Shin, Ajay Narayanan Sridhar, Sivani Devarapalli, Rui Zhang, Jack Sampson, Vijaykrishnan Narayanan ·

    When Relations Break: Analyzing Relation Hallucination in Vision-Language Model Under Rotation and Noise

    arXiv:2605.05045v1 Announce Type: cross Abstract: Vision-language models (VLMs) achieve strong multimodal performance but remain prone to relation hallucination, which requires accurate reasoning over inter-object interactions. We study the impact of visual perturbations, specifi…

  15. arXiv cs.CL TIER_1 · Mina Gabriel ·

    The First Token Knows: Single-Decode Confidence for Hallucination Detection

    Self-consistency detects hallucinations by generating multiple sampled answers to a question and measuring agreement, but this requires repeated decoding and can be sensitive to lexical variation. Semantic self-consistency improves this by clustering sampled answers by meaning us…

  16. arXiv cs.LG TIER_1 · Mohamed Akrout ·

    Low-Cost Black-Box Detection of LLM Hallucinations via Dynamical System Prediction

    Large Language Models (LLMs) frequently generate plausible but non-factual content, a phenomenon known as hallucination. While existing detection methods typically rely on computationally expensive sampling-based consistency checks or external knowledge retrieval, we propose a ne…

  17. arXiv cs.CL TIER_1 · Gijs van Dijk ·

    Detecting Hallucinations in Large Language Models via Internal Attention Divergence Signals

    We propose a lightweight and single-pass uncertainty quantification method for detecting hallucinations in Large Language Models. The method uses attention matrices to estimate uncertainty without requiring repeated sampling or external models. Specifically, we measure the Kullba…

  18. arXiv cs.AI TIER_1 · Ahmed Ibrahim ·

    Neuro-Symbolic Agents for Hallucination-Free Requirements Reuse

    arXiv:2605.01562v1 Announce Type: cross Abstract: The Object-Oriented Method for Requirements Authoring and Management (OOMRAM) is a requirements reuse framework that relies on exact identifier matching and rigid templates, limiting its ability to adapt specifications across dive…

  19. arXiv cs.CL TIER_1 · Hao Mi, Qiang Sheng, Shaofei Wang, Beizhe Hu, Yifan Sun, Zhengjia Wang, Hengqi Zeng, Yang Li, Danding Wang, Juan Cao ·

    Logical Consistency as a Bridge: Improving LLM Hallucination Detection via Label Constraint Modeling between Responses and Self-Judgments

    arXiv:2605.03971v1 Announce Type: new Abstract: Large Language Models (LLMs) are prone to factual hallucinations, risking their reliability in real-world applications. Existing hallucination detectors mainly extract micro-level intrinsic patterns for uncertainty quantification or…

  20. arXiv cs.CL TIER_1 · Severin Ye, Xiao Kong, Xiaopeng He, Guangsu Yan, Dongsuk Oh ·

    CuraView: A Multi-Agent Framework for Medical Hallucination Detection with GraphRAG-Enhanced Knowledge Verification

    arXiv:2605.03476v1 Announce Type: new Abstract: Discharge summaries require extracting critical information from lengthy electronic health records (EHRs), a process that is labor-intensive when performed manually. Large language models (LLMs) can improve generation efficiency; ho…

  21. arXiv cs.CL TIER_1 · Juan Cao ·

    Logical Consistency as a Bridge: Improving LLM Hallucination Detection via Label Constraint Modeling between Responses and Self-Judgments

    Large Language Models (LLMs) are prone to factual hallucinations, risking their reliability in real-world applications. Existing hallucination detectors mainly extract micro-level intrinsic patterns for uncertainty quantification or elicit macro-level self-judgments through verba…

  22. Hugging Face Daily Papers TIER_1 ·

    CuraView: A Multi-Agent Framework for Medical Hallucination Detection with GraphRAG-Enhanced Knowledge Verification

    Discharge summaries require extracting critical information from lengthy electronic health records (EHRs), a process that is labor-intensive when performed manually. Large language models (LLMs) can improve generation efficiency; however, they are prone to producing faithfulness …

  23. arXiv cs.CL TIER_1 · Dongsuk Oh ·

    CuraView: A Multi-Agent Framework for Medical Hallucination Detection with GraphRAG-Enhanced Knowledge Verification

    Discharge summaries require extracting critical information from lengthy electronic health records (EHRs), a process that is labor-intensive when performed manually. Large language models (LLMs) can improve generation efficiency; however, they are prone to producing faithfulness …

  24. arXiv cs.LG TIER_1 · Yee Zhing Liew, Andrew Huey Ping Tan, Anwar P. P Abdul Majeed ·

    From Flat Facts to Sharp Hallucinations: Detecting Stubborn Errors via Gradient Sensitivity

    arXiv:2605.00939v1 Announce Type: new Abstract: Traditional hallucination detection fails on "Stubborn Hallucinations" -- errors where LLMs are confidently wrong. We propose a geometric solution: Embedding-Perturbed Gradient Sensitivity (EPGS). We hypothesize that while robust fa…

  25. arXiv cs.LG TIER_1 · Jianxiong Zhang, Bing Guo, Yuming Jiang, Haobo Wang, Bo An, Sean Du ·

    Harnessing Reasoning Trajectories for Hallucination Detection via Answer-agreement Representation Shaping

    arXiv:2601.17467v2 Announce Type: replace Abstract: Large reasoning models (LRMs) often generate long, seemingly coherent reasoning traces yet still produce incorrect answers, making hallucination detection challenging. Although trajectories contain useful signals, directly using…

  26. arXiv cs.CL TIER_1 · Ahmed Cherif ·

    HalluScan: A Systematic Benchmark for Detecting and Mitigating Hallucinations in Instruction-Following LLMs

    arXiv:2605.02443v1 Announce Type: new Abstract: Large Language Models (LLMs) have demonstrated remarkable capabilities across diverse natural language processing tasks, yet they remain susceptible to hallucinations -- generating content that is factually incorrect, unfaithful to …

  27. arXiv cs.CL TIER_1 · Freja Thoresen, Dan Saattrup Smart ·

    A multilingual hallucination benchmark: MultiWikiQHalluA

    arXiv:2605.02504v1 Announce Type: new Abstract: Most hallucination evaluations focus on English, leaving it unclear whether findings transfer to lower-resource languages. We investigate faithfulness hallucinations, defined as model-generated content that is fluent and plausible b…

  28. arXiv cs.CL TIER_1 · Alexandra Bazarova, Aleksandr Yugay, Andrey Shulga, Alina Ermilova, Andrei Volodichev, Konstantin Polev, Julia Belikova, Rauf Parchiev, Dmitry Simakov, Maxim Savchenko, Andrey Savchenko, Serguei Barannikov, Alexey Zaytsev ·

    Hallucination Detection in LLMs with Topological Divergence on Attention Graphs

    arXiv:2504.10063v4 Announce Type: replace Abstract: Hallucination, i.e., generating factually incorrect content, remains a critical challenge for large language models (LLMs). We introduce TOHA, a TOpology-based HAllucination detector in the RAG setting, which leverages a topolog…

  29. arXiv cs.CL TIER_1 · Joseph Spracklen, Pedram Aghazadeh, Farinaz Koushanfar, Murtuza Jadliwala ·

    LLM Ghostbusters: Surgical Hallucination Suppression via Adaptive Unlearning

    arXiv:2605.01047v1 Announce Type: cross Abstract: Hallucinations, outputs that sound plausible but are factually incorrect, remain an open challenge for deployed LLMs. In code generation, models frequently hallucinate non-existent software packages, recommending imports and insta…

  30. arXiv cs.CL TIER_1 · Dan Saattrup Smart ·

    A multilingual hallucination benchmark: MultiWikiQHalluA

    Most hallucination evaluations focus on English, leaving it unclear whether findings transfer to lower-resource languages. We investigate faithfulness hallucinations, defined as model-generated content that is fluent and plausible but diverges from the provided input or is intern…

  31. arXiv cs.CL TIER_1 · Ahmed Cherif ·

    HalluScan: A Systematic Benchmark for Detecting and Mitigating Hallucinations in Instruction-Following LLMs

    Large Language Models (LLMs) have demonstrated remarkable capabilities across diverse natural language processing tasks, yet they remain susceptible to hallucinations -- generating content that is factually incorrect, unfaithful to provided context, or misaligned with user instru…

  32. Hugging Face Daily Papers TIER_1 ·

    Mitigating Multimodal LLMs Hallucinations via Relevance Propagation at Inference Time

    Multimodal large language models (MLLMs) have revolutionized the landscape of AI, demonstrating impressive capabilities in tackling complex vision and audio-language tasks. However, a critical challenge remains: these models often suffer from hallucinations, generating outputs th…

  33. arXiv cs.CL TIER_1 · Guoshenghui Zhao, Weijie Zhao, Tan Yu ·

    HIVE: Hidden-Evidence Verification for Hallucination Detection in Diffusion Large Language Models

    arXiv:2604.26139v1 Announce Type: new Abstract: Diffusion large language models generate text through multi-step denoising, where hallucination signals may emerge throughout the trajectory rather than only in the final output. Existing detectors mainly rely on output uncertainty …

  34. arXiv cs.CL TIER_1 · Jiawei Li, Akshayaa Magesh, Venugopal V. Veeravalli ·

    Principled Detection of Hallucinations in Large Language Models via Multiple Testing

    arXiv:2508.18473v3 Announce Type: replace Abstract: While Large Language Models (LLMs) have emerged as powerful foundational models to solve a variety of tasks, they have also been shown to be prone to hallucinations, i.e., generating responses that sound confident but are actual…

  35. arXiv cs.CL TIER_1 · Tan Yu ·

    HIVE: Hidden-Evidence Verification for Hallucination Detection in Diffusion Large Language Models

    Diffusion large language models generate text through multi-step denoising, where hallucination signals may emerge throughout the trajectory rather than only in the final output. Existing detectors mainly rely on output uncertainty or coarse trace statistics, which often fail to …

  36. arXiv cs.AI TIER_1 · Federico A. Kamelhar ·

    GSAR: Typed Grounding for Hallucination Detection and Recovery in Multi-Agent LLMs

    arXiv:2604.23366v1 Announce Type: new Abstract: Autonomous multi-agent LLM systems are increasingly deployed to investigate operational incidents and produce structured diagnostic reports. Their trustworthiness hinges on whether each claim is grounded in observed evidence rather …

  37. Hugging Face Daily Papers TIER_1 ·

    Global Context or Local Detail? Adaptive Visual Grounding for Hallucination Mitigation

    Vision-Language Models (VLMs) are frequently undermined by object hallucination--generating content that contradicts visual reality--due to an over-reliance on linguistic priors. We introduce Positive-and-Negative Decoding (PND), a training-free inference framework that intervene…

  38. Hugging Face Daily Papers TIER_1 ·

    GSAR: Typed Grounding for Hallucination Detection and Recovery in Multi-Agent LLMs

    Autonomous multi-agent LLM systems are increasingly deployed to investigate operational incidents and produce structured diagnostic reports. Their trustworthiness hinges on whether each claim is grounded in observed evidence rather than model-internal inference. Existing grounded…

  39. arXiv cs.CV TIER_1 · Aofan Liu ·

    Dual-Pathway Circuits of Object Hallucination in Vision-Language Models

    Vision-language models (VLMs) have demonstrated remarkable capabilities in bridging visual perception and natural language understanding, enabling a wide range of multimodal reasoning tasks. However, they often produce object hallucinations, describing content absent from the inp…

  40. arXiv cs.CV TIER_1 · Jing Li ·

    Vocabulary Hijacking in LVLMs: Unveiling Critical Attention Heads by Excluding Inert Tokens to Mitigate Hallucination

    Large Vision-Language Models (LVLMs) have achieved remarkable progress in multimodal tasks, yet their reliability is persistently undermined by hallucinations-generating text that contradicts visual input. Recent studies often attribute these errors to inadequate visual attention…

  41. arXiv stat.ML TIER_1 · Prabhat Kc, Rongping Zeng, Nirmal Soni, Aldo Badano ·

    sFRC for assessing hallucinations in medical image restoration

    arXiv:2603.04673v2 Announce Type: replace-cross Abstract: Deep learning (DL) methods are currently being explored to restore images from sparse-view-, limited-data-, and undersampled-based acquisitions in medical applications. Although outputs from DL may appear visually appealin…

  42. arXiv cs.CV TIER_1 · Vijaykrishnan Narayanan ·

    When Relations Break: Analyzing Relation Hallucination in Vision-Language Model Under Rotation and Noise

    Vision-language models (VLMs) achieve strong multimodal performance but remain prone to relation hallucination, which requires accurate reasoning over inter-object interactions. We study the impact of visual perturbations, specifically rotation and noise, and show that even mild …

  43. arXiv cs.CV TIER_1 · Itai Allouche, Joseph Keshet ·

    Mitigating Multimodal LLMs Hallucinations via Relevance Propagation at Inference Time

    arXiv:2605.01766v1 Announce Type: cross Abstract: Multimodal large language models (MLLMs) have revolutionized the landscape of AI, demonstrating impressive capabilities in tackling complex vision and audio-language tasks. However, a critical challenge remains: these models often…

  44. arXiv cs.CV TIER_1 · Jianfei Zhao, Feng Zhang, Xin Sun, Chong Feng, Zhixing Tan ·

    Tell Model Where to Look: Mitigating Hallucinations in MLLMs by Vision-Guided Attention

    arXiv:2511.20032v3 Announce Type: replace Abstract: Visual attention serves as the primary mechanism through which MLLMs interpret visual information; however, its limited localization capability often leads to hallucinations. We observe that although MLLMs can accurately extract…

  45. arXiv cs.CV TIER_1 · Chengsheng Zhang, Chenghao Sun, Xinyan Jiang, Wei Li, Xinmei Tian ·

    Prefill-Time Intervention for Mitigating Hallucination in Large Vision-Language Models

    arXiv:2604.25642v1 Announce Type: new Abstract: Large Vision-Language Models (LVLMs) have achieved remarkable progress in visual-textual understanding, yet their reliability is critically undermined by hallucinations, i.e., the generation of factually incorrect or inconsistent re…

  46. arXiv cs.CV TIER_1 · Xinmei Tian ·

    Prefill-Time Intervention for Mitigating Hallucination in Large Vision-Language Models

    Large Vision-Language Models (LVLMs) have achieved remarkable progress in visual-textual understanding, yet their reliability is critically undermined by hallucinations, i.e., the generation of factually incorrect or inconsistent responses. While recent studies using steering vec…

  47. arXiv cs.CV TIER_1 · Yubo Jiang, Xin Yang, Abudukelimu Wuerkaixi, Zheming Yuan, Xuxin Cheng, Fengying Xie, Zhiguo Jiang, Cao Liu, Ke Zeng, Haopeng Zhang ·

    Global Context or Local Detail? Adaptive Visual Grounding for Hallucination Mitigation

    arXiv:2604.24396v1 Announce Type: new Abstract: Vision-Language Models (VLMs) are frequently undermined by object hallucination--generating content that contradicts visual reality--due to an over-reliance on linguistic priors. We introduce Positive-and-Negative Decoding (PND), a …

  48. arXiv cs.CV TIER_1 · Zhiyuan Jiang, Weihao Hong, Xinlei Guan, Tejaswi Dhandu, Miles Q. Li, Meng Xu, Kuan Huang, Umamaheswara Rao Tida, Bingyu Shen, Daehan Kwak, Boyang Li ·

    LLM-as-Judge Framework for Evaluating Tone-Induced Hallucination in Vision-Language Models

    arXiv:2604.18803v3 Announce Type: replace Abstract: Vision-Language Models (VLMs) are increasingly deployed in settings where reliable visual grounding carries operational consequences, yet their behavior under progressively coercive prompt phrasing remains undercharacterized. Ex…

  49. arXiv cs.CV TIER_1 · Jiawei Chen, Dingkang Yang, Tong Wu, Yue Jiang, Xiaolu Hou, Mingcheng Li, Shunli Wang, Dongling Xiao, Ke Li, Lihua Zhang ·

    Detecting and Evaluating Medical Hallucinations in Large Vision Language Models

    arXiv:2406.10185v2 Announce Type: replace Abstract: Large Vision Language Models (LVLMs) are increasingly integral to healthcare applications, including medical visual question answering and imaging report generation. While these models inherit the robust capabilities of foundati…

  50. arXiv cs.CV TIER_1 · JiYang Wang, Jiawei Chen, Mengqi Xiao, Yu Cheng, Yangfu Li, Zhaoxia Yin ·

    DO-Bench: An Attributable Benchmark for Diagnosing Object Hallucination in Vision-Language Models

    arXiv:2604.22822v1 Announce Type: new Abstract: Object level hallucination remains a central reliability challenge for vision language models (VLMs), particularly in binary object existence verification. Existing benchmarks emphasize aggregate accuracy but rarely disentangle whet…

  51. arXiv cs.CV TIER_1 · Haopeng Zhang ·

    Global Context or Local Detail? Adaptive Visual Grounding for Hallucination Mitigation

    Vision-Language Models (VLMs) are frequently undermined by object hallucination--generating content that contradicts visual reality--due to an over-reliance on linguistic priors. We introduce Positive-and-Negative Decoding (PND), a training-free inference framework that intervene…

  52. Eugene Yan TIER_1 ·

    Out-of-Domain Finetuning to Bootstrap Hallucination Detection

    How to use open-source, permissive-use data and collect less labeled samples for our tasks.

  53. dev.to — LLM tag TIER_1 Português(PT) · Marcelo Cabral Ghilardi ·

    When AI Lies and Asks You to CALM DOWN: a straight talk about hallucinations

    <p> </p> <p>E aí, gurizada! Tudo tranquilo? Hoje eu quero trocar uma ideia com vocês sobre umas paradas que andei percebendo com as IAs, e que me motivaram a gravar um vídeo e até escrever um post lá no meu site, o marcelocabral.com.br. Sabe quando a inteligência artificial solta…