PulseAugur
LIVE 06:37:41
research · [8 sources] ·
0
research

Aligning with Your Own Voice: Self-Corrected Preference Learning for Hallucination Mitigation in LVLMs

Researchers are developing new frameworks to address hallucinations in large language models (LLMs). One approach, termed "LLM Psychosis," categorizes severe reality-boundary failures and proposes a diagnostic scale to evaluate them, with findings from ChatGPT 5 documented. Another method, KARL, uses reinforcement learning to align abstention behavior with a model's knowledge boundary, aiming to reduce hallucinations without sacrificing accuracy. Additionally, PRISM offers a benchmark to disentangle hallucinations into knowledge, reasoning, and instruction-following errors, aiding in understanding their origins. For vision-language models, AVES-DPO focuses on self-correction to mitigate hallucinations using in-distribution data. AI

Summary written by None from 8 sources. How we write summaries →

IMPACT New diagnostic tools and mitigation strategies for LLM hallucinations could improve the reliability and trustworthiness of deployed AI systems.

RANK_REASON Multiple academic papers introducing new frameworks and benchmarks for understanding and mitigating LLM hallucinations.

Read on arXiv cs.AI →

COVERAGE [8]

  1. arXiv cs.AI TIER_1 · Ashutosh Raj ·

    LLM Psychosis: A Theoretical and Diagnostic Framework for Reality-Boundary Failures in Large Language Models

    arXiv:2604.25934v1 Announce Type: cross Abstract: The deployment of large language models (LLMs) as interactive agents has exposed a category of behavioral failure that prevailing terminology, principally hallucination, fails to adequately characterize. This paper introduces LLM …

  2. arXiv cs.AI TIER_1 · Wentao Hu, Yanbo Zhai, Xiaohui Hu, Mingkuan Zhao, Shanhong yu, Xue Liu, Kaidong Yu, Shuangyong Song, Xuelong Li ·

    Awakening Dormant Experts:Counterfactual Routing to Mitigate MoE Hallucinations

    arXiv:2604.14246v2 Announce Type: replace-cross Abstract: Sparse Mixture-of-Experts (MoE) models have achieved remarkable scalability, yet they remain vulnerable to hallucinations, particularly when processing long-tail knowledge. We identify that this fragility stems from static…

  3. arXiv cs.CL TIER_1 · Cheng Gao, Cheng Huang, Kangyang Luo, Ziqing Qiao, Shuzheng Si, Huimin Chen, Chaojun Xiao, Maosong Sun ·

    KARL: Mitigating Hallucinations in LLMs via Knowledge-Boundary-Aware Reinforcement Learning

    arXiv:2604.22779v1 Announce Type: cross Abstract: Enabling large language models (LLMs) to appropriately abstain from answering questions beyond their knowledge is crucial for mitigating hallucinations. While existing reinforcement learning methods foster autonomous abstention, t…

  4. arXiv cs.CL TIER_1 · Yuhe Wu, Guangyu Wang, Yuran Chen, Jiatong Zhang, Yutong Zhang, Yujie Chen, Jiaming Shang, Guang Zhang, Zhuang Liu ·

    PRISM: Probing Reasoning, Instruction, and Source Memory in LLM Hallucinations

    arXiv:2604.16909v2 Announce Type: replace Abstract: As large language models (LLMs) evolve from conversational assistants into agents capable of handling complex tasks, they are increasingly deployed in high-risk domains. However, existing benchmarks largely rely on mixed queries…

  5. arXiv cs.AI TIER_1 · Byeonggeuk Lim, JungMin Yun, Junehyoung Kwon, Kyeonghyun Kim, YoungBin Kim ·

    Aligning with Your Own Voice: Self-Corrected Preference Learning for Hallucination Mitigation in LVLMs

    arXiv:2604.24395v1 Announce Type: new Abstract: Large Vision-Language Models (LVLMs) frequently suffer from hallucinations. Existing preference learning-based approaches largely rely on proprietary models to construct preference datasets. We identify that this reliance introduces…

  6. arXiv cs.AI TIER_1 · YoungBin Kim ·

    Aligning with Your Own Voice: Self-Corrected Preference Learning for Hallucination Mitigation in LVLMs

    Large Vision-Language Models (LVLMs) frequently suffer from hallucinations. Existing preference learning-based approaches largely rely on proprietary models to construct preference datasets. We identify that this reliance introduces a distributional mismatch between the proprieta…

  7. Mastodon — fosstodon.org TIER_1 · [email protected] ·

    KARL: RL Framework Cuts LLM Hallucinations Without Accuracy Loss KARL introduces a reinforcement learning framework that dynamically estimates an LLM's knowledg

    KARL: RL Framework Cuts LLM Hallucinations Without Accuracy Loss KARL introduces a reinforcement learning framework that dynamically estimates an LLM's knowledge boundary to reward abstention only when appropriate, achieving a superior accuracy-hallucination trade- https:// genti…

  8. Mastodon — fosstodon.org TIER_1 · [email protected] ·

    Japan Deploys Unitree G1 Robots at Haneda Airport Amid Labor Shortage Japan is testing Unitree G1 and taller humanoid robots at Tokyo Haneda Airport to tackle i

    Japan Deploys Unitree G1 Robots at Haneda Airport Amid Labor Shortage Japan is testing Unitree G1 and taller humanoid robots at Tokyo Haneda Airport to tackle its labor shortage crisis, marking a real-world deployment of AI-driven robotics. https:// gentic.news/article/japan-depl…