Researchers have developed a post-hoc framework called Conformal Abstention (CA) to help language models determine when they should abstain from answering a query. This method aims to reduce hallucinations by providing finite-sample guarantees on both the likelihood of participation and the correctness of responses. CA utilizes prediction confidence, calibrated by the model's internal representation geometry, to measure knowledge involvement in response generation. Experiments show this approach significantly improves selective answering capabilities, achieving 75 percent conditional correctness. AI
Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →
IMPACT Introduces a method to improve language model reliability by enabling them to admit ignorance, potentially reducing hallucinations and increasing trust in their outputs.
RANK_REASON This is a research paper detailing a new framework for language models.