PulseAugur
LIVE 03:41:53
research · [2 sources] ·
0
research

Researchers release Faithfulness-QA dataset to train context-faithful RAG models

Researchers have developed Faithfulness-QA, a new dataset containing nearly 100,000 samples designed to train Retrieval-Augmented Generation (RAG) models to prioritize retrieved context over their internal knowledge. The dataset was created by systematically replacing named entities in existing question-answering benchmarks with alternatives, thereby generating conflicts between context and parametric memory. This resource aims to improve the faithfulness of RAG systems and provides a benchmark for evaluating their context-grounding capabilities. AI

Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →

IMPACT Improves RAG model faithfulness by providing a dataset to train context-grounding capabilities.

RANK_REASON Release of a new dataset for training and evaluating RAG models.

Read on arXiv cs.CL →

COVERAGE [2]

  1. arXiv cs.CL TIER_1 · Li Ju, Junzhe Wang, Qi Zhang ·

    Faithfulness-QA: A Counterfactual Entity Substitution Dataset for Training Context-Faithful RAG Models

    arXiv:2604.25313v1 Announce Type: new Abstract: Retrieval-Augmented Generation (RAG) models frequently produce answers grounded in parametric memory rather than the retrieved context, undermining the core promise of retrieval augmentation. A fundamental obstacle to fixing this un…

  2. arXiv cs.CL TIER_1 · Qi Zhang ·

    Faithfulness-QA: A Counterfactual Entity Substitution Dataset for Training Context-Faithful RAG Models

    Retrieval-Augmented Generation (RAG) models frequently produce answers grounded in parametric memory rather than the retrieved context, undermining the core promise of retrieval augmentation. A fundamental obstacle to fixing this unfaithfulness is the lack of training data that e…