PulseAugur
LIVE 08:20:17
research · [2 sources] ·
0
research

Study reveals linguistic cues and annotator attitudes impact harmful language detection.

A new paper analyzes annotation variation in NLP datasets, focusing on harmful language detection. The research combines annotator characteristics with linguistic properties of the data to understand labeling discrepancies. Findings indicate that interactions between annotator traits and item features, particularly lexical cues and annotator attitudes, are crucial, but patterns vary significantly across different datasets, cautioning against overgeneralization. AI

Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →

IMPACT Highlights the importance of considering both annotator and data characteristics for reliable NLP model training.

RANK_REASON The cluster contains an academic paper published on arXiv.

Read on arXiv cs.CL →

COVERAGE [2]

  1. arXiv cs.CL TIER_1 · Maximilian Maurer, Maximilian Linde, Gabriella Lapesa ·

    Who and What? Using Linguistic Features and Annotator Characteristics to Analyze Annotation Variation

    arXiv:2605.06318v1 Announce Type: new Abstract: Human label variation has been established as a central phenomenon in NLP: the perspectives different annotators have on the same item need to be embraced. Data collection practices thus shifted towards increasing the annotator numb…

  2. arXiv cs.CL TIER_1 · Gabriella Lapesa ·

    Who and What? Using Linguistic Features and Annotator Characteristics to Analyze Annotation Variation

    Human label variation has been established as a central phenomenon in NLP: the perspectives different annotators have on the same item need to be embraced. Data collection practices thus shifted towards increasing the annotator numbers and releasing disaggregated datasets, harmfu…