A new research paper published on arXiv investigates the use of large language models (LLMs) as substitutes for human judgment in philosophical contexts. The study found that LLMs tend to over-correlate philosophical positions, creating an artificial consensus and collapsing the natural heterogeneity of human opinions. This effect was observed across both proprietary and open-source models and was partially attributed to models assuming specialists hold uniform views. The findings have implications for AI alignment, evaluation methods, and the reliability of using AI systems to replicate human decision-making. AI
Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →
IMPACT Highlights potential biases in LLMs for replicating human judgment, impacting AI alignment and evaluation.
RANK_REASON Academic paper analyzing LLM behavior on a specific task.