PulseAugur
LIVE 07:29:27
tool · [1 source] ·
3
tool

Many-shot CoT-ICL shows unstable scaling for reasoning tasks

Researchers have investigated the effectiveness of many-shot chain-of-thought in-context learning (CoT-ICL) for reasoning tasks, finding that standard many-shot approaches do not directly translate. Their study revealed that increasing CoT demonstrations can be unstable for non-reasoning models and primarily benefits reasoning-oriented LLMs. The research also indicated that similarity-based retrieval is effective for non-reasoning tasks but not for reasoning, and that performance variance increases with more CoT examples. To address these issues, they propose Curvilinear Demonstration Selection (CDS), an ordering method that improves performance by treating demonstrations as a structured curriculum for in-context test-time learning. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT Reframes in-context learning as test-time learning, suggesting new methods for ordering demonstrations to improve LLM reasoning.

RANK_REASON Academic paper detailing a new method for in-context learning. [lever_c_demoted from research: ic=1 ai=1.0]

Read on arXiv cs.AI →

COVERAGE [1]

  1. arXiv cs.AI TIER_1 · Dit-Yan Yeung ·

    Many-Shot CoT-ICL: Making In-Context Learning Truly Learn

    In-context learning (ICL) adapts large language models (LLMs) to new tasks by conditioning on demonstrations in the prompt without parameter updates. With long-context models, many-shot ICL can use dozens to hundreds of examples and achieve performance comparable to fine-tuning, …