PulseAugur
LIVE 23:16:08
research · [383 sources] ·
0
research

Google and OpenAI advance AI factuality, multilingualism, and safety

Google DeepMind has introduced the FACTS Benchmark Suite, a new set of evaluations designed to systematically assess the factuality of large language models across various use cases. This suite includes benchmarks for parametric knowledge, search-based information retrieval, and multimodal understanding, alongside an updated grounding benchmark. The initiative aims to provide a more comprehensive measure of LLM accuracy and is being launched with a public leaderboard on Kaggle to track progress across leading models. AI

Summary written by gemini-2.5-flash-lite from 383 sources. How we write summaries →

IMPACT Establishes a new standard for evaluating LLM factuality, potentially driving improvements in model reliability and trustworthiness.

RANK_REASON This is a research paper introducing a new benchmark suite for evaluating LLMs.

Read on OpenAI News →

Google and OpenAI advance AI factuality, multilingualism, and safety

COVERAGE [383]

  1. Google AI / Research TIER_1 ·

    ATLAS: Practical scaling laws for multilingual models

    Generative AI

  2. Google DeepMind TIER_1 ·

    FACTS Benchmark Suite: Systematically evaluating the factuality of large language models

    Systematically evaluating the factuality of large language models with the FACTS Benchmark Suite.

  3. Google AI / Research TIER_1 ·

    Simulating large systems with Regression Language Models

    Generative AI

  4. OpenAI News TIER_1 Dansk(DA) ·

    Deliberative alignment: reasoning enables safer language models

    Deliberative alignment: reasoning enables safer language models Introducing our new alignment strategy for o1 models, which are directly taught safety specifications and how to reason over them.

  5. Google DeepMind TIER_1 ·

    FACTS Grounding: A new benchmark for evaluating the factuality of large language models

    Our comprehensive benchmark and online leaderboard offer a much-needed measure of how accurately LLMs ground their responses in provided source material and avoid hallucinations

  6. OpenAI News TIER_1 ·

    Prover-Verifier Games improve legibility of language model outputs

    Discover how prover-verifier games improve the legibility of language model outputs, making AI solutions clearer, easier to verify, and more trustworthy for both humans and machines.

  7. OpenAI News TIER_1 ·

    Efficient training of language models to fill in the middle

  8. OpenAI News TIER_1 (CA) ·

    Best practices for deploying language models

    Cohere, OpenAI, and AI21 Labs have developed a preliminary set of best practices applicable to any organization developing or deploying large language models.

  9. OpenAI News TIER_1 ·

    Aligning language models to follow instructions

  10. OpenAI News TIER_1 ·

    WebGPT: Improving the factual accuracy of language models through web browsing

    We’ve fine-tuned GPT-3 to more accurately answer open-ended questions using a text-based web browser.

  11. OpenAI News TIER_1 ·

    Evaluating large language models trained on code

  12. OpenAI News TIER_1 ·

    Improving language model behavior by training on a curated dataset

    Our latest research finds we can improve language model behavior with respect to specific behavioral values by fine-tuning on a small, curated dataset.

  13. OpenAI News TIER_1 ·

    Understanding the capabilities, limitations, and societal impact of large language models

  14. OpenAI News TIER_1 ·

    Scaling laws for neural language models

  15. OpenAI News TIER_1 ·

    Better language models and their implications

    We’ve trained a large-scale unsupervised language model which generates coherent paragraphs of text, achieves state-of-the-art performance on many language modeling benchmarks, and performs rudimentary reading comprehension, machine translation, question answering, and summarizat…

  16. Apple Machine Learning Research TIER_1 ·

    Adaptive Thinking: Large Language Models Know When to Think in Latent Space

    Recent advances in large language models (LLMs) test-time computing have introduced the capability to perform intermediate chain-of-thought (CoT) reasoning (thinking) before generating answers. While increasing the thinking budget yields smooth performance improvements at inferen…

  17. Microsoft Research TIER_1 · Sidharth Sinha, Anson Bastos, Xuchao Zhang, Akshay Nambi, Rujia Wang, Chetan Bansal ·

    AutoAdapt: Automated domain adaptation for large language models

    <p>Deploying large language models (LLMs) in real-world, high-stakes settings is harder than it should be. In high-stakes settings like law, medicine, and cloud incident response, performance and reliability can quickly break down because adapting models to domain-specific requir…

  18. Hugging Face Blog TIER_1 ·

    Falcon 2: An 11B parameter pretrained language model and VLM, trained on over 5000B tokens and 11 languages

  19. Hugging Face Blog TIER_1 ·

    Red-Teaming Large Language Models

  20. Hugging Face Blog TIER_1 ·

    Very Large Language Models and How to Evaluate Them

  21. Hugging Face Blog TIER_1 ·

    Introducing The World's Largest Open Multilingual Language Model: BLOOM

  22. Hugging Face Blog TIER_1 ·

    BERT 101 - State Of The Art NLP Model Explained

  23. Hugging Face Blog TIER_1 ·

    Scaling up BERT-like model Inference on modern CPU - Part 2

  24. Hugging Face Blog TIER_1 ·

    Large Language Models: A New Moore's Law?

  25. Hugging Face Blog TIER_1 ·

    Scaling-up BERT Inference on CPU (Part 1)

  26. Hugging Face Blog TIER_1 ·

    Block Sparse Matrices for Smaller and Faster Language Models

  27. Hugging Face Blog TIER_1 ·

    The Reformer - Pushing the limits of language modeling

  28. 量子位 (QbitAI) TIER_1 中文(ZH) · henry ·

    He Kaiming's First Language Model: 105M Parameters, Not Following GPT's Autoregressive Path

    顶级CV大佬也来卷LLM?

  29. arXiv cs.AI TIER_1 · Alborz Geramifard ·

    Beyond GRPO and On-Policy Distillation: An Empirical Sparse-to-Dense Reward Principle for Language-Model Post-Training

    In settings where labeled verifiable training data is the binding constraint, each checked example should be allocated carefully. The standard practice is to use this data directly on the model that will be deployed, for example by running GRPO on the deployment student. We argue…

  30. arXiv cs.CL TIER_1 · Jonas Geiping ·

    Multi-Stream LLMs: Unblocking Language Models with Parallel Streams of Thoughts, Inputs and Outputs

    The continued improvements in language model capability have unlocked their widespread use as drivers of autonomous agents, for example in coding or computer use applications. However, the core of these systems has not changed much since early instruction-tuned models like ChatGP…

  31. arXiv cs.CL TIER_1 · Chao Chen ·

    ORCE: Order-Aware Alignment of Verbalized Confidence in Large Language Models

    Large language models (LLMs) often produce answers with high certainty even when they are incorrect, making reliable confidence estimation essential for deployment in real-world scenarios. Verbalized confidence, where models explicitly state their confidence in natural language, …

  32. arXiv cs.AI TIER_1 · Eric de la Clergerie ·

    A Causal Language Modeling Detour Improves Encoder Continued Pretraining

    When adapting an encoder to a new domain, the standard approach is to continue training with Masked Language Modeling (MLM). We show that temporarily switching to Causal Language Modeling (CLM) followed by a short MLM decay improves downstream performance. On biomedical texts wit…

  33. arXiv cs.CL TIER_1 · Adam Jatowt ·

    Question Difficulty Estimation for Large Language Models via Answer Plausibility Scoring

    Estimating question difficulty is a critical component in evaluating and improving large language models (LLMs) for question answering (QA). Existing approaches often rely on readability formulas, retrieval-based signals, or popularity statistics, which may not fully capture the …

  34. arXiv cs.AI TIER_1 · Soujanya Poria ·

    $δ$-mem: Efficient Online Memory for Large Language Models

    Large language models increasingly need to accumulate and reuse historical information in long-term assistants and agent systems. Simply expanding the context window is costly and often fails to ensure effective context utilization. We propose $δ$-mem, a lightweight memory mechan…

  35. arXiv cs.LG TIER_1 · Dan Alistarh ·

    Grid Games: The Power of Multiple Grids for Quantizing Large Language Models

    A major recent advance in quantization is given by microscaled 4-bit formats such as NVFP4 and MXFP4, quantizing values into small groups sharing a scale, assuming a fixed floating-point grid. In this paper, we study the following natural extension: assume that, for each group of…

  36. arXiv cs.AI TIER_1 · Yazhe Niu ·

    PriorZero: Bridging Language Priors and World Models for Decision Making

    Leveraging the rich world knowledge of Large Language Models (LLMs) to enhance Reinforcement Learning (RL) agents offers a promising path toward general intelligence. However, a fundamental prior-dynamics mismatch hinders existing approaches: static LLM knowledge cannot directly …

  37. arXiv cs.CL TIER_1 · Mahesh Viswanathan ·

    Correcting Selection Bias in Sparse User Feedback for Large Language Model Quality Estimation: A Multi-Agent Hierarchical Bayesian Approach

    [Abridged] Production LLM deployments receive feedback from a non-random fraction of users: thumbs sit mostly in the tails of the satisfaction distribution, and a naive average over them can land 40-50 percentage points away from true system quality. We treat this as a topic- and…

  38. arXiv cs.AI TIER_1 · Yong-eun Cho ·

    It's Not the Size: Harness Design Determines Operational Stability in Small Language Models

    This paper experimentally analyzes how the level of harness engineering affects the operational performance of small language models (SLMs, 2-3B parameters). Three harness conditions - model-only (raw prompt), minimal-shell (wrapper tags), and a 4-stage pipeline (plan->execute->v…

  39. arXiv cs.AI TIER_1 · Andrew M. Bean ·

    To Whom Do Language Models Align? Measuring Principal Hierarchies Under High-Stakes Competing Demands

    Language models deployed in high-stakes professional settings face conflicting demands from users, institutional authorities, and professional norms. How models act when these demands conflict reveals a principal hierarchy -- an implicit ordering over competing stakeholders that …

  40. arXiv cs.AI TIER_1 · Yong Liu ·

    Large Language Models as Amortized Pareto-Front Generators for Constrained Bi-Objective Convex Optimization

    Generating feasible Pareto fronts for constrained bi-objective continuous optimization is central to multi-criteria decision-making. Existing methods usually rely on iterative scalarization, evolutionary search, or problem-specific solvers, requiring repeated optimization for eac…

  41. arXiv cs.CL TIER_1 · Sebastian Padó ·

    Do Language Models Encode Knowledge of Linguistic Constraint Violations?

    Large Language Models (LLMs) achieve strong linguistic performance, yet their internal mechanisms for producing these predictions remain unclear. We investigate the hypothesis that LLMs encode representations of linguistic constraint violations within their parameters, which are …

  42. arXiv cs.CL TIER_1 · Jingren Zhou ·

    Qwen-Scope: Turning Sparse Features into Development Tools for Large Language Models

    Large language models have achieved remarkable capabilities across diverse tasks, yet their internal decision-making processes remain largely opaque, limiting our ability to inspect, control, and systematically improve them. This opacity motivates a growing body of research in me…

  43. arXiv cs.CL TIER_1 · Sarath Chandar ·

    Probabilistic Calibration Is a Trainable Capability in Language Models

    Language models are increasingly used in settings where outputs must satisfy user-specified randomness constraints, yet their generation probabilities are often poorly calibrated to those targets. We study whether this capability can be improved directly through fine-tuning. Conc…

  44. arXiv cs.AI TIER_1 · Kaiming He ·

    ELF: Embedded Language Flows

    Diffusion and flow-based models have become the de facto approaches for generating continuous data, e.g., in domains such as images and videos. Their success has attracted growing interest in applying them to language modeling. Unlike their image-domain counterparts, today's lead…

  45. arXiv cs.CL TIER_1 · Mohamed S. Abdelfattah ·

    Compute Where it Counts: Self Optimizing Language Models

    Efficient LLM inference research has largely focused on reducing the cost of each decoding step (e.g., using quantization, pruning, or sparse attention), typically applying a uniform computation budget to every generated token. In practice, token difficulty varies widely, so stat…

  46. arXiv cs.AI TIER_1 · Long Tran-Thanh ·

    Training-Free Cultural Alignment of Large Language Models via Persona Disagreement

    Large language models increasingly mediate decisions that turn on moral judgement, yet a growing body of evidence shows that their implicit preferences are not culturally neutral. Existing cultural alignment methods either require per-country preference data and fine-tuning budge…

  47. arXiv cs.LG TIER_1 · Omer Reingold ·

    Mistake-Bounded Language Generation

    We investigate the learning task of language generation in the limit, but shift focus from the traditional time-of-last-mistake metric of a generator's success to a new notion of "mistake-bounded generation." While existing results for language generation in the limit focus on gu…

  48. arXiv cs.AI TIER_1 · Peter West ·

    Can You Keep a Secret? Involuntary Information Leakage in Language Model Writing

    Language models are deployed in settings that require compartmentalization: system prompts should not be disclosed, chain-of-thought reasoning is hidden from users, and sensitive data passes through shared contexts. We test whether models can keep prompted information out of thei…

  49. arXiv cs.CL TIER_1 · Zanmin Wang ·

    A Single-Layer Model Can Do Language Modeling

    Modern language models scale depth by stacking layers, each holding its own state - a per-layer KV cache in transformers, a per-layer matrix in Mamba, Gated DeltaNet (GDN), RWKV, and xLSTM. Biological systems lean heavily on recurrence rather than on stacking. We ask how far that…

  50. arXiv cs.AI TIER_1 · Jun Xu ·

    Towards Understanding Continual Factual Knowledge Acquisition of Language Models: From Theory to Algorithm

    Continual Pre-Training (CPT) is essential for enabling Language Models (LMs) to integrate new knowledge without erasing old. While classical CPT techniques like data replay have become the standard paradigm, the mechanisms underlying how LMs acquire and retain facts over time, te…

  51. arXiv cs.CL TIER_1 · Pavlos Fafalios ·

    Can Language Models Analyze Data? Evaluating Large Language Models for Question Answering over Datasets

    This paper investigates the effectiveness of large language models (LLMs) in answering questions over datasets. We examine their performance in two scenarios: (a) directly answering questions given a dataset file as input, and (b) generating SQL queries to answer questions given …

  52. arXiv cs.CL TIER_1 · Qingqiang Wu ·

    ANCHOR: Abductive Network Construction with Hierarchical Orchestration for Reliable Probability Inference in Large Language Models

    A central challenge in large-scale decision-making under incomplete information is estimating reliable probabilities. Recent approaches leverage Large Language Models (LLMs) to generate explanatory factors and elicit coarse-grained probability estimates. Typically, an LLM perform…

  53. arXiv cs.CL TIER_1 · Qipeng Guo ·

    Synthetic Pre-Pre-Training Improves Language Model Robustness to Noisy Pre-Training Data

    Large language models (LLMs) rely on web-scale corpora for pre-training. The noise inherent in these datasets tends to obscure meaningful patterns and ultimately degrade model performance. Data curation mitigates but cannot eliminate such noise, so pre-training corpora remain noi…

  54. arXiv cs.AI TIER_1 · Bo Yan ·

    MicroWorld: Empowering Multimodal Large Language Models to Bridge the Microscopic Domain Gap with Multimodal Attribute Graph

    Multimodal large language models (MLLMs) show remarkable potential for scientific reasoning, yet their performance in specialized domains such as microscopy remains limited by the scarcity of domain-specific training data and the difficulty of encoding fine-grained expert knowled…

  55. arXiv cs.AI TIER_1 · Yinglun Zhu ·

    Active Testing of Large Language Models via Approximate Neyman Allocation

    Large language models (LLMs) require reliable evaluation from pre-training to test-time scaling, making evaluation a recurring rather than one-off cost. As model scales grow and target tasks increasingly demand expert annotators, both the compute and labeling costs needed for eac…

  56. arXiv cs.CL TIER_1 · Sietse Schelpe ·

    Merlin: Deterministic Byte-Exact Deduplication for Lossless Context Optimization in Large Language Model Inference

    Data-intensive applications, ranging from large-scale retrieval systems to advanced data pipelines, are increasingly bottlenecked by the processing of highly redundant text corpora. We present Merlin, a local-first, agnostic, high-throughput deduplication and context optimization…

  57. arXiv cs.CL TIER_1 · Xinlong Huang ·

    Evolving Knowledge Distillation for Lightweight Neural Machine Translation

    Recent advancements in Neural Machine Translation (NMT) have significantly improved translation quality. However, the increasing size and complexity of state-of-the-art models present significant challenges for deployment on resource-limited devices. Knowledge distillation (KD) i…

  58. arXiv cs.CL TIER_1 · Zhiyuan Su ·

    Beyond Language: Format-Agnostic Reasoning Subspaces in Large Language Models

    Large language models represent the same reasoning in vastly different surface forms -- English prose, Python code, mathematical notation -- yet whether they share a common internal substrate across these symbolic systems remains unknown. We introduce the TriForm Benchmark (18 co…

  59. arXiv cs.CL TIER_1 · Jiaji Zhong ·

    APCD: Adaptive Path-Contrastive Decoding for Reliable Large Language Model Generation

    Large language models (LLMs) often suffer from hallucinations due to error accumulation in autoregressive decoding, where suboptimal early token choices misguide subsequent generation. Although multi-path decoding can improve robustness by exploring alternative trajectories, exis…

  60. arXiv cs.AI TIER_1 · Shervin Malmasi ·

    Beyond Pairs: Your Language Model is Secretly Optimizing a Preference Graph

    Direct Preference Optimization (DPO) aligns language models using pairwise preference comparisons, offering a simple and effective alternative to Reinforcement Learning (RL) from human feedback. However, in many practical settings, training data consists of multiple rollouts per …

  61. arXiv cs.AI TIER_1 · Maria Perez-Ortiz ·

    Tool Calling is Linearly Readable and Steerable in Language Models

    When a tool-calling agent picks the wrong tool, the failure is invisible until execution: the email gets sent, the meeting gets missed. Probing 12 instruction-tuned models across Gemma 3, Qwen 3, Qwen 2.5, and Llama 3.1 (270M to 27B), we find the identity of the chosen tool is li…

  62. arXiv cs.AI TIER_1 · Nick Rui ·

    Where's the Plan? Locating Latent Planning in Language Models with Lightweight Mechanistic Interventions

    We study planning site formation in language models -- where internal representations of structurally-constrained future tokens form during the forward pass, and whether they causally drive generation. Using rhyming-couplet completion as a clean test of forward-looking constraint…

  63. arXiv cs.CL TIER_1 · Xin Geng ·

    Chain-based Distillation for Effective Initialization of Variable-Sized Small Language Models

    Large language models (LLMs) achieve strong performance but remain costly to deploy in resource-constrained settings. Training small language models (SLMs) from scratch is computationally expensive, while conventional knowledge distillation requires repeated access to large teach…

  64. arXiv cs.CL TIER_1 · Wangmeng Zuo ·

    TextLDM: Language Modeling with Continuous Latent Diffusion

    Diffusion Transformers (DiT) trained with flow matching in a VAE latent space have unified visual generation across images and videos. A natural next step toward a single architecture for both generation (visual synthesis) and understanding (text generation) is to apply this fram…

  65. arXiv cs.AI TIER_1 · Fabio Valerio Massoli ·

    Memory-Efficient Looped Transformer: Decoupling Compute from Memory in Looped Language Models

    Recurrent LLM architectures have emerged as a promising approach for improving reasoning, as they enable multi-step computation in the embedding space without generating intermediate tokens. Models such as Ouro perform reasoning by iteratively updating internal representations wh…

  66. arXiv cs.CL TIER_1 · Yves Scherrer ·

    Why do Large Language Models Fail in Low-resource Translation? Unraveling the Token Dynamics of Large Language Models for Machine Translation

    Large Language Models (LLMs) have recently demonstrated strong performance in machine translation (MT). However, most prior work focuses on improving or benchmarking translation quality, offering limited insight into when and why LLM-based translation fails. In this work, we syst…

  67. arXiv cs.CL TIER_1 · Xin Geng ·

    Understanding Performance Collapse in Layer-Pruned Large Language Models via Decision Representation Transitions

    Layer pruning efficiently reduces Large Language Model (LLM) computational costs but often triggers sudden performance collapse. Existing representation-based analyses struggle to explain this mechanism. We propose studying pruning through decision representation. Focusing on mul…

  68. arXiv cs.LG TIER_1 · Stanislav Budzinskiy, Marian Gloser, Tolunay Yilmaz, Ying Hong Tham, Yuanyi Lin, Wenyi Fang, Fan Wu, Philipp Petersen ·

    LAMP: Look-Ahead Mixed-Precision Inference of Large Language Models

    arXiv:2601.21623v2 Announce Type: replace Abstract: Mixed-precision computations are a hallmark of the current stage of AI, driving the progress in large language models towards efficient, locally deployable solutions. This article addresses the floating-point computation of comp…

  69. arXiv cs.LG TIER_1 · Beicheng Xu, Weitong Qian, Lingching Tung, Yupeng Lu, Bin Cui ·

    Tree-Structured Synergy of Large Language Models and Bayesian Optimization for Efficient CASH

    arXiv:2601.12355v2 Announce Type: replace Abstract: To lower the expertise barrier in machine learning, the AutoML community has focused on the CASH problem, which jointly automates algorithm selection and hyperparameter tuning. While traditional methods like Bayesian Optimizatio…

  70. arXiv cs.LG TIER_1 · Yuanming Zhang, Yan Lin, Arijit Khan, Huaiyu Wan ·

    Large Language Model Prompt Datasets: An In-depth Analysis and Insights

    arXiv:2510.09316v2 Announce Type: replace Abstract: We compile 129 heterogeneous LLM prompt datasets (>1.22 TB, >673M instances) into a structured taxonomy and conduct a multi-level linguistic analysis (lexical, syntactic, and semantic) on seven representative corpora, surfacing …

  71. arXiv cs.AI TIER_1 · Pedro Orvalho, Marta Kwiatkowska ·

    Are Large Language Models Robust in Understanding Code Against Semantics-Preserving Mutations?

    arXiv:2505.10443v3 Announce Type: replace-cross Abstract: With the widespread adoption of vibe coding, understanding the reasoning and robustness of Large Language Models (LLMs) is critical for their reliable use in programming tasks. While recent studies assess LLMs' ability to …

  72. arXiv cs.AI TIER_1 · Jazmia Henry ·

    Beyond Static Snapshots: A Grounded Evaluation Framework for Language Models at the Agentic Frontier

    arXiv:2604.17573v2 Announce Type: replace Abstract: We argue that current evaluation frameworks for large language models (LLMs) suffer from four systematic failures that make them structurally inadequate for deployed, agentic systems: distributional, temporal, scope, and process…

  73. arXiv cs.LG TIER_1 · Jiwan Chung, Saejin Kim, Yongrae Jo, Jaewoo Park, Dongjun Min, Youngjae Yu ·

    Teaching Metric Distance to Discrete Autoregressive Language Models

    arXiv:2503.02379v5 Announce Type: replace Abstract: Large language models (LLMs) operate as autoregressive predictors over discrete token vocabularies, a formulation that has enabled their adaptation far beyond natural language to vision, robotics, and multimodal reasoning. Howev…

  74. arXiv cs.LG TIER_1 · Yiqiao Jin, Yiyang Wang, Lucheng Fu, Yijia Xiao, Yinyi Luo, Haoxin Liu, B. Aditya Prakash, Josiah Hester, Jindong Wang, Srijan Kumar ·

    UniSD: Towards a Unified Self-Distillation Framework for Large Language Models

    arXiv:2605.06597v1 Announce Type: cross Abstract: Self-distillation (SD) offers a promising path for adapting large language models (LLMs) without relying on stronger external teachers. However, SD in autoregressive LLMs remains challenging because self-generated trajectories are…

  75. arXiv cs.LG TIER_1 Dansk(DA) · Yuping Lin, Pengfei He, Yue Xing, Yingqian Cui, Jiayuan Ding, Subhabrata Mukherjee, Hui Liu, Zhen Xiang ·

    Crafting Reversible SFT Behaviors in Large Language Models

    arXiv:2605.06632v1 Announce Type: new Abstract: Supervised fine-tuning (SFT) induces new behaviors in large language models, yet imposes no structural constraint on how these behaviors are distributed within the model. Existing behavior interpretation methods, such as circuit att…

  76. arXiv cs.LG TIER_1 · Agnibh Dasgupta, Abdullah Tanvir, Xin Zhong ·

    Invariant Features in Language Models: Geometric Characterization and Model Attribution

    arXiv:2605.06458v1 Announce Type: new Abstract: Language models exhibit strong robustness to paraphrasing, suggesting that semantic information may be encoded through stable internal representations, yet the structure and origin of such invariance remain unclear. We propose a loc…

  77. arXiv cs.LG TIER_1 · Masatsugu Yamada, Mahito Sugiyama ·

    When Graph Language Models Go Beyond Memorization

    arXiv:2605.06239v1 Announce Type: new Abstract: It remains unclear whether graph language models learn structural regularities or merely memorize training graphs; this cannot be resolved by current aggregate fidelity metrics alone. We develop a calibrated diagnostic protocol that…

  78. arXiv cs.LG TIER_1 · Muhammad Shahir Abdurrahman, Chun Deng, Azalia Mirhoseini, Philip Levis ·

    Federation of Experts: Communication Efficient Distributed Inference for Large Language Models

    arXiv:2605.06206v1 Announce Type: new Abstract: Mixture of experts has emerged as the primary mechanism for making Large Language Models (LLMs) computationally efficient. However, in distributed settings, communicating token embeddings between experts is a significant bottleneck.…

  79. arXiv cs.LG TIER_1 · Mingcheng Zhu, Yu Liu, Tingting Zhu ·

    Towards Generation-Efficient Uncertainty Estimation in Large Language Models

    arXiv:2605.06053v1 Announce Type: new Abstract: Uncertainty estimation is important for deploying LLMs in high-stakes applications such as healthcare and finance, where hallucinations can appear fluent and plausible while being factually incorrect, making it difficult for users t…

  80. arXiv cs.LG TIER_1 · Hua-Dong Xiong ·

    Hypothesis generation and updating in large language models

    arXiv:2605.05851v1 Announce Type: new Abstract: Large language models (LLMs) increasingly help people solve problems, from debugging code to repairing machinery. This process requires generating plausible hypotheses from partial descriptions, then updating them as more informatio…

  81. arXiv cs.LG TIER_1 · Yiwei Zhang, Jeremiah Birrell, Reza Ebrahimi, Rouzbeh Behnia, Jason Pacheco, Elisa Bertino ·

    Information Theoretic Adversarial Training of Large Language Models

    arXiv:2605.05415v1 Announce Type: new Abstract: Large language models (LLMs) remain vulnerable to adversarial prompting despite advances in alignment and safety, often exhibiting harmful behaviors under novel attack strategies. While adversarial training can improve robustness, e…

  82. arXiv cs.CL TIER_1 · Bing Wang, Ximing Li, Changchun Li, Jinjin Chi, Gang Niu, Masashi Sugiyama ·

    Decomposing the Basic Abilities of Large Language Models: Mitigating Cross-Task Interference in Multi-Task Instruct-Tuning

    arXiv:2605.05676v1 Announce Type: new Abstract: Recently, the prominent performance of large language models (LLMs) has been largely driven by multi-task instruct-tuning. Unfortunately, this training paradigm suffers from a key issue, named cross-task interference, due to conflic…

  83. arXiv cs.LG TIER_1 (CA) · Akhil Jindal, Harang Ju ·

    SMolLM: Small Language Models Learn Small Molecular Grammar

    arXiv:2605.06322v1 Announce Type: new Abstract: Language models for molecular design have scaled to hundreds of millions of parameters, yet how they learn chemical grammar is poorly understood. We train SMolLM, a 53K-parameter weight-shared transformer, to generate novel SMILES w…

  84. arXiv cs.CL TIER_1 · Sohan Venkatesh ·

    Negative Before Positive: Asymmetric Valence Processing in Large Language Models

    arXiv:2605.05653v1 Announce Type: new Abstract: Mechanistic interpretability has revealed how concepts are encoded in large language models (LLMs), but emotional content remains poorly understood at the mechanistic level. We study whether LLMs process emotional valence through de…

  85. arXiv cs.CL TIER_1 · Fabrice Harel-Canada, Amit Sahai ·

    SLAM: Structural Linguistic Activation Marking for Language Models

    arXiv:2605.05443v1 Announce Type: new Abstract: LLM watermarks must be detectable without compromising text quality, yet most existing schemes bias the next-token distribution and pay for detection with measurable quality loss. We present SLAM (Structural Linguistic Activation Ma…

  86. arXiv cs.LG TIER_1 · Tony Mason, Vaastav Anand ·

    Epistemic Observability in Language Models

    arXiv:2603.20531v2 Announce Type: replace-cross Abstract: We find that models report highest confidence precisely when they are fabricating. Across four model families (OLMo-3, Llama-3.1, Qwen3, Mistral), self-reported confidence inversely correlates with accuracy, with AUC rangi…

  87. arXiv cs.LG TIER_1 · Sayantan Dasgupta, Trevor Cohn, Timothy Baldwin ·

    Don't Ignore the Tail: Decoupling top-K Probabilities for Efficient Language Model Distillation

    arXiv:2602.20816v2 Announce Type: replace-cross Abstract: The core learning signal used in language model distillation is the standard Kullback-Leibler (KL) divergence between the student and teacher distributions. Traditional KL divergence tends to be dominated by the next token…

  88. arXiv cs.LG TIER_1 · Reza T. Batley, Sourav Saha ·

    Leviathan: Decoupling Input and Output Representations in Language Models

    arXiv:2601.22040v2 Announce Type: replace-cross Abstract: Modern language models use a single matrix for input embedding and output projection. This couples two distinct objectives: token representation and discrimination over a vocabulary. This work introduces Leviathan, a Trans…

  89. arXiv cs.LG TIER_1 · Xiaoyu Xu, Minxin Du, Kun Fang, Yaxin Xiao, Zhicong Huang, Cheng Hong, Qingqing Ye, Haibo Hu ·

    FIT to Forget: Robust Continual Unlearning for Large Language Models

    arXiv:2601.21682v2 Announce Type: replace-cross Abstract: While large language models (LLMs) exhibit remarkable capabilities, they increasingly face demands to unlearn memorized privacy-sensitive, copyrighted, or harmful content. Existing unlearning methods primarily focus on \em…

  90. arXiv cs.CL TIER_1 · Lin Yao ·

    Remask, Don't Replace: Token-to-Mask Refinement in Diffusion Large Language Models

    arXiv:2604.18738v2 Announce Type: replace Abstract: Diffusion large language models (dLLMs) gain speed by committing multiple tokens in parallel at each denoising step, but any erroneous commitment persists as conditioning context and biases every subsequent prediction. LLaDA2.1 …

  91. arXiv cs.CL TIER_1 · Omnilingual MT Team, Belen Alastruey, Niyati Bafna, Andrea Caciolai, Kevin Heffernan, Artyom Kozhevnikov, Christophe Ropers, Eduardo S\'anchez, Charles-Eric Saint-James, Ioannis Tsiamas, Xiang "Tony" Cao, Chierh Cheng, Joe Chuang, Paul-Ambroise Duquenne, ·

    Omnilingual MT: Machine Translation for 1,600 Languages

    arXiv:2603.16309v3 Announce Type: replace Abstract: High-quality machine translation (MT) can scale to hundreds of languages, setting a high bar for multilingual systems. However, compared to the world's 7,000 languages, current systems still offer only limited coverage: about 20…

  92. arXiv cs.CL TIER_1 · Messi H. J. Lee ·

    Token-Level Entropy Reveals Demographic Disparities in Language Models

    arXiv:2501.19337v3 Announce Type: replace Abstract: We ask whether demographic identity, signaled by a name alone, systematically reshapes the generative distribution of a language model. Measuring full-vocabulary Shannon entropy at temperature zero across six open-weight base mo…

  93. arXiv cs.AI TIER_1 · Yuhang Wang, Wenjie Mei, Junkai Zhang, Guangyu He, Zhenxing Niu, Haichang Gao ·

    ICU-Bench:Benchmarking Continual Unlearning in Multimodal Large Language Models

    arXiv:2605.05938v1 Announce Type: new Abstract: Although Multimodal Large Language Models (MLLMs) have achieved remarkable progress across many domains, their training on large-scale multimodal datasets raises serious privacy concerns, making effective machine unlearning increasi…

  94. arXiv cs.AI TIER_1 · Xiaomin Li, Andrzej Banburski-Fahey, Jaron Lanier ·

    DataDignity: Training Data Attribution for Large Language Models

    arXiv:2605.05687v1 Announce Type: new Abstract: Auditing language-model outputs often requires more than judging correctness: an auditor may need to identify which source document most likely supports the knowledge expressed in a response. We study this as pinpoint provenance: gi…

  95. arXiv cs.LG TIER_1 Dansk(DA) · Zhen Xiang ·

    Crafting Reversible SFT Behaviors in Large Language Models

    Supervised fine-tuning (SFT) induces new behaviors in large language models, yet imposes no structural constraint on how these behaviors are distributed within the model. Existing behavior interpretation methods, such as circuit attribution approaches, identify sparse subnetworks…

  96. arXiv cs.AI TIER_1 · Srijan Kumar ·

    UniSD: Towards a Unified Self-Distillation Framework for Large Language Models

    Self-distillation (SD) offers a promising path for adapting large language models (LLMs) without relying on stronger external teachers. However, SD in autoregressive LLMs remains challenging because self-generated trajectories are free-form, correctness is task-dependent, and pla…

  97. arXiv cs.CL TIER_1 · Xin Zhong ·

    Invariant Features in Language Models: Geometric Characterization and Model Attribution

    Language models exhibit strong robustness to paraphrasing, suggesting that semantic information may be encoded through stable internal representations, yet the structure and origin of such invariance remain unclear. We propose a local geometric framework in which semantically equ…

  98. arXiv cs.LG TIER_1 · Buu Phan, Ashish Khisti, Karen Ullrich ·

    Cross-Tokenizer Likelihood Scoring Algorithms for Language Model Distillation

    arXiv:2512.14954v2 Announce Type: replace-cross Abstract: Computing next-token likelihood ratios between two language models (LMs) is a standard task in training paradigms such as knowledge distillation. Since this requires both models to share the same probability space, it beco…

  99. arXiv cs.LG TIER_1 · Shu-Hao Zhang, Le-Tong Huang, Xiang-Sheng Deng, Xin-Yi Zou, Chen Wu, Nan Li, Shao-Qun Zhang ·

    EdgeRazor: A Lightweight Framework for Large Language Models via Mixed-Precision Quantization-Aware Distillation

    arXiv:2605.04062v1 Announce Type: new Abstract: Recent years have witnessed an increasing interest in deploying LLMs on resource-constrained devices, among which quantization has emerged as a promising lightweight technique that converts full-precision model weights and activatio…

  100. arXiv cs.LG TIER_1 · Tarun Kathuria, Sachin Kumar ·

    Leveraging Pretrained Language Models as Energy Functions for Glauber Dynamics Text Diffusion

    arXiv:2605.04291v1 Announce Type: new Abstract: We present a discrete diffusion-based language model using Glauber dynamics from statistical physics. Our main insight is that instead of trying to train a discrete state space diffusion model using Glauber dynamics with a uniform t…

  101. arXiv cs.LG TIER_1 · Christine Ye, Joe Benton ·

    Efficiently Aligning Language Models with Online Natural Language Feedback

    arXiv:2605.04356v1 Announce Type: new Abstract: Reinforcement learning with verifiable rewards has been used to elicit impressive performance from language models in many domains. But, broadly beneficial deployments of AI may require us to train models with strong capabilities in…

  102. arXiv cs.LG TIER_1 · Huatian Zhang, Zhendong Mao, Lei Zhang, Yongdong Zhang ·

    Uncertainty-Aware Exploratory Direct Preference Optimization for Multimodal Large Language Models

    arXiv:2605.04874v1 Announce Type: new Abstract: Direct Preference Optimization (DPO) has proven to be an effective solution for mitigating hallucination in Multimodal Large Language Models (MLLMs) by learning from preference pairs. One of its key challenges lies in how to transfe…

  103. arXiv cs.LG TIER_1 · Albert F. Modenbach ·

    A geometric relation of the error introduced by sampling a language model's output distribution to its internal state

    arXiv:2605.04899v1 Announce Type: new Abstract: GPT-style language models are sensitive to single-token changes at generation points where the predicted probability distribution is spread across multiple tokens. Viewing this sensitivity as a geometric property, we derive an $\mat…

  104. arXiv cs.LG TIER_1 · Zetai Cen, Jin Zhu, Xinwei Shen, Chengchun Shi ·

    Perturbation is All You Need for Extrapolating Language Models

    arXiv:2605.04344v1 Announce Type: cross Abstract: We introduce a simple yet powerful framework for training large language models. In contrast to the standard autoregressive next-token prediction based on an exact prefix, we propose a perturbation-based procedure that first trans…

  105. arXiv cs.LG TIER_1 · Jonathan von Rad, Yong Cao, Andreas Geiger ·

    UniComp: A Unified Evaluation of Large Language Model Compression via Pruning, Quantization and Distillation

    arXiv:2602.09130v4 Announce Type: replace Abstract: Model compression is increasingly essential for deploying large language models (LLMs), yet existing comparative studies largely focus on pruning and quantization evaluated primarily on knowledge-centric benchmarks. Thus, we int…

  106. arXiv cs.LG TIER_1 · Sarthak Munshi, Manish Bhatt, Vineeth Sai Narajala, Idan Habler, Ammar Al-Kahfah, Ken Huang, Blake Gatto ·

    Manifold of Failure: Behavioral Attraction Basins in Language Models

    arXiv:2602.22291v3 Announce Type: replace Abstract: While prior work has focused on projecting adversarial examples back onto the manifold of natural data to restore safety, we argue that a comprehensive understanding of AI safety requires characterizing the unsafe regions themse…

  107. arXiv cs.LG TIER_1 · Qiming Bao, Xiaoxuan Fu, Michael Witbrock ·

    Conflict-Aware Fusion: Mitigating Logic Inertia in Large Language Models via Structured Cognitive Priors

    arXiv:2512.06393v5 Announce Type: replace-cross Abstract: Large language models (LLMs) achieve high accuracy on many reasoning benchmarks but remain brittle under structural perturbations of rule-based systems. We introduce a diagnostic framework with four stress tests -- redunda…

  108. arXiv cs.CL TIER_1 · Jinju Kim, Haeji Jung, Youjeong Roh, Jong Hwan Ko, David R. Mortensen ·

    Harnessing Linguistic Dissimilarity for Language Generalization on Unseen Low-Resource Varieties

    arXiv:2605.04500v1 Announce Type: new Abstract: Low-resource language varieties used by specific groups remain neglected in the development of Multilingual Language Models. A great deal of cross-lingual research focuses on inter-lingual language transfer which strives to align al…

  109. arXiv cs.CL TIER_1 · Qiming Bao, Juho Leinonen, Paul Denny, Michael J. Witbrock ·

    RLearner-LLM: Balancing Logical Grounding and Fluency in Large Language Models via Hybrid Direct Preference Optimization

    arXiv:2605.04539v1 Announce Type: new Abstract: Direct Preference Optimization (DPO), the efficient alternative to PPO-based RLHF, falls short on knowledge-intensive generation: standard preference signals from human annotators or LLM judges exhibit a systematic verbosity bias th…

  110. arXiv cs.CL TIER_1 · Mingda Li, Rundong Lv, Xinyu Li, Weinan Zhang, Ting Liu ·

    Gradients with Respect to Semantics Preserving Embeddings Tell the Uncertainty of Large Language Models

    arXiv:2605.04638v1 Announce Type: new Abstract: Uncertainty quantification (UQ) is an important technique for ensuring the trustworthiness of LLMs, given their tendency to hallucinate. Existing state-of-the-art UQ approaches for free-form generation rely heavily on sampling, whic…

  111. arXiv cs.CL TIER_1 · Mullosharaf K. Arabov, Svetlana S. Khaybullina ·

    Adapting Large Language Models to a Low-Resource Agglutinative Language: A Comparative Study of LoRA and QLoRA for Bashkir

    arXiv:2605.04948v1 Announce Type: new Abstract: This paper presents a comparative study of parameter-efficient fine-tuning (PEFT) methods, including LoRA and QLoRA, applied to the task of adapting large language models to the Bashkir language, a low-resource agglutinative languag…

  112. arXiv cs.CL TIER_1 · Quintin Pope, Ajay Hayagreeve Balaji, Jacques Thibodeau, Xiaoli Fern ·

    Automatically Finding and Validating Unexpected Side-Effects of Interventions on Language Models

    arXiv:2605.05090v1 Announce Type: new Abstract: We present an automated, contrastive evaluation pipeline for auditing the behavioral impact of interventions on large language models. Given a base model $M_1$ and an intervention model $M_2$, our method compares their free-form, mu…

  113. arXiv cs.CL TIER_1 · Yingshan Susan Wang, Linlu Qiu, Zhaofeng Wu, Roger P. Levy, Yoon Kim ·

    Implicit Representations of Grammaticality in Language Models

    arXiv:2605.05197v1 Announce Type: new Abstract: Grammaticality and likelihood are distinct notions in human language. Pretrained language models (LMs), which are probabilistic models of language fitted to maximize corpus likelihood, generate grammatically well-formed text and dis…

  114. arXiv cs.CL TIER_1 · Yukin Zhang, Qi Dong, Kemu Xu ·

    Emergent Hierarchical Structure in Large Language Models: An Information-Theoretic Framework for Multi-Scale Representation

    arXiv:2505.18244v3 Announce Type: replace Abstract: Why do language models from different architecture families respond so differently to the same perturbation? We argue that the answer is not scale, but \emph{how architecture shapes information compression}. Analyzing eight Tran…

  115. arXiv cs.CL TIER_1 · Mikhail L. Arbuzov, Sisong Bei, Ziwei Dong, Dmitri Kalaev, Alexey A. Shvets ·

    Beyond Exponential Decay: Rethinking Error Accumulation in Large Language Models

    arXiv:2505.24187v2 Announce Type: replace Abstract: The prevailing assumption of an exponential decay in large language model (LLM) reliability with sequence length, predicated on independent per-token error probabilities, posits an inherent limitation for long autoregressive out…

  116. arXiv cs.CL TIER_1 · Yuanhao Shen, Daniel Xavier de Sousa, Ricardo Mar\c{c}al, Hongyu Guo, Xiaodan Zhu ·

    IDRBench: Understanding the Capability of Large Language Models on Interdisciplinary Research

    arXiv:2507.15736v2 Announce Type: replace Abstract: Innovation is a key driving force of human civilization. As the body of knowledge has grown considerably, bridging knowledge across different disciplines, where significant innovation often emerges, has become increasingly chall…

  117. arXiv cs.CL TIER_1 · Hao Fang, Tianyi Zhang, Tianqu Zhuang, Jiawei Kong, Kuofeng Gao, Bin Chen, Leqi Zheng, Shu-Tao Xia, Ke Xu ·

    Towards Distillation-Resistant Large Language Models: An Information-Theoretic Perspective

    arXiv:2602.03396v3 Announce Type: replace Abstract: Proprietary large language models (LLMs) embody substantial economic value and are generally exposed only as black-box APIs, yet adversaries can still exploit their outputs to extract knowledge via distillation. Existing defense…

  118. arXiv cs.CL TIER_1 · Yingtao Shen, An Zou ·

    River-LLM: Large Language Model Seamless Exit Based on KV Share

    arXiv:2604.18396v2 Announce Type: replace Abstract: Large Language Models (LLMs) have demonstrated exceptional performance across diverse domains but are increasingly constrained by high inference latency. Early Exit has emerged as a promising solution to accelerate inference by …

  119. arXiv cs.AI TIER_1 · Bo-Wen Zhang, Jin Ye, Peng-Yu Hua, Jia-Wei Cao, Jie-Jing Shao, Yu-Feng Li, Lan-Zhe Guo ·

    Revisiting the Travel Planning Capabilities of Large Language Models

    arXiv:2605.03308v1 Announce Type: new Abstract: Travel planning serves as a critical task for long-horizon reasoning, exposing significant deficits in LLMs. However, existing benchmarks and evaluations primarily assess final plans in an end-to-end manner, which lacks interpretabi…

  120. arXiv cs.AI TIER_1 · Konstantinos Papaioannou, Thaleia Dimitra Doudali ·

    TCM-Serve: Modality-aware Scheduling for Multimodal Large Language Model Inference

    arXiv:2603.26498v2 Announce Type: replace-cross Abstract: Multimodal Large Language Models (MLLMs) power platforms like ChatGPT, Gemini, and Copilot, enabling richer interactions with text, images, and videos. These heterogeneous workloads introduce additional inference stages, s…

  121. arXiv cs.CL TIER_1 · Yoon Kim ·

    Implicit Representations of Grammaticality in Language Models

    Grammaticality and likelihood are distinct notions in human language. Pretrained language models (LMs), which are probabilistic models of language fitted to maximize corpus likelihood, generate grammatically well-formed text and discriminate well between grammatical and ungrammat…

  122. arXiv cs.CL TIER_1 · Xiaoli Fern ·

    Automatically Finding and Validating Unexpected Side-Effects of Interventions on Language Models

    We present an automated, contrastive evaluation pipeline for auditing the behavioral impact of interventions on large language models. Given a base model $M_1$ and an intervention model $M_2$, our method compares their free-form, multi-token generations across aligned prompt cont…

  123. arXiv cs.CL TIER_1 · Svetlana S. Khaybullina ·

    Adapting Large Language Models to a Low-Resource Agglutinative Language: A Comparative Study of LoRA and QLoRA for Bashkir

    This paper presents a comparative study of parameter-efficient fine-tuning (PEFT) methods, including LoRA and QLoRA, applied to the task of adapting large language models to the Bashkir language, a low-resource agglutinative language of the Turkic family. Experimental evaluation …

  124. arXiv cs.LG TIER_1 · Albert F. Modenbach ·

    A geometric relation of the error introduced by sampling a language model's output distribution to its internal state

    GPT-style language models are sensitive to single-token changes at generation points where the predicted probability distribution is spread across multiple tokens. Viewing this sensitivity as a geometric property, we derive an $\mathfrak{so}(n)$-valued 1-form that depends only on…

  125. arXiv cs.CL TIER_1 · Ting Liu ·

    Gradients with Respect to Semantics Preserving Embeddings Tell the Uncertainty of Large Language Models

    Uncertainty quantification (UQ) is an important technique for ensuring the trustworthiness of LLMs, given their tendency to hallucinate. Existing state-of-the-art UQ approaches for free-form generation rely heavily on sampling, which incurs high computational cost and variance. I…

  126. arXiv cs.CL TIER_1 · Michael J. Witbrock ·

    RLearner-LLM: Balancing Logical Grounding and Fluency in Large Language Models via Hybrid Direct Preference Optimization

    Direct Preference Optimization (DPO), the efficient alternative to PPO-based RLHF, falls short on knowledge-intensive generation: standard preference signals from human annotators or LLM judges exhibit a systematic verbosity bias that rewards fluency over logical correctness. Thi…

  127. arXiv cs.CL TIER_1 · David R. Mortensen ·

    Harnessing Linguistic Dissimilarity for Language Generalization on Unseen Low-Resource Varieties

    Low-resource language varieties used by specific groups remain neglected in the development of Multilingual Language Models. A great deal of cross-lingual research focuses on inter-lingual language transfer which strives to align allied varieties and minimize differences between …

  128. arXiv cs.CL TIER_1 · Mullosharaf K. Arabov ·

    Benchmarking Parameter-Efficient Fine-Tuning of Large Language Models for Low-Resource Tajik Text Generation with the Tajik Web Corpus

    arXiv:2605.03742v1 Announce Type: new Abstract: This paper is devoted to the adaptation of generative large language models for the Tajik language, a low-resource language with Cyrillic script. To overcome the shortage of digital text resources, the author created and publicly re…

  129. arXiv cs.CL TIER_1 · Pei Zhang, Yiming Wang, Jialong Tang, Baosong Yang, Rui Wang, Derek F. Wong, Fei Huang ·

    Direct Simultaneous Translation Activation for Large Audio-Language Models

    arXiv:2509.15692v2 Announce Type: replace-cross Abstract: Simultaneous speech-to-text translation (Simul-S2TT) aims to translate speech into target text in real time, outputting translations while receiving source speech input, rather than waiting for the entire utterance to be s…

  130. arXiv cs.CL TIER_1 · Hao Yu, Tianyi Xu, Michael A. Hedderich, Wassim Hamidouche, Syed Waqas Zamir, David Ifeoluwa Adelani ·

    AfriqueLLM: How Data Mixing and Model Architecture Impact Continued Pre-training for African Languages

    arXiv:2601.06395v2 Announce Type: replace Abstract: Large language models (LLMs) are increasingly multilingual, yet open models continue to underperform relative to proprietary systems, with the gap most pronounced for African languages. Continued pre-training (CPT) offers a prac…

  131. arXiv cs.CL TIER_1 · Junichiro Niimi ·

    Hierarchical Memorization in Large Language Models: Evidence from Citation Generation

    arXiv:2511.08877v2 Announce Type: replace Abstract: Large language models (LLMs) generate fluent text across a wide range of tasks, but the fabrication of non-existent academic citations remains a critical and well-documented failure mode. Building on prior work that frames hallu…

  132. arXiv cs.LG TIER_1 · Francesco Sovrano, Gabriele Dominici, Marc Langheinrich ·

    Neuron-Anchored Rule Extraction for Large Language Models via Contrastive Hierarchical Ablation

    arXiv:2605.03058v1 Announce Type: new Abstract: A key goal of explainable AI (XAI) is to express the decision logic of large language models (LLMs) in symbolic form and link it to internal mechanisms. Global rule-extraction methods typically learn symbolic surrogates without grou…

  133. arXiv cs.LG TIER_1 · Ping Wang, Yan-Qi Du ·

    Finite-Size Gradient Transport in Large Language Model Pretraining: From Cascade Size to Intensive Transport Efficiency

    arXiv:2605.02968v1 Announce Type: new Abstract: We introduce a finite-size gradient-transport framework for real language-model training, based on five observables $(D,z,\beta,\delta,v_{\mathrm{rel}})$ that separate cascade size, duration, absolute transport, and intensive transp…

  134. arXiv cs.LG TIER_1 · Jingkai He, Pengfei Chen, Chenghui Wu, Shuang Liang, Ye Li, Gou Tan, Xiadao Wen, Chuanfu Zhang ·

    An End-to-End Framework for Building Large Language Models for Software Operations

    arXiv:2605.02906v1 Announce Type: new Abstract: In the field of software operations, Large Language Models (LLMs) have attracted increasing attention. However, existing research has not yet achieved efficient and effective end-to-end intelligent operations due to low-quality data…

  135. arXiv cs.AI TIER_1 · Xiyuan Wang, Yi Hu, Yanbo Wang, Chuan Shi, Muhan Zhang ·

    Position: How can Graphs Help Large Language Models?

    arXiv:2605.02452v1 Announce Type: new Abstract: With the rapid advancement of large language models (LLMs), classic graph learning tasks have greatly benefited from LLMs, including improved encoding of textual features, more efficient construction of graphs from text, and enhance…

  136. arXiv cs.LG TIER_1 · Jinbin Bai, Yixuan Li, Yuchen Zhu, Yi Xin, Qingyu Shi, Aosong Feng, Xiaohong Liu, Molei Tao, Jianru Xue, Xiangtai Li, Ming-Hsuan Yang ·

    Prism: Efficient Test-Time Scaling via Hierarchical Search and Self-Verification for Discrete Diffusion Language Models

    arXiv:2602.01842v3 Announce Type: replace Abstract: Inference-time compute has re-emerged as a practical way to improve LLM reasoning. Most test-time scaling (TTS) algorithms rely on autoregressive decoding, which is ill-suited to discrete diffusion language models (dLLMs) due to…

  137. arXiv cs.LG TIER_1 · Sebastian Wind, Tri-Thien Nguyen, Jeta Sopa, Mahshad Lotfinia, Sebastian Bickelhaup, Michael Uder, Harald K\"ostler, Gerhard Wellein, Sven Nebelung, Daniel Truhn, Andreas Maier, Soroosh Tayebi Arasteh ·

    Safety and accuracy follow different scaling laws in clinical large language models

    arXiv:2605.04039v1 Announce Type: cross Abstract: Clinical LLMs are often scaled by increasing model size, context length, retrieval complexity, or inference-time compute, with the implicit expectation that higher accuracy implies safer behavior. This assumption is incomplete in …

  138. arXiv cs.LG TIER_1 · Jiaxi Li, Lu Yin, Li Shen, Jinjin Xu, Yuhui Liu, Wenwu Wang, Shiwei Liu, Xilu Wang ·

    ELAS: Efficient Pre-Training of Low-Rank Large Language Models via 2:4 Activation Sparsity

    arXiv:2605.03667v1 Announce Type: new Abstract: Large Language Models (LLMs) have achieved remarkable capabilities, but their immense computational demands during training remain a critical bottleneck for widespread adoption. Low-rank training has received attention in recent yea…

  139. arXiv cs.LG TIER_1 · Akshat Singh Jaswal, Ashish Baghel, Paras Chopra ·

    Discovering Reinforcement Learning Interfaces with Large Language Models

    arXiv:2605.03408v1 Announce Type: new Abstract: Reinforcement learning systems rely on environment interfaces that specify observations and reward functions, yet constructing these interfaces for new tasks often requires substantial manual effort. While recent work has automated …

  140. Hugging Face Daily Papers TIER_1 ·

    Safety and accuracy follow different scaling laws in clinical large language models

    Clinical LLMs are often scaled by increasing model size, context length, retrieval complexity, or inference-time compute, with the implicit expectation that higher accuracy implies safer behavior. This assumption is incomplete in medicine, where a few confident, high-risk, or evi…

  141. arXiv cs.CL TIER_1 · Soroosh Tayebi Arasteh ·

    Safety and accuracy follow different scaling laws in clinical large language models

    Clinical LLMs are often scaled by increasing model size, context length, retrieval complexity, or inference-time compute, with the implicit expectation that higher accuracy implies safer behavior. This assumption is incomplete in medicine, where a few confident, high-risk, or evi…

  142. arXiv cs.CL TIER_1 · Mullosharaf K. Arabov ·

    Benchmarking Parameter-Efficient Fine-Tuning of Large Language Models for Low-Resource Tajik Text Generation with the Tajik Web Corpus

    This paper is devoted to the adaptation of generative large language models for the Tajik language, a low-resource language with Cyrillic script. To overcome the shortage of digital text resources, the author created and publicly released the Tajik Web Corpus, the largest open-ac…

  143. arXiv cs.AI TIER_1 · Xilu Wang ·

    ELAS: Efficient Pre-Training of Low-Rank Large Language Models via 2:4 Activation Sparsity

    Large Language Models (LLMs) have achieved remarkable capabilities, but their immense computational demands during training remain a critical bottleneck for widespread adoption. Low-rank training has received attention in recent years due to its ability to significantly reduce tr…

  144. arXiv cs.CL TIER_1 · Hyeontaek Hwang, Nguyen Dinh Son, Daeyoung Kim ·

    Model-Dowser: Data-Free Importance Probing to Mitigate Catastrophic Forgetting in Multimodal Large Language Models

    arXiv:2602.04509v4 Announce Type: replace Abstract: Fine-tuning Multimodal Large Language Models (MLLMs) on task-specific data is an effective way to improve performance on downstream applications. However, such adaptation often leads to a degradation in generalization on pretrai…

  145. arXiv cs.LG TIER_1 (CA) · Nicholas T. Runcie, Fergus Imrie, Charlotte M. Deane ·

    Molecular Representations for Large Language Models

    arXiv:2605.01822v1 Announce Type: new Abstract: Large Language Models (LLMs) are increasingly being used to support scientific discovery. In chemistry, tasks such as reaction prediction and structure elucidation require reasoning about the structures of molecules. As such, LLM-ba…

  146. arXiv cs.LG TIER_1 · Yan Jiang, Ruihong Qiu, Zi Huang ·

    Break the Block: Dynamic-size Reasoning Blocks for Diffusion Large Language Models via Monotonic Entropy Descent with Reinforcement Learning

    arXiv:2605.02263v1 Announce Type: new Abstract: Recent diffusion large language models (dLLMs) have demonstrated both effectiveness and efficiency in reasoning via a block-based semi-autoregressive generation paradigm. Despite their progress, the fixed-size block generations rema…

  147. arXiv cs.LG TIER_1 · Michael Helcig, Eldar Kurtic, Dan Alistarh ·

    Statistically-Lossless Quantization of Large Language Models

    arXiv:2605.02404v1 Announce Type: new Abstract: Model quantization has become essential for efficient large language model deployment, yet existing approaches involve clear trade-offs: methods such as GPTQ and AWQ achieve practical compression but are lossy, while lossless techni…

  148. arXiv cs.LG TIER_1 · Inoussa Mouiche ·

    Gradient-Gated DPO: Stabilizing Preference Optimization in Language Models

    arXiv:2605.02626v1 Announce Type: new Abstract: Preference optimization has become a central paradigm for aligning large language models with human feedback. Direct Preference Optimization (DPO) simplifies reinforcement learning from human feedback by directly optimizing pairwise…

  149. arXiv cs.LG TIER_1 · Sunghwan Kim, Junhee Cho, Beong-woo Kwak, Taeyoon Kwon, Liang Wang, Nan Yang, Xingxing Zhang, Furu Wei, Jinyoung Yeo ·

    On Training Large Language Models for Long-Horizon Tasks: An Empirical Study of Horizon Length

    arXiv:2605.02572v1 Announce Type: cross Abstract: Large language models (LLMs) have shown promise as interactive agents that solve tasks through extended sequences of environment interactions. While prior work has primarily focused on system-level optimizations or algorithmic imp…

  150. arXiv cs.LG TIER_1 · Yuxiang Chen, Yifan Liu, Xiaoming Xu, Pengle Zhang, Michael Beyer, Martin Rapp, Jun Zhu, Jianfei Chen ·

    TetraJet-v2: Accurate NVFP4 Training for Large Language Models with Oscillation Suppression and Outlier Control

    arXiv:2510.27527v2 Announce Type: replace Abstract: Large Language Models (LLMs) training is prohibitively expensive, driving interest in low-precision fully-quantized training (FQT). While novel 4-bit formats like NVFP4 offer substantial efficiency gains, achieving near-lossless…

  151. arXiv cs.LG TIER_1 · Chen Liu, Xingzhi Sun, Xi Xiao, Alexandre Van Tassel, Ke Xu, Kristof Reimann, Danqi Liao, Mark Gerstein, Tianyang Wang, Xiao Wang, Smita Krishnaswamy ·

    Dispersion Loss Counteracts Embedding Condensation and Improves Generalization in Small Language Models

    arXiv:2602.00217v2 Announce Type: replace Abstract: Large language models (LLMs) achieve remarkable performance through ever-increasing parameter counts, but scaling incurs steep computational costs. To better understand LLM scaling, we study representational differences between …

  152. arXiv cs.LG TIER_1 · Nils Strassenburg, Boris Glavic, Tilmann Rabl ·

    Poodle: Seamlessly Scaling Down Large Language Models with Just-in-Time Model Replacement

    arXiv:2512.05525v2 Announce Type: replace-cross Abstract: Businesses increasingly rely on large language models (LLMs) to automate simple repetitive tasks instead of developing custom machine learning models. LLMs require few, if any, training examples and can be utilized by user…

  153. arXiv cs.LG TIER_1 · Matthias Mertens, Natalia Fischl-Lanzoni, Neil Thompson ·

    Is there "Secret Sauce'' in Large Language Model Development?

    arXiv:2602.07238v2 Announce Type: replace-cross Abstract: Do leading LLM developers possess a proprietary `"secret sauce'', or is LLM performance driven by scaling up compute? Using training and benchmark data for 809 models released between 2022 and 2025, we estimate scaling-law…

  154. arXiv cs.AI TIER_1 · Sydney Johns, Heng Jin, Chaoyu Zhang, Y. Thomas Hou, Wenjing Lou ·

    ARMOR 2025: A Military-Aligned Benchmark for Evaluating Large Language Model Safety Beyond Civilian Contexts

    arXiv:2605.00245v1 Announce Type: new Abstract: Large language models (LLMs) are now being explored for defense applications that require reliable and legally compliant decision support. They also hold significant potential to enhance decision making, coordination, and operationa…

  155. arXiv cs.AI TIER_1 · Damiano Fornasiere, Mirko Bronzi, Spencer Kitts, Alessandro Palmas, Yoshua Bengio, Oliver Richardson ·

    Language models recognize dropout and Gaussian noise applied to their activations

    arXiv:2604.17465v2 Announce Type: replace Abstract: We provide evidence that language models can detect, localize and, to a certain degree, verbalize the difference between perturbations applied to their activations. More precisely, we either (a) mask activations, simulating drop…

  156. arXiv cs.CL TIER_1 · Cutter Dawes, Aryan Sharma, Angelos Ioannis Lagos, Shivam Raval ·

    H-Probes: Extracting Hierarchical Structures From Latent Representations of Language Models

    arXiv:2605.00847v1 Announce Type: new Abstract: Representing and navigating hierarchy is a fundamental primitive of reasoning. Large language models have demonstrated proficiency in a wide variety of tasks requiring hierarchical reasoning, but there exists limited analysis on how…

  157. arXiv cs.CL TIER_1 · Quoc Phong Dao, Hoang Son Nguyen, Pham Khanh Chi, Tung Nguyen, Linh Ngo Van, Nguyen Thi Ngoc Diep, Trung Le ·

    SRA: Span Representation Alignment for Large Language Model Distillation

    arXiv:2605.01205v1 Announce Type: new Abstract: Cross-Tokenizer Knowledge Distillation (CTKD) enables knowledge transfer between a large language model and a smaller student, even when they employ different tokenizers. While existing approaches mainly focus on token-level alignme…

  158. arXiv cs.CL TIER_1 · Zhiwen Ruan, Yichao Du, Jianjie Zheng, Longyue Wang, Yun Chen, Peng Li, Jinsong Su, Yang Liu, Guanhua Chen ·

    GIFT: Guided Fine-Tuning and Transfer for Enhancing Instruction-Tuned Language Models

    arXiv:2605.01256v1 Announce Type: new Abstract: A promising paradigm for adapting instruction-tuned language models is to learn task-specific updates on a pretrained base model and subsequently merge them into the instruction-tuned model. However, existing approaches typically tr…

  159. arXiv cs.CL TIER_1 · Jinyuan Feng, Xin Yu, Yiqun Chen, Xiaochi Wei, Yan Gao, Yi Wu, Yao Hu, Zhiqiang Pu ·

    Focus on the Core: Empowering Diffusion Large Language Models by Self-Contrast

    arXiv:2605.01373v1 Announce Type: new Abstract: The iterative denoising paradigm of Diffusion Large Language Models (DLMs) endows them with a distinct advantage in global context modeling. However, current decoding strategies fail to leverage this capability, typically exhibiting…

  160. arXiv cs.CL TIER_1 · Pham Khanh Chi, Quoc Phong Dao, Thuat Nguyen, Linh Ngo Van, Trung Le, Thanh Hong Nguyen ·

    MTA: Multi-Granular Trajectory Alignment for Large Language Model Distillation

    arXiv:2605.01374v1 Announce Type: new Abstract: Knowledge distillation is a key technique for compressing large language models (LLMs), but most existing methods align representations at fixed layers or token-level outputs, ignoring how representations evolve across depth. As a r…

  161. arXiv cs.CL TIER_1 · Arnau Marin-Llobet, Javier Ferrando ·

    Automated Interpretability and Feature Discovery in Language Models with Agents

    arXiv:2605.01555v1 Announce Type: new Abstract: We introduce an autonomous multiagent framework for mechanistic interpretability that automates both explaining and finding internal features in large language models. The system runs two coupled loops: (1) explanation refinement, w…

  162. arXiv cs.CL TIER_1 · Zhuoyun Li, Boxuan Wang, Jinwei Hu, Zhenglin Huang, Qisong He, Xinmiao Huang, Guangliang Cheng, Xiaowei Huang, Yi Dong ·

    Where Do Prompt Perturbations Break Generation? A Segment-Level View of Robustness in LoRA-Tuned Language Models

    arXiv:2605.01605v1 Announce Type: new Abstract: Large language models are sensitive to minor prompt perturbations, yet existing robustness methods usually enforce consistency at the whole-sequence level. This holistic view can hide an important failure mode: a perturbed response …

  163. arXiv cs.CL TIER_1 · Lang Gao, Jinghui Zhang, Wei Liu, Fengxian Ji, Chenxi Wang, Zirui Song, Akash Ghosh, Youssef Mohamed, Preslav Nakov, Xiuying Chen ·

    The Cylindrical Representation Hypothesis for Language Model Steering

    arXiv:2605.01844v1 Announce Type: new Abstract: Steering is a widely used technique for controlling large language models, yet its effects are often unstable and hard to predict. Existing theoretical accounts are largely based on the Linear Representation Hypothesis (LRH). While …

  164. arXiv cs.CL TIER_1 · Kotaro Furuya, Takahito Tanimura ·

    Spatiotemporal Hidden-State Dynamics as a Signature of Internal Reasoning in Large Language Models

    arXiv:2605.01853v1 Announce Type: new Abstract: Large reasoning models (LRMs) generate extended solutions, yet it remains unclear whether these traces reflect substantive internal computation or merely verbosity and overthinking. Although recent hidden-state analyses suggest that…

  165. arXiv cs.CL TIER_1 · Tianxiang Dai, Jonathan Fan ·

    Counting as a minimal probe of language model reliability

    arXiv:2605.02028v1 Announce Type: new Abstract: Large language models perform strongly on benchmarks in mathematical reasoning, coding and document analysis, suggesting a broad ability to follow instructions. However, it remains unclear whether such success reflects general logic…

  166. arXiv cs.CL TIER_1 · Ranit Karmakar, Jayita Chatterjee ·

    What Single-Prompt Accuracy Misses: A Multi-Variant Reliability Audit of Language Models

    arXiv:2605.02038v1 Announce Type: new Abstract: Single-prompt accuracy is the dominant way to benchmark language models, but it can miss reliability failures that matter. We evaluate a 15-model open-weight corpus, with the main reliability analyses focused on 10 instruct models a…

  167. arXiv cs.CL TIER_1 · Cosimo Galeone, Minsu Park, Giuseppe Ettorre, Daniele Ligorio ·

    When Correct Isn't Usable: Improving Structured Output Reliability in Small Language Models

    arXiv:2605.02363v1 Announce Type: new Abstract: Deployed language models must produce outputs that are both correct and format-compliant. We study this structured-output reliability gap using two mathematical benchmarks -- GSM8K and MATH -- as a controlled testbed: ground truth i…

  168. arXiv cs.CL TIER_1 · Fengze Liu, Weidong Zhou, Binbin Liu, Ping Guo, Zijun Wang, Bingni Zhang, Yifan Zhang, Yifeng Yu, Xiaohuan Zhou, Taifeng Wang ·

    InfoLaw: Information Scaling Laws for Large Language Models with Quality-Weighted Mixture Data and Repetition

    arXiv:2605.02364v1 Announce Type: new Abstract: Upweighting high-quality data in LLM pretraining often improves performance, but in datalimited regimes, especially under overtraining, stronger upweighting increases repetition and can degrade performance. However, standard scaling…

  169. arXiv cs.CL TIER_1 · Hillary Mutisya, John Mugane ·

    Attention Sinks in Massively Multilingual Neural Machine Translation:Discovery, Analysis, and Mitigation

    arXiv:2605.01229v1 Announce Type: cross Abstract: Cross-attention patterns in neural machine translation (NMT) are widely used to study how multilingual models align linguistic structure. We report a systematic artifact in cross-attention analysis of NLLB-200 (600M): non-content …

  170. arXiv cs.CL TIER_1 · Yukun Zhang, Qi Dong, Mengkang Li ·

    Latent Trajectory Dynamics in Large Language Models: A Manifold Evolution Framework with Empirical Validation

    arXiv:2505.20340v3 Announce Type: replace Abstract: Understanding how latent representations evolve during generation is a central open problem in large language model interpretability. We introduce \textbf{Dynamical Manifold Evolution Theory} (DMET), a phenomenological framework…

  171. arXiv cs.CL TIER_1 · Chen Xiong, Zihao Wang, Rui Zhu, Tsung-Yi Ho, Pin-Yu Chen, Jingwei Xiong, Haixu Tang ·

    Hey, That's My Data! Token-Only Dataset Inference in Large Language Models

    arXiv:2506.06057v2 Announce Type: replace Abstract: Large Language Models (LLMs) rely on massive training datasets, often including proprietary data, which raises concerns about unauthorized usage and copyright infringement. Existing dataset inference methods typically require ac…

  172. arXiv cs.CL TIER_1 · Jiaqi Chen, Yanzhe Zhang, Yutong Zhang, Yijia Shao, Diyi Yang ·

    Generative Interfaces for Language Models

    arXiv:2508.19227v3 Announce Type: replace Abstract: Large language models (LLMs) are increasingly seen as assistants, copilots, and consultants, capable of supporting a wide range of tasks through natural conversation. However, most systems remain constrained by a linear request-…

  173. arXiv cs.CL TIER_1 · Kai R. Larsen, Sen Yan, Roland M. Mueller, Lan Sang, Mikko R\"onkk\"o, Ravi Starzl, Donald Edmondson ·

    ALIGNS: Unlocking nomological networks in psychological measurement through a large language model

    arXiv:2509.09723v3 Announce Type: replace Abstract: Psychological measurement is critical to many disciplines. Despite advances in measurement, building nomological networks, theoretical maps of how concepts and measures relate to establish validity, remains a challenge 70 years …

  174. arXiv cs.CL TIER_1 · Bryan E. Tuck, Rakesh M. Verma ·

    Orthographic Constraint Satisfaction and Human Difficulty Alignment in Large Language Models

    arXiv:2511.21086v2 Announce Type: replace Abstract: Large language models must satisfy hard orthographic constraints during controlled text generation, yet systematic cross-family evaluation remains limited. We evaluate 39 configurations spanning three model families (Qwen3, Clau…

  175. arXiv cs.CL TIER_1 · Amit Dhurandhar, Vijil Chenthamarakshan, Dennis Wei, Tejaswini Pedapati, Karthikeyan Natesan Ramamurthy, Rahul Nair ·

    CoFrGeNet: Continued Fraction Architectures for Language Generation

    arXiv:2601.21766v3 Announce Type: replace Abstract: Transformers are arguably the preferred architecture for language generation. In this paper, inspired by continued fractions, we introduce a new function class for generative modeling. The architecture family implementing this f…

  176. arXiv cs.LG TIER_1 · Daniel Agyei Asante, Ernie Chang, Yang Li ·

    Importance-Guided Basis Selection for Low-Rank Decomposition of Large Language Models

    arXiv:2605.01627v1 Announce Type: new Abstract: Low-rank decomposition is a compelling approach for compressing large language models, but its effectiveness hinges on selecting which singular-vector bases to retain for a target task. Existing methods such as Basel adapt singular-…

  177. arXiv cs.LG TIER_1 · Inoussa Mouiche ·

    Gradient-Gated DPO: Stabilizing Preference Optimization in Language Models

    Preference optimization has become a central paradigm for aligning large language models with human feedback. Direct Preference Optimization (DPO) simplifies reinforcement learning from human feedback by directly optimizing pairwise preferences, removing the need for reward model…

  178. arXiv cs.AI TIER_1 · Jinyoung Yeo ·

    On Training Large Language Models for Long-Horizon Tasks: An Empirical Study of Horizon Length

    Large language models (LLMs) have shown promise as interactive agents that solve tasks through extended sequences of environment interactions. While prior work has primarily focused on system-level optimizations or algorithmic improvements, the role of task horizon length in shap…

  179. arXiv cs.AI TIER_1 · Muhan Zhang ·

    Position: How can Graphs Help Large Language Models?

    With the rapid advancement of large language models (LLMs), classic graph learning tasks have greatly benefited from LLMs, including improved encoding of textual features, more efficient construction of graphs from text, and enhanced reasoning over knowledge graphs. In this paper…

  180. Hugging Face Daily Papers TIER_1 ·

    Position: How can Graphs Help Large Language Models?

    With the rapid advancement of large language models (LLMs), classic graph learning tasks have greatly benefited from LLMs, including improved encoding of textual features, more efficient construction of graphs from text, and enhanced reasoning over knowledge graphs. In this paper…

  181. Hugging Face Daily Papers TIER_1 ·

    Statistically-Lossless Quantization of Large Language Models

    Model quantization has become essential for efficient large language model deployment, yet existing approaches involve clear trade-offs: methods such as GPTQ and AWQ achieve practical compression but are lossy, while lossless techniques preserve fidelity but typically do not acce…

  182. arXiv cs.LG TIER_1 · Dan Alistarh ·

    Statistically-Lossless Quantization of Large Language Models

    Model quantization has become essential for efficient large language model deployment, yet existing approaches involve clear trade-offs: methods such as GPTQ and AWQ achieve practical compression but are lossy, while lossless techniques preserve fidelity but typically do not acce…

  183. arXiv cs.CL TIER_1 · Taifeng Wang ·

    InfoLaw: Information Scaling Laws for Large Language Models with Quality-Weighted Mixture Data and Repetition

    Upweighting high-quality data in LLM pretraining often improves performance, but in datalimited regimes, especially under overtraining, stronger upweighting increases repetition and can degrade performance. However, standard scaling laws do not reliably extrapolate across mixture…

  184. arXiv cs.CL TIER_1 · Daniele Ligorio ·

    When Correct Isn't Usable: Improving Structured Output Reliability in Small Language Models

    Deployed language models must produce outputs that are both correct and format-compliant. We study this structured-output reliability gap using two mathematical benchmarks -- GSM8K and MATH -- as a controlled testbed: ground truth is unambiguous and the output contract is strict …

  185. arXiv cs.LG TIER_1 · Yufei Guo, Muzhe Guo, Juntao Su, Zhou Yang, Mengqiu Zhu, Hongfei Li, Mengyang Qiu, Shuo Shuo Liu ·

    Bias in Large Language Models: Origin, Evaluation, and Mitigation

    arXiv:2411.10915v2 Announce Type: replace-cross Abstract: Large Language Models (LLMs) have revolutionized natural language processing, but their susceptibility to biases poses significant challenges. This comprehensive review examines the landscape of bias in LLMs, from its orig…

  186. arXiv cs.LG TIER_1 · Cameron Yetman ·

    Representation in large language models

    arXiv:2501.00885v2 Announce Type: replace-cross Abstract: The extraordinary success of recent Large Language Models (LLMs) on a diverse array of tasks has led to an explosion of scientific and philosophical theorizing aimed at explaining how they do what they do. Unfortunately, d…

  187. arXiv cs.LG TIER_1 · Zhaoyi Li, Jiatong Li, Gangwei Jiang, Linqi Song, Defu Lian, Ying Wei ·

    Scaling Reasoning Hop Exposes Weaknesses: Demystifying and Improving Hop Generalization in Large Language Models

    arXiv:2601.21214v2 Announce Type: replace-cross Abstract: Chain-of-thought (CoT) reasoning has become the standard paradigm for enabling Large Language Models (LLMs) to solve complex problems. However, recent studies reveal a sharp performance drop in reasoning hop generalization…

  188. arXiv cs.LG TIER_1 · Gregory N. Frank ·

    How Alignment Routes: Localizing, Scaling, and Controlling Policy Circuits in Language Models

    arXiv:2604.04385v4 Announce Type: replace-cross Abstract: We localize the policy routing mechanism in alignment-trained language models. An intermediate-layer attention gate reads detected content and triggers deeper amplifier heads that boost the signal toward refusal. In smalle…

  189. arXiv cs.CL TIER_1 · Enzo S. N. Silva, Pablo B. Costa, Raphael C. Vlasman, Rosimeire P. Costa, Henrique L. P. Silva, Lucas F. A. O. Pellicer, Guilherme Rinaldo, Renato A. Almeida, Darian S. R. Rabbani, Cinthya O. Oestreich, Vinicius F. Carid\'a ·

    NorBERTo: A ModernBERT Model Trained for Portuguese with 331 Billion Tokens Corpus

    arXiv:2605.00086v1 Announce Type: new Abstract: High-quality corpora are essential for advancing Natural Language Processing (NLP) in Portuguese. Building on previous encoder-only models such as BERTimbau and Albertina PT-BR, we introduce NorBERTo, a modern encoder based on the M…

  190. arXiv cs.CL TIER_1 · Kenneth J. K. Ong ·

    Impact of Task Phrasing on Presumptions in Large Language Models

    arXiv:2605.00436v1 Announce Type: new Abstract: Concerns with the safety and reliability of applying large-language models (LLMs) in unpredictable real-world applications motivate this study, which examines how task phrasing can lead to presumptions in LLMs, making it difficult f…

  191. arXiv cs.CL TIER_1 · Jiawei Wu, DouDou Zhou ·

    Unlearning What Matters: Token-Level Attribution for Precise Language Model Unlearning

    arXiv:2605.00364v1 Announce Type: new Abstract: Machine unlearning has emerged as a critical capability for addressing privacy, safety, and regulatory concerns in large language models (LLMs). Existing methods operate at the sequence level, applying uniform updates across all tok…

  192. arXiv cs.CL TIER_1 · Michael A. Lepori, Tal Linzen, Ann Yuan, Katja Filippova ·

    Language Models Struggle to Use Representations Learned In-Context

    arXiv:2602.04212v2 Announce Type: replace Abstract: Though large language models (LLMs) have enabled great success across a wide variety of tasks, they still appear to fall short of one of the loftier goals of artificial intelligence research: creating an artificial system that c…

  193. arXiv cs.CL TIER_1 · Peng Yu, Zeyuan Zhao, Shao Zhang, Luoyi Fu, Xinbing Wang, Ying Wen ·

    Structured In-context Environment Scaling for Large Language Model Reasoning

    arXiv:2509.23330v3 Announce Type: replace Abstract: Large language models (LLMs) have achieved significant advancements in reasoning capabilities through reinforcement learning (RL) via environmental exploration. As the intrinsic properties of the environment determine the abilit…

  194. arXiv cs.CL TIER_1 · Yunhan Zhao, Zhaorun Chen, Xingjun Ma, Yu-Gang Jiang, Bo Li ·

    ML-Bench&Guard: Policy-Grounded Multilingual Safety Benchmark and Guardrail for Large Language Models

    arXiv:2605.00689v1 Announce Type: new Abstract: As Large Language Models (LLMs) are increasingly deployed in cross-linguistic contexts, ensuring safety in diverse regulatory and cultural environments has become a critical challenge. However, existing multilingual benchmarks large…

  195. arXiv cs.CL TIER_1 · Gaofei Shen, Martijn Bentum, Tom Lentz, Afra Alishahi, Grzegorz Chrupa{\l}a ·

    Beyond Decodability: Reconstructing Language Model Representations with an Encoding Probe

    arXiv:2605.00607v1 Announce Type: new Abstract: Probing is widely used to study which features can be decoded from language model representations. However, the common decoding probe approach has two limitations that we aim to solve with our new encoding probe approach: contributi…

  196. arXiv cs.LG TIER_1 · Roman Klypa, Oleksandr Cherednichenko ·

    Diversity in Large Language Models under Supervised Fine-Tuning

    arXiv:2605.00195v1 Announce Type: new Abstract: Supervised Fine-Tuning (SFT) is essential for aligning Large Language Models (LLMs) with user intent, yet it is believed to suppress generative diversity. Although this reduction is frequently referenced, formal empirical testing of…

  197. arXiv cs.LG TIER_1 · Hamidreza Saghir ·

    How Language Models Process Out-of-Distribution Inputs: A Two-Pathway Framework

    arXiv:2605.00269v1 Announce Type: cross Abstract: Recent white-box OOD detection methods for LLMs -- including CED, RAUQ, and WildGuard confidence scores -- appear effective, but we show they are structurally confounded by sequence length (|r| >= 0.61) and collapse to near-chance…

  198. arXiv cs.LG TIER_1 · Haotian Xu, Jiannan Yang, Tian Gao, Tsui-Wei Weng, Tengfei Ma ·

    Resting Neurons, Active Insights: Robustify Activation Sparsity for Large Language Models

    arXiv:2512.12744v3 Announce Type: replace Abstract: Activation sparsity offers a compelling route to accelerate large language model (LLM) inference by selectively suppressing hidden activations, yet existing approaches exhibit severe accuracy degradation at high sparsity. We sho…

  199. arXiv cs.CL TIER_1 · Jayita Chatterjee ·

    What Single-Prompt Accuracy Misses: A Multi-Variant Reliability Audit of Language Models

    Single-prompt accuracy is the dominant way to benchmark language models, but it can miss reliability failures that matter. We evaluate a 15-model open-weight corpus, with the main reliability analyses focused on 10 instruct models across five classification and reasoning benchmar…

  200. arXiv cs.CL TIER_1 · Jonathan Fan ·

    Counting as a minimal probe of language model reliability

    Large language models perform strongly on benchmarks in mathematical reasoning, coding and document analysis, suggesting a broad ability to follow instructions. However, it remains unclear whether such success reflects general logical competence, repeated application of learned p…

  201. arXiv cs.CL TIER_1 · Takahito Tanimura ·

    Spatiotemporal Hidden-State Dynamics as a Signature of Internal Reasoning in Large Language Models

    Large reasoning models (LRMs) generate extended solutions, yet it remains unclear whether these traces reflect substantive internal computation or merely verbosity and overthinking. Although recent hidden-state analyses suggest that internal representations carry correctness-rela…

  202. arXiv cs.CL TIER_1 · Xiuying Chen ·

    The Cylindrical Representation Hypothesis for Language Model Steering

    Steering is a widely used technique for controlling large language models, yet its effects are often unstable and hard to predict. Existing theoretical accounts are largely based on the Linear Representation Hypothesis (LRH). While LRH assumes that concepts can be orthogonalized …

  203. arXiv cs.CL TIER_1 · Bo Li ·

    ML-Bench&Guard: Policy-Grounded Multilingual Safety Benchmark and Guardrail for Large Language Models

    As Large Language Models (LLMs) are increasingly deployed in cross-linguistic contexts, ensuring safety in diverse regulatory and cultural environments has become a critical challenge. However, existing multilingual benchmarks largely rely on general risk taxonomies and machine t…

  204. arXiv cs.CL TIER_1 · Grzegorz Chrupała ·

    Beyond Decodability: Reconstructing Language Model Representations with an Encoding Probe

    Probing is widely used to study which features can be decoded from language model representations. However, the common decoding probe approach has two limitations that we aim to solve with our new encoding probe approach: contributions of different features to model representatio…

  205. arXiv cs.CL TIER_1 · Kenneth J. K. Ong ·

    Impact of Task Phrasing on Presumptions in Large Language Models

    Concerns with the safety and reliability of applying large-language models (LLMs) in unpredictable real-world applications motivate this study, which examines how task phrasing can lead to presumptions in LLMs, making it difficult for them to adapt when the task deviates from the…

  206. arXiv cs.CL TIER_1 · Jaap Jumelet, Leonie Weissweiler, Joakim Nivre, Arianna Bisazza ·

    MultiBLiMP 1.0: A Massively Multilingual Benchmark of Linguistic Minimal Pairs

    arXiv:2504.02768v4 Announce Type: replace Abstract: We introduce MultiBLiMP 1.0, a massively multilingual benchmark of linguistic minimal pairs, covering 101 languages and 2 types of subject-verb agreement, containing more than 128,000 minimal pairs. Our minimal pairs are created…

  207. arXiv cs.AI TIER_1 · Jinquan Zheng, Jia Yuan, Jiacheng Yao, Chenyang Gu, Pujun Zheng, Guoxiu He ·

    Mitigating Selection Bias in Large Language Models via Permutation-Aware GRPO

    arXiv:2603.21016v2 Announce Type: replace-cross Abstract: Large language models (LLMs) used for multiple-choice and pairwise evaluation tasks often exhibit selection bias due to non-semantic factors like option positions and label symbols. Existing inference-time debiasing is cos…

  208. arXiv cs.CL TIER_1 · Shihan Dou, Yujiong Shen, Chenhao Huang, Junjie Ye, Jiayi Chen, Junzhe Wang, Qianyu He, Shichun Liu, Changze Lv, Jiahang Lin, Jiazheng Zhang, Ming Zhang, Shaofan Liu, Tao Ji, Zhangyue Yin, Cheng Zhang, Huaibing Xie, Jianglu Hu, Jingcheng Deng, Lincheng Li ·

    CL-bench Life: Can Language Models Learn from Real-Life Context?

    arXiv:2604.27043v1 Announce Type: new Abstract: Today's AI assistants such as OpenClaw are designed to handle context effectively, making context learning an increasingly important capability for models. As these systems move beyond professional settings into everyday life, the n…

  209. arXiv cs.CL TIER_1 · M. K. Khalidi Siam, Md. Tausif-Ul-Islam, Md. Reshad Romim Khan, Mohammed Ali Hossain, Mushfiqul Amin, Labib Hasan Khan, Niloy Farhan, Farig Sadeque ·

    Exploring the Limits of Pruning: Task-Specific Neurons, Model Collapse, and Recovery in Task-Specific Large Language Models

    arXiv:2604.27115v1 Announce Type: new Abstract: Neuron pruning is widely used to reduce the computational cost and parameter footprint of large language models, yet it remains unclear whether neurons in task-specific models contribute uniformly to task performance. In this work, …

  210. arXiv cs.CL TIER_1 · Camelia Baluta ·

    Cross-Lingual Response Consistency in Large Language Models: An ILR-Informed Evaluation of Claude Across Six Languages

    arXiv:2604.27137v1 Announce Type: new Abstract: This paper introduces a systematic evaluation framework grounded in the Interagency Language Roundtable (ILR) Skill Level Descriptions and applies it to Claude (Sonnet 4.6) across six languages: English, French, Romanian, Spanish, I…

  211. arXiv cs.CL TIER_1 · Austin C. Kozlowski, Andrei Boutyline ·

    Semantic Structure of Feature Space in Large Language Models

    arXiv:2604.27169v1 Announce Type: new Abstract: We show that the geometric relations between semantic features in large language models' hidden states closely mirror human psychological associations. We construct feature vectors corresponding to 360 words and project them on 32 s…

  212. arXiv cs.CL TIER_1 · Th\'eo Gigant, Bowen Peng, Jeffrey Quesnelle ·

    Decoupling the Benefits of Subword Tokenization for Language Model Training via Byte-level Simulation

    arXiv:2604.27263v1 Announce Type: new Abstract: Subword tokenization is an essential part of modern large language models (LLMs), yet its specific contributions to training efficiency and model performance remain poorly understood. In this work, we decouple the effects of subword…

  213. arXiv cs.CL TIER_1 · Minori Noguchi ·

    Exploring Applications of Transfer-State Large Language Models: Cognitive Profiling and Socratic AI Tutoring

    arXiv:2604.27454v1 Announce Type: new Abstract: Large language models (LLMs) sometimes exhibit qualitative shifts in response style under sustained self-referential dialogue conditions (Berg et al., 2025). This study refers to this phenomenon as "transfer" and explores the applic…

  214. arXiv cs.CL TIER_1 · Thibault Ba\~neras-Roux, Micka\"el Rouvier, Jane Wottawa, Richard Dufour ·

    Qualitative Evaluation of Language Model Rescoring in Automatic Speech Recognition

    arXiv:2604.27533v1 Announce Type: new Abstract: Evaluating automatic speech recognition (ASR) systems is a classical but difficult and still open problem, which often boils down to focusing only on the word error rate (WER). However, this metric suffers from many limitations and …

  215. arXiv cs.LG TIER_1 · Lei Li, Xingwen Yu, Jianguo Ni, Junxuan Zhu, Jieqiong Zhang, Jian Zhao, Zhi Liu ·

    ChipLingo: A Systematic Training Framework for Large Language Models in EDA

    arXiv:2604.27415v1 Announce Type: new Abstract: With the rapid advancement of semiconductor technology, Electronic Design Automation (EDA) has become an increasingly knowledge-intensive and document-driven engineering domain. Although large language models (LLMs) have shown stron…

  216. arXiv cs.AI TIER_1 · Wenhao Yuan, Chenchen Lin, Jian Chen, Jinfeng Xu, Shuo Yang, Edith Cheuk Han Ngai ·

    Belief-Guided Inference Control for Large Language Model Services via Verifiable Observations

    arXiv:2604.27536v1 Announce Type: new Abstract: In black-box large language model (LLM) services, response reliability is often only partially observable at decision time, while stronger inference pathways incur substantial computational cost, inducing a budgeted sequential decis…

  217. arXiv cs.AI TIER_1 · Shuzheng Si, Haozhe Zhao, Yu Lei, Qingyi Wang, Dingwei Chen, Zhitong Wang, Zhenhailong Wang, Kangyang Luo, Zheng Wang, Gang Chen, Fanchao Qi, Minjia Zhang, Maosong Sun ·

    From Context to Skills: Can Language Models Learn from Context Skillfully?

    arXiv:2604.27660v1 Announce Type: new Abstract: Many real-world tasks require language models (LMs) to reason over complex contexts that exceed their parametric knowledge. This calls for context learning, where LMs directly learn relevant knowledge from the given context. An intu…

  218. arXiv cs.AI TIER_1 · Ansar Aynetdinov, Patrick Haller, Alan Akbik ·

    Repetition over Diversity: High-Signal Data Filtering for Sample-Efficient German Language Modeling

    arXiv:2604.28075v1 Announce Type: cross Abstract: Recent research has shown that filtering massive English web corpora into high-quality subsets significantly improves training efficiency. However, for high-resource non-English languages like German, French, or Japanese, aggressi…

  219. arXiv cs.AI TIER_1 · Juan Manuel Contreras ·

    Policy-Grounded Safety Evaluation of 20 Large Language Models

    arXiv:2507.14719v2 Announce Type: replace Abstract: As large language models (LLMs) become increasingly integrated into real-world applications, scalable and rigorous safety evaluation is essential. This paper introduces Aymara AI, a programmatic platform for generating and admin…

  220. arXiv cs.AI TIER_1 · Nuno Fachada, Daniel Fernandes, Carlos M. Fernandes, Jo\~ao P. Matos-Carvalho ·

    Can Large Language Models Implement Agent-Based Models? An ODD-based Replication Study

    arXiv:2602.10140v2 Announce Type: replace-cross Abstract: Large language models (LLMs) can now synthesize non-trivial executable code from textual descriptions, raising an important question: can LLMs reliably implement agent-based models from standardized specifications in a way…

  221. arXiv cs.AI TIER_1 · Lingwei Gu, Nour Jedidi, Jimmy Lin ·

    NanoKnow: How to Know What Your Language Model Knows

    arXiv:2602.20122v2 Announce Type: replace-cross Abstract: How do large language models (LLMs) know what they know? Answering this question has been difficult because pre-training data is often a "black box" - unknown or inaccessible. The recent release of nanochat - a family of s…

  222. arXiv cs.CL TIER_1 · DouDou Zhou ·

    Unlearning What Matters: Token-Level Attribution for Precise Language Model Unlearning

    Machine unlearning has emerged as a critical capability for addressing privacy, safety, and regulatory concerns in large language models (LLMs). Existing methods operate at the sequence level, applying uniform updates across all tokens despite only a subset encoding the knowledge…

  223. arXiv cs.CL TIER_1 · Hamidreza Saghir ·

    How Language Models Process Out-of-Distribution Inputs: A Two-Pathway Framework

    Recent white-box OOD detection methods for LLMs -- including CED, RAUQ, and WildGuard confidence scores -- appear effective, but we show they are structurally confounded by sequence length (|r| >= 0.61) and collapse to near-chance under length-matched evaluation. Even raw attenti…

  224. arXiv cs.AI TIER_1 · Wenjing Lou ·

    ARMOR 2025: A Military-Aligned Benchmark for Evaluating Large Language Model Safety Beyond Civilian Contexts

    Large language models (LLMs) are now being explored for defense applications that require reliable and legally compliant decision support. They also hold significant potential to enhance decision making, coordination, and operational efficiency in military contexts. These uses de…

  225. arXiv cs.CL TIER_1 · Vinicius F. Caridá ·

    NorBERTo: A ModernBERT Model Trained for Portuguese with 331 Billion Tokens Corpus

    High-quality corpora are essential for advancing Natural Language Processing (NLP) in Portuguese. Building on previous encoder-only models such as BERTimbau and Albertina PT-BR, we introduce NorBERTo, a modern encoder based on the ModernBERT architecture, featuring long-context s…

  226. arXiv cs.CL TIER_1 · Alan Akbik ·

    Repetition over Diversity: High-Signal Data Filtering for Sample-Efficient German Language Modeling

    Recent research has shown that filtering massive English web corpora into high-quality subsets significantly improves training efficiency. However, for high-resource non-English languages like German, French, or Japanese, aggressive filtering creates a strategic dilemma: should p…

  227. arXiv cs.CL TIER_1 · Alan Akbik ·

    Repetition over Diversity: High-Signal Data Filtering for Sample-Efficient German Language Modeling

    Recent research has shown that filtering massive English web corpora into high-quality subsets significantly improves training efficiency. However, for high-resource non-English languages like German, French, or Japanese, aggressive filtering creates a strategic dilemma: should p…

  228. Hugging Face Daily Papers TIER_1 ·

    Qualitative Evaluation of Language Model Rescoring in Automatic Speech Recognition

    Evaluating automatic speech recognition (ASR) systems is a classical but difficult and still open problem, which often boils down to focusing only on the word error rate (WER). However, this metric suffers from many limitations and does not allow an in-depth analysis of automatic…

  229. arXiv cs.CL TIER_1 · Richard Dufour ·

    Qualitative Evaluation of Language Model Rescoring in Automatic Speech Recognition

    Evaluating automatic speech recognition (ASR) systems is a classical but difficult and still open problem, which often boils down to focusing only on the word error rate (WER). However, this metric suffers from many limitations and does not allow an in-depth analysis of automatic…

  230. arXiv cs.CL TIER_1 · Minori Noguchi ·

    Exploring Applications of Transfer-State Large Language Models: Cognitive Profiling and Socratic AI Tutoring

    Large language models (LLMs) sometimes exhibit qualitative shifts in response style under sustained self-referential dialogue conditions (Berg et al., 2025). This study refers to this phenomenon as "transfer" and explores the application potential of LLMs in a transfer state. As …

  231. arXiv cs.CL TIER_1 · Jinho Choo, JunSeung Lee, Jimyeong Kim, Yeeho Song, S. K. Hong, Yeong-Dae Kwon ·

    TLPO: Token-Level Policy Optimization for Mitigating Language Confusion in Large Language Models

    arXiv:2604.26553v1 Announce Type: new Abstract: Large language models (LLMs) demonstrate strong multilingual capabilities, yet often fail to consistently generate responses in the intended language, exhibiting a phenomenon known as language confusion. Prior mitigation approaches …

  232. arXiv cs.CL TIER_1 · Mengya Hu, Qiong Wei, Sandeep Atluri ·

    From Prompt Risk to Response Risk: Paired Analysis of Safety Behavior of Large Language Model

    arXiv:2604.26052v1 Announce Type: new Abstract: Safety evaluations of large language models (LLMs) typically report binary outcomes such as attack success rate, refusal rate, or harmful/not-harmful response classification. While useful, these can hide how risk changes between a u…

  233. arXiv cs.CL TIER_1 · Rei Emura, Saku Sugawara ·

    A Dual-Task Paradigm to Investigate Sentence Comprehension Strategies in Language Models

    arXiv:2604.26351v1 Announce Type: new Abstract: Language models (LMs) behave more like humans when their cognitive resources are restricted, particularly in predicting sentence processing costs such as reading times. However, it remains unclear whether such constraints similarly …

  234. arXiv cs.CL TIER_1 · Gongbo Zhang, Wen Wang, Ye Tian, Li Yuan ·

    Turning the TIDE: Cross-Architecture Distillation for Diffusion Large Language Models

    arXiv:2604.26951v1 Announce Type: new Abstract: Diffusion large language models (dLLMs) offer parallel decoding and bidirectional context, but state-of-the-art dLLMs require billions of parameters for competitive performance. While existing distillation methods for dLLMs reduce i…

  235. arXiv cs.CL TIER_1 · Bao Pham, Mohammed J. Zaki, Luca Ambrogioni, Dmitry Krotov, Matteo Negri ·

    Language Diffusion Models are Associative Memories Capable of Retrieving Unseen Data

    arXiv:2604.26841v1 Announce Type: cross Abstract: When do language diffusion models memorize their training data, and how to quantitatively assess their true generative regime? We address these questions by showing that Uniform-based Discrete Diffusion Models (UDDMs) fundamentall…

  236. arXiv cs.CL TIER_1 · Chahat Raj, Mahika Banerjee, Jinhao Pan, Aylin Caliskan, Antonios Anastasopoulos, Ziwei Zhu ·

    Talent or Luck? Evaluating Attribution Bias in Large Language Models

    arXiv:2505.22910v2 Announce Type: replace Abstract: When a student fails an exam, do we tend to blame their effort or the test's difficulty? Attribution, defined as how reasons are assigned to event outcomes, shapes perceptions, reinforces stereotypes, and influences decisions. A…

  237. arXiv cs.CL TIER_1 · Davyd Naveriani, Albert Zeyer, Ralf Schl\"uter, Hermann Ney ·

    Diffusion Language Models for Speech Recognition

    arXiv:2604.14001v2 Announce Type: replace Abstract: Diffusion language models have recently emerged as a leading alternative to standard language models, due to their ability for bidirectional attention and parallel text generation. In this work, we explore variants for their use…

  238. arXiv cs.CL TIER_1 · Thibault Ba\~neras-Roux, Shashi Kumar, Driss Khalil, Sergio Burdisso, Petr Motlicek, Shiran Liu, Mickael Rouvier, Jane Wottawa, Richard Dufour ·

    Evaluation of Automatic Speech Recognition Using Generative Large Language Models

    arXiv:2604.21928v2 Announce Type: replace Abstract: Automatic Speech Recognition (ASR) is traditionally evaluated using Word Error Rate (WER), a metric that is insensitive to meaning. Embedding-based semantic metrics are better correlated with human perception, but decoder-based …

  239. arXiv cs.CL TIER_1 · Wenxuan Wang, Yuk-Kit Chan, Zixuan Ling, Juluan Shi, Youliang Yuan, Jen-tse Huang, Yifei Zhang, Wenxiang Jiao, Zhaopeng Tu, Michael R. Lyu ·

    Identifying the Achilles' Heel: An Iterative Method for Dynamically Uncovering Factual Errors in Large Language Models

    arXiv:2401.00761v2 Announce Type: replace-cross Abstract: Large Language Models (LLMs) like ChatGPT are foundational in various applications due to their extensive knowledge from pre-training and fine-tuning. Despite this, they are prone to generating factual and commonsense erro…

  240. arXiv cs.AI TIER_1 · Xinru Yan, Boxi Cao, Yaojie Lu, Hongyu Lin, Weixiang Zhou, Le Sun, Xianpei Han ·

    Beyond Text-Dominance: Understanding Modality Preference of Omni-modal Large Language Models

    arXiv:2604.16902v3 Announce Type: replace Abstract: Native Omni-modal Large Language Models (OLLMs) have shifted from pipeline architectures to unified representation spaces. However, this native integration gives rise to a critical yet underexplored phenomenon: modality preferen…

  241. arXiv cs.LG TIER_1 · Chao Jiang, Dugang Liu, Cheng Wen, Zhiwu Xu, Hua Zheng, Muhammad Sadiq, Jawwad Ahmed Shamsi, Shengchao Qin, Zhong Ming ·

    Large Language Models for Multilingual Code Intelligence: A Survey

    arXiv:2604.25960v1 Announce Type: cross Abstract: Large language models have transformed AI-assisted software engineering, but current research remains biased toward high-resource languages such as Python, with weaker performance in languages like Rust and OCaml. Since real-world…

  242. arXiv cs.CL TIER_1 · Jeffrey Quesnelle ·

    Decoupling the Benefits of Subword Tokenization for Language Model Training via Byte-level Simulation

    Subword tokenization is an essential part of modern large language models (LLMs), yet its specific contributions to training efficiency and model performance remain poorly understood. In this work, we decouple the effects of subword tokenization by isolating them within a control…

  243. arXiv cs.CL TIER_1 · Andrei Boutyline ·

    Semantic Structure of Feature Space in Large Language Models

    We show that the geometric relations between semantic features in large language models' hidden states closely mirror human psychological associations. We construct feature vectors corresponding to 360 words and project them on 32 semantic axes (e.g. beautiful-ugly, soft-hard), a…

  244. arXiv cs.CL TIER_1 · Camelia Baluta ·

    Cross-Lingual Response Consistency in Large Language Models: An ILR-Informed Evaluation of Claude Across Six Languages

    This paper introduces a systematic evaluation framework grounded in the Interagency Language Roundtable (ILR) Skill Level Descriptions and applies it to Claude (Sonnet 4.6) across six languages: English, French, Romanian, Spanish, Italian, and German. We administer a battery of 1…

  245. Hugging Face Daily Papers TIER_1 ·

    Cross-Lingual Response Consistency in Large Language Models: An ILR-Informed Evaluation of Claude Across Six Languages

    This paper introduces a systematic evaluation framework grounded in the Interagency Language Roundtable (ILR) Skill Level Descriptions and applies it to Claude (Sonnet 4.6) across six languages: English, French, Romanian, Spanish, Italian, and German. We administer a battery of 1…

  246. arXiv cs.CL TIER_1 · Farig Sadeque ·

    Exploring the Limits of Pruning: Task-Specific Neurons, Model Collapse, and Recovery in Task-Specific Large Language Models

    Neuron pruning is widely used to reduce the computational cost and parameter footprint of large language models, yet it remains unclear whether neurons in task-specific models contribute uniformly to task performance. In this work, we provide empirical evidence for the existence …

  247. arXiv cs.CL TIER_1 · Li Yuan ·

    Turning the TIDE: Cross-Architecture Distillation for Diffusion Large Language Models

    Diffusion large language models (dLLMs) offer parallel decoding and bidirectional context, but state-of-the-art dLLMs require billions of parameters for competitive performance. While existing distillation methods for dLLMs reduce inference steps within a single architecture, non…

  248. arXiv cs.CL TIER_1 · Matteo Negri ·

    Language Diffusion Models are Associative Memories Capable of Retrieving Unseen Data

    When do language diffusion models memorize their training data, and how to quantitatively assess their true generative regime? We address these questions by showing that Uniform-based Discrete Diffusion Models (UDDMs) fundamentally behave as Associative Memories (AMs) $\textit{wi…

  249. arXiv cs.CL TIER_1 · Yeong-Dae Kwon ·

    TLPO: Token-Level Policy Optimization for Mitigating Language Confusion in Large Language Models

    Large language models (LLMs) demonstrate strong multilingual capabilities, yet often fail to consistently generate responses in the intended language, exhibiting a phenomenon known as language confusion. Prior mitigation approaches based on sequence-level fine-tuning, such as DPO…

  250. arXiv cs.CL TIER_1 · Saku Sugawara ·

    A Dual-Task Paradigm to Investigate Sentence Comprehension Strategies in Language Models

    Language models (LMs) behave more like humans when their cognitive resources are restricted, particularly in predicting sentence processing costs such as reading times. However, it remains unclear whether such constraints similarly affect sentence comprehension strategies. Beside…

  251. arXiv cs.CL TIER_1 Deutsch(DE) · Shu Yang, Shenzhe Zhu, Hao Zhu, Jos\'e Ram\'on Enr\'iquez, Di Wang, Alex Pentland, Michiel A. Bakker, Jiaxin Pei ·

    Multi-User Large Language Model Agents

    arXiv:2604.08567v2 Announce Type: replace Abstract: Large language models (LLMs) and LLM-based agents are increasingly deployed as assistants in planning and decision making, yet most existing systems are implicitly optimized for a single-principal interaction paradigm, in which …

  252. arXiv cs.CL TIER_1 · Feng Gu, Zongxia Li, Carlos Rafael Colon, Benjamin Evans, Ishani Mondal, Jordan Lee Boyd-Graber ·

    Large Language Models Are Effective Human Annotation Assistants, But Not Good Independent Annotators

    arXiv:2503.06778v3 Announce Type: replace Abstract: Event annotation is important for identifying market changes, monitoring breaking news, and understanding sociological trends. Although expert annotators set the gold standards, human coding is expensive and inefficient. Unlike …

  253. arXiv cs.CL TIER_1 · Alexandra Dragomir, Ioana Pintilie, Antonio Barbalau, Marius Dragoi, Florin Brad, Cristian Daniel Paduraru, Alexandru Tifrea, Elena Burceanu, Radu Tudor Ionescu ·

    JumpLoRA: Sparse Adapters for Continual Learning in Large Language Models

    arXiv:2604.16171v3 Announce Type: replace-cross Abstract: Adapter-based methods have become a cost-effective approach to continual learning (CL) for Large Language Models (LLMs), by sequentially learning a low-rank update matrix for each task. To mitigate catastrophic forgetting,…

  254. arXiv cs.CL TIER_1 · Ziyan Wang, Enmao Diao, Qi Le, Pu Wang, Minwoo Lee, Shu-ping Yeh, Evgeny Stupachenko, Hao Feng, Li Yang ·

    From Local to Global: Revisiting Structured Pruning Paradigms for Large Language Models

    arXiv:2510.18030v2 Announce Type: replace Abstract: Structured pruning is a practical approach to deploying large language models (LLMs) efficiently, as it yields compact, hardware-friendly architectures. However, the dominant local paradigm is task-agnostic: by optimizing layer-…

  255. arXiv cs.CL TIER_1 · Seok Hwan Song, Mohna Chakraborty, Qi Li, Wallapak Tavanapong ·

    Is Large Language Model Performance on Reasoning Tasks Impacted by Different Ways Questions Are Asked?

    arXiv:2507.15707v2 Announce Type: replace Abstract: Large Language Models (LLMs) have been evaluated using diverse question types, e.g., multiple-choice, true/false, and short/long answers. This study answers an unexplored question about the impact of different question types on …

  256. arXiv cs.CL TIER_1 · Nirmalendu Prakash, Yeo Wei Jie, Amir Abdullah, Ranjan Satapathy, Erik Cambria, Roy Ka Wei Lee ·

    Beyond I'm Sorry, I Can't: Dissecting Large Language Model Refusal

    arXiv:2509.09708v3 Announce Type: replace Abstract: Refusal on harmful prompts is a key safety behaviour in instruction-tuned large language models (LLMs), yet the internal causes of this behaviour remain poorly understood. We study two public instruction-tuned models, Gemma-2-2B…

  257. arXiv cs.LG TIER_1 · Ajmain Inqiad Alam, Palash Roy, Chanchal K. Roy, Banani Roy, Kevin A. Schneider ·

    Carbon-Taxed Transformers: A Green Compression Pipeline for Overgrown Language Models

    arXiv:2604.25903v1 Announce Type: cross Abstract: The accelerating adoption of Large Language Models (LLMs) in software engineering (SE) has brought with it a silent crisis: unsustainable computational cost. While these models demonstrate remarkable capabilities in different SE t…

  258. arXiv cs.CL TIER_1 · Yuanhao Zeng, Ao Lu, Lufei Li, Zheng Zhang, Yexin Li, Kan Ren ·

    Large Language Models Explore by Latent Distilling

    arXiv:2604.24927v1 Announce Type: new Abstract: Generating diverse responses is crucial for test-time scaling of large language models (LLMs), yet standard stochastic sampling mostly yields surface-level lexical variation, limiting semantic exploration. In this paper, we propose …

  259. arXiv cs.CL TIER_1 · Abhinav Kumar Singh, Harsha Vardhan Khurdula, Yoeven D Khemlani, Vineet Agarwal ·

    The Structured Output Benchmark: A Multi-Source Benchmark for Evaluating Structured Output Quality in Large Language Models

    arXiv:2604.25359v1 Announce Type: new Abstract: Large Language Models are increasingly being deployed to extract structured data from unstructured and semi-structured sources: parsing invoices, medical records, and converting PDF documents to database entries. Yet existing benchm…

  260. arXiv cs.CL TIER_1 · Fan Jiang, Yu Zhao, Chenyang Lyu, Tianqi Shi, Yichao Du, Feihu Jiang, Longyue Wang, Weihua Luo ·

    Marco-MoE: Open Multilingual Mixture-of-Expert Language Models with Efficient Upcycling

    arXiv:2604.25578v1 Announce Type: new Abstract: We present Marco-MoE, a suite of fully open multilingual sparse Mixture-of-Experts (MoE) models. Marco-MoE features a highly sparse design in which only around 5\% of the total parameters are activated per input token. This extreme …

  261. arXiv cs.CL TIER_1 · Sharma Aditya, Agarwal Vinti, Kumar Rajesh ·

    G-Loss: Graph-Guided Fine-Tuning of Language Models

    arXiv:2604.25853v1 Announce Type: new Abstract: Traditional loss functions, including cross-entropy, contrastive, triplet, and su pervised contrastive losses, used for fine-tuning pre-trained language models such as BERT, operate only within local neighborhoods and fail to accoun…

  262. arXiv cs.CL TIER_1 · Minkyu Kim, Vincent-Daniel Yun, Youngrae Kim, Youngjin Heo, Suin Cho, Seong-hun Kim, Woosang Lim, Gaeul Kwon ·

    Rethinking Layer Redundancy in Large Language Models: Calibration Objectives and Search for Depth Pruning

    arXiv:2604.24938v1 Announce Type: cross Abstract: Depth pruning improves the inference efficiency of large language models by removing Transformer blocks. Prior work has focused on importance criteria and search algorithms, often treating layer redundancy as an inherent structura…

  263. arXiv cs.CL TIER_1 · Chun-Yi Kuan, Wei-Ping Huang, Hung-yi Lee ·

    Walking Through Uncertainty: An Empirical Study of Uncertainty Estimation for Audio-Aware Large Language Models

    arXiv:2604.25591v1 Announce Type: cross Abstract: Recent audio-aware large language models (ALLMs) have demonstrated strong capabilities across diverse audio understanding and reasoning tasks, but they still frequently produce hallucinated or overly confident outputs. While uncer…

  264. arXiv cs.CL TIER_1 · Inderjeet Nair, Jie Ruan, Lu Wang ·

    Value-Conflict Diagnostics Reveal Widespread Alignment Faking in Language Models

    arXiv:2604.20995v2 Announce Type: replace-cross Abstract: Alignment faking, where a model behaves aligned with developer policy when monitored but reverts to its own preferences when unobserved, is a concerning yet poorly understood phenomenon, in part because current diagnostic …

  265. arXiv cs.CL TIER_1 · Sandeep Atluri ·

    From Prompt Risk to Response Risk: Paired Analysis of Safety Behavior of Large Language Model

    Safety evaluations of large language models (LLMs) typically report binary outcomes such as attack success rate, refusal rate, or harmful/not-harmful response classification. While useful, these can hide how risk changes between a user's input and the model's response. We present…

  266. arXiv cs.LG TIER_1 · Kevin A. Schneider ·

    Carbon-Taxed Transformers: A Green Compression Pipeline for Overgrown Language Models

    The accelerating adoption of Large Language Models (LLMs) in software engineering (SE) has brought with it a silent crisis: unsustainable computational cost. While these models demonstrate remarkable capabilities in different SE tasks, they are unmanageably large, slow to deploy,…

  267. arXiv cs.CL TIER_1 · Kumar Rajesh ·

    G-Loss: Graph-Guided Fine-Tuning of Language Models

    Traditional loss functions, including cross-entropy, contrastive, triplet, and su pervised contrastive losses, used for fine-tuning pre-trained language models such as BERT, operate only within local neighborhoods and fail to account for the global semantic structure. We present …

  268. arXiv cs.AI TIER_1 · Simon Thorne ·

    Large language models eroding science understanding: an experimental study

    This paper is under review in AI and Ethics This study examines whether large language models (LLMs) can reliably answer scientific questions and demonstrates how easily they can be influenced by fringe scientific material. The authors modified custom LLMs to prioritise knowledge…

  269. arXiv cs.CL TIER_1 · Hung-yi Lee ·

    Walking Through Uncertainty: An Empirical Study of Uncertainty Estimation for Audio-Aware Large Language Models

    Recent audio-aware large language models (ALLMs) have demonstrated strong capabilities across diverse audio understanding and reasoning tasks, but they still frequently produce hallucinated or overly confident outputs. While uncertainty estimation has been extensively studied in …

  270. arXiv cs.CL TIER_1 · Weihua Luo ·

    Marco-MoE: Open Multilingual Mixture-of-Expert Language Models with Efficient Upcycling

    We present Marco-MoE, a suite of fully open multilingual sparse Mixture-of-Experts (MoE) models. Marco-MoE features a highly sparse design in which only around 5\% of the total parameters are activated per input token. This extreme sparsity, combined with upcycling from dense mod…

  271. Hugging Face Daily Papers TIER_1 ·

    The Structured Output Benchmark: A Multi-Source Benchmark for Evaluating Structured Output Quality in Large Language Models

    Large Language Models are increasingly being deployed to extract structured data from unstructured and semi-structured sources: parsing invoices, medical records, and converting PDF documents to database entries. Yet existing benchmarks for structured output generation either foc…

  272. arXiv cs.CL TIER_1 · Vineet Agarwal ·

    The Structured Output Benchmark: A Multi-Source Benchmark for Evaluating Structured Output Quality in Large Language Models

    Large Language Models are increasingly being deployed to extract structured data from unstructured and semi-structured sources: parsing invoices, medical records, and converting PDF documents to database entries. Yet existing benchmarks for structured output generation either foc…

  273. arXiv cs.AI TIER_1 · Priyal Deep, Shane Emmons, Amy Fox, Kyle Bacon, Kelley McAllister, Krisztian Flautner ·

    Evaluation of Prompt Injection Defenses in Large Language Models

    arXiv:2604.23887v1 Announce Type: cross Abstract: LLM-powered applications routinely embed secrets in system prompts, yet models can be tricked into revealing them. We built an adaptive attacker that evolves its strategies over hundreds of rounds and tested it against nine defens…

  274. arXiv cs.CL TIER_1 · Nay Myat Min, Long H. Pham, Jun Sun ·

    Layerwise Convergence Fingerprints for Runtime Misbehavior Detection in Large Language Models

    arXiv:2604.24542v1 Announce Type: cross Abstract: Large language models deployed at runtime can misbehave in ways that clean-data validation cannot anticipate: training-time backdoors lie dormant until triggered, jailbreaks subvert safety alignment, and prompt injections override…

  275. arXiv cs.CL TIER_1 · Shiping Yang, Jie Wu, Wenbiao Ding, Ning Wu, Shining Liang, Ming Gong, Hongzhi Li, Hengyuan Zhang, Angel X. Chang, Dongmei Zhang ·

    Quantifying and Improving the Robustness of Retrieval-Augmented Language Models Against Spurious Features in Grounding Data

    arXiv:2503.05587v3 Announce Type: replace Abstract: Robustness has become a critical attribute for the deployment of RAG systems in real-world applications. Existing research focuses on robustness to explicit noise (e.g., document semantics) but overlooks implicit noise (spurious…

  276. arXiv cs.CL TIER_1 · Lovisa Hagstr\"om, Youna Kim, Haeun Yu, Sang-goo Lee, Richard Johansson, Hyunsoo Cho, Isabelle Augenstein ·

    CUB: Benchmarking Context Utilisation Techniques for Language Models

    arXiv:2505.16518v3 Announce Type: replace Abstract: Incorporating external knowledge is crucial for knowledge-intensive tasks, such as question answering and fact checking. However, language models (LMs) may ignore relevant information that contradicts outdated parametric memory …

  277. arXiv cs.CL TIER_1 · Danny Wang, Ruihong Qiu, Zi Huang ·

    When to Commit? Towards Variable-Size Self-Contained Blocks for Discrete Diffusion Language Models

    arXiv:2604.23994v1 Announce Type: cross Abstract: Discrete diffusion language models (dLLMs) enable parallel token updates with bidirectional attention, yet practical generation typically adopts blockwise semi-autoregressive decoding. This switch creates a training-inference mism…

  278. arXiv cs.CL TIER_1 · Jack King, Evelina Fedorenko, Eghbal A. Hosseini ·

    Representational Curvature Modulates Behavioral Uncertainty in Large Language Models

    arXiv:2604.23985v1 Announce Type: cross Abstract: In autoregressive large language models (LLMs), temporal straightening offers an account of how the next-token prediction objective shapes representations. Models learn to progressively straighten the representational trajectory o…

  279. arXiv cs.CL TIER_1 · Yu Wang, Leyi Lao, Langchu Huang, Gabriel Skantze, Yang Xu, Hendrik Buschmeier ·

    Investigating the Representation of Backchannels and Fillers in Fine-tuned Language Models

    arXiv:2509.20237v2 Announce Type: replace Abstract: Backchannels and fillers are important linguistic expressions in dialogue, but often treated as 'noise' to be bypassed in modern transformer-based language models (LMs). Here, we study how they are represented in LMs using three…

  280. arXiv cs.CL TIER_1 · Dikran Hovagimian ·

    Evolve: A Persistent Knowledge Lifecycle for Small Language Models

    arXiv:2604.23424v1 Announce Type: cross Abstract: Evolve pairs a small local language model with a persistent, teacher-compiled knowledge store -- refined through sleep consolidation and usage-driven refresh -- to deliver substantial accuracy gains over the model's parametric bas…

  281. arXiv cs.CL TIER_1 · Xulin Fan, Vishal Sunder, Samuel Thomas, Mark Hasegawa-Johnson, Brian Kingsbury, George Saon ·

    In-Sync: Adaptation of Speech Aware Large Language Models for ASR with Word Level Timestamp Predictions

    arXiv:2604.22817v1 Announce Type: cross Abstract: Recent advances in speech-aware language models have coupled strong acoustic encoders with large language models, enabling systems that move beyond transcription to produce richer outputs. Among these, word-level timestamp predict…

  282. arXiv cs.CL TIER_1 · Yunze Xiao, Vivienne J. Zhang, Chenghao Yang, Ningshan Ma, Weihao Xuan, Jen-tse Huang ·

    The Chameleon's Limit: Investigating Persona Collapse and Homogenization in Large Language Models

    arXiv:2604.24698v1 Announce Type: new Abstract: Applications based on large language models (LLMs), such as multi-agent simulations, require population diversity among agents. We identify a pervasive failure mode we term \emph{Persona Collapse}: agents each assigned a distinct pr…

  283. arXiv cs.CL TIER_1 · Riley Grossman, Yi Chen ·

    Zero-shot Large Language Models for Automatic Readability Assessment

    arXiv:2604.24470v1 Announce Type: new Abstract: Unsupervised automatic readability assessment (ARA) methods have important practical and research applications (e.g., ensuring medical or educational materials are suitable for their target audiences). In this paper, we propose a ne…

  284. arXiv cs.CL TIER_1 · Lisa Korver, Mohamed Mostagir, Sherief Reda ·

    A Multi-Dimensional Audit of Politically Aligned Large Language Models

    arXiv:2604.24429v1 Announce Type: new Abstract: As the application of Large Language Models (LLMs) spreads across various industries, there are increasing concerns about the potential for their misuse, especially in sensitive areas such as political discourse. Deliberately aligni…

  285. arXiv cs.CL TIER_1 · Jason Ramapuram, Eeshan Gunesh Dhekane, Amitis Shidani, Dan Busbridge, Bogdan Mazoure, Zijin Gu, Russ Webb, Tatiana Likhomanenko, Navdeep Jaitly ·

    Scaling Properties of Continuous Diffusion Spoken Language Models

    arXiv:2604.24416v1 Announce Type: new Abstract: Speech-only spoken language models (SLMs) lag behind text and text-speech models in performance, with recent discrete autoregressive (AR) SLMs indicating significant computational and data demands to match text models. Since discret…

  286. arXiv cs.CL TIER_1 · Zekun Yuan, Yangfan Ye, Xiaocheng Feng, Baohang Li, Qichen Hong, Yunfei Lu, Dandan Tu, Bing Qin ·

    Culture-Aware Machine Translation in Large Language Models: Benchmarking and Investigation

    arXiv:2604.24361v1 Announce Type: new Abstract: Large language models (LLMs) have achieved strong performance in general machine translation, yet their ability in culture-aware scenarios remains poorly understood. To bridge this gap, we introduce CanMT, a Culture-Aware Novel-Driv…

  287. arXiv cs.CL TIER_1 · Yimin Deng, Yejing Wang, Zhenxi Lin, Zichuan Fu, Guoshuai Zhao, Derong Xu, Yefeng Zheng, Xiangyu Zhao, Xian Wu, Li Zhu, Xueming Qian ·

    AdapTime: Enabling Adaptive Temporal Reasoning in Large Language Models

    arXiv:2604.24175v1 Announce Type: new Abstract: Large language models have demonstrated strong reasoning capabilities in general knowledge question answering. However, their ability to handle temporal information remains limited. To address this limitation, existing approaches of…

  288. arXiv cs.CL TIER_1 · Robert Litschko, Barbara Plank, Diego Frassinelli ·

    Resource-Lean Lexicon Induction for German Dialects

    arXiv:2604.23824v1 Announce Type: new Abstract: Automatic induction of high-quality dictionaries is essential for building lexical resources, yet low-resource languages and dialects pose several challenges: limited access to annotators, high degree of spelling variations, and poo…

  289. arXiv cs.CL TIER_1 · Bishwamittra Ghosh, Soumi Das, Till Speicher, Qinyuan Wu, Mohammad Aflah Khan, Deepak Garg, Krishna P. Gummadi, Evimaria Terzi ·

    Fine-tuning vs. In-context Learning in Large Language Models: A Formal Language Learning Perspective

    arXiv:2604.23267v1 Announce Type: new Abstract: Large language models (LLMs) operate in two fundamental learning modes - fine-tuning (FT) and in-context learning (ICL) - raising key questions about which mode yields greater language proficiency and whether they differ in their in…

  290. arXiv cs.CL TIER_1 · Brandon Hsu, Daniel Beaglehole, Adityanarayanan Radhakrishnan, Mikhail Belkin ·

    Contextual Linear Activation Steering of Language Models

    arXiv:2604.24693v1 Announce Type: new Abstract: Linear activation steering is a powerful approach for eliciting the capabilities of large language models and specializing their behavior using limited labeled data. While effective, existing methods often apply a fixed steering str…

  291. arXiv cs.CL TIER_1 · Harry Lu ·

    Measuring Temporal Linguistic Emergence in Diffusion Language Models

    arXiv:2604.23235v1 Announce Type: new Abstract: Diffusion language models expose an explicit denoising trajectory, making it possible to ask when different kinds of information become measurable during generation. We study three independent 32-step runs of LLaDA-8B-Base on masked…

  292. arXiv cs.CL TIER_1 · Yash Kumar Atri, Steven L. Johnson, Tom Hartvigsen ·

    Evaluating Temporal Consistency in Multi-Turn Language Models

    arXiv:2604.23051v1 Announce Type: new Abstract: Language models are increasingly deployed in interactive settings where users reason about facts over time rather than in isolation. In such scenarios, correct behavior requires models to maintain and update implicit temporal assump…

  293. arXiv cs.CL TIER_1 · Pouya Pezeshkpour, Estevam Hruschka ·

    AutoPyVerifier: Learning Compact Executable Verifiers for Large Language Model Outputs

    arXiv:2604.22937v1 Announce Type: new Abstract: Verification is becoming central to both reinforcement-learning-based training and inference-time control of large language models (LLMs). Yet current verifiers face a fundamental trade-off: LLM-based verifiers are expressive but ha…

  294. arXiv cs.CL TIER_1 · Jaros{\l}aw Hryszko ·

    The Randomness Floor: Measuring Intrinsic Non-Randomness in Language Model Token Distributions

    arXiv:2604.22771v1 Announce Type: new Abstract: Language models cannot be random. This paper introduces Entropic Deviation (ED), the normalised KL divergence between a model's token distribution and the uniform distribution, and measures it systematically across 31,200 generation…

  295. arXiv cs.CL TIER_1 · Ahmed M. Hussain, Salahuddin Salahuddin ·

    Beyond Context: Large Language Models' Failure to Grasp Users' Intent

    arXiv:2512.21110v3 Announce Type: replace-cross Abstract: Current Large Language Models (LLMs) safety approaches focus on explicitly harmful content while overlooking a critical vulnerability: the inability to understand context and recognize user intent. This creates exploitable…

  296. arXiv cs.CL TIER_1 · Xue Jiang, Ge Li, Jiaru Qian, Xianjie Shi, Chenjie Li, Hao Zhu, Ziyu Wang, Jielun Zhang, Zheyu Zhao, Lingwei Wu, Kechi Zhang, Jia Li, Wenpin Jiao, Zhi Jin, Yihong Dong ·

    KOCO-BENCH: Can Large Language Models Leverage Domain Knowledge in Software Development?

    arXiv:2601.13240v3 Announce Type: replace-cross Abstract: Large language models (LLMs) excel at general programming but struggle with domain-specific software development, necessitating domain specialization methods for LLMs to learn and utilize domain knowledge and data. However…

  297. arXiv cs.CL TIER_1 · Kristian Schwethelm, Daniel Rueckert, Georgios Kaissis ·

    How Much Is One Recurrence Worth? Iso-Depth Scaling Laws for Looped Language Models

    arXiv:2604.21106v2 Announce Type: replace-cross Abstract: We measure how much one extra recurrence is worth to a looped (depth-recurrent) language model, in equivalent unique parameters. From an iso-depth sweep of 116 pretraining runs across recurrence counts $r \in \{1, 2, 4, 8\…

  298. arXiv cs.LG TIER_1 · Divakar Kumar Yadav, Tian Zhao ·

    Hybrid JIT-CUDA Graph Optimization for Low-Latency Large Language Model Inference

    arXiv:2604.23467v1 Announce Type: new Abstract: Large Language Models (LLMs) have achieved strong performance across natural language and multimodal tasks, yet their practical deployment remains constrained by inference latency and kernel launch overhead, particularly in interact…

  299. arXiv cs.LG TIER_1 · Ziqing Wen, Ping Luo, Jiahuan Wang, Kun Yuan, Dongsheng Li, Tao Sun ·

    GWT: Scalable Optimizer State Compression for Large Language Model Training

    arXiv:2501.07237v5 Announce Type: replace Abstract: Large Language Models (LLMs) have demonstrated exceptional capabilities across diverse natural language processing benchmarks. However, the escalating scale of model parameters imposes prohibitive memory overheads during trainin…

  300. arXiv cs.LG TIER_1 · Ryan Chen, Youngmin Ko, Zeyu Zhang, Catherine Cho, Sunny Chung, Mauro Giuffr\'e, Dennis L. Shung, Bradly C. Stadie ·

    LAMP: Extracting Local Decision Surfaces From Large Language Models

    arXiv:2505.11772v3 Announce Type: replace Abstract: We introduce LAMP (Local Attribution Mapping Probe), a method that shines light onto a black-box language model's decision surface and studies how reliably a model maps its stated reasons to its reported predictions by approxima…

  301. arXiv cs.LG TIER_1 · Jiawei Chen, Zhengwei Fang, Yu Tian, Jiawei Du, Chao Yu, Zhaoxia Yin, Hang Su ·

    Exploring the Secondary Risks of Large Language Models

    arXiv:2506.12382v5 Announce Type: replace Abstract: Ensuring the safety and alignment of Large Language Models is a significant challenge with their growing integration into critical applications and societal functions. While prior research has primarily focused on jailbreak atta…

  302. arXiv cs.AI TIER_1 · Basel Shbita, Anna Lisa Gentile, Bing Zhang, Sungeun An, Shailja Thakur, Shubhi Asthana, Yi Zhou, Saptha Surendran, Farhan Ahmed, Rohan Kulkarni, Yuya Jeremy Ong, Chad DeLuca, Hima Patel ·

    A Systematic Approach for Large Language Models Debugging

    arXiv:2604.23027v1 Announce Type: new Abstract: Large language models (LLMs) have become central to modern AI workflows, powering applications from open-ended text generation to complex agent-based reasoning. However, debugging these models remains a persistent challenge due to t…

  303. arXiv cs.CL TIER_1 · Hikmat Karimov, Rahid Zahid Alekberli ·

    An Information-Geometric Framework for Stability Analysis of Large Language Models under Entropic Stress

    arXiv:2604.24076v1 Announce Type: cross Abstract: As large language models (LLMs) are increasingly deployed in high-stakes and operational settings, evaluation strategies based solely on aggregate accuracy are often insucient to characterize system reliability. This study propose…

  304. arXiv cs.AI TIER_1 · Yuxuan Jiang, Francis Ferraro ·

    SCRIBE: Structured Mid-Level Supervision for Tool-Using Language Models

    arXiv:2601.03555v2 Announce Type: replace Abstract: Training reliable tool-augmented agents remains a significant challenge, largely due to the difficulty of credit assignment in multi-step reasoning. While process-level reward models offer a promising direction, existing LLM-bas…

  305. arXiv cs.AI TIER_1 · Yifan Qian, Zhe Wen, Alexander C. Furnas, Yue Bai, Erzhuo Shao, Dashun Wang ·

    The Rise of Large Language Models and the Direction and Impact of US Federal Research Funding

    arXiv:2601.15485v2 Announce Type: replace-cross Abstract: Federal research funding shapes the direction, diversity, and impact of the US scientific enterprise. Large language models (LLMs) are rapidly diffusing into scientific practice, holding substantial promise while raising w…

  306. arXiv cs.CL TIER_1 · Zhijun Chen, Zeyu Ji, Qianren Mao, Hao Wu, Jinhuan Song, Junhang Cheng, Bangjie Qin, Zhuoran Li, Jingzheng Li, Kai Sun, Zizhe Wang, Yikun Ban, Zhu Sun, Xiangyang Ji, Hailong Sun ·

    Scoring, Reasoning, and Selecting the Best! Ensembling Large Language Models via a Peer-Review Process

    arXiv:2512.23213v3 Announce Type: replace Abstract: We propose LLM-PeerReview, an unsupervised LLM Ensemble method that selects the most ideal response from multiple LLM-generated candidates for each query, harnessing the collective wisdom of multiple models with diverse strength…

  307. arXiv cs.CL TIER_1 · Jingyi Sun, Pepa Atanasova, Sagnik Ray Choudhury, Sekh Mainul Islam, Isabelle Augenstein ·

    Evaluation Framework for Highlight Explanations of Context Utilisation in Language Models

    arXiv:2510.02629v3 Announce Type: replace Abstract: Context utilisation, the ability of Language Models (LMs) to incorporate relevant information from the provided context when generating responses, remains largely opaque to users, who cannot determine whether models draw from pa…

  308. arXiv cs.CL TIER_1 · Shi Feng, Hanlin Zhang, Fan Nie, Sham Kakade, Yiling Chen ·

    Peer-Predictive Self-Training for Language Model Reasoning

    arXiv:2604.13356v2 Announce Type: replace Abstract: Mechanisms for continued self-improvement of language models without external supervision remain an open challenge. We propose Peer-Predictive Self-Training (PST), a label-free fine-tuning framework in which multiple language mo…

  309. arXiv cs.CL TIER_1 · Chih-Kai Yang, Neo S. Ho, Hung-yi Lee ·

    Towards Holistic Evaluation of Large Audio-Language Models: A Comprehensive Survey

    arXiv:2505.15957v4 Announce Type: replace-cross Abstract: With advancements in large audio-language models (LALMs), which enhance large language models (LLMs) with auditory capabilities, these models are expected to demonstrate universal proficiency across various auditory tasks.…

  310. arXiv cs.CL TIER_1 · Kai-Wei Chang, En-Pei Hu, Chun-Yi Kuan, Wenze Ren, Wei-Chih Chen, Guan-Ting Lin, Yu Tsao, Shao-Hua Sun, Hung-yi Lee, James Glass ·

    Game-Time: Evaluating Temporal Dynamics in Spoken Language Models

    arXiv:2509.26388v3 Announce Type: replace-cross Abstract: Conversational Spoken Language Models (SLMs) are emerging as a promising paradigm for real-time speech interaction. However, their capacity of temporal dynamics, including the ability to manage timing, tempo and simultaneo…

  311. arXiv cs.CL TIER_1 · Gaeul Kwon ·

    Rethinking Layer Redundancy in Large Language Models: Calibration Objectives and Search for Depth Pruning

    Depth pruning improves the inference efficiency of large language models by removing Transformer blocks. Prior work has focused on importance criteria and search algorithms, often treating layer redundancy as an inherent structural property of pretrained networks. In contrast, we…

  312. arXiv cs.CL TIER_1 · Kan Ren ·

    Large Language Models Explore by Latent Distilling

    Generating diverse responses is crucial for test-time scaling of large language models (LLMs), yet standard stochastic sampling mostly yields surface-level lexical variation, limiting semantic exploration. In this paper, we propose Exploratory Sampling (ESamp), a decoding approac…

  313. arXiv cs.CL TIER_1 · Jen-tse Huang ·

    The Chameleon's Limit: Investigating Persona Collapse and Homogenization in Large Language Models

    Applications based on large language models (LLMs), such as multi-agent simulations, require population diversity among agents. We identify a pervasive failure mode we term \emph{Persona Collapse}: agents each assigned a distinct profile nonetheless converge into a narrow behavio…

  314. arXiv cs.CL TIER_1 · Mikhail Belkin ·

    Contextual Linear Activation Steering of Language Models

    Linear activation steering is a powerful approach for eliciting the capabilities of large language models and specializing their behavior using limited labeled data. While effective, existing methods often apply a fixed steering strength to all tokens, resulting in inconsistent s…

  315. arXiv cs.CL TIER_1 · Jun Sun ·

    Layerwise Convergence Fingerprints for Runtime Misbehavior Detection in Large Language Models

    Large language models deployed at runtime can misbehave in ways that clean-data validation cannot anticipate: training-time backdoors lie dormant until triggered, jailbreaks subvert safety alignment, and prompt injections override the deployer's instructions. Existing runtime def…

  316. Hugging Face Daily Papers TIER_1 ·

    Layerwise Convergence Fingerprints for Runtime Misbehavior Detection in Large Language Models

    Large language models deployed at runtime can misbehave in ways that clean-data validation cannot anticipate: training-time backdoors lie dormant until triggered, jailbreaks subvert safety alignment, and prompt injections override the deployer's instructions. Existing runtime def…

  317. arXiv cs.CL TIER_1 · Yi Chen ·

    Zero-shot Large Language Models for Automatic Readability Assessment

    Unsupervised automatic readability assessment (ARA) methods have important practical and research applications (e.g., ensuring medical or educational materials are suitable for their target audiences). In this paper, we propose a new zero-shot prompting methodology for ARA and pr…

  318. arXiv cs.CL TIER_1 · Sherief Reda ·

    A Multi-Dimensional Audit of Politically Aligned Large Language Models

    As the application of Large Language Models (LLMs) spreads across various industries, there are increasing concerns about the potential for their misuse, especially in sensitive areas such as political discourse. Deliberately aligning LLMs with specific political ideologies, thro…

  319. arXiv cs.CL TIER_1 · Navdeep Jaitly ·

    Scaling Properties of Continuous Diffusion Spoken Language Models

    Speech-only spoken language models (SLMs) lag behind text and text-speech models in performance, with recent discrete autoregressive (AR) SLMs indicating significant computational and data demands to match text models. Since discretizing continuous speech for AR creates bottlenec…

  320. Hugging Face Daily Papers TIER_1 ·

    Culture-Aware Machine Translation in Large Language Models: Benchmarking and Investigation

    Large language models (LLMs) have achieved strong performance in general machine translation, yet their ability in culture-aware scenarios remains poorly understood. To bridge this gap, we introduce CanMT, a Culture-Aware Novel-Driven Parallel Dataset for Machine Translation, tog…

  321. arXiv cs.CL TIER_1 · Bing Qin ·

    Culture-Aware Machine Translation in Large Language Models: Benchmarking and Investigation

    Large language models (LLMs) have achieved strong performance in general machine translation, yet their ability in culture-aware scenarios remains poorly understood. To bridge this gap, we introduce CanMT, a Culture-Aware Novel-Driven Parallel Dataset for Machine Translation, tog…

  322. arXiv cs.CL TIER_1 · Xueming Qian ·

    AdapTime: Enabling Adaptive Temporal Reasoning in Large Language Models

    Large language models have demonstrated strong reasoning capabilities in general knowledge question answering. However, their ability to handle temporal information remains limited. To address this limitation, existing approaches often involve external tools or manual verificatio…

  323. arXiv cs.CL TIER_1 · Rahid Zahid Alekberli ·

    An Information-Geometric Framework for Stability Analysis of Large Language Models under Entropic Stress

    As large language models (LLMs) are increasingly deployed in high-stakes and operational settings, evaluation strategies based solely on aggregate accuracy are often insucient to characterize system reliability. This study proposes a thermodynamic inspired modeling framework for …

  324. arXiv cs.CL TIER_1 · Ryoma Kumon, Hitomi Yanaka ·

    Fine-Grained Analysis of Shared Syntactic Mechanisms in Language Models

    arXiv:2604.22166v1 Announce Type: new Abstract: While language models demonstrate sophisticated syntactic capabilities, the extent to which their internal mechanisms align with cross-constructional principles studied in linguistics remains poorly understood. This study investigat…

  325. arXiv cs.CL TIER_1 · Shuowei Li, Haoxin Li, Wenda Chu, Yi Fang ·

    How Large Language Models Balance Internal Knowledge with User and Document Assertions

    arXiv:2604.22193v1 Announce Type: new Abstract: Large language models (LLMs) often need to balance their internal parametric knowledge with external information, such as user beliefs and content from retrieved documents, in real-world scenarios like RAG or chat-based systems. A m…

  326. arXiv cs.CL TIER_1 · Ayan Datta, Zhixue Zhao, Bhuvanesh Verma, Radhika Mamidi, Mounika Marreddy, Alexander Mehler ·

    Large Language Models Decide Early and Explain Later

    arXiv:2604.22266v1 Announce Type: new Abstract: Large Language Models often achieve strong performance by generating long intermediate chain-of-thought reasoning. However, it remains unclear when a model's final answer is actually determined during generation. If the answer is al…

  327. arXiv cs.CL TIER_1 · Weixu Zhang, Ye Yuan, Changjiang Han, Yuxing Tian, Zipeng Sun, Linfeng Du, Jikun Kang, Hong Kang, Xue Liu, Haolun Wu ·

    Preference Heads in Large Language Models: A Mechanistic Framework for Interpretable Personalization

    arXiv:2604.22345v1 Announce Type: new Abstract: Large Language Models (LLMs) exhibit strong implicit personalization ability, yet most existing approaches treat this behavior as a black box, relying on prompt engineering or fine tuning on user data. In this work, we adopt a mecha…

  328. arXiv cs.CL TIER_1 · Alberto Messina, Stefano Scotta ·

    Introducing Background Temperature to Characterise Hidden Randomness in Large Language Models

    arXiv:2604.22411v1 Announce Type: cross Abstract: Even when decoding with temperature $T=0$, large language models (LLMs) can produce divergent outputs for identical inputs. Recent work by Thinking Machines Lab highlights implementation-level sources of nondeterminism, including …

  329. arXiv cs.CL TIER_1 · Ishika Agarwal, Nimet Beyza Bozdag, Nisval Patel, Dilek Hakkani-T\"ur ·

    Language Specific Knowledge: Do Models Know Better in X than in English?

    arXiv:2505.14990v3 Announce Type: replace Abstract: Often, multilingual language models are trained with the objective to map semantically similar content (in different languages) in the same latent space. In this paper, we show a nuance in this training objective, and find that …

  330. arXiv cs.CL TIER_1 · Marco Baroni, Emily Cheng, Iria de-Dios-Flores, Francesca Franzon ·

    Tracing the complexity profiles of different linguistic phenomena through the intrinsic dimension of LLM representations

    arXiv:2601.03779v2 Announce Type: replace Abstract: We explore intrinsic dimension (ID) of LLM representations as a marker of linguistic complexity. Specifically, we test whether ID differences across model layers reflect well-known complexity contrasts established in (psycho)lin…

  331. arXiv cs.CL TIER_1 · Chao Xue, Yao Wang, Mengqiao Liu, Di Liang, Xingsheng Han, Peiyang Liu, Xianjie Wu, Chenyao Lu, Lei Jiang, Yu Lu, Haibo Shi, Shuang Liang, Minlong Peng, Flora D. Salim ·

    Why Supervised Fine-Tuning Fails to Learn: A Systematic Study of Incomplete Learning in Large Language Models

    arXiv:2604.10079v4 Announce Type: replace Abstract: Supervised Fine-Tuning (SFT) is the standard approach for adapting large language models (LLMs) to downstream tasks. However, we observe a persistent failure mode: even after convergence, models often fail to correctly reproduce…

  332. arXiv cs.CL TIER_1 · Bruno Puri, Jim Berend, Sebastian Lapuschkin, Wojciech Samek ·

    Atlas-Alignment: Making Interpretability Transferable Across Language Models

    arXiv:2510.27413v2 Announce Type: replace-cross Abstract: Interpretability is crucial for building safe, reliable, and controllable language models, yet existing interpretability pipelines remain costly and difficult to scale. Interpreting a new model typically requires training …

  333. arXiv cs.LG TIER_1 · Rico Angell, Raghav Singhal, Zachary Horvitz, Zhou Yu, Rajesh Ranganath, Kathleen McKeown, He He ·

    Estimating Tail Risks in Language Model Output Distributions

    arXiv:2604.22167v1 Announce Type: new Abstract: Language models are increasingly capable and are being rapidly deployed on a population-level scale. As a result, the safety of these models is increasingly high-stakes. Fortunately, advances in alignment have significantly reduced …

  334. arXiv cs.LG TIER_1 · Yijia Dai, Zhaolin Gao, Yahya Sattar, Sarah Dean, Jennifer J. Sun ·

    Pre-trained Large Language Models Learn Hidden Markov Models In-context

    arXiv:2506.07298v3 Announce Type: replace Abstract: Hidden Markov Models (HMMs) are foundational tools for modeling sequential data with latent Markovian structure, yet fitting them to real-world data remains computationally challenging. In this work, we show that pre-trained lar…

  335. arXiv cs.AI TIER_1 · Zewen Liu, Juntong Ni, Xianfeng Tang, Max S. Y. Lau, Qi He, Wenpeng Yin, Wei Jin ·

    Can Large Language Models Adequately Perform Symbolic Reasoning Over Time Series?

    arXiv:2508.03963v4 Announce Type: replace Abstract: Uncovering hidden symbolic laws from time series data, as an aspiration dating back to Kepler's discovery of planetary motion, remains a core challenge in scientific discovery and artificial intelligence. While Large Language Mo…

  336. arXiv cs.AI TIER_1 · Brandon Yee, Krishna Sharma ·

    Calibrating Behavioral Parameters with Large Language Models

    arXiv:2602.01022v2 Announce Type: replace-cross Abstract: Behavioral parameters such as loss aversion, herding, and extrapolation are central to asset pricing models but remain difficult to measure reliably. We develop a framework that treats large language models (LLMs) as calib…

  337. arXiv cs.CL TIER_1 · Zi Huang ·

    When to Commit? Towards Variable-Size Self-Contained Blocks for Discrete Diffusion Language Models

    Discrete diffusion language models (dLLMs) enable parallel token updates with bidirectional attention, yet practical generation typically adopts blockwise semi-autoregressive decoding. This switch creates a training-inference mismatch: training denoises with full-sequence context…

  338. arXiv cs.CL TIER_1 · Eghbal A. Hosseini ·

    Representational Curvature Modulates Behavioral Uncertainty in Large Language Models

    In autoregressive large language models (LLMs), temporal straightening offers an account of how the next-token prediction objective shapes representations. Models learn to progressively straighten the representational trajectory of input sequences across layers, potentially facil…

  339. arXiv cs.CL TIER_1 · Diego Frassinelli ·

    Resource-Lean Lexicon Induction for German Dialects

    Automatic induction of high-quality dictionaries is essential for building lexical resources, yet low-resource languages and dialects pose several challenges: limited access to annotators, high degree of spelling variations, and poor performance of large language models (LLMs). W…

  340. arXiv cs.CL TIER_1 · Stefano Scotta ·

    Introducing Background Temperature to Characterise Hidden Randomness in Large Language Models

    Even when decoding with temperature $T=0$, large language models (LLMs) can produce divergent outputs for identical inputs. Recent work by Thinking Machines Lab highlights implementation-level sources of nondeterminism, including batch-size variation, kernel non-invariance, and f…

  341. arXiv cs.CL TIER_1 · Haolun Wu ·

    Preference Heads in Large Language Models: A Mechanistic Framework for Interpretable Personalization

    Large Language Models (LLMs) exhibit strong implicit personalization ability, yet most existing approaches treat this behavior as a black box, relying on prompt engineering or fine tuning on user data. In this work, we adopt a mechanistic interpretability perspective and hypothes…

  342. arXiv cs.CL TIER_1 · Alexander Mehler ·

    Large Language Models Decide Early and Explain Later

    Large Language Models often achieve strong performance by generating long intermediate chain-of-thought reasoning. However, it remains unclear when a model's final answer is actually determined during generation. If the answer is already fixed at an intermediate stage, subsequent…

  343. arXiv cs.CL TIER_1 · Yi Fang ·

    How Large Language Models Balance Internal Knowledge with User and Document Assertions

    Large language models (LLMs) often need to balance their internal parametric knowledge with external information, such as user beliefs and content from retrieved documents, in real-world scenarios like RAG or chat-based systems. A model's ability to reliably process these sources…

  344. arXiv cs.AI TIER_1 · He He ·

    Estimating Tail Risks in Language Model Output Distributions

    Language models are increasingly capable and are being rapidly deployed on a population-level scale. As a result, the safety of these models is increasingly high-stakes. Fortunately, advances in alignment have significantly reduced the likelihood of harmful model outputs. However…

  345. arXiv cs.CL TIER_1 · Hitomi Yanaka ·

    Fine-Grained Analysis of Shared Syntactic Mechanisms in Language Models

    While language models demonstrate sophisticated syntactic capabilities, the extent to which their internal mechanisms align with cross-constructional principles studied in linguistics remains poorly understood. This study investigates whether models employ shared neural mechanism…

  346. arXiv cs.CL TIER_1 · Richard Dufour ·

    Evaluation of Automatic Speech Recognition Using Generative Large Language Models

    Automatic Speech Recognition (ASR) is traditionally evaluated using Word Error Rate (WER), a metric that is insensitive to meaning. Embedding-based semantic metrics are better correlated with human perception, but decoder-based Large Language Models (LLMs) remain underexplored fo…

  347. arXiv cs.CL TIER_1 · Taro Watanabe ·

    Revisiting Non-Verbatim Memorization in Large Language Models: The Role of Entity Surface Forms

    Understanding what kinds of factual knowledge large language models (LLMs) memorize is essential for evaluating their reliability and limitations. Entity-based QA is a common framework for analyzing non-verbatim memorization, but typical evaluations query each entity using a sing…

  348. arXiv cs.CL TIER_1 · Maziar Kianimoghadam Jouneghani ·

    MKJ at SemEval-2026 Task 9: A Comparative Study of Generalist, Specialist, and Ensemble Strategies for Multilingual Polarization

    We present a systematic study of multilingual polarization detection across 22 languages for SemEval-2026 Task 9 (Subtask 1), contrasting multilingual generalists with language-specific specialists and hybrid ensembles. While a standard generalist like XLM-RoBERTa suffices when i…

  349. Hugging Face Daily Papers TIER_1 ·

    Convergent Evolution: How Different Language Models Learn Similar Number Representations

    Language models trained on natural text learn to represent numbers using periodic features with dominant periods at $T=2, 5, 10$. In this paper, we identify a two-tiered hierarchy of these features: while Transformers, Linear RNNs, LSTMs, and classical word embeddings trained in …

  350. Hugging Face Daily Papers TIER_1 ·

    On the Quantization Robustness of Diffusion Language Models in Coding Benchmarks

    Auto-regressive Large Language Models (LLMs) achieve strong performance on coding tasks, but incur high memory and inference costs. Diffusion-based language models (d-LLMs) offer bounded inference cost via iterative denoising, but their behavior under post-training quantization (…

  351. Hugging Face Daily Papers TIER_1 ·

    Heterogeneity in Formal Linguistic Competence of Language Models: Is Data the Real Bottleneck?

    Large Language Models (LLMs) exhibit a puzzling disparity in their formal linguistic competence: while they learn some linguistic phenomena with near-perfect mastery, they often perform below chance on others, even after training on trillions of tokens. In this work, we investiga…

  352. Hugging Face Daily Papers TIER_1 ·

    LEPO: Latent Reasoning Policy Optimization for Large Language Models

    Recently, latent reasoning has been introduced into large language models (LLMs) to leverage rich information within a continuous space. However, without stochastic sampling, these methods inevitably collapse to deterministic inference, failing to discover diverse reasoning paths…

  353. Bounded Regret (Jacob Steinhardt) TIER_1 · Ruiqi Zhong ·

    Augmenting Statistical Models with Natural Language Parameters

    <p><em>This is a guest post by my student <a href="https://ruiqi-zhong.github.io/?ref=bounded-regret.ghost.io">Ruiqi Zhong</a>, who has some <a href="https://arxiv.org/abs/2409.08466?ref=bounded-regret.ghost.io">very exciting work</a> defining new families of statistical models t…

  354. arXiv stat.ML TIER_1 · Mohsen Hariri, Amirhossein Samandar, Michael Hinczewski, Vipin Chaudhary ·

    Don't Pass@k: A Bayesian Framework for Large Language Model Evaluation

    arXiv:2510.04265v4 Announce Type: replace-cross Abstract: Pass$@k$ is widely used to report the reasoning performance of LLMs, but it often produces unstable and potentially misleading rankings, especially when the number of trials (samples) is limited and computational resources…

  355. arXiv stat.ML TIER_1 · Xinhao Qu, Qiang Heng, Hao Zeng, Xiaoqian Liu ·

    An Interpretable and Scalable Framework for Evaluating Large Language Models

    arXiv:2605.07046v1 Announce Type: new Abstract: Evaluation of large language models (LLMs) is increasingly critical, yet standard benchmarking methods rely on average accuracy, overlooking both the inherent stochasticity of LLM outputs and the heterogeneity of benchmark items. It…

  356. LessWrong (AI tag) TIER_1 · Noah Weinberger ·

    Claude Does Not Actually Taste Bananas: Potassium-Based Synthetic Phenomenology In Language Models

    <p><a href="https://huggingface.co/blog/Clock070303/claude-does-not-actually-taste-bananas" rel="noreferrer"><span>I originally published this on Hugging Face: </span></a></p><p><span>For those of you who read my semi-serious musings seriously, you know I love a good benign adver…

  357. arXiv cs.CV TIER_1 · Jiatao Gu ·

    STARFlow2: Bridging Language Models and Normalizing Flows for Unified Multimodal Generation

    Deep generative models have advanced rapidly across text and vision, motivating unified multimodal systems that can understand, reason over, and generate interleaved text-image sequences. Most existing approaches combine autoregressive language modeling with diffusion-based image…

  358. arXiv stat.ML TIER_1 · Xiaoqian Liu ·

    An Interpretable and Scalable Framework for Evaluating Large Language Models

    Evaluation of large language models (LLMs) is increasingly critical, yet standard benchmarking methods rely on average accuracy, overlooking both the inherent stochasticity of LLM outputs and the heterogeneity of benchmark items. Item Response Theory (IRT) offers a principled fra…

  359. arXiv cs.CV TIER_1 · Yongdong Zhang ·

    Uncertainty-Aware Exploratory Direct Preference Optimization for Multimodal Large Language Models

    Direct Preference Optimization (DPO) has proven to be an effective solution for mitigating hallucination in Multimodal Large Language Models (MLLMs) by learning from preference pairs. One of its key challenges lies in how to transfer the sequence-level preference into fine-graine…

  360. arXiv cs.CV TIER_1 · Sucheng Ren, Chen Chen, Zhenbang Wang, Liangchen Song, Xiangxin Zhu, Alan Yuille, Liang-Chieh Chen, Jiasen Lu ·

    Large Language Models are Universal Reasoners for Visual Generation

    arXiv:2605.04040v1 Announce Type: new Abstract: Text-to-image generation has advanced rapidly with diffusion models, progressing from CLIP and T5 conditioning to unified systems where a single LLM backbone handles both visual understanding and generation. Despite the architectura…

  361. arXiv stat.ML TIER_1 · Chengchun Shi ·

    Perturbation is All You Need for Extrapolating Language Models

    We introduce a simple yet powerful framework for training large language models. In contrast to the standard autoregressive next-token prediction based on an exact prefix, we propose a perturbation-based procedure that first transforms the prefix into a semantic neighbor and then…

  362. arXiv cs.CV TIER_1 · Jiasen Lu ·

    Large Language Models are Universal Reasoners for Visual Generation

    Text-to-image generation has advanced rapidly with diffusion models, progressing from CLIP and T5 conditioning to unified systems where a single LLM backbone handles both visual understanding and generation. Despite the architectural unification, these systems frequently fail to …

  363. arXiv stat.ML TIER_1 · Jikai Jin, Vasilis Syrgkanis ·

    The Partial Testimony of Logs: Evaluation of Language Model Generation under Confounded Model Choice

    arXiv:2605.01311v1 Announce Type: cross Abstract: Offline evaluation of language models from usage logs is biased when model choice is confounded: the same user-side factors that influence which model is used can also influence how its output is judged, so raw comparisons of logg…

  364. arXiv cs.CV TIER_1 · Lei Lei, Jie Gu, Xiaokang Ma, Chu Tang, Jingmin Chen, Tong Xu ·

    Task-Related Token Compression in Multimodal Large Language Models from an Explainability Perspective

    arXiv:2506.01097v2 Announce Type: replace Abstract: Existing Multimodal Large Language Models (MLLMs) process a large number of visual tokens, leading to significant computational costs and inefficiency. Instruction-related visual token compression demonstrates strong task releva…

  365. LessWrong (AI tag) TIER_1 · Oliver Sourbut ·

    How did ‘large’ language models get that way? The role of Transformers and Pretraining in GPT

    <p><span>Large language models are </span><i><span>really</span></i><span> large. They’re among the largest machine learning projects ever, and set to be (perhaps already are by some measures) some of the </span><a href="https://www.mckinsey.com/industries/technology-media-and-te…

  366. arXiv stat.ML TIER_1 · Vasilis Syrgkanis ·

    The Partial Testimony of Logs: Evaluation of Language Model Generation under Confounded Model Choice

    Offline evaluation of language models from usage logs is biased when model choice is confounded: the same user-side factors that influence which model is used can also influence how its output is judged, so raw comparisons of logged scores mix self-selected populations rather tha…

  367. arXiv cs.CV TIER_1 · Fujun Han, Junan Chen, Xintong Zhu, Jingqi Ye, Xuanjie Mao, Tao Chen, Peng Ye ·

    Can Multimodal Large Language Models Truly Understand Small Objects?

    arXiv:2604.22884v1 Announce Type: new Abstract: Multimodal Large Language Models (MLLMs) have shown promising potential in diverse understanding tasks, e.g., image and video analysis, math and physics olympiads. However, they remain blank and unexplored for Small Object Understan…

  368. arXiv stat.ML TIER_1 · Minda Zhao, Yilun Du, Mengyu Wang ·

    Large Language Models Are Bad Dice Players: LLMs Struggle to Generate Random Numbers from Statistical Distributions

    arXiv:2601.05414v3 Announce Type: cross Abstract: As large language models (LLMs) transition from chat interfaces to integral components of stochastic pipelines and systems approaching general intelligence, the ability to faithfully sample from specified probability distributions…

  369. Smol AINews TIER_1 Français(FR) ·

    LLaDA: Large Language Diffusion Models

    **LLaDA (Large Language Diffusion Model) 8B** is a breakthrough diffusion-based language model that rivals **LLaMA 3 8B** while training on **7x fewer tokens (2 trillion tokens)** and using **0.13 million H800 GPU hours**. It introduces a novel text generation approach by predict…

  370. Smol AINews TIER_1 ·

    Mixture of Depths: Dynamically allocating compute in transformer-based language models

    **DeepMind** introduces the Mixture-of-Depths (MoD) technique, dynamically allocating FLOPs across transformer layers to optimize compute usage, achieving over **50% faster** forward passes without training impact. MoD selectively processes tokens using top-k routing, improving e…

  371. Smol AINews TIER_1 ·

    ReALM: Reference Resolution As Language Modeling

    **Apple** is advancing in AI with a new approach called **ReALM: Reference Resolution As Language Modeling**, which improves understanding of ambiguous references using three contexts and finetunes a smaller **FLAN-T5** model that outperforms **GPT-4** on this task. In Reddit AI …

  372. Hacker News — AI stories ≥50 points TIER_1 · giuliomagnifico ·

    Study: Back-to-basics approach can match or outperform AI in language analysis

  373. Practical AI TIER_1 · Practical AI LLC ·

    NLP for the world's 7000+ languages

    <p>Expanding AI technology to the local languages of emerging markets presents huge challenges. Good data is scarce or non-existent. Users often have bandwidth or connectivity issues. Existing platforms target only a small number of high-resource languages.</p><p>Our own Daniel W…

  374. Practical AI TIER_1 · Practical AI LLC ·

    BERT: one NLP model to rule them all

    <p><strong><em>Fully Connected</em></strong><em> – a series where Chris and Daniel keep you up to date with everything that’s happening in the AI community.</em></p><p>This week we discuss BERT, a new method of pre-training language representations from Google for natural languag…

  375. Towards AI TIER_1 · Gowtham Boyina ·

    Interpreting the Black Box: Natural Language Autoencoders Translate LLM Activations into Readable…

    <div class="medium-feed-item"><p class="medium-feed-image"><a href="https://pub.towardsai.net/interpreting-the-black-box-natural-language-autoencoders-translate-llm-activations-into-readable-c28f646a6ed5?source=rss----98111c9905da---4"><img src="https://cdn-images-1.medium.com/ma…

  376. Medium — fine-tuning tag TIER_1 · Jeya ·

    What are the popular methods for fine-tuning Large Language Models (LLMs)?

    <div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/@jeya.lakshmi/what-are-the-popular-methods-for-fine-tuning-large-language-models-llms-c4eee43bef07?source=rss------fine_tuning-5"><img src="https://cdn-images-1.medium.com/max/680/1*HxGfI8skix2…

  377. Medium — fine-tuning tag TIER_1 · Bhavika Pawani ·

    Generative AI from First Principles Article 11 : How Language Models Learn to Follow Instructions

    <div class="medium-feed-item"><p class="medium-feed-snippet">1. Where We Are in the Story</p><p class="medium-feed-link"><a href="https://medium.com/@bhavikapawani4444/generative-ai-from-first-principles-article-11-how-language-models-learn-to-follow-instructions-5860fb2316c5?sou…

  378. Medium — MLOps tag TIER_1 · Jpranav ·

    How to Fine-Tune a Small Language Model in 2026 — Without Wasting a Weekend

    <div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/@jpranav97/how-to-fine-tune-a-small-language-model-in-2026-without-wasting-a-weekend-c8cb8ba7297e?source=rss------mlops-5"><img src="https://cdn-images-1.medium.com/max/1344/1*4I2LMiE2EPa8HdFC9…

  379. Lobsters — AI tag TIER_1 · arxiv.org via darccio ·

    On the Limits of Self-Improving in Large Language Models: The Singularity Is Not Near Without Symbolic Model Synthesis

    <p><a href="https://lobste.rs/s/jgsiqa/on_limits_self_improving_large_language">Comments</a></p>

  380. HN — machine learning stories TIER_1 · 2bit ·

    Controlling Language and Diffusion Models by Transporting Activations

  381. dev.to — LLM tag TIER_1 · zeromathai ·

    How Large Language Models Work — From Transformers to Conversational AI

    <p>LLMs can look like magic from the outside.</p> <p>You type a prompt.</p> <p>The model generates language.</p> <p>But underneath that behavior is a clear architecture.</p> <h2> Core Idea </h2> <p>A Large Language Model is a neural network trained to understand and generate text…

  382. Mastodon — fosstodon.org TIER_1 · [email protected] ·

    On the Limits of Self-Improving in Large Language Models: The Singularity Is Not Near Without Symbolic Model Synthesis https:// arxiv.org/html/2601.05280v2 Anyo

    On the Limits of Self-Improving in Large Language Models: The Singularity Is Not Near Without Symbolic Model Synthesis https:// arxiv.org/html/2601.05280v2 Anyone out there in # AI expert land able to translate this into lower dimensional ordinary human speak? Bonus points if you…

  383. Mastodon — mastodon.social TIER_1 · lobsters ·

    On the Limits of Self-Improving in Large Language Models: The Singularity Is Not Near Without Symbolic Model Synthesis https:// lobste.rs/s/jgsiqa # ai https://

    On the Limits of Self-Improving in Large Language Models: The Singularity Is Not Near Without Symbolic Model Synthesis https:// lobste.rs/s/jgsiqa # ai https:// arxiv.org/html/2601.05280v2