PulseAugur
LIVE 00:09:15
ENTITY Pythia

Pythia

PulseAugur coverage of Pythia — every cluster mentioning Pythia across labs, papers, and developer communities, ranked by signal.

Total · 30d
10
10 over 90d
Releases · 30d
0
0 over 90d
Papers · 30d
10
10 over 90d
TIER MIX · 90D
SENTIMENT · 30D

1 day(s) with sentiment data

RECENT · PAGE 1/1 · 7 TOTAL
  1. RESEARCH · CL_22182 ·

    Language model surprisal may not predict metaphor novelty as thought

    A new paper published on arXiv suggests that language model surprisal, often used as a proxy for contextual predictability and metaphor novelty, may be misleading. The research indicates that lexical frequency is a stro…

  2. TOOL · CL_26990 ·

    New AEN-SAE architecture tackles feature starvation in LLM interpretability

    Researchers have introduced Adaptive Elastic Net Sparse Autoencoders (AEN-SAEs) to address feature starvation in sparse autoencoders used for interpreting LLM representations. Traditional methods struggle with dead neur…

  3. RESEARCH · CL_18265 ·

    Researchers find Transformers know counts but struggle to output them

    A new paper identifies a specific bottleneck in Transformer models that hinders their ability to perform counting tasks. Researchers found that while models like Pythia, Qwen3, and Mistral store count information accura…

  4. RESEARCH · CL_15547 ·

    HeadQ: Model-Visible Distortion and Score-Space Correction for KV-Cache Quantization

    Researchers are developing several novel methods to optimize the Key-Value (KV) cache in large language models, which is a major bottleneck for long-context processing. These approaches include training models to inhere…

  5. RESEARCH · CL_09277 ·

    AI model evaluations are becoming a costly bottleneck, surpassing training expenses

    AI model evaluations are becoming prohibitively expensive, with recent benchmarks costing tens of thousands of dollars and consuming thousands of GPU hours. This high cost is particularly pronounced for agent-based eval…

  6. RESEARCH · CL_08642 ·

    Transformer architecture significantly impacts model error detection capabilities

    A new paper reveals that a transformer model's architecture significantly impacts its ability to signal decision quality through internal activations, a property termed 'observability.' This observability is crucial for…

  7. RESEARCH · CL_06772 ·

    Transformer research probes security flaws, training dynamics, and in-context learning limits

    Researchers have identified vulnerabilities in the shuffling defense mechanism used to secure Transformer models during inference, demonstrating an attack that can extract model weights by aligning permuted activations.…