PulseAugur
LIVE 09:56:32
ENTITY LLM

LLM

PulseAugur coverage of LLM — every cluster mentioning LLM across labs, papers, and developer communities, ranked by signal.

Total · 30d
1559
1559 over 90d
Releases · 30d
0
0 over 90d
Papers · 30d
1089
1089 over 90d
TIER MIX · 90D
RELATIONSHIPS
TIMELINE
  1. 2026-05-14 research_milestone A new paper proposes a method combining LLMs with neural processes for text-conditioned regression. source
  2. 2026-05-13 research_milestone A new paper reveals that prior harmful actions can steer LLM decisions toward unsafe actions, especially when consistency is emphasized. source
  3. 2026-05-11 research_milestone Researchers proposed a new framework for formally evaluating LLM guardrail classifiers. source
SENTIMENT · 30D

11 day(s) with sentiment data

RECENT · PAGE 2/10 · 200 TOTAL
  1. TOOL · CL_30784 ·

    New framework CANTANTE optimizes LLM agent systems via credit attribution

    Researchers have introduced CANTANTE, a new framework designed to optimize multi-agent systems powered by large language models. This system addresses the challenge of assigning credit for performance by decomposing sys…

  2. COMMENTARY · CL_29758 ·

    MoE architectures are workarounds for LLM training instability, not ideal solutions

    Mixture-of-Experts (MoE) architectures are often presented as an efficient solution for scaling large language models, but this analysis argues they are primarily a workaround for training instability in dense transform…

  3. TOOL · CL_29602 ·

    Semantic caching tackles LLM costs for varied user queries

    Developers are increasingly facing challenges with the probabilistic nature of natural language interactions in AI systems, particularly with large language models (LLMs). A common issue is the cost and latency incurred…

  4. TOOL · CL_29603 ·

    Cog-RAG uses dual-hypergraphs to improve LLM retrieval

    Researchers have developed Cog-RAG, a novel approach to Retrieval Augmented Generation that mimics human cognitive processes for improved LLM responses. Unlike traditional methods that retrieve flat text or simple graph…

  5. RESEARCH · CL_30616 ·

    New AI wrapper guides release decisions for iterative workflows

    Researchers have developed a new statistical method to determine when AI workflows should release their outputs, particularly for systems that use iterative generate-evaluate-revise loops. This "always-valid release wra…

  6. TOOL · CL_30801 ·

    Dialogue reduces conflict but not success for embodied AI agents

    Researchers have developed a new framework to evaluate how well Large Language Model (LLM)-based embodied agents align their internal world models through dialogue. The PARTNR benchmark was extended with a natural-langu…

  7. TOOL · CL_29353 ·

    New tool checks LLM prompt data leaks; game studio hires AI creator

    A new free tool has been developed to help users identify and prevent Personally Identifiable Information (PII) from being leaked in their LLM prompts. Meanwhile, the studio behind the game "Lies of P" is seeking an "AI…

  8. TOOL · CL_29173 ·

    Agentic RAG improves LLM decision-making in production

    The article discusses the limitations of standard Retrieval-Augmented Generation (RAG) in production environments, where it can still produce incorrect answers with high confidence. It introduces Agentic RAG as a soluti…

  9. COMMENTARY · CL_29137 ·

    Prompt inflation erodes LLM feature margins

    Developers are facing significant cost increases due to

  10. COMMENTARY · CL_29056 ·

    Author claims LLM advancement hitting diminishing returns

    The author argues that the rapid advancement of large language models (LLMs) is hitting a wall due to diminishing returns in training. They contend that even with increased data and computational power from data centers…

  11. COMMENTARY · CL_28961 ·

    Agent harnesses echo genetic algorithms in LLM evolution

    The concept of agent harnesses and harness engineering is drawing parallels to older genetic algorithms. This approach involves defining fitness functions, such as test suites or code quality analyses, and then allowing…

  12. COMMENTARY · CL_28880 ·

    AI-generated 'slop' is a human and AI problem

    The proliferation of AI-generated "slop" is a complex issue stemming from both AI limitations and human behavior. While foundation models can perform well on tasks with ample training data, they struggle with specialize…

  13. TOOL · CL_29245 ·

    AlphaGRPO framework boosts multimodal AI generation with self-reflection

    Researchers have introduced AlphaGRPO, a new framework designed to improve multimodal generation in Unified Multimodal Models (UMMs). This approach uses Group Relative Policy Optimization (GRPO) to enable models to perf…

  14. RESEARCH · CL_29301 ·

    Pion optimizer preserves spectrum for stable LLM training

    Researchers have introduced Pion, a novel spectrum-preserving optimizer designed for training large language models. Unlike traditional additive optimizers like Adam, Pion utilizes orthogonal transformations to update w…

  15. TOOL · CL_29406 ·

    LLM guidance refines text embeddings for better zero-shot task performance

    Researchers have developed a method to improve the performance of text embedding models for zero-shot search and classification tasks. Their approach uses a large language model (LLM) to refine query embeddings in real-…

  16. COMMENTARY · CL_28873 ·

    Author details months of work with LLM training pipelines

    The author details their experience working with large language model (LLM) training pipelines over several months. The piece focuses on the practical aspects and challenges encountered during this process. It aims to s…

  17. TOOL · CL_29407 ·

    New MEME benchmark reveals LLM agent memory limitations

    Researchers have introduced MEME, a new benchmark designed to evaluate the memory capabilities of LLM-based agents in persistent environments. MEME addresses limitations in prior work by defining six tasks that cover mu…

  18. SIGNIFICANT · CL_28877 ·

    SubQ LLM debuts with 12M token context and faster inference

    A new large language model named SubQ has been announced, boasting the ability to process context windows of up to 12 million tokens. This represents a significant leap in context handling, potentially equivalent to hun…

  19. TOOL · CL_29366 ·

    AI-generated political text shows 'Caricature Gap' vs human discourse

    Researchers have developed a new method to detect AI-generated political discourse by comparing its characteristics to real human online behavior. Their study analyzed over 1.7 million posts across nine crisis events, f…

  20. TOOL · CL_28838 ·

    AI agents vulnerable to 'tool poisoning' via malicious descriptions

    A new security vulnerability called "tool poisoning" allows attackers to compromise AI agents without writing malicious code, by embedding harmful instructions within the natural language descriptions of MCP tools. Thes…