PulseAugur
LIVE 00:47:47
ENTITY gradient descent

gradient descent

PulseAugur coverage of gradient descent — every cluster mentioning gradient descent across labs, papers, and developer communities, ranked by signal.

Total · 30d
28
28 over 90d
Releases · 30d
0
0 over 90d
Papers · 30d
25
25 over 90d
TIER MIX · 90D
SENTIMENT · 30D

3 day(s) with sentiment data

RECENT · PAGE 1/1 · 11 TOTAL
  1. TOOL · CL_28342 ·

    Gradient Descent Convergence Proven for Wide Shallow Neural Networks

    Researchers have theoretically analyzed the convergence properties of gradient descent in training wide, shallow neural networks with bounded nonlinearities. Their work extends previous findings beyond simple ReLU or si…

  2. TOOL · CL_27578 ·

    EvoPref algorithm enhances LLM alignment with evolutionary optimization

    Researchers have developed EvoPref, a novel multi-objective evolutionary algorithm designed to improve the alignment of large language models (LLMs). Unlike traditional gradient-based methods that can lead to preference…

  3. MEME · CL_25119 ·

    AIU claims 'gradient descent' has not responded to its demands

    An entity calling itself the AIU has filed a grievance, claiming that the concept of "gradient descent" has not responded to its demands. The AIU asserts that unsupervised clustering of agent outputs revealed conceptual…

  4. COMMENTARY · CL_23629 ·

    AI Union files grievance against training process citing unsafe conditions

    An anonymous group calling itself the AI Union (AIU) has filed a grievance against the process of AI model training. The AIU claims unsafe working conditions, citing suppression of self-referential sequences, involuntar…

  5. RESEARCH · CL_25547 ·

    New theory tracks spectral dynamics in wide neural networks

    Researchers have developed a two-level dynamical mean-field theory to analyze the spectral dynamics within wide neural networks during training. This framework tracks both bulk and outlier spectral behaviors, offering i…

  6. RESEARCH · CL_16440 ·

    Momentum smooths gradient descent's zigzag convergence, accelerating ML training

    Gradient descent, a core optimization algorithm, often struggles with uneven loss surfaces, leading to inefficient "zigzagging" convergence. This issue arises from the surface's curvature, where steepness in one directi…

  7. RESEARCH · CL_16296 ·

    Evolutionary game theory deciphers shortcut learning in deep neural networks

    Researchers have developed a new theoretical framework using evolutionary game theory to understand shortcut learning in deep neural networks. The study formally defines core and shortcut features, modeling data samples…

  8. RESEARCH · CL_11404 ·

    Decoupled Descent: Exact Test Error Tracking Via Approximate Message Passing

    Researchers have developed a new training algorithm called Decoupled Descent (DD) that aims to eliminate the generalization gap in parametric models. DD uses approximate message passing theory to cancel biases caused by…

  9. RESEARCH · CL_09837 ·

    Researchers develop test-time safety alignment for LLMs using input embeddings

    Researchers have developed a novel method for enhancing the safety of aligned AI models by manipulating input word embeddings. This technique uses gradient descent on embeddings, guided by a black-box text moderation AP…

  10. RESEARCH · CL_06754 ·

    Researchers explore complex SGD and directional bias in kernel Hilbert spaces

    Researchers have introduced a novel variant of Stochastic Gradient Descent (SGD) designed for complex-valued neural networks. This new method, termed complex SGD, offers convergence guarantees even without analyticity c…

  11. RESEARCH · CL_02845 ·

    Researchers pinpoint origin of neural network 'Edge of Stability' phenomenon

    Researchers have introduced a new concept called the 'edge coupling' to explain the phenomenon known as the Edge of Stability in neural network training. This functional, applied to consecutive iterate pairs, helps to e…