PulseAugur
LIVE 04:19:42
tool · [1 source] ·
0
tool

New R-SGD-Mini method tackles heavy-tailed noise in optimization

Researchers have introduced a new optimization method called Robust Stochastic Gradient Descent with medoid mini-batch gradient sampling (R-SGD-Mini). This method is designed to handle heavy-tailed noise in gradient calculations, which can have infinite variances. R-SGD-Mini works by dividing data batches into smaller chunks, calculating gradients for each, and then using the medoid of these gradients to update the solution estimate. The approach has demonstrated improved performance over existing methods like standard SGD and Median-of-Means in experimental settings. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT Introduces a novel method for handling noisy gradients in optimization, potentially improving the stability and performance of machine learning training processes.

RANK_REASON The cluster contains an arXiv paper detailing a new optimization method. [lever_c_demoted from research: ic=1 ai=1.0]

Read on arXiv cs.LG →

COVERAGE [1]

  1. arXiv cs.LG TIER_1 · Dusan Jakovetic ·

    Robust stochastic first order methods in heavy-tailed noise via medoid mini-batch gradient sampling

    We consider a first order stochastic optimization framework where, at each iteration, $K$ independent identically distributed (i.i.d.) data point samples are drawn, based on which stochastic gradients can be queried. We allow gradient noise to be heavy-tailed, with possibly infin…