LRU
PulseAugur coverage of LRU — every cluster mentioning LRU across labs, papers, and developer communities, ranked by signal.
-
Apple researchers unveil SpecMD for faster MoE model inference
Apple's machine learning research team has published a paper detailing SpecMD, a new framework for evaluating Mixture-of-Experts (MoE) model caching policies. Their experiments show that traditional caching assumptions …
-
New ML-based GPU caching algorithm LCR boosts LLM inference speed
Researchers have developed a new GPU caching algorithm called Learning-Augmented LRU (LALRU) designed to improve efficiency during AI inference. This algorithm integrates learned predictions with caching policies to ens…
-
Memristor-based AI systems show promise for efficient learning and neuromorphic computing
Researchers are exploring Self-Organising Memristive Networks (SOMNs) as a physical alternative to conventional hardware for artificial intelligence, aiming for energy-efficient, brain-like continual learning. These net…