PulseAugur
LIVE 08:15:03
research · [3 sources] ·
0
research

New methods QFlash and ELSA boost Vision Transformer attention efficiency

Researchers have developed two new methods to improve the efficiency of attention mechanisms in vision transformers. QFlash focuses on enabling integer-only operations for FlashAttention, achieving significant speedups and reduced energy consumption without accuracy loss on certain models. ELSA, on the other hand, reformulates attention to preserve exact softmax semantics in real arithmetic, offering hardware-agnostic performance gains and memory reduction across various platforms and precisions. AI

Summary written by gemini-2.5-flash-lite from 3 sources. How we write summaries →

IMPACT New attention algorithms offer significant speedups and memory efficiency, potentially lowering inference costs and enabling deployment on resource-constrained devices.

RANK_REASON Two academic papers introduce novel algorithmic approaches to optimize attention mechanisms in vision transformers.

Read on arXiv cs.CV →

COVERAGE [3]

  1. arXiv cs.LG TIER_1 · Sehyeon Oh, Yongin Kwon, Jemin Lee ·

    QFlash: Bridging Quantization and Memory Efficiency in Vision Transformer Attention

    arXiv:2604.25306v1 Announce Type: new Abstract: FlashAttention improves efficiency through tiling, but its online softmax still relies on floating-point arithmetic for numerical stability, making full quantization difficult. We identify three main obstacles to integer-only FlashA…

  2. arXiv cs.AI TIER_1 · Jemin Lee ·

    QFlash: Bridging Quantization and Memory Efficiency in Vision Transformer Attention

    FlashAttention improves efficiency through tiling, but its online softmax still relies on floating-point arithmetic for numerical stability, making full quantization difficult. We identify three main obstacles to integer-only FlashAttention: (1) scale explosion during tile-wise a…

  3. arXiv cs.CV TIER_1 · Chih-Chung Hsu, Xin-Di Ma, Wo-Ting Liao, Chia-Ming Lee ·

    ELSA: Exact Linear-Scan Attention for Fast and Memory-Light Vision Transformers

    arXiv:2604.23798v1 Announce Type: cross Abstract: Existing attention accelerators often trade exact softmax semantics, depend on fused Tensor Core kernels, or incur sequential depth that limits FP32 throughput on long sequences. We present \textbf{ELSA}, an algorithmic reformulat…