PulseAugur
LIVE 06:48:25
research · [1 source] ·
0
research

New FADE method improves ASR model quantization for edge devices

Researchers have developed FADE, a novel framework for improving post-training quantization of encoder-decoder Automatic Speech Recognition (ASR) models. This method addresses the issue of error accumulation across layers by assigning adaptive compensation coefficients to each layer. FADE combines intrinsic vulnerability scores from weight geometry with data-driven calibration reliability scores to balance local fidelity and cross-layer error correction. Experiments on models like Whisper and Qwen3-ASR demonstrated consistent improvements in Word Error Rate at 3- and 4-bit precision. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT Enables more efficient deployment of ASR models on memory-constrained edge devices by improving quantization accuracy.

RANK_REASON This is a research paper detailing a new framework for model quantization.

Read on arXiv cs.CL →

COVERAGE [1]

  1. arXiv cs.CL TIER_1 · Xinyu Wang, Ziyu Zhao, Yajie Luo, Yihong Wu, Liheng Ma, Jingrui Tian, Lei Ding, Xiao-Wen Chang, Peng Lu ·

    Diagnostic-Driven Layer-Wise Compensation for Post-Training Quantization of Encoder-Decoder ASR Models

    arXiv:2601.02455v2 Announce Type: replace-cross Abstract: Deploying Automatic Speech Recognition (ASR) models on memory-constrained edge devices requires aggressive low-bit weight quantization. Layer-wise post-training quantization is practical and effective, but it suffers from …