Pruning as Regularization: Sensitivity-Aware One-Shot Pruning in ASR

arXiv — cs.CLWednesday, November 12, 2025 at 5:00:00 AM
The publication of the study on arXiv highlights a shift in understanding neural network pruning, particularly in the context of automatic speech recognition (ASR). Traditionally viewed as a mere compression technique, the research reveals that one-shot magnitude pruning can act as a powerful implicit regularizer. By utilizing the Whisper-small model, the team combined gradient- and Fisher-based sensitivity diagnostics with targeted pruning, uncovering that decoder feedforward networks (FFNs) are particularly fragile to pruning. This insight led to significant improvements in generalization, with pruning 50% of decoder self-attention yielding a 2.38% absolute reduction in word error rates on the LibriSpeech test set. Additionally, pruning the last four encoder layers at the same rate resulted in a 1.72% absolute improvement. The gains were consistent across other datasets like Common Voice and TED-LIUM. Notably, the sensitivity-aware approach allows for maintaining near-baseline accura…
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
Speech-Aware Long Context Pruning and Integration for Contextualized Automatic Speech Recognition
PositiveArtificial Intelligence
The paper titled 'Speech-Aware Long Context Pruning and Integration for Contextualized Automatic Speech Recognition' presents a novel framework called SAP² aimed at improving automatic speech recognition (ASR) systems. These systems typically perform well under standard conditions but face challenges in utilizing long-context information, particularly in specialized scenarios like conference presentations. The SAP² method employs a two-stage process to dynamically prune and integrate relevant contextual keywords, demonstrating significant improvements in word error rates on the SlideSpeech and LibriSpeech datasets.