Partitioned Gradient Matching-based Data Subset Selection for Compute-Efficient Robust ASR Training

10/30/2022
by   Ashish Mittal, et al.
0

Training state-of-the-art ASR systems such as RNN-T often has a high associated financial and environmental cost. Training with a subset of training data could mitigate this problem if the subset selected could achieve on-par performance with training with the entire dataset. Although there are many data subset selection(DSS) algorithms, direct application to the RNN-T is difficult, especially the DSS algorithms that are adaptive and use learning dynamics such as gradients, as RNN-T tend to have gradients with a significantly larger memory footprint. In this paper, we propose Partitioned Gradient Matching (PGM) a novel distributable DSS algorithm, suitable for massive datasets like those used to train RNN-T. Through extensive experiments on Librispeech 100H and Librispeech 960H, we show that PGM achieves between 3x to 6x speedup with only a very small accuracy degradation (under 1 addition, we demonstrate similar results for PGM even in settings where the training data is corrupted with noise.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset