Learning ASR pathways: A sparse multilingual ASR model

09/13/2022
by   Mu Yang, et al.
8

Neural network pruning can be effectively applied to compress automatic speech recognition (ASR) models. However, in multilingual ASR, performing language-agnostic pruning may lead to severe performance degradation on some languages because language-agnostic pruning masks may not fit all languages and discard important language-specific parameters. In this work, we present ASR pathways, a sparse multilingual ASR model that activates language-specific sub-networks ("pathways"), such that the parameters for each language are learned explicitly. With the overlapping sub-networks, the shared parameters can also enable knowledge transfer for lower resource languages via joint multilingual training. We propose a novel algorithm to learn ASR pathways, and evaluate the proposed method on 4 languages with a streaming RNN-T model. Our proposed ASR pathways outperform both dense models (-5.0 language-agnostically pruned model (-21.4 performance on low-resource languages compared to the monolingual sparse models.

READ FULL TEXT
research
07/06/2020

Massively Multilingual ASR: 50 Languages, 1 Model, 1 Billion Parameters

We study training a single acoustic model for multiple languages with th...
research
11/05/2022

LAMASSU: Streaming Language-Agnostic Multilingual Speech Recognition and Translation Using Neural Transducers

End-to-end formulation of automatic speech recognition (ASR) and speech ...
research
04/20/2020

Language-agnostic Multilingual Modeling

Multilingual Automated Speech Recognition (ASR) systems allow for the jo...
research
09/14/2019

Multilingual ASR with Massive Data Augmentation

Towards developing high-performing ASR for low-resource languages, appro...
research
11/25/2020

Bootstrap an end-to-end ASR system by multilingual training, transfer learning, text-to-text mapping and synthetic audio

Bootstrapping speech recognition on limited data resources has been an a...
research
07/04/2017

Multilingual Hierarchical Attention Networks for Document Classification

Hierarchical attention networks have recently achieved remarkable perfor...
research
03/09/2022

Language Adaptive Cross-lingual Speech Representation Learning with Sparse Sharing Sub-networks

Unsupervised cross-lingual speech representation learning (XLSR) has rec...

Please sign up or login with your details

Forgot password? Click here to reset