Continual learning using lattice-free MMI for speech recognition

10/13/2021
by   Hossein Hadian, et al.
0

Continual learning (CL), or domain expansion, recently became a popular topic for automatic speech recognition (ASR) acoustic modeling because practical systems have to be updated frequently in order to work robustly on types of speech not observed during initial training. While sequential adaptation allows tuning a system to a new domain, it may result in performance degradation on the old domains due to catastrophic forgetting. In this work we explore regularization-based CL for neural network acoustic models trained with the lattice-free maximum mutual information (LF-MMI) criterion. We simulate domain expansion by incrementally adapting the acoustic model on different public datasets that include several accents and speaking styles. We investigate two well-known CL techniques, elastic weight consolidation (EWC) and learning without forgetting (LWF), which aim to reduce forgetting by preserving model weights or network outputs. We additionally introduce a sequence-level LWF regularization, which exploits posteriors from the denominator graph of LF-MMI to further reduce forgetting. Empirical results show that the proposed sequence-level LWF can improve the best average word error rate across all domains by up to 9.4

READ FULL TEXT
research
06/19/2023

Rehearsal-Free Online Continual Learning for Automatic Speech Recognition

Fine-tuning an Automatic Speech Recognition (ASR) model to new domains r...
research
12/17/2021

Continual Learning for Monolingual End-to-End Automatic Speech Recognition

Adapting Automatic Speech Recognition (ASR) models to new domains leads ...
research
10/01/2019

Domain Expansion in DNN-based Acoustic Models for Robust Speech Recognition

Training acoustic models with sequentially incoming data – while both le...
research
05/12/2023

Continual Learning for End-to-End ASR by Averaging Domain Experts

Continual learning for end-to-end automatic speech recognition has to co...
research
10/27/2022

Weight Averaging: A Simple Yet Effective Method to Overcome Catastrophic Forgetting in Automatic Speech Recognition

Adapting a trained Automatic Speech Recognition (ASR) model to new tasks...
research
12/02/2022

Continual Learning for On-Device Speech Recognition using Disentangled Conformers

Automatic speech recognition research focuses on training and evaluating...
research
08/07/2023

Do You Remember? Overcoming Catastrophic Forgetting for Fake Audio Detection

Current fake audio detection algorithms have achieved promising performa...

Please sign up or login with your details

Forgot password? Click here to reset