A Novel Self-training Approach for Low-resource Speech Recognition

08/10/2023
by   Satwinder Singh, et al.
0

In this paper, we propose a self-training approach for automatic speech recognition (ASR) for low-resource settings. While self-training approaches have been extensively developed and evaluated for high-resource languages such as English, their applications to low-resource languages like Punjabi have been limited, despite the language being spoken by millions globally. The scarcity of annotated data has hindered the development of accurate ASR systems, especially for low-resource languages (e.g., Punjabi and Māori languages). To address this issue, we propose an effective self-training approach that generates highly accurate pseudo-labels for unlabeled low-resource speech. Our experimental analysis demonstrates that our approach significantly improves word error rate, achieving a relative improvement of 14.94 baseline model across four real speech datasets. Further, our proposed approach reports the best results on the Common Voice Punjabi dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/04/2021

Voice Conversion Can Improve ASR in Very Low-Resource Settings

Voice conversion (VC) has been proposed to improve speech recognition sy...
research
05/19/2020

Generative Adversarial Training Data Adaptation for Very Low-resource Automatic Speech Recognition

It is important to transcribe and archive speech data of endangered lang...
research
06/02/2021

Improving low-resource ASR performance with untranscribed out-of-domain data

Semi-supervised training (SST) is a common approach to leverage untransc...
research
12/22/2020

Applying wav2vec2.0 to Speech Recognition in various low-resource languages

Several domains own corresponding widely used feature extractors, such a...
research
04/05/2022

Combining Spectral and Self-Supervised Features for Low Resource Speech Recognition and Translation

Self-Supervised Learning (SSL) models have been successfully applied in ...
research
10/05/2021

Disambiguation-BERT for N-best Rescoring in Low-Resource Conversational ASR

We study the inclusion of past conversational context through BERT langu...
research
02/05/2023

MAC: A unified framework boosting low resource automatic speech recognition

We propose a unified framework for low resource automatic speech recogni...

Please sign up or login with your details

Forgot password? Click here to reset