Self-Transriber: Few-shot Lyrics Transcription with Self-training

11/18/2022
by   Xiaoxue Gao, et al.
0

The current lyrics transcription approaches heavily rely on supervised learning with labeled data, but such data are scarce and manual labeling of singing is expensive. How to benefit from unlabeled data and alleviate limited data problem have not been explored for lyrics transcription. We propose the first semi-supervised lyrics transcription paradigm, Self-Transcriber, by leveraging on unlabeled data using self-training with noisy student augmentation. We attempt to demonstrate the possibility of lyrics transcription with a few amount of labeled data. Self-Transcriber generates pseudo labels of the unlabeled singing using teacher model, and augments pseudo-labels to the labeled data for student model update with both self-training and supervised training losses. This work closes the gap between supervised and semi-supervised learning as well as opens doors for few-shot learning of lyrics transcription. Our experiments show that our approach using only 12.7 hours of labeled data achieves competitive performance compared with the supervised approaches trained on 149.1 hours of labeled data for lyrics transcription.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/18/2023

Enhancing Self-Training Methods

Semi-supervised learning approaches train on small sets of labeled data ...
research
08/14/2020

Semi-supervised learning using teacher-student models for vocal melody extraction

The lack of labeled data is a major obstacle in many music information r...
research
03/03/2022

Semi-supervised Learning using Robust Loss

The amount of manually labeled data is limited in medical applications, ...
research
06/01/2023

Doubly Robust Self-Training

Self-training is an important technique for solving semi-supervised lear...
research
05/06/2019

MixMatch: A Holistic Approach to Semi-Supervised Learning

Semi-supervised learning has proven to be a powerful paradigm for levera...
research
02/05/2022

LST: Lexicon-Guided Self-Training for Few-Shot Text Classification

Self-training provides an effective means of using an extremely small am...
research
02/05/2022

Exemplar-Based Contrastive Self-Supervised Learning with Few-Shot Class Incremental Learning

Humans are capable of learning new concepts from only a few (labeled) ex...

Please sign up or login with your details

Forgot password? Click here to reset