Exploring Heterogeneous Characteristics of Layers in ASR Models for More Efficient Training

10/08/2021
by   Lillian Zhou, et al.
0

Transformer-based architectures have been the subject of research aimed at understanding their overparameterization and the non-uniform importance of their layers. Applying these approaches to Automatic Speech Recognition, we demonstrate that the state-of-the-art Conformer models generally have multiple ambient layers. We study the stability of these layers across runs and model sizes, propose that group normalization may be used without disrupting their formation, and examine their correlation with model weight updates in each layer. Finally, we apply these findings to Federated Learning in order to improve the training procedure, by targeting Federated Dropout to layers by importance. This allows us to reduce the model size optimized by clients without quality degradation, and shows potential for future exploration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/07/2021

Enabling On-Device Training of Speech Recognition Models with Federated Dropout

Federated learning can be used to train machine learning models on the e...
research
02/08/2021

Federated Acoustic Modeling For Automatic Speech Recognition

Data privacy and protection is a crucial issue for any automatic speech ...
research
11/06/2021

Privacy attacks for automatic speech recognition acoustic models in a federated learning framework

This paper investigates methods to effectively retrieve speaker informat...
research
06/06/2022

FedNST: Federated Noisy Student Training for Automatic Speech Recognition

Federated Learning (FL) enables training state-of-the-art Automatic Spee...
research
10/26/2020

Optimal Importance Sampling for Federated Learning

Federated learning involves a mixture of centralized and decentralized p...
research
12/01/2021

Compare Where It Matters: Using Layer-Wise Regularization To Improve Federated Learning on Heterogeneous Data

Federated Learning is a widely adopted method to train neural networks o...
research
04/03/2023

FedIN: Federated Intermediate Layers Learning for Model Heterogeneity

Federated learning (FL) facilitates edge devices to cooperatively train ...

Please sign up or login with your details

Forgot password? Click here to reset