Ultra-Lightweight Speech Separation via Group Communication

11/17/2020
by   Yi Luo, et al.
0

Model size and complexity remain the biggest challenges in the deployment of speech enhancement and separation systems on low-resource devices such as earphones and hearing aids. Although methods such as compression, distillation and quantization can be applied to large models, they often come with a cost on the model performance. In this paper, we provide a simple model design paradigm that explicitly designs ultra-lightweight models without sacrificing the performance. Motivated by the sub-band frequency-LSTM (F-LSTM) architectures, we introduce the group communication (GroupComm), where a feature vector is split into smaller groups and a small processing block is used to perform inter-group communication. Unlike standard F-LSTM models where the sub-band outputs are concatenated, an ultra-small module is applied on all the groups in parallel, which allows a significant decrease on the model size. Experiment results show that comparing with a strong baseline model which is already lightweight, GroupComm can achieve on par performance with 35.6 times fewer parameters and 2.3 times fewer operations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/14/2020

Group Communication with Context Codec for Ultra-Lightweight Source Separation

Ultra-lightweight model design is an important topic for the deployment ...
research
04/18/2023

Neural Speech Enhancement with Very Low Algorithmic Latency and Complexity via Integrated Full- and Sub-Band Modeling

We propose FSB-LSTM, a novel long short-term memory (LSTM) based archite...
research
11/07/2021

LiMuSE: Lightweight Multi-modal Speaker Extraction

The past several years have witnessed significant progress in modeling t...
research
05/29/2020

Sub-band Knowledge Distillation Framework for Speech Enhancement

In single-channel speech enhancement, methods based on full-band spectra...
research
05/09/2023

Inter-SubNet: Speech Enhancement with Subband Interaction

Subband-based approaches process subbands in parallel through the model ...
research
11/17/2020

Rethinking the Separation Layers in Speech Separation Networks

Modules in all existing speech separation networks can be categorized in...
research
10/27/2022

Too Brittle To Touch: Comparing the Stability of Quantization and Distillation Towards Developing Lightweight Low-Resource MT Models

Leveraging shared learning through Massively Multilingual Models, state-...

Please sign up or login with your details

Forgot password? Click here to reset