A Model Compression Method with Matrix Product Operators for Speech Enhancement

10/10/2020
by   Xingwei Sun, et al.
0

The deep neural network (DNN) based speech enhancement approaches have achieved promising performance. However, the number of parameters involved in these methods is usually enormous for the real applications of speech enhancement on the device with the limited resources. This seriously restricts the applications. To deal with this issue, model compression techniques are being widely studied. In this paper, we propose a model compression method based on matrix product operators (MPO) to substantially reduce the number of parameters in DNN models for speech enhancement. In this method, the weight matrices in the linear transformations of neural network model are replaced by the MPO decomposition format before training. In experiment, this process is applied to the causal neural network models, such as the feedforward multilayer perceptron (MLP) and long short-term memory (LSTM) models. Both MLP and LSTM models with/without compression are then utilized to estimate the ideal ratio mask for monaural speech enhancement. The experimental results show that our proposed MPO-based method outperforms the widely-used pruning method for speech enhancement under various compression rates, and further improvement can be achieved with respect to low compression rates. Our proposal provides an effective model compression method for speech enhancement, especially in cloud-free application.

READ FULL TEXT
research
02/14/2020

Real-time speech enhancement using equilibriated RNN

We propose a speech enhancement method using a causal deep neural networ...
research
12/25/2018

Tensor-Train Long Short-Term Memory for Monaural Speech Enhancement

In recent years, Long Short-Term Memory (LSTM) has become a popular choi...
research
08/28/2019

Convolutional Recurrent Neural Network Based Progressive Learning for Monaural Speech Enhancement

Recently, progressive learning has shown its capacity of improving speec...
research
12/22/2020

Compressing LSTM Networks by Matrix Product Operators

Long Short-Term Memory (LSTM) models are the building blocks of many sta...
research
11/08/2021

SEOFP-NET: Compression and Acceleration of Deep Neural Networks for Speech Enhancement Using Sign-Exponent-Only Floating-Points

Numerous compression and acceleration strategies have achieved outstandi...
research
08/28/2017

Integrated Speech Enhancement Method Based on Weighted Prediction Error and DNN for Dereverberation and Denoising

Both reverberation and additive noises degrade the speech quality and in...
research
11/03/2021

Weight, Block or Unit? Exploring Sparsity Tradeoffs for Speech Enhancement on Tiny Neural Accelerators

We explore network sparsification strategies with the aim of compressing...

Please sign up or login with your details

Forgot password? Click here to reset