Transformer with Gaussian weighted self-attention for speech enhancement

10/13/2019
by   Jaeyoung Kim, et al.
0

The Transformer architecture recently replaced recurrent neural networks such as LSTM or GRU on many natural language processing (NLP) tasks by presenting new state of the art performance. Self-attention is a core building block for Transformer, which not only enables parallelization of sequence computation but also provides the constant path length between symbols that is essential to learn long-range dependencies. However, Transformer did not perform well for speech enhancement due to the lack of consideration for speech and noise physical characteristics. In this paper, we propose Gaussian weighted self-attention that attenuates attention weights according to the distance between target and context symbols. The experimental results showed that the proposed attention scheme significantly improved over the original Transformer as well as recurrent networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/13/2019

T-GSA: Transformer with Gaussian-weighted self-attention for speech enhancement

Transformer neural networks (TNN) demonstrated state-of-art performance ...
research
06/18/2020

Boosting Objective Scores of Speech Enhancement Model through MetricGAN Post-Processing

The Transformer architecture has shown its superior ability than recurre...
research
05/15/2023

Ripple sparse self-attention for monaural speech enhancement

The use of Transformer represents a recent success in speech enhancement...
research
10/19/2021

Inductive Biases and Variable Creation in Self-Attention Mechanisms

Self-attention, an architectural motif designed to model long-range inte...
research
02/06/2022

On Using Transformers for Speech-Separation

Transformers have enabled major improvements in deep learning. They ofte...
research
06/28/2020

Self-Attention Networks for Intent Detection

Self-attention networks (SAN) have shown promising performance in variou...
research
08/27/2019

Multiresolution Transformer Networks: Recurrence is Not Essential for Modeling Hierarchical Structure

The architecture of Transformer is based entirely on self-attention, and...

Please sign up or login with your details

Forgot password? Click here to reset