Seq-U-Net: A One-Dimensional Causal U-Net for Efficient Sequence Modelling

11/14/2019
by   Daniel Stoller, et al.
0

Convolutional neural networks (CNNs) with dilated filters such as the Wavenet or the Temporal Convolutional Network (TCN) have shown good results in a variety of sequence modelling tasks. However, efficiently modelling long-term dependencies in these sequences is still challenging. Although the receptive field of these models grows exponentially with the number of layers, computing the convolutions over very long sequences of features in each layer is time and memory-intensive, prohibiting the use of longer receptive fields in practice. To increase efficiency, we make use of the "slow feature" hypothesis stating that many features of interest are slowly varying over time. For this, we use a U-Net architecture that computes features at multiple time-scales and adapt it to our auto-regressive scenario by making convolutions causal. We apply our model ("Seq-U-Net") to a variety of tasks including language and audio generation. In comparison to TCN and Wavenet, our network consistently saves memory and computation time, with speed-ups for training and inference of over 4x in the audio generation experiment in particular, while achieving a comparable performance in all tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/14/2019

Temporal FiLM: Capturing Long-Range Sequence Dependencies with Feature-Wise Modulations

Learning representations that accurately capture long-range dependencies...
research
03/12/2017

Tree Memory Networks for Modelling Long-term Temporal Dependencies

In the domain of sequence modelling, Recurrent Neural Networks (RNN) hav...
research
11/23/2022

Rega-Net:Retina Gabor Attention for Deep Convolutional Neural Networks

Extensive research works demonstrate that the attention mechanism in con...
research
10/01/2016

Very Deep Convolutional Neural Networks for Raw Waveforms

Learning acoustic models directly from the raw waveform data with minima...
research
06/23/2021

Should You Go Deeper? Optimizing Convolutional Neural Network Architectures without Training by Receptive Field Analysis

Applying artificial neural networks (ANN) to specific tasks, researchers...
research
02/02/2020

Sound Event Detection with Depthwise Separable and Dilated Convolutions

State-of-the-art sound event detection (SED) methods usually employ a se...
research
04/06/2020

Residual Shuffle-Exchange Networks for Fast Processing of Long Sequences

Attention is a commonly used mechanism in sequence processing, but it is...

Please sign up or login with your details

Forgot password? Click here to reset