TUNet: A Block-online Bandwidth Extension Model based on Transformers and Self-supervised Pretraining

10/26/2021
by   Viet Anh Nguyen, et al.
0

We introduce a block-online variant of the temporal feature-wise linear modulation (TFiLM) model to achieve bandwidth extension. The proposed architecture simplifies the UNet backbone of the TFiLM to reduce inference time and employs an efficient transformer at the bottleneck to alleviate performance degradation. We also utilize self-supervised pretraining and data augmentation to enhance the quality of bandwidth extended signals and reduce the sensitivity with respect to downsampling methods. Experiment results on the VCTK dataset show that the proposed method outperforms several recent baselines in terms of spectral distance and source-to-distortion ratio. Pretraining and filter augmentation also help stabilize and enhance the overall performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/23/2021

Self-Supervised Pretraining Improves Self-Supervised Pretraining

While self-supervised pretraining has proven beneficial for many compute...
research
03/02/2023

Evolutionary Augmentation Policy Optimization for Self-supervised Learning

Self-supervised learning (SSL) is a Machine Learning algorithm for pretr...
research
05/15/2023

Self-Supervised Pretraining on Paired Sequences of fMRI Data for Transfer Learning to Brain Decoding Tasks

In this work we introduce a self-supervised pretraining framework for tr...
research
05/10/2023

XTab: Cross-table Pretraining for Tabular Transformers

The success of self-supervised learning in computer vision and natural l...
research
10/23/2022

Adversarial Pretraining of Self-Supervised Deep Networks: Past, Present and Future

In this paper, we review adversarial pretraining of self-supervised deep...
research
07/28/2023

SimDETR: Simplifying self-supervised pretraining for DETR

DETR-based object detectors have achieved remarkable performance but are...
research
06/01/2023

Understanding Augmentation-based Self-Supervised Representation Learning via RKHS Approximation

Good data augmentation is one of the key factors that lead to the empiri...

Please sign up or login with your details

Forgot password? Click here to reset