Adversarial Permutation Invariant Training for Universal Sound Separation

10/21/2022
by   Emilian Postolache, et al.
0

Universal sound separation consists of separating mixes with arbitrary sounds of different types, and permutation invariant training (PIT) is used to train source agnostic models that do so. In this work, we complement PIT with adversarial losses but find it challenging with the standard formulation used in speech source separation. We overcome this challenge with a novel I-replacement context-based adversarial loss, and by training with multiple discriminators. Our experiments show that by simply improving the loss (keeping the same model and dataset) we obtain a non-negligible improvement of 1.4 dB SI-SNRi in the reverberant FUSS dataset. We also find adversarial PIT to be effective at reducing spectral holes, ubiquitous in mask-based separation models, which highlights the potential relevance of adversarial losses for source separation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/11/2023

Universal Source Separation with Weakly Labelled Data

Universal source separation (USS) is a fundamental research task for com...
research
02/09/2021

On permutation invariant training for speech source separation

We study permutation invariant training (PIT), which targets at the perm...
research
05/08/2019

Universal Sound Separation

Recent deep learning approaches have achieved impressive performance on ...
research
11/11/2020

Surrogate Source Model Learning for Determined Source Separation

We propose to learn surrogate functions of universal speech priors for d...
research
11/11/2022

Optimal Condition Training for Target Source Separation

Recent research has shown remarkable performance in leveraging multiple ...
research
04/07/2022

Heterogeneous Target Speech Separation

We introduce a new paradigm for single-channel target source separation ...

Please sign up or login with your details

Forgot password? Click here to reset