Active Learning for Sound Event Detection

02/12/2020
by   Shuyang Zhao, et al.
0

This paper proposes an active learning system for sound event detection (SED). It aims at maximizing the accuracy of a learned SED model with limited annotation effort. The proposed system analyzes an initially unlabeled audio dataset, from which it selects sound segments for manual annotation. The candidate segments are generated based on a proposed change point detection approach, and the selection is based on the principle of mismatch-first farthest-traversal. During the training of SED models, recordings are used as training inputs, preserving the long-term context for annotated segments. The proposed system clearly outperforms reference methods in the two datasets used for evaluation (TUT Rare Sound 2017 and TAU Spatial Sound 2019). Training with recordings as context outperforms training with only annotated segments. Mismatch-first farthest-traversal outperforms reference sample selection methods based on random sampling and uncertainty sampling. Remarkably, the required annotation effort can be greatly reduced on the dataset where target sound events are rare: by annotating only 2 SED performance is similar to annotating all the training data.

READ FULL TEXT

page 3

page 4

page 5

research
06/04/2022

STARSS22: A dataset of spatial recordings of real scenes with spatiotemporal annotations of sound events

This report presents the Sony-TAu Realistic Spatial Soundscapes 2022 (ST...
research
01/10/2019

Cosine-similarity penalty to discriminate sound classes in weakly-supervised sound event detection

The design of new methods and models when only weakly-labeled data are a...
research
06/09/2021

Audiovisual transfer learning for audio tagging and sound event detection

We study the merit of transfer learning for two sound recognition proble...
research
03/26/2020

Incremental Learning Algorithm for Sound Event Detection

This paper presents a new learning strategy for the Sound Event Detectio...
research
12/19/2021

Detect what you want: Target Sound Detection

Human beings can perceive a target sound that we are interested in from ...
research
11/03/2021

A Strongly-Labelled Polyphonic Dataset of Urban Sounds with Spatiotemporal Context

This paper introduces SINGA:PURA, a strongly labelled polyphonic urban s...
research
01/17/2018

NELS - Never-Ending Learner of Sounds

Sounds are essential to how humans perceive and interact with the world ...

Please sign up or login with your details

Forgot password? Click here to reset