Enhancing Automatically Discovered Multi-level Acoustic Patterns Considering Context Consistency With Applications in Spoken Term Detection

09/07/2015
by   Cheng-Tao Chung, et al.
0

This paper presents a novel approach for enhancing the multiple sets of acoustic patterns automatically discovered from a given corpus. In a previous work it was proposed that different HMM configurations (number of states per model, number of distinct models) for the acoustic patterns form a two-dimensional space. Multiple sets of acoustic patterns automatically discovered with the HMM configurations properly located on different points over this two-dimensional space were shown to be complementary to one another, jointly capturing the characteristics of the given corpus. By representing the given corpus as sequences of acoustic patterns on different HMM sets, the pattern indices in these sequences can be relabeled considering the context consistency across the different sequences. Good improvements were observed in preliminary experiments of pattern spoken term detection (STD) performed on both TIMIT and Mandarin Broadcast News with such enhanced patterns.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/07/2015

Unsupervised Spoken Term Detection with Spoken Queries by Multi-level Acoustic Patterns with Varying Model Granularity

This paper presents a new approach for unsupervised Spoken Term Detectio...
research
07/17/2017

Unsupervised Iterative Deep Learning of Speech Features and Acoustic Tokens with Applications to Spoken Term Detection

In this paper we aim to automatically discover high quality frame-level ...
research
06/20/2016

A Nonparametric Bayesian Approach for Spoken Term detection by Example Query

State of the art speech recognition systems use data-intensive context-d...
research
11/28/2017

Unsupervised Discovery of Structured Acoustic Tokens with Applications to Spoken Term Detection

In this paper, we compare two paradigms for unsupervised discovery of st...
research
04/01/2018

Completely Unsupervised Phoneme Recognition by Adversarially Learning Mapping Relationships from Audio Embeddings

Unsupervised discovery of acoustic tokens from audio corpora without ann...
research
06/19/2023

Comparison of L2 Korean pronunciation error patterns from five L1 backgrounds by using automatic phonetic transcription

This paper presents a large-scale analysis of L2 Korean pronunciation er...

Please sign up or login with your details

Forgot password? Click here to reset