Recurrent Deep Stacking Networks for Speech Recognition

12/14/2016
by   Peidong Wang, et al.
0

This paper presented our work on applying Recurrent Deep Stacking Networks (RDSNs) to Robust Automatic Speech Recognition (ASR) tasks. In the paper, we also proposed a more efficient yet comparable substitute to RDSN, Bi- Pass Stacking Network (BPSN). The main idea of these two models is to add phoneme-level information into acoustic models, transforming an acoustic model to the combination of an acoustic model and a phoneme-level N-gram model. Experiments showed that RDSN and BPsn can substantially improve the performances over conventional DNNs.

READ FULL TEXT

page 1

page 2

page 3

research
12/14/2016

Incorporating Language Level Information into Acoustic Models

This paper proposed a class of novel Deep Recurrent Neural Networks whic...
research
02/18/2021

Echo State Speech Recognition

We propose automatic speech recognition (ASR) models inspired by echo st...
research
05/05/2021

Accent Recognition with Hybrid Phonetic Features

The performance of voice-controlled systems is usually influenced by acc...
research
10/11/2013

A Bayesian Network View on Acoustic Model-Based Techniques for Robust Speech Recognition

This article provides a unifying Bayesian network view on various approa...
research
04/15/2018

Twin Regularization for online speech recognition

Online speech recognition is crucial for developing natural human-machin...
research
11/30/2022

Preliminary Study on SSCF-derived Polar Coordinate for ASR

The transition angles are defined to describe the vowel-to-vowel transit...
research
06/16/2020

Quantization of Acoustic Model Parameters in Automatic Speech Recognition Framework

Robust automatic speech recognition (ASR) system exploits state-of-the-a...

Please sign up or login with your details

Forgot password? Click here to reset