Neuro-SERKET: Development of Integrative Cognitive System through the Composition of Deep Probabilistic Generative Models

10/20/2019
by   Tadahiro Taniguchi, et al.
0

This paper describes a framework for the development of an integrative cognitive system based on probabilistic generative models (PGMs) called Neuro-SERKET. Neuro-SERKET is an extension of SERKET, which can compose elemental PGMs developed in a distributed manner and provide a scheme that allows the composed PGMs to learn throughout the system in an unsupervised way. In addition to the head-to-tail connection supported by SERKET, Neuro-SERKET supports tail-to-tail and head-to-head connections, as well as neural network-based modules, i.e., deep generative models. As an example of a Neuro-SERKET application, an integrative model was developed by composing a variational autoencoder (VAE), a Gaussian mixture model (GMM), latent Dirichlet allocation (LDA), and automatic speech recognition (ASR). The model is called VAE+GMM+LDA+ASR. The performance of VAE+GMM+LDA+ASR and the validity of Neuro-SERKET were demonstrated through a multimodal categorization task using image data and a speech signal of numerical digits.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/25/2017

Investigation of Using VAE for i-Vector Speaker Verification

New system for i-vector speaker recognition based on variational autoenc...
research
11/29/2016

Deep Quantization: Encoding Convolutional Activations with Deep Generative Model

Deep convolutional neural networks (CNNs) have proven highly effective f...
research
05/24/2022

Symbol Emergence as Inter-personal Categorization with Head-to-head Latent Word

In this study, we propose a head-to-head type (H2H-type) inter-personal ...
research
05/06/2017

A Generative Model of a Pronunciation Lexicon for Hindi

Voice browser applications in Text-to- Speech (TTS) and Automatic Speech...
research
02/06/2020

Generating diverse and natural text-to-speech samples using a quantized fine-grained VAE and auto-regressive prosody prior

Recent neural text-to-speech (TTS) models with fine-grained latent featu...
research
05/11/2020

Exploring TTS without T Using Biologically/Psychologically Motivated Neural Network Modules (ZeroSpeech 2020)

In this study, we reported our exploration of Text-To-Speech without Tex...
research
05/28/2020

Variational Autoencoder with Embedded Student-t Mixture Model for Authorship Attribution

Traditional computational authorship attribution describes a classificat...

Please sign up or login with your details

Forgot password? Click here to reset