FRMDN: Flow-based Recurrent Mixture Density Network

08/05/2020
by   Seyedeh Fatemeh Razavi, et al.
0

Recurrent Mixture Density Networks (RMDNs) are consisted of two main parts: a Recurrent Neural Network (RNN) and a Gaussian Mixture Model (GMM), in which a kind of RNN (almost LSTM) is used to find the parameters of a GMM in every time step. While available RMDNs have been faced with different difficulties. The most important of them is high-dimensional problems. Since estimating the covariance matrix for the high-dimensional problems is more difficult, due to existing correlation between dimensions and satisfying the positive definition condition. Consequently, the available methods have usually used RMDN with a diagonal covariance matrix for high-dimensional problems by supposing independence among dimensions. Hence, in this paper with inspiring a common approach in the literature of GMM, we consider a tied configuration for each precision matrix (inverse of the covariance matrix) in RMDN as (Σ _k^ - 1 = UD_kU) to enrich GMM rather than considering a diagonal form for it. But due to simplicity, we assume U be an Identity matrix and D_k is a specific diagonal matrix for k^th component. Until now, we only have a diagonal matrix and it does not differ with available diagonal RMDNs. Besides, Flow-based neural networks are a new group of generative models that are able to transform a distribution to a simpler distribution and vice versa, through a sequence of invertible functions. Therefore, we applied a diagonal GMM on transformed observations. At every time step, the next observation, y_t + 1, has been passed through a flow-based neural network to obtain a much simpler distribution. Experimental results for a reinforcement learning problem verify the superiority of the proposed method to the base-line method in terms of Negative Log-Likelihood (NLL) for RMDN and the cumulative reward for a controller with fewer population size.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/15/2018

Estimating the Mean and Variance of a High-dimensional Normal Distribution Using a Mixture Prior

This paper provides a framework for estimating the mean and variance of ...
research
02/08/2023

Estimation of Gaussian Bi-Clusters with General Block-Diagonal Covariance Matrix and Applications

Bi-clustering is a technique that allows for the simultaneous clustering...
research
09/20/2023

Drift Control of High-Dimensional RBM: A Computational Method Based on Neural Networks

Motivated by applications in queueing theory, we consider a stochastic c...
research
02/16/2018

High-dimensional covariance matrix estimation using a low-rank and diagonal decomposition

We study high-dimensional covariance/precision matrix estimation under t...
research
08/17/2018

Inconsistency of diagonal scaling under high-dimensional limit: a replica approach

In this note, we claim that diagonal scaling of a sample covariance matr...
research
07/17/2021

Sparse Bayesian Learning with Diagonal Quasi-Newton Method For Large Scale Classification

Sparse Bayesian Learning (SBL) constructs an extremely sparse probabilis...
research
10/26/2020

Distributed Node-Specific Block-Diagonal LCMV Beamforming in Wireless Acoustic Sensor Networks

This paper derives the analytical solution of a novel distributed node-s...

Please sign up or login with your details

Forgot password? Click here to reset