A Confident Information First Principle for Parametric Reduction and Model Selection of Boltzmann Machines

02/05/2015
by   Xiaozhao Zhao, et al.
0

Typical dimensionality reduction (DR) methods are often data-oriented, focusing on directly reducing the number of random variables (features) while retaining the maximal variations in the high-dimensional data. In unsupervised situations, one of the main limitations of these methods lies in their dependency on the scale of data features. This paper aims to address the problem from a new perspective and considers model-oriented dimensionality reduction in parameter spaces of binary multivariate distributions. Specifically, we propose a general parameter reduction criterion, called Confident-Information-First (CIF) principle, to maximally preserve confident parameters and rule out less confident parameters. Formally, the confidence of each parameter can be assessed by its contribution to the expected Fisher information distance within the geometric manifold over the neighbourhood of the underlying real distribution. We then revisit Boltzmann machines (BM) from a model selection perspective and theoretically show that both the fully visible BM (VBM) and the BM with hidden units can be derived from the general binary multivariate distribution using the CIF principle. This can help us uncover and formalize the essential parts of the target density that BM aims to capture and the non-essential parts that BM should discard. Guided by the theoretical analysis, we develop a sample-specific CIF for model selection of BM that is adaptive to the observed samples. The method is studied in a series of density estimation experiments and has been shown effective in terms of the estimate accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/16/2013

Understanding Boltzmann Machine and Deep Learning via A Confident Information First Principle

Typical dimensionality reduction methods focus on directly reducing the ...
research
10/10/2019

A Multi-view Dimensionality Reduction Algorithm Based on Smooth Representation Model

Over the past few decades, we have witnessed a large family of algorithm...
research
03/17/2018

Large-Scale Model Selection with Misspecification

Model selection is crucial to high-dimensional learning and inference fo...
research
04/28/2014

Conditional Density Estimation with Dimensionality Reduction via Squared-Loss Conditional Entropy Minimization

Regression aims at estimating the conditional mean of output given input...
research
11/05/2010

Model Selection by Loss Rank for Classification and Unsupervised Learning

Hutter (2007) recently introduced the loss rank principle (LoRP) as a ge...
research
11/21/2019

TMI: Thermodynamic inference of data manifolds

The Gibbs-Boltzmann distribution offers a physically interpretable way t...
research
08/09/2012

Algorithmic Simplicity and Relevance

The human mind is known to be sensitive to complexity. For instance, the...

Please sign up or login with your details

Forgot password? Click here to reset