Learning Restricted Boltzmann Machines with Few Latent Variables
Restricted Boltzmann Machines (RBMs) are a common family of undirected graphical models with latent variables. An RBM is described by a bipartite graph, with all observed variables in one layer and all latent variables in the other. We consider the task of learning an RBM given samples generated according to it. The best algorithms for this task currently have time complexity Õ(n^2) for ferromagnetic RBMs (i.e., with attractive potentials) but Õ(n^d) for general RBMs, where n is the number of observed variables and d is the maximum degree of a latent variable. Let the MRF neighborhood of an observed variable be its neighborhood in the Markov Random Field of the marginal distribution of the observed variables. In this paper, we give an algorithm for learning general RBMs with time complexity Õ(n^2^s+1), where s is the maximum number of latent variables connected to the MRF neighborhood of an observed variable. This represents an improvement when s < log_2 (d-1), which is satisfied by many classes of RBMs with "few latent variables”. Furthermore, we give a version of this learning algorithm that recovers a model with small prediction error and whose sample complexity is independent of the minimum potential in the Markov Random Field of the observed variables. This is of interest because the sample complexity of current algorithms scales with the inverse of the minimum potential, which cannot be controlled in terms of natural properties of the RBM.
READ FULL TEXT