Truncated Variational EM for Semi-Supervised Neural Simpletrons

02/07/2017
by   Dennis Forster, et al.
0

Inference and learning for probabilistic generative networks is often very challenging and typically prevents scalability to as large networks as used for deep discriminative approaches. To obtain efficiently trainable, large-scale and well performing generative networks for semi-supervised learning, we here combine two recent developments: a neural network reformulation of hierarchical Poisson mixtures (Neural Simpletrons), and a novel truncated variational EM approach (TV-EM). TV-EM provides theoretical guarantees for learning in generative networks, and its application to Neural Simpletrons results in particularly compact, yet approximately optimal, modifications of learning equations. If applied to standard benchmarks, we empirically find, that learning converges in fewer EM iterations, that the complexity per EM iteration is reduced, and that final likelihood values are higher on average. For the task of classification on data sets with few labels, learning improvements result in consistently lower error rates if compared to applications without truncation. Experiments on the MNIST data set herein allow for comparison to standard and state-of-the-art models in the semi-supervised setting. Further experiments on the NIST SD19 data set show the scalability of the approach when a manifold of additional unlabeled data is available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/28/2015

Neural Simpletrons - Minimalistic Directed Generative Networks for Learning with Few Labels

Classifiers for the semi-supervised setting often combine strong supervi...
research
12/06/2016

Semi-Supervised Learning with the Deep Rendering Mixture Model

Semi-supervised learning algorithms reduce the high cost of acquiring la...
research
10/02/2020

Deep Expectation-Maximization for Semi-Supervised Lung Cancer Screening

We present a semi-supervised algorithm for lung cancer screening in whic...
research
05/31/2022

Improvements to Supervised EM Learning of Shared Kernel Models by Feature Space Partitioning

Expectation maximisation (EM) is usually thought of as an unsupervised l...
research
11/09/2017

Can clustering scale sublinearly with its clusters? A variational EM acceleration of GMMs and k-means

One iteration of k-means or EM for Gaussian mixture models (GMMs) scales...
research
10/01/2018

Accelerated Training of Large-Scale Gaussian Mixtures by a Merger of Sublinear Approaches

We combine two recent lines of research on sublinear clustering to signi...
research
06/06/2014

Small Sample Learning of Superpixel Classifiers for EM Segmentation- Extended Version

Pixel and superpixel classifiers have become essential tools for EM segm...

Please sign up or login with your details

Forgot password? Click here to reset