Deep learning with t-exponential Bayesian kitchen sinks

02/10/2018
by   Harris Partaourides, et al.
0

Bayesian learning has been recently considered as an effective means of accounting for uncertainty in trained deep network parameters. This is of crucial importance when dealing with small or sparse training datasets. On the other hand, shallow models that compute weighted sums of their inputs, after passing them through a bank of arbitrary randomized nonlinearities, have been recently shown to enjoy good test error bounds that depend on the number of nonlinearities. Inspired from these advances, in this paper we examine novel deep network architectures, where each layer comprises a bank of arbitrary nonlinearities, linearly combined using multiple alternative sets of weights. We effect model training by means of approximate inference based on a t-divergence measure; this generalizes the Kullback-Leibler divergence in the context of the t-exponential family of distributions. We adopt the t-exponential family since it can more flexibly accommodate real-world data, that entail outliers and distributions with fat tails, compared to conventional Gaussian model assumptions. We extensively evaluate our approach using several challenging benchmarks, and provide comparative results to related state-of-the-art techniques.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/04/2018

t-Exponential Memory Networks for Question-Answering Machines

Recent advances in deep learning have brought to the fore models that ca...
research
03/18/2019

Approximating exponential family models (not single distributions) with a two-network architecture

Recently much attention has been paid to deep generative models, since t...
research
02/18/2018

Bayesian Uncertainty Estimation for Batch Normalized Deep Networks

Deep neural networks have led to a series of breakthroughs, dramatically...
research
11/02/2016

Natural-Parameter Networks: A Class of Probabilistic Neural Networks

Neural networks (NN) have achieved state-of-the-art performance in vario...
research
04/05/2016

Comparative Deep Learning of Hybrid Representations for Image Recommendations

In many image-related tasks, learning expressive and discriminative repr...
research
03/04/2020

Maximal Causes for Exponential Family Observables

The data model of standard sparse coding assumes a weighted linear summa...
research
01/06/2018

Design Exploration of Hybrid CMOS-OxRAM Deep Generative Architectures

Deep Learning and its applications have gained tremendous interest recen...

Please sign up or login with your details

Forgot password? Click here to reset