Generative learning for deep networks

09/25/2017
by   Boris Flach, et al.
0

Learning, taking into account full distribution of the data, referred to as generative, is not feasible with deep neural networks (DNNs) because they model only the conditional distribution of the outputs given the inputs. Current solutions are either based on joint probability models facing difficult estimation problems or learn two separate networks, mapping inputs to outputs (recognition) and vice-versa (generation). We propose an intermediate approach. First, we show that forward computation in DNNs with logistic sigmoid activations corresponds to a simplified approximate Bayesian inference in a directed probabilistic multi-layer model. This connection allows to interpret DNN as a probabilistic model of the output and all hidden units given the input. Second, we propose that in order for the recognition and generation networks to be more consistent with the joint model of the data, weights of the recognition and generator network should be related by transposition. We demonstrate in a tentative experiment that such a coupled pair can be learned generatively, modelling the full distribution of the data, and has enough capacity to perform well in both recognition and generation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/06/2021

Disentangling deep neural networks with rectified linear units using duality

Despite their success deep neural networks (DNNs) are still largely cons...
research
04/16/2021

Fast ABC with joint generative modelling and subset simulation

We propose a novel approach for solving inverse-problems with high-dimen...
research
07/07/2020

SpinalNet: Deep Neural Network with Gradual Input

Over the past few years, deep neural networks (DNNs) have garnered remar...
research
02/21/2018

Detecting Learning vs Memorization in Deep Neural Networks using Shared Structure Validation Sets

The roles played by learning and memorization represent an important top...
research
04/22/2023

Lookahead Diffusion Probabilistic Models for Refining Mean Estimation

We propose lookahead diffusion probabilistic models (LA-DPMs) to exploit...
research
10/09/2020

Multichannel Generative Language Model: Learning All Possible Factorizations Within and Across Channels

A channel corresponds to a viewpoint or transformation of an underlying ...
research
12/08/2020

Deep Energy-Based NARX Models

This paper is directed towards the problem of learning nonlinear ARX mod...

Please sign up or login with your details

Forgot password? Click here to reset