Dual Training of Energy-Based Models with Overparametrized Shallow Neural Networks

07/11/2021
by   Carles Domingo Enrich, et al.
6

Energy-based models (EBMs) are generative models that are usually trained via maximum likelihood estimation. This approach becomes challenging in generic situations where the trained energy is nonconvex, due to the need to sample the Gibbs distribution associated with this energy. Using general Fenchel duality results, we derive variational principles dual to maximum likelihood EBMs with shallow overparametrized neural network energies, both in the active (aka feature-learning) and lazy regimes. In the active regime, this dual formulation leads to a training algorithm in which one updates concurrently the particles in the sample space and the neurons in the parameter space of the energy. We also consider a variant of this algorithm in which the particles are sometimes restarted at random samples drawn from the data set, and show that performing these restarts at every iteration step corresponds to score matching training. Using intermediate parameter setups in our dual algorithm thereby gives a way to interpolate between maximum likelihood and score matching training. These results are illustrated in simple numerical experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/15/2021

On Energy-Based Models with Overparametrized Shallow Neural Networks

Energy-based models (EBMs) are a simple yet powerful framework for gener...
research
05/24/2023

Training Energy-Based Normalizing Flow with Score-Matching Objectives

In this paper, we establish a connection between the parameterization of...
research
10/03/2022

Statistical Efficiency of Score Matching: The View from Isoperimetry

Deep generative models parametrized up to a normalizing constant (e.g. e...
research
12/01/2021

Maximum Likelihood Estimation for Brownian Motion Tree Models Based on One Sample

We study the problem of maximum likelihood estimation given one data sam...
research
10/21/2019

On Predictive Information Sub-optimality of RNNs

Certain biological neurons demonstrate a remarkable capability to optima...
research
02/23/2021

EBMs Trained with Maximum Likelihood are Generator Models Trained with a Self-adverserial Loss

Maximum likelihood estimation is widely used in training Energy-based mo...
research
01/04/2019

Coverage and Quality Driven Training of Generative Image Models

Generative modeling of natural images has been extensively studied in re...

Please sign up or login with your details

Forgot password? Click here to reset