Guiding Neural Network Initialization via Marginal Likelihood Maximization

12/17/2020
by   Anthony S. Tai, et al.
0

We propose a simple, data-driven approach to help guide hyperparameter selection for neural network initialization. We leverage the relationship between neural network and Gaussian process models having corresponding activation and covariance functions to infer the hyperparameter values desirable for model initialization. Our experiment shows that marginal likelihood maximization provides recommendations that yield near-optimal prediction performance on MNIST classification task under experiment constraints. Furthermore, our empirical results indicate consistency in the proposed technique, suggesting that computation cost for the procedure could be significantly reduced with smaller training sets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/07/2023

Improving Hyperparameter Learning under Approximate Inference in Gaussian Process Models

Approximate inference in Gaussian process (GP) models with non-conjugate...
research
02/03/2016

Single-Solution Hypervolume Maximization and its use for Improving Generalization of Neural Networks

This paper introduces the hypervolume maximization with a single solutio...
research
05/24/2022

Imposing Gaussian Pre-Activations in a Neural Network

The goal of the present work is to propose a way to modify both the init...
research
02/23/2022

Bayesian Model Selection, the Marginal Likelihood, and Generalization

How do we compare between hypotheses that are entirely consistent with o...
research
02/18/2020

Observational nonidentifiability, generalized likelihood and free energy

We study the parameter estimation problem in mixture models with observa...
research
02/22/2011

Predictive Active Set Selection Methods for Gaussian Processes

We propose an active set selection framework for Gaussian process classi...

Please sign up or login with your details

Forgot password? Click here to reset