The Geometry of Neural Nets' Parameter Spaces Under Reparametrization

02/14/2023
by   Agustinus Kristiadi, et al.
0

Model reparametrization – transforming the parameter space via a bijective differentiable map – is a popular way to improve the training of neural networks. But reparametrizations have also been problematic since they induce inconsistencies in, e.g., Hessian-based flatness measures, optimization trajectories, and modes of probability density functions. This complicates downstream analyses, e.g. one cannot make a definitive statement about the connection between flatness and generalization. In this work, we study the invariance quantities of neural nets under reparametrization from the perspective of Riemannian geometry. We show that this notion of invariance is an inherent property of any neural net, as long as one acknowledges the assumptions about the metric that is always present, albeit often implicitly, and uses the correct transformation rules under reparametrization. We present discussions on measuring the flatness of minima, in optimization, and in probability-density maximization, along with applications in studying the biases of optimizers and in Bayesian inference.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/04/2018

Accelerating Natural Gradient with Higher-Order Invariance

An appealing property of the natural gradient is that it is invariant to...
research
06/30/2022

Invariance Properties of the Natural Gradient in Overparametrised Systems

The natural gradient field is a vector field that lives on a model equip...
research
05/19/2023

Complexity of Feed-Forward Neural Networks from the Perspective of Functional Equivalence

In this paper, we investigate the complexity of feed-forward neural netw...
research
03/15/2017

Sharp Minima Can Generalize For Deep Nets

Despite their overwhelming capacity to overfit, deep learning architectu...
research
10/11/2019

Bayesian Optimization Meets Riemannian Manifolds in Robot Learning

Bayesian optimization (BO) recently became popular in robotics to optimi...
research
08/16/2020

A Functional Perspective on Learning Symmetric Functions with Neural Networks

Symmetric functions, which take as input an unordered, fixed-size set, a...
research
05/25/2017

Latent Geometry and Memorization in Generative Models

It can be difficult to tell whether a trained generative model has learn...

Please sign up or login with your details

Forgot password? Click here to reset