Phase Transitions, Optimal Errors and Optimality of Message-Passing in Generalized Linear Models

08/10/2017
by   Jean Barbier, et al.
0

We consider generalized linear models (GLMs) where an unknown n-dimensional signal vector is observed through the application of a random matrix and a non-linear (possibly probabilistic) componentwise output function. We consider the models in the high-dimensional limit, where the observation consists of m points, and m/n→α where α stays finite in the limit m,n→∞. This situation is ubiquitous in applications ranging from supervised machine learning to signal processing. A substantial amount of theoretical work analyzed the model-case when the observation matrix has i.i.d. elements and the components of the ground-truth signal are taken independently from some known distribution. While statistical physics provided number of explicit conjectures for special cases of this model, results existing for non-linear output functions were so far non-rigorous. At the same time GLMs with non-linear output functions are used as a basic building block of powerful multilayer feedforward neural networks. Therefore rigorously establishing the formulas conjectured for the mutual information is a key open problem that we solve in this paper. We also provide an explicit asymptotic formula for the optimal generalization error, and confirm the prediction of phase transitions in GLMs. Analyzing the resulting formulas for several non-linear output functions, including the rectified linear unit or modulus functions, we obtain quantitative descriptions of information-theoretic limitations of high-dimensional inference. Our proof technique relies on a new version of the interpolation method with an adaptive interpolation path and is of independent interest. Furthermore we show that a polynomial-time algorithm referred to as generalized approximate message-passing reaches the optimal generalization error for a large set of parameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/12/2017

Additivity of Information in Multilayer Networks via Additive Gaussian Noise Transforms

Multilayer (or deep) networks are powerful probabilistic models based on...
research
12/04/2019

Landscape Complexity for the Empirical Risk of Generalized Linear Models

We present a method to obtain the average and the typical value of the n...
research
07/03/2019

Understanding Phase Transitions via Mutual Information and MMSE

The ability to understand and solve high-dimensional inference problems ...
research
06/09/2020

Phase retrieval in high dimensions: Statistical and computational phase transitions

We consider the phase retrieval problem of reconstructing a n-dimensiona...
research
05/13/2019

Generalized Approximate Survey Propagation for High-Dimensional Estimation

In Generalized Linear Estimation (GLE) problems, we seek to estimate a s...
research
06/14/2018

The committee machine: Computational to statistical gaps in learning a two-layers neural network

Heuristic tools from statistical physics have been used in the past to l...
research
09/11/2019

Goodness-of-fit tests on manifolds

We develop a general theory for the goodness-of-fit test to non-linear m...

Please sign up or login with your details

Forgot password? Click here to reset