Statistical limits of dictionary learning: random matrix theory and the spectral replica method

09/14/2021
by   Jean Barbier, et al.
0

We consider increasingly complex models of matrix denoising and dictionary learning in the Bayes-optimal setting, in the challenging regime where the matrices to infer have a rank growing linearly with the system size. This is in contrast with most existing literature concerned with the low-rank (i.e., constant-rank) regime. We first consider a class of rotationally invariant matrix denoising problems whose mutual information and minimum mean-square error are computable using standard techniques from random matrix theory. Next, we analyze the more challenging models of dictionary learning. To do so we introduce a novel combination of the replica method from statistical mechanics together with random matrix theory, coined spectral replica method. It allows us to conjecture variational formulas for the mutual information between hidden representations and the noisy data as well as for the overlaps quantifying the optimal reconstruction error. The proposed methods reduce the number of degrees of freedom from Θ(N^2) (matrix entries) to Θ(N) (eigenvalues or singular values), and yield Coulomb gas representations of the mutual information which are reminiscent of matrix models in physics. The main ingredients are the use of HarishChandra-Itzykson-Zuber spherical integrals combined with a new replica symmetric decoupling ansatz at the level of the probability distributions of eigenvalues (or singular values) of certain overlap matrices.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/10/2015

Matrix Completion from Fewer Entries: Spectral Detectability and Rank Estimation

The completion of low rank matrices from few entries is a task with many...
research
06/07/2023

Bayesian Extensive-Rank Matrix Factorization with Rotational Invariant Priors

We consider a statistical model for matrix factorization in a regime whe...
research
05/16/2020

Information-theoretic limits of a multiview low-rank symmetric spiked matrix model

We consider a generalization of an important class of high-dimensional i...
research
06/02/2023

Matrix Inference in Growing Rank Regimes

The inference of a large symmetric signal-matrix 𝐒∈ℝ^N× N corrupted by a...
research
07/15/2019

Concentration of the matrix-valued minimum mean-square error in optimal Bayesian inference

We consider Bayesian inference of signals with vector-valued entries. Ex...
research
02/25/2018

The Mutual Information in Random Linear Estimation Beyond i.i.d. Matrices

There has been definite progress recently in proving the variational sin...
research
03/16/2023

Gradient flow on extensive-rank positive semi-definite matrix denoising

In this work, we present a new approach to analyze the gradient flow for...

Please sign up or login with your details

Forgot password? Click here to reset