Noise Regularizes Over-parameterized Rank One Matrix Recovery, Provably

02/07/2022
by   Tianyi Liu, et al.
0

We investigate the role of noise in optimization algorithms for learning over-parameterized models. Specifically, we consider the recovery of a rank one matrix Y^*∈ R^d× d from a noisy observation Y using an over-parameterization model. We parameterize the rank one matrix Y^* by XX^⊤, where X∈ R^d× d. We then show that under mild conditions, the estimator, obtained by the randomly perturbed gradient descent algorithm using the square loss function, attains a mean square error of O(σ^2/d), where σ^2 is the variance of the observational noise. In contrast, the estimator obtained by gradient descent without random perturbation only attains a mean square error of O(σ^2). Our result partially justifies the implicit regularization effect of noise when learning over-parameterized models, and provides new understanding of training over-parameterized neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/26/2017

Algorithmic Regularization in Over-parameterized Matrix Recovery

We study the problem of recovering a low-rank matrix X^ from linear meas...
research
06/16/2020

Robust Recovery via Implicit Bias of Discrepant Learning Rates for Double Over-parameterization

Recent advances have shown that implicit bias of gradient descent on ove...
research
07/16/2020

Understanding Implicit Regularization in Over-Parameterized Nonlinear Statistical Model

We study the implicit regularization phenomenon induced by simple optimi...
research
09/12/2019

Analysis of Regression Tree Fitting Algorithms in Learning to Rank

In learning to rank area, industry-level applications have been dominate...
research
01/18/2019

On Estimation under Noisy Order Statistics

This paper presents an estimation framework to assess the performance of...
research
01/27/2021

On the computational and statistical complexity of over-parameterized matrix sensing

We consider solving the low rank matrix sensing problem with Factorized ...
research
07/09/2021

Batch Inverse-Variance Weighting: Deep Heteroscedastic Regression

Heteroscedastic regression is the task of supervised learning where each...

Please sign up or login with your details

Forgot password? Click here to reset