Nice latent variable models have log-rank

05/21/2017
by   Madeleine Udell, et al.
0

Matrices of low rank are pervasive in big data, appearing in recommender systems, movie preferences, topic models, medical records, and genomics. While there is a vast literature on how to exploit low rank structure in these datasets, there is less attention on explaining why the low rank structure appears in the first place. We explain the abundance of low rank matrices in big data by proving that certain latent variable models associated to piecewise analytic functions are of log-rank. A large matrix from such a latent variable model can be approximated, up to a small error, by a low rank matrix.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/23/2015

Bayesian Learning for Low-Rank matrix reconstruction

We develop latent variable models for Bayesian learning based low-rank m...
research
04/08/2023

Statistical and computational rates in high rank tensor estimation

Higher-order tensor datasets arise commonly in recommendation systems, n...
research
12/13/2017

Stochastic Low-Rank Bandits

Many problems in computer vision and recommender systems involve low-ran...
research
11/23/2021

Locality Sensitive Hash Aggregated Nonlinear Neighbourhood Matrix Factorization for Online Sparse Big Data Analysis

Matrix factorization (MF) can extract the low-rank features and integrat...
research
11/02/2019

Rank-deficiencies in a reduced information latent variable model

Latent variable models are well-known to suffer from rank deficiencies, ...
research
09/18/2020

Entanglement Properties of Quantum Superpositions of Smooth, Differentiable Functions

We present an entanglement analysis of quantum superpositions correspond...
research
05/17/2021

Learning a Latent Simplex in Input-Sparsity Time

We consider the problem of learning a latent k-vertex simplex K⊂ℝ^d, giv...

Please sign up or login with your details

Forgot password? Click here to reset