Understanding Trainable Sparse Coding via Matrix Factorization

09/01/2016
by   Thomas Moreau, et al.
0

Sparse coding is a core building block in many data analysis and machine learning pipelines. Typically it is solved by relying on generic optimization techniques, that are optimal in the class of first-order methods for non-smooth, convex functions, such as the Iterative Soft Thresholding Algorithm and its accelerated version (ISTA, FISTA). However, these methods don't exploit the particular structure of the problem at hand nor the input data distribution. An acceleration using neural networks was proposed in Gregor10, coined LISTA, which showed empirically that one could achieve high quality estimates with few iterations by modifying the parameters of the proximal splitting appropriately. In this paper we study the reasons for such acceleration. Our mathematical analysis reveals that it is related to a specific matrix factorization of the Gram kernel of the dictionary, which attempts to nearly diagonalise the kernel with a basis that produces a small perturbation of the ℓ_1 ball. When this factorization succeeds, we prove that the resulting splitting algorithm enjoys an improved convergence bound with respect to the non-adaptive version. Moreover, our analysis also shows that conditions for acceleration occur mostly at the beginning of the iterative process, consistent with numerical experiments. We further validate our analysis by showing that on dictionaries where this factorization does not exist, adaptive acceleration fails.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/02/2017

Understanding the Learned Iterative Soft Thresholding Algorithm with matrix factorization

Sparse coding is a core building block in many data analysis and machine...
research
07/09/2021

Block Alternating Bregman Majorization Minimization with Extrapolation

In this paper, we consider a class of nonsmooth nonconvex optimization p...
research
12/05/2022

Matrix factorization with neural networks

Matrix factorization is an important mathematical problem encountered in...
research
10/20/2021

A Row-Wise Update Algorithm for Sparse Stochastic Matrix Factorization

Nonnegative matrix factorization arises widely in machine learning and d...
research
03/31/2017

Catalyst Acceleration for Gradient-Based Non-Convex Optimization

We introduce a generic scheme to solve nonconvex optimization problems u...
research
08/23/2019

QuicK-means: Acceleration of K-means by learning a fast transform

K-means -- and the celebrated Lloyd algorithm -- is more than the cluste...
research
01/23/2020

Ada-LISTA: Learned Solvers Adaptive to Varying Models

Neural networks that are based on unfolding of an iterative solver, such...

Please sign up or login with your details

Forgot password? Click here to reset