Accelerated and Inexact Soft-Impute for Large-Scale Matrix and Tensor Completion

03/16/2017
by   Quanming Yao, et al.
0

Matrix and tensor completion aim to recover a low-rank matrix / tensor from limited observations and have been commonly used in applications such as recommender systems and multi-relational data mining. A state-of-the-art matrix completion algorithm is Soft-Impute, which exploits the special "sparse plus low-rank" structure of the matrix iterates to allow efficient SVD in each iteration. Though Soft-Impute is a proximal algorithm, it is generally believed that acceleration destroys the special structure and is thus not useful. In this paper, we show that Soft-Impute can indeed be accelerated without comprising this structure. To further reduce the iteration time complexity, we propose an approximate singular value thresholding scheme based on the power method. Theoretical analysis shows that the proposed algorithm still enjoys the fast O(1/T^2) convergence rate of accelerated proximal algorithms. We further extend the proposed algorithm to tensor completion with the scaled latent nuclear norm regularizer. We show that a similar "sparse plus low-rank" structure also exists, leading to low iteration complexity and fast O(1/T^2) convergence rate. Extensive experiments demonstrate that the proposed algorithm is much faster than Soft-Impute and other state-of-the-art matrix and tensor completion algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/01/2017

Large-Scale Low-Rank Matrix Learning with Nonconvex Regularizers

Low-rank modeling has many important applications in computer vision and...
research
02/18/2022

A two-phase rank-based Soft-Impute algorithm for low-rank matrix completion

Matrix completion aims to recover an unknown low-rank matrix from a smal...
research
12/03/2015

Fast Low-Rank Matrix Learning with Nonconvex Regularization

Low-rank modeling has a lot of important applications in machine learnin...
research
04/04/2014

Orthogonal Rank-One Matrix Pursuit for Low Rank Matrix Completion

In this paper, we propose an efficient and scalable low rank matrix comp...
research
07/23/2018

Scalable Tensor Completion with Nonconvex Regularization

Low-rank tensor completion problem aims to recover a tensor from limited...
research
03/02/2019

Matrix Completion via Nonconvex Regularization: Convergence of the Proximal Gradient Algorithm

Matrix completion has attracted much interest in the past decade in mach...
research
11/29/2019

Sparse and Low-Rank Tensor Regression via Parallel Proximal Method

Motivated by applications in various scientific fields having demand of ...

Please sign up or login with your details

Forgot password? Click here to reset