iALS++: Speeding up Matrix Factorization with Subspace Optimization

10/26/2021
by   Steffen Rendle, et al.
0

iALS is a popular algorithm for learning matrix factorization models from implicit feedback with alternating least squares. This algorithm was invented over a decade ago but still shows competitive quality compared to recent approaches like VAE, EASE, SLIM, or NCF. Due to a computational trick that avoids negative sampling, iALS is very efficient especially for large item catalogues. However, iALS does not scale well with large embedding dimensions, d, due to its cubic runtime dependency on d. Coordinate descent variations, iCD, have been proposed to lower the complexity to quadratic in d. In this work, we show that iCD approaches are not well suited for modern processors and can be an order of magnitude slower than a careful iALS implementation for small to mid scale embedding sizes (d   100) and only perform better than iALS on large embeddings d   1000. We propose a new solver iALS++ that combines the advantages of iALS in terms of vector processing with a low computational complexity as in iCD. iALS++ is an order of magnitude faster than iCD both for small and large embedding dimensions. It can solve benchmark problems like Movielens 20M or Million Song Dataset even for 1000 dimensional embedding vectors in a few minutes.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/20/2020

On Application of Block Kaczmarz Methods in Matrix Factorization

Matrix factorization techniques compute low-rank product approximations ...
research
12/03/2021

ALX: Large Scale Matrix Factorization on TPUs

We present ALX, an open-source library for distributed matrix factorizat...
research
11/05/2018

Fast Non-Bayesian Poisson Factorization for Implicit-Feedback Recommendations

This work explores non-negative matrix factorization based on regularize...
research
06/26/2019

NetSMF: Large-Scale Network Embedding as Sparse Matrix Factorization

We study the problem of large-scale network embedding, which aims to lea...
research
07/05/2011

Distributed Matrix Completion and Robust Factorization

If learning methods are to scale to the massive sizes of modern datasets...
research
07/18/2018

Efficient Training on Very Large Corpora via Gramian Estimation

We study the problem of learning similarity functions over very large co...
research
07/19/2014

Tight convex relaxations for sparse matrix factorization

Based on a new atomic norm, we propose a new convex formulation for spar...

Please sign up or login with your details

Forgot password? Click here to reset