Adaptive Dense-to-Sparse Paradigm for Pruning Online Recommendation System with Non-Stationary Data

10/16/2020
by   Mao Ye, et al.
16

Large scale deep learning provides a tremendous opportunity to improve the quality of content recommendation systems by employing both wider and deeper models, but this comes at great infrastructural cost and carbon footprint in modern data centers. Pruning is an effective technique that reduces both memory and compute demand for model inference. However, pruning for online recommendation systems is challenging due to the continuous data distribution shift (a.k.a non-stationary data). Although incremental training on the full model is able to adapt to the non-stationary data, directly applying it on the pruned model leads to accuracy loss. This is because the sparsity pattern after pruning requires adjustment to learn new patterns. To the best of our knowledge, this is the first work to provide in-depth analysis and discussion of applying pruning to online recommendation systems with non-stationary data distribution. Overall, this work makes the following contributions: 1) We present an adaptive dense to sparse paradigm equipped with a novel pruning algorithm for pruning a large scale recommendation system with non-stationary data distribution; 2) We design the pruning algorithm to automatically learn the sparsity across layers to avoid repeating hand-tuning, which is critical for pruning the heterogeneous architectures of recommendation systems trained with non-stationary data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/04/2021

Alternate Model Growth and Pruning for Efficient Training of Recommendation Systems

Deep learning recommendation systems at scale have provided remarkable g...
research
05/21/2018

Non-Oscillatory Pattern Learning for Non-Stationary Signals

This paper proposes a novel non-oscillatory pattern (NOP) learning schem...
research
08/08/2016

Online Adaptation of Deep Architectures with Reinforcement Learning

Online learning has become crucial to many problems in machine learning....
research
02/08/2023

Non-Stationary Bandits with Knapsack Problems with Advice

We consider a non-stationary Bandits with Knapsack problem. The outcome ...
research
09/02/2022

Future Gradient Descent for Adapting the Temporal Shifting Data Distribution in Online Recommendation Systems

One of the key challenges of learning an online recommendation model is ...
research
07/17/2020

A Review of Meta-level Learning in the Context of Multi-component, Multi-level Evolving Prediction Systems

The exponential growth of volume, variety and velocity of data is raisin...
research
08/16/2021

Do Proportionate Algorithms Exploit Sparsity?

Adaptive filters exploiting sparsity have been a very active research fi...

Please sign up or login with your details

Forgot password? Click here to reset