Future Gradient Descent for Adapting the Temporal Shifting Data Distribution in Online Recommendation Systems

09/02/2022
by   Mao Ye, et al.
15

One of the key challenges of learning an online recommendation model is the temporal domain shift, which causes the mismatch between the training and testing data distribution and hence domain generalization error. To overcome, we propose to learn a meta future gradient generator that forecasts the gradient information of the future data distribution for training so that the recommendation model can be trained as if we were able to look ahead at the future of its deployment. Compared with Batch Update, a widely used paradigm, our theory suggests that the proposed algorithm achieves smaller temporal domain generalization error measured by a gradient variation term in a local regret. We demonstrate the empirical advantage by comparing with various representative baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2018

Towards Understanding the Generalization Bias of Two Layer Convolutional Linear Classifiers with Gradient Descent

A major challenge in understanding the generalization of deep learning i...
research
07/21/2023

General regularization in covariate shift adaptation

Sample reweighting is one of the most widely used methods for correcting...
research
02/15/2021

How to Learn when Data Reacts to Your Model: Performative Gradient Descent

Performative distribution shift captures the setting where the choice of...
research
10/16/2020

Adaptive Dense-to-Sparse Paradigm for Pruning Online Recommendation System with Non-Stationary Data

Large scale deep learning provides a tremendous opportunity to improve t...
research
06/14/2023

Distribution Shift Inversion for Out-of-Distribution Prediction

Machine learning society has witnessed the emergence of a myriad of Out-...
research
03/21/2018

Incremental Learning-to-Learn with Statistical Guarantees

In learning-to-learn the goal is to infer a learning algorithm that work...
research
08/15/2021

Training for the Future: A Simple Gradient Interpolation Loss to Generalize Along Time

In several real world applications, machine learning models are deployed...

Please sign up or login with your details

Forgot password? Click here to reset