A Generic Network Compression Framework for Sequential Recommender Systems

04/21/2020
by   Yang Sun, et al.
0

Sequential recommender systems (SRS) have become the key technology in capturing user's dynamic interests and generating high-quality recommendations. Current state-of-the-art sequential recommender models are typically based on a sandwich-structured deep neural network, where one or more middle (hidden) layers are placed between the input embedding layer and output softmax layer. In general, these models require a large number of parameters (such as using a large embedding dimension or a deep network architecture) to obtain their optimal performance. Despite the effectiveness, at some point, further increasing model size may be harder for model deployment in resource-constraint devices, resulting in longer responding time and larger memory footprint. To resolve the issues, we propose a compressed sequential recommendation framework, termed as CpRec, where two generic model shrinking techniques are employed. Specifically, we first propose a block-wise adaptive decomposition to approximate the input and softmax matrices by exploiting the fact that items in SRS obey a long-tailed distribution. To reduce the parameters of the middle layers, we introduce three layer-wise parameter sharing schemes. We instantiate CpRec using deep convolutional neural network with dilated kernels given consideration to both recommendation accuracy and efficiency. By the extensive ablation studies, we demonstrate that the proposed CpRec can achieve up to 4∼8 times compression rates in real-world SRS datasets. Meanwhile, CpRec is faster during training, and in most cases outperforms its uncompressed counterpart.

READ FULL TEXT

page 2

page 3

page 5

page 6

page 7

page 9

page 10

page 11

research
06/23/2023

Review of compressed embedding layers and their applications for recommender systems

We review the literature on trainable, compressed embedding layers and d...
research
07/15/2021

Scene-adaptive Knowledge Distillation for Sequential Recommendation via Differentiable Architecture Search

Sequential recommender systems (SRS) have become a research hotspot due ...
research
06/18/2018

GroupReduce: Block-Wise Low-Rank Approximation for Neural Language Model Shrinking

Model compression is essential for serving large deep neural nets on dev...
research
04/23/2022

On-Device Next-Item Recommendation with Self-Supervised Knowledge Distillation

Modern recommender systems operate in a fully server-based fashion. To c...
research
10/31/2021

Enhancing Top-N Item Recommendations by Peer Collaboration

Deep neural networks (DNN) have achieved great success in the recommende...
research
05/18/2023

Less is More! A slim architecture for optimal language translation

The softmax attention mechanism has emerged as a noteworthy development ...
research
11/05/2022

One Person, One Model–Learning Compound Router for Sequential Recommendation

Deep learning has brought significant breakthroughs in sequential recomm...

Please sign up or login with your details

Forgot password? Click here to reset