AdaMCT: Adaptive Mixture of CNN-Transformer for Sequential Recommendation

05/18/2022
by   Juyong Jiang, et al.
0

Sequential recommendation (SR) aims to model users' dynamic preferences from their historical interactions. Recently, Transformer and convolution neural networks (CNNs) have shown great success in learning representations for SR. Nevertheless, Transformer mainly focus on capturing content-based global interactions, while CNNs effectively exploit local features in practical recommendation scenarios. Thus, how to effectively aggregate CNNs and Transformer to model both local and global dependencies of historical item sequence still remains an open challenge and is rarely studied in SR. To this regard, we inject locality inductive bias into Transformer by combining its global attention mechanism with a local convolutional filter, and adaptively determine the mixing importance on a personalized basis through a module and layer-aware adaptive mixture units, named AdaMCT. Moreover, considering that softmax-based attention may encourage unimodal activation, we introduce the Squeeze-Excitation Attention (with sigmoid activation) into sequential recommendation to capture multiple relevant items (keys) simultaneously. Extensive experiments on three widely used benchmark datasets demonstrate that AdaMCT significantly outperforms the previous Transformer and CNNs-based models by an average of 18.46 NDCG@5 and achieves state-of-the-art performance.

READ FULL TEXT

page 8

page 9

research
05/16/2020

Conformer: Convolution-augmented Transformer for Speech Recognition

Recently Transformer and Convolution neural network (CNN) based models h...
research
08/18/2023

Attention Calibration for Transformer-based Sequential Recommendation

Transformer-based sequential recommendation (SR) has been booming in rec...
research
09/16/2022

Recursive Attentive Methods with Reused Item Representations for Sequential Recommendation

Sequential recommendation aims to recommend the next item of users' inte...
research
05/07/2023

Contrastive Enhanced Slide Filter Mixer for Sequential Recommendation

Sequential recommendation (SR) aims to model user preferences by capturi...
research
06/23/2021

Improving Transformer-based Sequential Recommenders through Preference Editing

One of the key challenges in Sequential Recommendation (SR) is how to ex...
research
05/16/2022

Poincaré Heterogeneous Graph Neural Networks for Sequential Recommendation

Sequential recommendation (SR) learns users' preferences by capturing th...
research
06/25/2023

G-STO: Sequential Main Shopping Intention Detection via Graph-Regularized Stochastic Transformer

Sequential recommendation requires understanding the dynamic patterns of...

Please sign up or login with your details

Forgot password? Click here to reset