Getting deep recommenders fit: Bloom embeddings for sparse binary input/output networks

06/13/2017
by   Joan Serrà, et al.
0

Recommendation algorithms that incorporate techniques from deep learning are becoming increasingly popular. Due to the structure of the data coming from recommendation domains (i.e., one-hot-encoded vectors of item preferences), these algorithms tend to have large input and output dimensionalities that dominate their overall size. This makes them difficult to train, due to the limited memory of graphical processing units, and difficult to deploy on mobile devices with limited hardware. To address these difficulties, we propose Bloom embeddings, a compression technique that can be applied to the input and output of neural network models dealing with sparse high-dimensional binary-coded instances. Bloom embeddings are computationally efficient, and do not seriously compromise the accuracy of the model up to 1/5 compression ratios. In some cases, they even improve over the original accuracy, with relative increases up to 12 alternative methods, obtaining favorable results. We also discuss a number of further advantages of Bloom embeddings, such as 'on-the-fly' constant-time operation, zero or marginal space requirements, training time speedups, or the fact that they do not require any change to the core model architecture or training configuration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/08/2022

Beyond Double Ascent via Recurrent Neural Tangent Kernel in Sequential Recommendation

Overfitting has long been considered a common issue to large neural netw...
research
01/01/2020

Lossless Compression of Deep Neural Networks

Deep neural networks have been successful in many predictive modeling ta...
research
09/27/2022

Efficient On-Device Session-Based Recommendation

On-device session-based recommendation systems have been achieving incre...
research
11/03/2017

Compressing Word Embeddings via Deep Compositional Code Learning

Natural language processing (NLP) models often require a massive number ...
research
01/22/2020

Normalization of Input-output Shared Embeddings in Text Generation Models

Neural Network based models have been state-of-the-art models for variou...
research
07/27/2019

Modeling Winner-Take-All Competition in Sparse Binary Projections

Inspired by the advances in biological science, the study of sparse bina...
research
02/27/2020

Entangled Watermarks as a Defense against Model Extraction

Machine learning involves expensive data collection and training procedu...

Please sign up or login with your details

Forgot password? Click here to reset