DynamicEmbedding: Extending TensorFlow for Colossal-Scale Applications

04/17/2020
by   Yun Zeng, et al.
6

One of the limitations of deep learning models with sparse features today stems from the predefined nature of their input, which requires a dictionary be defined prior to the training. With this paper we propose both a theory and a working system design which remove this limitation, and show that the resulting models are able to perform better and efficiently run at a much larger scale. Specifically, we achieve this by decoupling a model's content from its form to tackle architecture evolution and memory growth separately. To efficiently handle model growth, we propose a new neuron model, called DynamicCell, drawing inspiration from from the free energy principle [15] to introduce the concept of reaction to discharge non-digestive energy, which also subsumes gradient descent based approaches as its special cases. We implement DynamicCell by introducing a new server into TensorFlow to take over most of the work involving model growth. Consequently, it enables any existing deep learning models to efficiently handle arbitrary number of distinct sparse features (e.g., search queries), and grow incessantly without redefining the model. Most notably, one of our models, which has been reliably running in production for over a year, is capable of suggesting high quality keywords for advertisers of Google Smart Campaigns and achieved significant accuracy gains based on a challenging metric – evidence that data-driven, self-evolving systems can potentially exceed the performance of traditional rule-based approaches.

READ FULL TEXT

page 6

page 7

page 8

page 10

page 13

page 18

page 19

page 20

research
01/08/2019

CROSSBOW: Scaling Deep Learning with Small Batch Sizes on Multi-GPU Servers

Deep learning models are trained on servers with many GPUs, and training...
research
09/20/2021

Scaling TensorFlow to 300 million predictions per second

We present the process of transitioning machine learning models to the T...
research
08/08/2020

Why to "grow" and "harvest" deep learning models?

Current expectations from training deep learning models with gradient-ba...
research
09/04/2019

Performance Analysis and Comparison of Distributed Machine Learning Systems

Deep learning has permeated through many aspects of computing/processing...
research
09/03/2019

Beyond Human-Level Accuracy: Computational Challenges in Deep Learning

Deep learning (DL) research yields accuracy and product improvements fro...
research
04/08/2021

A Design Space Study for LISTA and Beyond

In recent years, great success has been witnessed in building problem-sp...
research
01/10/2019

Context Aware Machine Learning

We propose a principle for exploring context in machine learning models....

Please sign up or login with your details

Forgot password? Click here to reset