Dynamic Deep Neural Networks: Optimizing Accuracy-Efficiency Trade-offs by Selective Execution

01/02/2017
by   Lanlan Liu, et al.
0

We introduce Dynamic Deep Neural Networks (D2NN), a new type of feed-forward deep neural network that allows selective execution. Given an input, only a subset of D2NN neurons are executed, and the particular subset is determined by the D2NN itself. By pruning unnecessary computation depending on input, D2NNs provide a way to improve computational efficiency. To achieve dynamic selective execution, a D2NN augments a feed-forward deep neural network (directed acyclic graph of differentiable modules) with controller modules. Each controller module is a sub-network whose output is a decision that controls whether other modules can execute. A D2NN is trained end to end. Both regular and controller modules in a D2NN are learnable and are jointly trained to optimize both accuracy and efficiency. Such training is achieved by integrating backpropagation with reinforcement learning. With extensive experiments of various D2NN architectures on image classification tasks, we demonstrate that D2NNs are general and flexible, and can effectively optimize accuracy-efficiency trade-offs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/24/2021

Neural Function Modules with Sparse Arguments: A Dynamic Approach to Integrating Information across Layers

Feed-forward neural networks consist of a sequence of layers, in which e...
research
07/25/2018

Conditional Information Gain Networks

Deep neural network models owe their representational power to the high ...
research
03/15/2022

LDP: Learnable Dynamic Precision for Efficient Deep Neural Network Training and Inference

Low precision deep neural network (DNN) training is one of the most effe...
research
07/01/2023

Sparsity-aware generalization theory for deep neural networks

Deep artificial neural networks achieve surprising generalization abilit...
research
05/07/2021

ResMLP: Feedforward networks for image classification with data-efficient training

We present ResMLP, an architecture built entirely upon multi-layer perce...
research
11/05/2018

Efficient Inference on Deep Neural Networks by Dynamic Representations and Decision Gates

The current trade-off between depth and computational cost makes it diff...
research
01/31/2019

A Data Driven Method of Optimizing Feedforward Compensator for Autonomous Vehicle

A reliable controller is critical and essential for the execution of saf...

Please sign up or login with your details

Forgot password? Click here to reset