FLOPs as a Direct Optimization Objective for Learning Sparse Neural Networks

11/07/2018
by   Raphael Tang, et al.
0

There exists a plethora of techniques for inducing structured sparsity in parametric models during the optimization process, with the final goal of resource-efficient inference. However, to the best of our knowledge, none target a specific number of floating-point operations (FLOPs) as part of a single end-to-end optimization objective, despite reporting FLOPs as part of the results. Furthermore, a one-size-fits-all approach ignores realistic system constraints, which differ significantly between, say, a GPU and a mobile phone -- FLOPs on the former incur less latency than on the latter; thus, it is important for practitioners to be able to specify a target number of FLOPs during model compression. In this work, we extend a state-of-the-art technique to directly incorporate FLOPs as part of the optimization objective and show that, given a desired FLOPs requirement, different neural networks can be successfully trained for image classification.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/08/2022

Controlled Sparsity via Constrained Optimization or: How I Learned to Stop Tuning Penalties and Love Constraints

The performance of trained neural networks is robust to harsh levels of ...
research
07/20/2020

Achieving Real-Time Execution of 3D Convolutional Neural Networks on Mobile Devices

Mobile devices are becoming an important carrier for deep learning tasks...
research
06/15/2022

Efficient Adaptive Ensembling for Image Classification

In recent times, except for sporadic cases, the trend in Computer Vision...
research
11/06/2019

A Programmable Approach to Model Compression

Deep neural networks frequently contain far more weights, represented at...
research
06/24/2020

Ramanujan Bipartite Graph Products for Efficient Block Sparse Neural Networks

Sparse neural networks are shown to give accurate predictions competitiv...
research
05/23/2023

Augmented Random Search for Multi-Objective Bayesian Optimization of Neural Networks

Deploying Deep Neural Networks (DNNs) on tiny devices is a common trend ...
research
05/03/2019

Compressibility Loss for Neural Network Weights

In this paper we apply a compressibility loss that enables learning high...

Please sign up or login with your details

Forgot password? Click here to reset