Compressing Representations for Embedded Deep Learning

11/23/2019
by   Juliano S. Assine, et al.
16

Despite recent advances in architectures for mobile devices, deep learning computational requirements remains prohibitive for most embedded devices. To address that issue, we envision sharing the computational costs of inference between local devices and the cloud, taking advantage of the compression performed by the first layers of the networks to reduce communication costs. Inference in such distributed setting would allow new applications, but requires balancing a triple trade-off between computation cost, communication bandwidth, and model accuracy. We explore that trade-off by studying the compressibility of representations at different stages of MobileNetV2, showing those results agree with theoretical intuitions about deep learning, and that an optimal splitting layer for network can be found with a simple PCA-based compression scheme.

READ FULL TEXT
research
10/21/2018

To Compress, or Not to Compress: Characterizing Deep Learning Model Compression for Embedded Inference

The recent advances in deep neural networks (DNNs) make them attractive ...
research
02/15/2019

AutoQB: AutoML for Network Quantization and Binarization on Mobile Devices

In this paper, we propose a hierarchical deep reinforcement learning (DR...
research
10/11/2021

ProgFed: Effective, Communication, and Computation Efficient Federated Learning by Progressive Training

Federated learning is a powerful distributed learning scheme that allows...
research
01/27/2020

The Final Frontier: Deep Learning in Space

Machine learning, particularly deep learning, is being increasing utilis...
research
05/22/2018

Sparse Binary Compression: Towards Distributed Deep Learning with minimal Communication

Currently, progressively larger deep neural networks are trained on ever...
research
10/13/2018

Embedded deep learning in ophthalmology: Making ophthalmic imaging smarter

Deep learning has recently gained high interest in ophthalmology, due to...
research
11/17/2017

Improved Bayesian Compression

Compression of Neural Networks (NN) has become a highly studied topic in...

Please sign up or login with your details

Forgot password? Click here to reset