Towards Collaborative Intelligence Friendly Architectures for Deep Learning

02/01/2019
by   Amir Erfan Eshratifar, et al.
0

Modern mobile devices are equipped with high-performance hardware resources such as graphics processing units (GPUs), making the end-side intelligent services more feasible. Even recently, specialized silicons as neural engines are being used for mobile devices. However, most mobile devices are still not capable of performing real-time inference using very deep models. Computations associated with deep models for today's intelligent applications are typically performed solely on the cloud. This cloud-only approach requires significant amounts of raw data to be uploaded to the cloud over the mobile wireless network and imposes considerable computational and communication load on the cloud server. Recent studies have shown that the latency and energy consumption of deep neural networks in mobile applications can be notably reduced by splitting the workload between the mobile device and the cloud. In this approach, referred to as collaborative intelligence, intermediate features computed on the mobile device are offloaded to the cloud instead of the raw input data of the network, reducing the size of the data needed to be sent to the cloud. In this paper, we design a new collaborative intelligence friendly architecture by introducing a unit responsible for reducing the size of the feature data needed to be offloaded to the cloud to a greater extent, where this unit is placed after a selected layer of a deep model. Our proposed method, across different wireless networks, achieves on average 53x improvements for end-to-end latency and 68x improvements for mobile energy consumption compared to the status quo cloud-only approach for ResNet-50, while the accuracy loss is less than 2

READ FULL TEXT
research
02/04/2019

BottleNet: A Deep Learning Architecture for Intelligent Mobile Cloud Computing Services

Recent studies have shown the latency and energy consumption of deep neu...
research
07/03/2019

On-Device Neural Net Inference with Mobile GPUs

On-device inference of machine learning models for mobile phones is desi...
research
06/24/2023

Mobile-Cloud Inference for Collaborative Intelligence

As AI applications for mobile devices become more prevalent, there is an...
research
02/01/2020

Shared Mobile-Cloud Inference for Collaborative Intelligence

As AI applications for mobile devices become more prevalent, there is an...
research
04/26/2018

Near-Lossless Deep Feature Compression for Collaborative Intelligence

Collaborative intelligence is a new paradigm for efficient deployment of...
research
01/25/2018

JointDNN: An Efficient Training and Inference Engine for Intelligent Mobile Cloud Computing Services

Deep neural networks are among the most influential architectures of dee...
research
02/13/2019

Structured Bayesian Compression for Deep models in mobile enabled devices for connected healthcare

Deep Models, typically Deep neural networks, have millions of parameters...

Please sign up or login with your details

Forgot password? Click here to reset