On-Device Neural Net Inference with Mobile GPUs

07/03/2019
by   Juhyun Lee, et al.
0

On-device inference of machine learning models for mobile phones is desirable due to its lower latency and increased privacy. Running such a compute-intensive task solely on the mobile CPU, however, can be difficult due to limited computing power, thermal constraints, and energy consumption. App developers and researchers have begun exploiting hardware accelerators to overcome these challenges. Recently, device manufacturers are adding neural processing units into high-end phones for on-device inference, but these account for only a small fraction of hand-held devices. In this paper, we present how we leverage the mobile GPU, a ubiquitous hardware accelerator on virtually every phone, to run inference of deep neural networks in real-time for both Android and iOS devices. By describing our architecture, we also discuss how to design networks that are mobile GPU-friendly. Our state-of-the-art mobile GPU inference engine is integrated into the open-source project TensorFlow Lite and publicly available at https://tensorflow.org/lite.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/01/2019

Towards Collaborative Intelligence Friendly Architectures for Deep Learning

Modern mobile devices are equipped with high-performance hardware resour...
research
12/17/2021

AI-Assisted Verification of Biometric Data Collection

Recognizing actions from a video feed is a challenging task to automate,...
research
05/06/2023

Energy-Latency Attacks to On-Device Neural Networks via Sponge Poisoning

In recent years, on-device deep learning has gained attention as a means...
research
12/09/2021

GPU backed Data Mining on Android Devices

Choosing an appropriate programming paradigm for high-performance comput...
research
11/04/2021

Safe and Practical GPU Acceleration in TrustZone

We present a holistic design for GPU-accelerated computation in TrustZon...
research
01/03/2019

HG-Caffe: Mobile and Embedded Neural Network GPU (OpenCL) Inference Engine with FP16 Supporting

Breakthroughs in the fields of deep learning and mobile system-on-chips ...
research
02/21/2022

Enabling On-Device Smartphone GPU based Training: Lessons Learned

Deep Learning (DL) has shown impressive performance in many mobile appli...

Please sign up or login with your details

Forgot password? Click here to reset