MobiRNN: Efficient Recurrent Neural Network Execution on Mobile GPU

06/03/2017
by   Qingqing Cao, et al.
0

In this paper, we explore optimizations to run Recurrent Neural Network (RNN) models locally on mobile devices. RNN models are widely used for Natural Language Processing, Machine Translation, and other tasks. However, existing mobile applications that use RNN models do so on the cloud. To address privacy and efficiency concerns, we show how RNN models can be run locally on mobile devices. Existing work on porting deep learning models to mobile devices focus on Convolution Neural Networks (CNNs) and cannot be applied directly to RNN models. In response, we present MobiRNN, a mobile-specific optimization framework that implements GPU offloading specifically for mobile GPUs. Evaluations using an RNN model for activity recognition shows that MobiRNN does significantly decrease the latency of running RNN models on phones.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/19/2020

RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition

Recurrent neural networks (RNNs) based automatic speech recognition has ...
research
06/20/2016

Fast, Compact, and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers for Mobile Devices

Acoustic models based on long short-term memory recurrent neural network...
research
11/12/2022

PriMask: Cascadable and Collusion-Resilient Data Masking for Mobile Cloud Inference

Mobile cloud offloading is indispensable for inference tasks based on la...
research
07/04/2022

Sustainable AI Processing at the Edge

Edge computing is a popular target for accelerating machine learning alg...
research
04/28/2019

Application of Autoencoder-Assisted Recurrent Neural Networks to Prevent Cases of Sudden Infant Death Syndrome

This project develops and trains a Recurrent Neural Network (RNN) that m...
research
11/09/2020

SplitEasy: A Practical Approach for Training ML models on Mobile Devices in a split second

Modern mobile devices, although resourceful, cannot train state-of-the-a...
research
01/10/2021

Adversarially robust and explainable model compression with on-device personalization for NLP applications

On-device Deep Neural Networks (DNNs) have recently gained more attentio...

Please sign up or login with your details

Forgot password? Click here to reset