Discovering Multi-Hardware Mobile Models via Architecture Search

08/18/2020
by   Grace Chu, et al.
0

Developing efficient models for mobile phones or other on-device deployments has been a popular topic in both industry and academia. In such scenarios, it is often convenient to deploy the same model on a diverse set of hardware devices owned by different end users to minimize the costs of development, deployment and maintenance. Despite the importance, designing a single neural network that can perform well on multiple devices is difficult as each device has its own specialty and restrictions: A model optimized for one device may not perform well on another. While most existing work proposes different models optimized for each single hardware, this paper is the first which explores the problem of finding a single model that performs well on multiple hardware. Specifically, we leverage architecture search to help us find the best model, where given a set of diverse hardware to optimize for, we first introduce a multi-hardware search space that is compatible with all examined hardware. Then, to measure the performance of a neural network over multiple hardware, we propose metrics that can characterize the overall latency performance in an average case and worst case scenario. With the multi-hardware search space and new metrics applied to Pixel4 CPU, GPU, DSP and EdgeTPU, we found models that perform on par or better than state-of-the-art (SOTA) models on each of our target accelerators and generalize well on many un-targeted hardware. Comparing with single-hardware searches, multi-hardware search gives a better trade-off between computation cost and model performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/17/2021

Rethinking Co-design of Neural Architectures and Hardware Accelerators

Neural architectures and hardware accelerators have been two driving for...
research
03/03/2023

Bespoke: A Block-Level Neural Network Optimization Framework for Low-Cost Deployment

As deep learning models become popular, there is a lot of need for deplo...
research
10/07/2019

Impact of Inference Accelerators on hardware selection

As opportunities for AI-assisted healthcare grow steadily, model deploym...
research
09/25/2021

Profiling Neural Blocks and Design Spaces for Mobile Neural Architecture Search

Neural architecture search automates neural network design and has achie...
research
04/27/2022

MAPLE-Edge: A Runtime Latency Predictor for Edge Devices

Neural Architecture Search (NAS) has enabled automatic discovery of more...
research
08/25/2022

SONAR: Joint Architecture and System Optimization Search

There is a growing need to deploy machine learning for different tasks o...
research
03/29/2022

AutoCoMet: Smart Neural Architecture Search via Co-Regulated Shaping Reinforcement

Designing suitable deep model architectures, for AI-driven on-device app...

Please sign up or login with your details

Forgot password? Click here to reset