AutoQB: AutoML for Network Quantization and Binarization on Mobile Devices

02/15/2019
by   Qian Lou, et al.
0

In this paper, we propose a hierarchical deep reinforcement learning (DRL)-based AutoML framework, AutoQB, to automatically explore the design space of channel-level network quantization and binarization for hardware-friendly deep learning on mobile devices. Compared to prior DDPG-based quantization techniques, on the various CNN models, AutoQB automatically achieves the same inference accuracy by ∼79% less computing overhead, or improves the inference accuracy by ∼2% with the same computing cost.

READ FULL TEXT

page 3

page 7

research
12/15/2017

Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference

The rising popularity of intelligent mobile devices and the daunting com...
research
06/17/2022

PICO: Pipeline Inference Framework for Versatile CNNs on Diverse Mobile Devices

Recent researches in artificial intelligence have proposed versatile con...
research
12/01/2021

Hardware-friendly Deep Learning by Network Quantization and Binarization

Quantization is emerging as an efficient approach to promote hardware-fr...
research
10/28/2020

INT8 Winograd Acceleration for Conv1D Equipped ASR Models Deployed on Mobile Devices

The intensive computation of Automatic Speech Recognition (ASR) models o...
research
03/22/2018

A Quantization-Friendly Separable Convolution for MobileNets

As deep learning (DL) is being rapidly pushed to edge computing, researc...
research
11/23/2019

Compressing Representations for Embedded Deep Learning

Despite recent advances in architectures for mobile devices, deep learni...
research
11/23/2018

Joint Neural Architecture Search and Quantization

Designing neural architectures is a fundamental step in deep learning ap...

Please sign up or login with your details

Forgot password? Click here to reset