Restructurable Activation Networks

08/17/2022
by   Kartikeya Bhardwaj, et al.
9

Is it possible to restructure the non-linear activation functions in a deep network to create hardware-efficient models? To address this question, we propose a new paradigm called Restructurable Activation Networks (RANs) that manipulate the amount of non-linearity in models to improve their hardware-awareness and efficiency. First, we propose RAN-explicit (RAN-e) – a new hardware-aware search space and a semi-automatic search algorithm – to replace inefficient blocks with hardware-aware blocks. Next, we propose a training-free model scaling method called RAN-implicit (RAN-i) where we theoretically prove the link between network topology and its expressivity in terms of number of non-linear units. We demonstrate that our networks achieve state-of-the-art results on ImageNet at different scales and for several types of hardware. For example, compared to EfficientNet-Lite-B0, RAN-e achieves a similar accuracy while improving Frames-Per-Second (FPS) by 1.5x on Arm micro-NPUs. On the other hand, RAN-i demonstrates up to 2x reduction in #MACs over ConvNexts with a similar or better accuracy. We also show that RAN-i achieves nearly 40 Finally, RAN-i based object detection networks achieve a similar or higher mAP and up to 33 The code to train and evaluate RANs and the pretrained networks are available at https://github.com/ARM-software/ML-restructurable-activation-networks.

READ FULL TEXT

page 3

page 6

research
07/15/2019

Padé Activation Units: End-to-end Learning of Flexible Activation Functions in Deep Networks

The performance of deep network learning strongly depends on the choice ...
research
09/22/2018

Design Space Exploration of Neural Network Activation Function Circuits

The widespread application of artificial neural networks has prompted re...
research
07/15/2020

Attention as Activation

Activation functions and attention mechanisms are typically treated as h...
research
06/02/2022

DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks

Efficient deep neural network (DNN) models equipped with compact operato...
research
08/12/2021

MicroNet: Improving Image Recognition with Extremely Low FLOPs

This paper aims at addressing the problem of substantial performance deg...
research
09/26/2021

Efficient Non-linear Calculators

A novel algorithm for producing smooth nonlinearities on digital hardwar...
research
05/17/2022

ShiftAddNAS: Hardware-Inspired Search for More Accurate and Efficient Neural Networks

Neural networks (NNs) with intensive multiplications (e.g., convolutions...

Please sign up or login with your details

Forgot password? Click here to reset