Tiered Pruning for Efficient Differentialble Inference-Aware Neural Architecture Search

09/23/2022
by   Slawomir Kierat, et al.
0

We propose three novel pruning techniques to improve the cost and results of inference-aware Differentiable Neural Architecture Search (DNAS). First, we introduce , a stochastic bi-path building block for DNAS, which can search over inner hidden dimensions with memory and compute complexity. Second, we present an algorithm for pruning blocks within a stochastic layer of the SuperNet during the search. Third, we describe a novel technique for pruning unnecessary stochastic layers during the search. The optimized models resulting from the search are called PruNet and establishes a new state-of-the-art Pareto frontier for NVIDIA V100 in terms of inference latency for ImageNet Top-1 image classification accuracy. PruNet as a backbone also outperforms GPUNet and EfficientNet on the COCO object detection task on inference latency relative to mean Average Precision (mAP).

READ FULL TEXT
research
06/17/2020

Fine-Grained Stochastic Architecture Search

State-of-the-art deep networks are often too large to deploy on mobile d...
research
12/21/2020

Searching for Controllable Image Restoration Networks

Diverse user preferences over images have recently led to a great amount...
research
03/30/2020

DHP: Differentiable Meta Pruning via HyperNetworks

Network pruning has been the driving force for the efficient inference o...
research
02/19/2020

Knapsack Pruning with Inner Distillation

Neural network pruning reduces the computational cost of an over-paramet...
research
05/20/2019

DARC: Differentiable ARchitecture Compression

In many learning situations, resources at inference time are significant...
research
04/25/2021

Balancing Accuracy and Latency in Multipath Neural Networks

The growing capacity of neural networks has strongly contributed to thei...
research
03/23/2023

DetOFA: Efficient Training of Once-for-All Networks for Object Detection by Using Pre-trained Supernet and Path Filter

We address the challenge of training a large supernet for the object det...

Please sign up or login with your details

Forgot password? Click here to reset