Knapsack Pruning with Inner Distillation

02/19/2020
by   Yonathan Aflalo, et al.
12

Neural network pruning reduces the computational cost of an over-parameterized network to improve its efficiency. Popular methods vary from ℓ_1-norm sparsification to Neural Architecture Search (NAS). In this work, we propose a novel pruning method that optimizes the final accuracy of the pruned network and distills knowledge from the over-parameterized parent network's inner layers. To enable this approach, we formulate the network pruning as a Knapsack Problem which optimizes the trade-off between the importance of neurons and their associated computational cost. Then we prune the network channels while maintaining the high-level structure of the network. The pruned network is fine-tuned under the supervision of the parent network using its inner network knowledge, a technique we refer to as the Inner Knowledge Distillation. Our method leads to state-of-the-art pruning results on ImageNet, CIFAR-10 and CIFAR-100 using ResNet backbones. To prune complex network structures such as convolutions with skip-links and depth-wise convolutions, we propose a block grouping approach to cope with these structures. Through this we produce compact architectures with the same FLOPs as EfficientNet-B0 and MobileNetV3 but with higher accuracy, by 1% and 0.3% respectively on ImageNet, and faster runtime on GPU.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2019

Network Pruning via Transformable Architecture Search

Network pruning reduces the computation costs of an over-parameterized n...
research
09/23/2022

Tiered Pruning for Efficient Differentialble Inference-Aware Neural Architecture Search

We propose three novel pruning techniques to improve the cost and result...
research
06/15/2020

Multi-fidelity Neural Architecture Search with Knowledge Distillation

Neural architecture search (NAS) targets at finding the optimal architec...
research
09/19/2020

Neural Architecture Search Using Stable Rank of Convolutional Layers

In Neural Architecture Search (NAS), Differentiable ARchiTecture Search ...
research
01/22/2021

Hessian-Aware Pruning and Optimal Neural Implant

Pruning is an effective method to reduce the memory footprint and FLOPs ...
research
09/27/2019

Pruning from Scratch

Network pruning is an important research field aiming at reducing comput...
research
01/31/2021

AACP: Model Compression by Accurate and Automatic Channel Pruning

Channel pruning is formulated as a neural architecture search (NAS) prob...

Please sign up or login with your details

Forgot password? Click here to reset