Compression-aware Projection with Greedy Dimension Reduction for Convolutional Neural Network Activations

10/17/2021
by   Yu-Shan Tai, et al.
0

Convolutional neural networks (CNNs) achieve remarkable performance in a wide range of fields. However, intensive memory access of activations introduces considerable energy consumption, impeding deployment of CNNs on resourceconstrained edge devices. Existing works in activation compression propose to transform feature maps for higher compressibility, thus enabling dimension reduction. Nevertheless, in the case of aggressive dimension reduction, these methods lead to severe accuracy drop. To improve the trade-off between classification accuracy and compression ratio, we propose a compression-aware projection system, which employs a learnable projection to compensate for the reconstruction loss. In addition, a greedy selection metric is introduced to optimize the layer-wise compression ratio allocation by considering both accuracy and #bits reduction simultaneously. Our test results show that the proposed methods effectively reduce 2.91x 5.97x memory access with negligible accuracy drop on MobileNetV2/ResNet18/VGG16.

READ FULL TEXT
research
07/16/2022

Learnable Mixed-precision and Dimension Reduction Co-design for Low-storage Activation

Recently, deep convolutional neural networks (CNNs) have achieved many e...
research
06/08/2020

EDCompress: Energy-Aware Model Compression with Dataflow

Edge devices demand low energy consumption, cost and small form factor. ...
research
09/25/2019

CAT: Compression-Aware Training for bandwidth reduction

Convolutional neural networks (CNNs) have become the dominant neural net...
research
10/07/2021

Optimal (Euclidean) Metric Compression

We study the problem of representing all distances between n points in ℝ...
research
10/01/2018

Extended Bit-Plane Compression for Convolutional Neural Network Accelerators

After the tremendous success of convolutional neural networks in image c...
research
02/19/2020

Model-Agnostic Structured Sparsification with Learnable Channel Shuffle

Recent advances in convolutional neural networks (CNNs) usually come wit...
research
09/17/2019

Thanks for Nothing: Predicting Zero-Valued Activations with Lightweight Convolutional Neural Networks

Convolutional neural networks (CNNs) introduce state-of-the-art results ...

Please sign up or login with your details

Forgot password? Click here to reset