Revisiting Data Augmentation in Model Compression: An Empirical and Comprehensive Study

05/22/2023
by   Muzhou Yu, et al.
0

The excellent performance of deep neural networks is usually accompanied by a large number of parameters and computations, which have limited their usage on the resource-limited edge devices. To address this issue, abundant methods such as pruning, quantization and knowledge distillation have been proposed to compress neural networks and achieved significant breakthroughs. However, most of these compression methods focus on the architecture or the training method of neural networks but ignore the influence from data augmentation. In this paper, we revisit the usage of data augmentation in model compression and give a comprehensive study on the relation between model sizes and their optimal data augmentation policy. To sum up, we mainly have the following three observations: (A) Models in different sizes prefer data augmentation with different magnitudes. Hence, in iterative pruning, data augmentation with varying magnitudes leads to better performance than data augmentation with a consistent magnitude. (B) Data augmentation with a high magnitude may significantly improve the performance of large models but harm the performance of small models. Fortunately, small models can still benefit from strong data augmentations by firstly learning them with "additional parameters" and then discard these "additional parameters" during inference. (C) The prediction of a pre-trained large model can be utilized to measure the difficulty of data augmentation. Thus it can be utilized as a criterion to design better data augmentation policies. We hope this paper may promote more research on the usage of data augmentation in model compression.

READ FULL TEXT

page 1

page 4

page 6

page 7

research
01/01/2022

Role of Data Augmentation Strategies in Knowledge Distillation for Wearable Sensor Data

Deep neural networks are parametrized by several thousands or millions o...
research
06/22/2021

Data Augmentation for Opcode Sequence Based Malware Detection

Data augmentation has been successfully used in many areas of deep-learn...
research
06/21/2023

Quantifying lottery tickets under label noise: accuracy, calibration, and complexity

Pruning deep neural networks is a widely used strategy to alleviate the ...
research
07/21/2023

Model Compression Methods for YOLOv5: A Review

Over the past few years, extensive research has been devoted to enhancin...
research
10/23/2019

Occlusions for Effective Data Augmentation in Image Classification

Deep networks for visual recognition are known to leverage "easy to reco...
research
10/21/2022

Augmentation with Projection: Towards an Effective and Efficient Data Augmentation Paradigm for Distillation

Knowledge distillation is one of the primary methods of transferring kno...
research
06/02/2021

Data augmentation and pre-trained networks for extremely low data regimes unsupervised visual inspection

The use of deep features coming from pre-trained neural networks for uns...

Please sign up or login with your details

Forgot password? Click here to reset