Reconstruction Task Finds Universal Winning Tickets

02/23/2022
by   Ruichen Li, et al.
0

Pruning well-trained neural networks is effective to achieve a promising accuracy-efficiency trade-off in computer vision regimes. However, most of existing pruning algorithms only focus on the classification task defined on the source domain. Different from the strong transferability of the original model, a pruned network is hard to transfer to complicated downstream tasks such as object detection arXiv:arch-ive/2012.04643. In this paper, we show that the image-level pretrain task is not capable of pruning models for diverse downstream tasks. To mitigate this problem, we introduce image reconstruction, a pixel-level task, into the traditional pruning framework. Concretely, an autoencoder is trained based on the original model, and then the pruning process is optimized with both autoencoder and classification losses. The empirical study on benchmark downstream tasks shows that the proposed method can outperform state-of-the-art results explicitly.

READ FULL TEXT
research
07/10/2023

One-Shot Pruning for Fast-adapting Pre-trained Models on Devices

Large-scale pre-trained models have been remarkably successful in resolv...
research
01/27/2023

Voting from Nearest Tasks: Meta-Vote Pruning of Pre-trained Models for Downstream Tasks

As a few large-scale pre-trained models become the major choices of vari...
research
02/19/2020

Compressing BERT: Studying the Effects of Weight Pruning on Transfer Learning

Universal feature extractors, such as BERT for natural language processi...
research
01/10/2020

Pruning Convolutional Neural Networks with Self-Supervision

Convolutional neural networks trained without supervision come close to ...
research
12/08/2020

The Lottery Ticket Hypothesis for Object Recognition

Recognition tasks, such as object recognition and keypoint estimation, h...
research
12/10/2021

Pruning Pretrained Encoders with a Multitask Objective

The sizes of pretrained language models make them challenging and expens...
research
01/22/2019

CAE-P: Compressive Autoencoder with Pruning Based on ADMM

Since compressive autoencoder (CAE) was proposed, autoencoder, as a simp...

Please sign up or login with your details

Forgot password? Click here to reset