Pre-training also Transfers Non-Robustness

06/21/2021
by   Jiaming Zhang, et al.
0

Pre-training has enabled many state-of-the-art results on many tasks. In spite of its recognized contribution to generalization, we observed in this study that pre-training also transfers the non-robustness from pre-trained model into the fine-tuned model. Using image classification as an example, we first conducted experiments on various datasets and network backbones to explore the factors influencing robustness. Further analysis is conducted on examining the difference between the fine-tuned model and standard model to uncover the reason leading to the non-robustness transfer. Finally, we introduce a simple robust pre-training solution by regularizing the difference between target and source tasks. Results validate the effectiveness in alleviating non-robustness and preserving generalization.

READ FULL TEXT

page 2

page 3

page 9

research
07/24/2023

On the Connection between Pre-training Data Diversity and Fine-tuning Robustness

Pre-training has been widely adopted in deep learning to improve model p...
research
01/24/2019

Is Pretraining Necessary for Hyperspectral Image Classification?

We address two questions for training a convolutional neural network (CN...
research
07/28/2021

An Evaluation of Generative Pre-Training Model-based Therapy Chatbot for Caregivers

With the advent of off-the-shelf intelligent home products and broader i...
research
04/25/2022

On-demand compute reduction with stochastic wav2vec 2.0

Squeeze and Efficient Wav2vec (SEW) is a recently proposed architecture ...
research
06/04/2019

Color Constancy Convolutional Autoencoder

In this paper, we study the importance of pre-training for the generaliz...
research
08/10/2022

Quality Not Quantity: On the Interaction between Dataset Design and Robustness of CLIP

Web-crawled datasets have enabled remarkable generalization capabilities...
research
04/16/2021

Ego-Exo: Transferring Visual Representations from Third-person to First-person Videos

We introduce an approach for pre-training egocentric video models using ...

Please sign up or login with your details

Forgot password? Click here to reset