CL-MAE: Curriculum-Learned Masked Autoencoders

08/31/2023
by   Neelu Madan, et al.
0

Masked image modeling has been demonstrated as a powerful pretext task for generating robust representations that can be effectively generalized across multiple downstream tasks. Typically, this approach involves randomly masking patches (tokens) in input images, with the masking strategy remaining unchanged during training. In this paper, we propose a curriculum learning approach that updates the masking strategy to continually increase the complexity of the self-supervised reconstruction task. We conjecture that, by gradually increasing the task complexity, the model can learn more sophisticated and transferable representations. To facilitate this, we introduce a novel learnable masking module that possesses the capability to generate masks of different complexities, and integrate the proposed module into masked autoencoders (MAE). Our module is jointly trained with the MAE, while adjusting its behavior during training, transitioning from a partner to the MAE (optimizing the same reconstruction loss) to an adversary (optimizing the opposite loss), while passing through a neutral state. The transition between these behaviors is smooth, being regulated by a factor that is multiplied with the reconstruction loss of the masking module. The resulting training procedure generates an easy-to-hard curriculum. We train our Curriculum-Learned Masked Autoencoder (CL-MAE) on ImageNet and show that it exhibits superior representation learning capabilities compared to MAE. The empirical results on five downstream tasks confirm our conjecture, demonstrating that curriculum learning can be successfully used to self-supervise masked autoencoders.

READ FULL TEXT

page 2

page 12

research
05/12/2016

Learning the Curriculum with Bayesian Optimization for Task-Specific Word Representation Learning

We use Bayesian optimization to learn curricula for word representation ...
research
01/30/2023

PointSmile: Point Self-supervised Learning via Curriculum Mutual Information

Self-supervised learning is attracting wide attention in point cloud pro...
research
07/04/2022

Masked Autoencoders in 3D Point Cloud Representation Learning

Transformer-based Self-supervised Representation Learning methods learn ...
research
04/25/2023

Curriculum Modeling the Dependence among Targets with Multi-task Learning for Financial Marketing

Multi-task learning for various real-world applications usually involves...
research
07/14/2023

A Quantitative Approach to Predicting Representational Learning and Performance in Neural Networks

A key property of neural networks (both biological and artificial) is ho...
research
03/30/2023

Mixed Autoencoder for Self-supervised Visual Representation Learning

Masked Autoencoder (MAE) has demonstrated superior performance on variou...
research
10/20/2022

i-MAE: Are Latent Representations in Masked Autoencoders Linearly Separable?

Masked image modeling (MIM) has been recognized as a strong and popular ...

Please sign up or login with your details

Forgot password? Click here to reset