How Severe is Benchmark-Sensitivity in Video Self-Supervised Learning?

03/27/2022
by   Fida Mohammad Thoker, et al.
2

Despite the recent success of video self-supervised learning, there is much still to be understood about their generalization capability. In this paper, we investigate how sensitive video self-supervised learning is to the currently used benchmark convention and whether methods generalize beyond the canonical evaluation setting. We do this across four different factors of sensitivity: domain, samples, actions and task. Our comprehensive set of over 500 experiments, which encompasses 7 video datasets, 9 self-supervised methods and 6 video understanding tasks, reveals that current benchmarks in video self-supervised learning are not a good indicator of generalization along these sensitivity factors. Further, we find that self-supervised methods considerably lag behind vanilla supervised pre-training, especially when domain shift is large and the amount of available downstream samples are low. From our analysis we distill the SEVERE-benchmark, a subset of our experiments, and discuss its implication for evaluating the generalizability of representations obtained by existing and future self-supervised video learning methods.

READ FULL TEXT

page 2

page 5

page 7

page 8

page 9

page 10

page 23

page 25

research
06/18/2022

Self-Supervised Learning for Videos: A Survey

The remarkable success of deep learning in various domains relies on the...
research
11/19/2020

Robot Gaining Accurate Pouring Skills through Self-Supervised Learning and Generalization

Pouring is one of the most commonly executed tasks in humans' daily live...
research
06/02/2023

Masked Autoencoder for Unsupervised Video Summarization

Summarizing a video requires a diverse understanding of the video, rangi...
research
06/03/2023

Uncovering the Hidden Dynamics of Video Self-supervised Learning under Distribution Shifts

Video self-supervised learning (VSSL) has made significant progress in r...
research
06/23/2023

Bring Your Own Data! Self-Supervised Evaluation for Large Language Models

With the rise of Large Language Models (LLMs) and their ubiquitous deplo...
research
06/22/2020

Don't Wait, Just Weight: Improving Unsupervised Representations by Learning Goal-Driven Instance Weights

In the absence of large labelled datasets, self-supervised learning tech...
research
06/02/2022

Expressiveness and Learnability: A Unifying View for Evaluating Self-Supervised Learning

We propose a unifying view to analyze the representation quality of self...

Please sign up or login with your details

Forgot password? Click here to reset