Learning Temporal Consistency for Source-Free Video Domain Adaptation

03/09/2022
by   Yuecong Xu, et al.
0

Video-based Unsupervised Domain Adaptation (VUDA) methods improve the robustness of video models, enabling them to be applied to action recognition tasks across different environments. However, these methods require constant access to source data during the adaptation process. Yet in many real-world applications, subjects and scenes in the source video domain should be irrelevant to those in the target video domain. With the increasing emphasis on data privacy, such methods that require source data access would raise serious privacy issues. Therefore, to cope with such concern, a more practical domain adaptation scenario is formulated as the Source-Free Video-based Domain Adaptation (SFVDA). Though there are a few methods for Source-Free Domain Adaptation (SFDA) on image data, these methods yield degenerating performance in SFVDA due to the multi-modality nature of videos, with the existence of additional temporal features. In this paper, we propose a novel Attentive Temporal Consistent Network (ATCoN) to address SFVDA by learning temporal consistency, guaranteed by two novel consistency objectives, namely feature consistency and source prediction consistency, performed across local temporal features. ATCoN further constructs effective overall temporal features by attending to local temporal features based on prediction confidence. Empirical results demonstrate the state-of-the-art performance of ATCoN across various cross-domain action recognition benchmarks.

READ FULL TEXT

page 6

page 16

research
08/10/2022

EXTERN: Leveraging Endo-Temporal Regularization for Black-box Video Domain Adaptation

To enable video models to be applied seamlessly across video tasks in di...
research
03/18/2023

Augmenting and Aligning Snippets for Few-Shot Video Domain Adaptation

For video models to be transferred and applied seamlessly across video t...
research
08/17/2023

The Unreasonable Effectiveness of Large Language-Vision Models for Source-free Video Domain Adaptation

Source-Free Video Unsupervised Domain Adaptation (SFVUDA) methods consis...
research
03/30/2022

CycDA: Unsupervised Cycle Domain Adaptation from Image to Video

Although action recognition has achieved impressive results over recent ...
research
09/21/2021

Multi-Source Video Domain Adaptation with Temporal Attentive Moment Alignment

Multi-Source Domain Adaptation (MSDA) is a more practical domain adaptat...
research
08/17/2021

Channel-Temporal Attention for First-Person Video Domain Adaptation

Unsupervised Domain Adaptation (UDA) can transfer knowledge from labeled...
research
08/21/2019

Preserving Semantic and Temporal Consistency for Unpaired Video-to-Video Translation

In this paper, we investigate the problem of unpaired video-to-video tra...

Please sign up or login with your details

Forgot password? Click here to reset