DeepAI AI Chat
Log In Sign Up

Semi-Supervised Learning of Semantic Correspondence with Pseudo-Labels

by   Jiwon Kim, et al.
Korea University

Establishing dense correspondences across semantically similar images remains a challenging task due to the significant intra-class variations and background clutters. Traditionally, a supervised learning was used for training the models, which required tremendous manually-labeled data, while some methods suggested a self-supervised or weakly-supervised learning to mitigate the reliance on the labeled data, but with limited performance. In this paper, we present a simple, but effective solution for semantic correspondence that learns the networks in a semi-supervised manner by supplementing few ground-truth correspondences via utilization of a large amount of confident correspondences as pseudo-labels, called SemiMatch. Specifically, our framework generates the pseudo-labels using the model's prediction itself between source and weakly-augmented target, and uses pseudo-labels to learn the model again between source and strongly-augmented target, which improves the robustness of the model. We also present a novel confidence measure for pseudo-labels and data augmentation tailored for semantic correspondence. In experiments, SemiMatch achieves state-of-the-art performance on various benchmarks, especially on PF-Willow by a large margin.


page 1

page 4

page 5

page 7


Barely-Supervised Learning: Semi-Supervised Learning with very few labeled images

This paper tackles the problem of semi-supervised learning when the set ...

Joint Learning of Feature Extraction and Cost Aggregation for Semantic Correspondence

Establishing dense correspondences across semantically similar images is...

FuRPE: Learning Full-body Reconstruction from Part Experts

Full-body reconstruction is a fundamental but challenging task. Owing to...

GAN-Supervised Dense Visual Alignment

We propose GAN-Supervised Learning, a framework for learning discriminat...

Diffusion Models and Semi-Supervised Learners Benefit Mutually with Few Labels

We propose a three-stage training strategy called dual pseudo training (...

ConMatch: Semi-Supervised Learning with Confidence-Guided Consistency Regularization

We present a novel semi-supervised learning framework that intelligently...

Recurrent Transformer Networks for Semantic Correspondence

We present recurrent transformer networks (RTNs) for obtaining dense cor...