Learning Transformations To Reduce the Geometric Shift in Object Detection

01/13/2023
by   Vidit Vidit, et al.
0

The performance of modern object detectors drops when the test distribution differs from the training one. Most of the methods that address this focus on object appearance changes caused by, e.g., different illumination conditions, or gaps between synthetic and real images. Here, by contrast, we tackle geometric shifts emerging from variations in the image capture process, or due to the constraints of the environment causing differences in the apparent geometry of the content itself. We introduce a self-training approach that learns a set of geometric transformations to minimize these shifts without leveraging any labeled data in the new domain, nor any information about the cameras. We evaluate our method on two different shifts, i.e., a camera's field of view (FoV) change and a viewpoint change. Our results evidence that learning geometric transformations helps detectors to perform better in the target domains.

READ FULL TEXT

page 4

page 8

page 12

page 13

page 14

page 17

page 18

page 20

research
03/27/2022

Towards Domain Generalization in Object Detection

Despite the striking performance achieved by modern detectors when train...
research
09/28/2022

The Change You Want to See

We live in a dynamic world where things change all the time. Given two i...
research
09/01/2021

DPA: Learning Robust Physical Adversarial Camouflages for Object Detectors

Adversarial attacks are feasible in the real world for object detection....
research
07/30/2020

Epipolar-Guided Deep Object Matching for Scene Change Detection

This paper describes a viewpoint-robust object-based change detection ne...
research
12/09/2022

Contrastive View Design Strategies to Enhance Robustness to Domain Shifts in Downstream Object Detection

Contrastive learning has emerged as a competitive pretraining method for...
research
04/23/2021

Co-training for Deep Object Detection: Comparing Single-modal and Multi-modal Approaches

Top-performing computer vision models are powered by convolutional neura...
research
06/30/2021

Small in-distribution changes in 3D perspective and lighting fool both CNNs and Transformers

Neural networks are susceptible to small transformations including 2D ro...

Please sign up or login with your details

Forgot password? Click here to reset