Associative embeddings for large-scale knowledge transfer with self-assessment

12/11/2013
by   Alexander Vezhnevets, et al.
0

We propose a method for knowledge transfer between semantically related classes in ImageNet. By transferring knowledge from the images that have bounding-box annotations to the others, our method is capable of automatically populating ImageNet with many more bounding-boxes and even pixel-level segmentations. The underlying assumption that objects from semantically related classes look alike is formalized in our novel Associative Embedding (AE) representation. AE recovers the latent low-dimensional space of appearance variations among image windows. The dimensions of AE space tend to correspond to aspects of window appearance (e.g. side view, close up, background). We model the overlap of a window with an object using Gaussian Processes (GP) regression, which spreads annotation smoothly through AE space. The probabilistic nature of GP allows our method to perform self-assessment, i.e. assigning a quality estimate to its own output. It enables trading off the amount of returned annotations for their quality. A large scale experiment on 219 classes and 0.5 million images demonstrates that our method outperforms state-of-the-art methods and baselines for both object localization and segmentation. Using self-assessment we can automatically return bounding-box annotations for 30 average overlap with ground-truth).

READ FULL TEXT

page 3

page 5

page 6

research
04/28/2021

Segmentation-Based Bounding Box Generation for Omnidirectional Pedestrian Detection

We propose a segmentation-based bounding box generation method for omnid...
research
11/11/2022

Bounding Box Priors for Cell Detection with Point Annotations

The size of an individual cell type, such as a red blood cell, does not ...
research
07/16/2018

Leveraging Pre-Trained 3D Object Detection Models For Fast Ground Truth Generation

Training 3D object detectors for autonomous driving has been limited to ...
research
01/06/2015

Object localization in ImageNet by looking out of the window

We propose a method for annotating the location of objects in ImageNet. ...
research
12/07/2020

A New Window Loss Function for Bone Fracture Detection and Localization in X-ray Images with Point-based Annotation

Object detection methods are widely adopted for computer-aided diagnosis...
research
08/30/2022

PanorAMS: Automatic Annotation for Detecting Objects in Urban Context

Large collections of geo-referenced panoramic images are freely availabl...
research
02/18/2020

FeatureNMS: Non-Maximum Suppression by Learning Feature Embeddings

Most state of the art object detectors output multiple detections per ob...

Please sign up or login with your details

Forgot password? Click here to reset