Distance in Latent Space as Novelty Measure

03/31/2020
by   Mark Philip Philipsen, et al.
0

Deep Learning performs well when training data densely covers the experience space. For complex problems this makes data collection prohibitively expensive. We propose to intelligently select samples when constructing data sets in order to best utilize the available labeling budget. The selection methodology is based on the presumption that two dissimilar samples are worth more than two similar samples in a data set. Similarity is measured based on the Euclidean distance between samples in the latent space produced by a DNN. By using a self-supervised method to construct the latent space, it is ensured that the space fits the data well and that any upfront labeling effort can be avoided. The result is more efficient, diverse, and balanced data set, which produce equal or superior results with fewer labeled examples.

READ FULL TEXT

page 1

page 2

page 3

research
11/07/2020

Identifying Mislabeled Images in Supervised Learning Utilizing Autoencoder

Supervised learning is based on the assumption that the ground truth in ...
research
03/20/2019

OCGAN: One-class Novelty Detection Using GANs with Constrained Latent Representations

We present a novel model called OCGAN for the classical problem of one-c...
research
03/31/2020

Learning from Small Data Through Sampling an Implicit Conditional Generative Latent Optimization Model

We revisit the long-standing problem of learning from small sample. In r...
research
09/10/2020

Self-Supervised Annotation of Seismic Images using Latent Space Factorization

Annotating seismic data is expensive, laborious and subjective due to th...
research
10/25/2022

SepLL: Separating Latent Class Labels from Weak Supervision Noise

In the weakly supervised learning paradigm, labeling functions automatic...
research
05/23/2019

A COLD Approach to Generating Optimal Samples

Optimising discrete data for a desired characteristic using gradient-bas...
research
05/23/2023

Self-Supervised Gaussian Regularization of Deep Classifiers for Mahalanobis-Distance-Based Uncertainty Estimation

Recent works show that the data distribution in a network's latent space...

Please sign up or login with your details

Forgot password? Click here to reset