Are All Training Examples Created Equal? An Empirical Study

11/30/2018
by   Kailas Vodrahalli, et al.
0

Modern computer vision algorithms often rely on very large training datasets. However, it is conceivable that a carefully selected subsample of the dataset is sufficient for training. In this paper, we propose a gradient-based importance measure that we use to empirically analyze relative importance of training images in four datasets of varying complexity. We find that in some cases, a small subsample is indeed sufficient for training. For other datasets, however, the relative differences in importance are negligible. These results have important implications for active learning on deep networks. Additionally, our analysis method can be used as a general tool to better understand diversity of training examples in datasets.

READ FULL TEXT

page 3

page 4

page 11

research
01/17/2019

Diverse mini-batch Active Learning

We study the problem of reducing the amount of labeled training data req...
research
06/03/2013

Learning from networked examples in a k-partite graph

Many machine learning algorithms are based on the assumption that traini...
research
02/27/2018

Adversarial Active Learning for Deep Networks: a Margin Based Approach

We propose a new active learning strategy designed for deep neural netwo...
research
11/25/2013

Are all training examples equally valuable?

When learning a new concept, not all training examples may prove equally...
research
07/15/2021

Deep Learning on a Data Diet: Finding Important Examples Early in Training

The recent success of deep learning has partially been driven by trainin...
research
06/26/2016

Fast Incremental Learning for Off-Road Robot Navigation

A promising approach to autonomous driving is machine learning. In such ...
research
06/05/2023

NLU on Data Diets: Dynamic Data Subset Selection for NLP Classification Tasks

Finetuning large language models inflates the costs of NLU applications ...

Please sign up or login with your details

Forgot password? Click here to reset