Towards Accelerated Model Training via Bayesian Data Selection

08/21/2023
by   Zhijie Deng, et al.
0

Mislabeled, duplicated, or biased data in real-world scenarios can lead to prolonged training and even hinder model convergence. Traditional solutions prioritizing easy or hard samples lack the flexibility to handle such a variety simultaneously. Recent work has proposed a more reasonable data selection principle by examining the data's impact on the model's generalization loss. However, its practical adoption relies on less principled approximations and additional clean holdout data. This work solves these problems by leveraging a lightweight Bayesian treatment and incorporating off-the-shelf zero-shot predictors built on large-scale pre-trained models. The resulting algorithm is efficient and easy-to-implement. We perform extensive empirical studies on challenging benchmarks with considerable data noise and imbalance in the online batch selection scenario, and observe superior training efficiency over competitive baselines. Notably, on the challenging WebVision benchmark, our method can achieve similar predictive performance with significantly fewer training iterations than leading data selection methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/13/2022

MAPL: Parameter-Efficient Adaptation of Unimodal Pre-Trained Models for Vision-Language Few-Shot Prompting

Large pre-trained models have proved to be remarkable zero- and (prompt-...
research
05/13/2020

A Biologically Inspired Feature Enhancement Framework for Zero-Shot Learning

Most of the Zero-Shot Learning (ZSL) algorithms currently use pre-traine...
research
04/20/2023

Learning Sample Difficulty from Pre-trained Models for Reliable Prediction

Large-scale pre-trained models have achieved remarkable success in a var...
research
09/14/2023

Efficiently Robustify Pre-trained Models

A recent trend in deep learning algorithms has been towards training lar...
research
04/12/2022

ReCLIP: A Strong Zero-Shot Baseline for Referring Expression Comprehension

Training a referring expression comprehension (ReC) model for a new visu...
research
09/12/2021

Leveraging Table Content for Zero-shot Text-to-SQL with Meta-Learning

Single-table text-to-SQL aims to transform a natural language question i...
research
03/28/2023

Large-scale Pre-trained Models are Surprisingly Strong in Incremental Novel Class Discovery

Discovering novel concepts from unlabelled data and in a continuous mann...

Please sign up or login with your details

Forgot password? Click here to reset