Impact of Training Dataset Size on Neural Answer Selection Models

01/29/2019
by   Trond Linjordet, et al.
0

It is held as a truism that deep neural networks require large datasets to train effective models. However, large datasets, especially with high-quality labels, can be expensive to obtain. This study sets out to investigate (i) how large a dataset must be to train well-performing models, and (ii) what impact can be shown from fractional changes to the dataset size. A practical method to investigate these questions is to train a collection of deep neural answer selection models using fractional subsets of varying sizes of an initial dataset. We observe that dataset size has a conspicuous lack of effect on the training of some of these models, bringing the underlying algorithms into question.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/29/2019

Less is More: An Exploration of Data Redundancy with Active Dataset Subsampling

Deep Neural Networks (DNNs) often rely on very large datasets for traini...
research
10/07/2021

A Data-Centric Approach for Training Deep Neural Networks with Less Data

While the availability of large datasets is perceived to be a key requir...
research
03/12/2022

A Proposal to Study "Is High Quality Data All We Need?"

Even though deep neural models have achieved superhuman performance on m...
research
03/25/2022

Impact of Dataset on Acoustic Models for Automatic Speech Recognition

In Automatic Speech Recognition, GMM-HMM had been widely used for acoust...
research
11/12/2021

A Simple and Fast Baseline for Tuning Large XGBoost Models

XGBoost, a scalable tree boosting algorithm, has proven effective for ma...
research
06/01/2015

Blocks and Fuel: Frameworks for deep learning

We introduce two Python frameworks to train neural networks on large dat...
research
03/27/2023

Curriculum Learning for Compositional Visual Reasoning

Visual Question Answering (VQA) is a complex task requiring large datase...

Please sign up or login with your details

Forgot password? Click here to reset