Can we learn better with hard samples?

04/07/2023
by   Subin Sahayam, et al.
0

In deep learning, mini-batch training is commonly used to optimize network parameters. However, the traditional mini-batch method may not learn the under-represented samples and complex patterns in the data, leading to a longer time for generalization. To address this problem, a variant of the traditional algorithm has been proposed, which trains the network focusing on mini-batches with high loss. The study evaluates the effectiveness of the proposed training using various deep neural networks trained on three benchmark datasets (CIFAR-10, CIFAR-100, and STL-10). The deep neural networks used in the study are ResNet-18, ResNet-50, Efficient Net B4, EfficientNetV2-S, and MobilenetV3-S. The experimental results showed that the proposed method can significantly improve the test accuracy and speed up the convergence compared to the traditional mini-batch training method. Furthermore, we introduce a hyper-parameter delta (δ) that decides how many mini-batches are considered for training. Experiments on various values of δ found that the performance of the proposed method for smaller δ values generally results in similar test accuracy and faster generalization. We show that the proposed method generalizes in 26.47 traditional mini-batch method in EfficientNet-B4 on STL-10. The proposed method also improves the test top-1 accuracy by 7.26

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/29/2018

Second-order Optimization Method for Large Mini-batch: Training ResNet-50 on ImageNet in 35 Epochs

Large-scale distributed training of deep neural networks suffer from the...
research
06/08/2020

Passive Batch Injection Training Technique: Boosting Network Performance by Injecting Mini-Batches from a different Data Distribution

This work presents a novel training technique for deep neural networks t...
research
09/30/2021

Impact of Channel Variation on One-Class Learning for Spoof Detection

The value of Spoofing detection in increasing the reliability of the ASV...
research
10/25/2021

Some like it tough: Improving model generalization via progressively increasing the training difficulty

In this work, we propose to progressively increase the training difficul...
research
11/19/2019

Carpe Diem, Seize the Samples Uncertain "At the Moment" for Adaptive Batch Selection

The performance of deep neural networks is significantly affected by how...
research
01/13/2022

Recursive Least Squares Policy Control with Echo State Network

The echo state network (ESN) is a special type of recurrent neural netwo...
research
07/14/2021

Disparity Between Batches as a Signal for Early Stopping

We propose a metric for evaluating the generalization ability of deep ne...

Please sign up or login with your details

Forgot password? Click here to reset