Tunable Sensitivity to Large Errors in Neural Network Training

11/23/2016
by   Gil Keren, et al.
0

When humans learn a new concept, they might ignore examples that they cannot make sense of at first, and only later focus on such examples, when they are more useful for learning. We propose incorporating this idea of tunable sensitivity for hard examples in neural network learning, using a new generalization of the cross-entropy gradient step, which can be used in place of the gradient in any gradient-based training method. The generalized gradient is parameterized by a value that controls the sensitivity of the training process to harder training examples. We tested our method on several benchmark datasets. We propose, and corroborate in our experiments, that the optimal level of sensitivity to hard example is positively correlated with the depth of the network. Moreover, the test prediction error obtained by our method is generally lower than that of the vanilla cross-entropy gradient learner. We therefore conclude that tunable sensitivity can be helpful for neural network learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/08/2020

Reinforcement Based Learning on Classification Task Could Yield Better Generalization and Adversarial Accuracy

Deep Learning has become interestingly popular in computer vision, mostl...
research
07/29/2019

Multi-Frame Cross-Entropy Training for Convolutional Neural Networks in Speech Recognition

We introduce Multi-Frame Cross-Entropy training (MFCE) for convolutional...
research
01/24/2019

Cross-Entropy Loss and Low-Rank Features Have Responsibility for Adversarial Examples

State-of-the-art neural networks are vulnerable to adversarial examples;...
research
12/25/2020

Adaptively Solving the Local-Minimum Problem for Deep Neural Networks

This paper aims to overcome a fundamental problem in the theory and appl...
research
01/03/2023

Effective and Efficient Training for Sequential Recommendation Using Cumulative Cross-Entropy Loss

Increasing research interests focus on sequential recommender systems, a...
research
11/05/2020

Teaching with Commentaries

Effective training of deep neural networks can be challenging, and there...
research
05/27/2022

Standalone Neural ODEs with Sensitivity Analysis

This paper presents the Standalone Neural ODE (sNODE), a continuous-dept...

Please sign up or login with your details

Forgot password? Click here to reset