Reinforcement Based Learning on Classification Task Could Yield Better Generalization and Adversarial Accuracy

12/08/2020
by   Shashi Kant Gupta, et al.
13

Deep Learning has become interestingly popular in computer vision, mostly attaining near or above human-level performance in various vision tasks. But recent work has also demonstrated that these deep neural networks are very vulnerable to adversarial examples (adversarial examples - inputs to a model which are naturally similar to original data but fools the model in classifying it into a wrong class). Humans are very robust against such perturbations; one possible reason could be that humans do not learn to classify based on an error between "target label" and "predicted label" but possibly due to reinforcements that they receive on their predictions. In this work, we proposed a novel method to train deep learning models on an image classification task. We used a reward-based optimization function, similar to the vanilla policy gradient method used in reinforcement learning, to train our model instead of conventional cross-entropy loss. An empirical evaluation on the cifar10 dataset showed that our method learns a more robust classifier than the same model architecture trained using cross-entropy loss function (on adversarial training). At the same time, our method shows a better generalization with the difference in test accuracy and train accuracy < 2% for most of the time compared to the cross-entropy one, whose difference most of the time remains > 2%.

READ FULL TEXT

page 2

page 8

page 9

page 10

research
01/24/2019

Cross-Entropy Loss and Low-Rank Features Have Responsibility for Adversarial Examples

State-of-the-art neural networks are vulnerable to adversarial examples;...
research
11/23/2016

Tunable Sensitivity to Large Errors in Neural Network Training

When humans learn a new concept, they might ignore examples that they ca...
research
12/01/2022

Noisy Label Classification using Label Noise Selection with Test-Time Augmentation Cross-Entropy and NoiseMix Learning

As the size of the dataset used in deep learning tasks increases, the no...
research
07/28/2021

United We Learn Better: Harvesting Learning Improvements From Class Hierarchies Across Tasks

Attempts of learning from hierarchical taxonomies in computer vision hav...
research
06/12/2020

Evaluation of Neural Architectures Trained with Square Loss vs Cross-Entropy in Classification Tasks

Modern neural architectures for classification tasks are trained using t...
research
05/01/2019

Introducing Graph Smoothness Loss for Training Deep Learning Architectures

We introduce a novel loss function for training deep learning architectu...
research
05/24/2023

Fantastic DNN Classifiers and How to Identify them without Data

Current algorithms and architecture can create excellent DNN classifier ...

Please sign up or login with your details

Forgot password? Click here to reset