Alternate Loss Functions Can Improve the Performance of Artificial Neural Networks

03/17/2023
by   Mathew Mithra Noel, et al.
0

All machine learning algorithms use a loss, cost, utility or reward function to encode the learning objective and oversee the learning process. This function that supervises learning is a frequently unrecognized hyperparameter that determines how incorrect outputs are penalized and can be tuned to improve performance. This paper shows that training speed and final accuracy of neural networks can significantly depend on the loss function used to train neural networks. In particular derivative values can be significantly different with different loss functions leading to significantly different performance after gradient descent based Backpropagation (BP) training. This paper explores the effect on performance of new loss functions that are more liberal or strict compared to the popular Cross-entropy loss in penalizing incorrect outputs. Eight new loss functions are proposed and a comparison of performance with different loss functions is presented. The new loss functions presented in this paper are shown to outperform Cross-entropy loss on computer vision and NLP benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/26/2022

Hybridised Loss Functions for Improved Neural Network Generalisation

Loss functions play an important role in the training of artificial neur...
research
06/04/2021

A novel multi-scale loss function for classification problems in machine learning

We introduce two-scale loss functions for use in various gradient descen...
research
11/28/2015

Loss Functions for Neural Networks for Image Processing

Neural networks are becoming central in several areas of computer vision...
research
06/29/2021

Fast and Accurate Road Crack Detection Based on Adaptive Cost-Sensitive Loss Function

Numerous detection problems in computer vision, including road crack det...
research
01/20/2021

Component Tree Loss Function: Definition and Optimization

In this article, we propose a method to design loss functions based on c...
research
01/31/2018

Optimizing Non-decomposable Measures with Deep Networks

We present a class of algorithms capable of directly training deep neura...
research
01/15/2022

Hyperplane bounds for neural feature mappings

Deep learning methods minimise the empirical risk using loss functions s...

Please sign up or login with your details

Forgot password? Click here to reset