Active Inference-Based Optimization of Discriminative Neural Network Classifiers

06/04/2023
by   Faezeh Fallah, et al.
0

Commonly used objective functions (losses) for a supervised optimization of discriminative neural network classifiers were either distribution-based or metric-based. The distribution-based losses could compromise the generalization or cause classification biases towards the dominant classes of an imbalanced class-sample distribution. The metric-based losses could make the network model independent of any distribution and thus improve its generalization. However, they could still be biased towards the dominant classes and could suffer from discrepancies when a class was absent in both the reference (ground truth) and the predicted labels. In this paper, we proposed a novel optimization process which not only tackled the unbalancedness of the class-sample distribution of the training samples but also provided a mechanism to tackle errors in the reference labels of the training samples. This was achieved by proposing a novel algorithm to find candidate classification labels of the training samples from their prior probabilities and the currently estimated posteriors on the network and a novel objective function for the optimizations. The algorithm was the result of casting the generalized Kelly criterion for optimal betting into a multiclass classification problem. The proposed objective function was the expected free energy of a prospective active inference and could incorporate the candidate labels, the original reference labels, and the priors of the training samples while still being distribution-based. The incorporation of the priors into the optimization not only helped to tackle errors in the reference labels but also allowed to reduce classification biases towards the dominant classes by focusing the attention of the neural network on important but minority foreground classes.

READ FULL TEXT

page 13

page 14

page 15

page 18

page 28

page 32

research
06/19/2022

Gray Learning from Non-IID Data with Out-of-distribution Samples

The quality of the training data annotated by experts cannot be guarante...
research
03/15/2021

Discriminative Learning for Probabilistic Context-Free Grammars based on Generalized H-Criterion

We present a formal framework for the development of a family of discrim...
research
09/17/2020

An Algorithm to Attack Neural Network Encoder-based Out-Of-Distribution Sample Detector

Deep neural network (DNN), especially convolutional neural network, has ...
research
12/13/2021

WOOD: Wasserstein-based Out-of-Distribution Detection

The training and test data for deep-neural-network-based classifiers are...
research
09/29/2022

Regularizing Neural Network Training via Identity-wise Discriminative Feature Suppression

It is well-known that a deep neural network has a strong fitting capabil...
research
08/19/2021

A Unified Objective for Novel Class Discovery

In this paper, we study the problem of Novel Class Discovery (NCD). NCD ...

Please sign up or login with your details

Forgot password? Click here to reset