Conditional Mutual Information Constrained Deep Learning for Classification

09/17/2023
by   En-hui Yang, et al.
0

The concepts of conditional mutual information (CMI) and normalized conditional mutual information (NCMI) are introduced to measure the concentration and separation performance of a classification deep neural network (DNN) in the output probability distribution space of the DNN, where CMI and the ratio between CMI and NCMI represent the intra-class concentration and inter-class separation of the DNN, respectively. By using NCMI to evaluate popular DNNs pretrained over ImageNet in the literature, it is shown that their validation accuracies over ImageNet validation data set are more or less inversely proportional to their NCMI values. Based on this observation, the standard deep learning (DL) framework is further modified to minimize the standard cross entropy function subject to an NCMI constraint, yielding CMI constrained deep learning (CMIC-DL). A novel alternating learning algorithm is proposed to solve such a constrained optimization problem. Extensive experiment results show that DNNs trained within CMIC-DL outperform the state-of-the-art models trained within the standard DL and other loss functions in the literature in terms of both accuracy and robustness against adversarial attacks. In addition, visualizing the evolution of learning process through the lens of CMI and NCMI is also advocated.

READ FULL TEXT

page 1

page 6

research
09/21/2022

Mutual Information Learned Classifiers: an Information-theoretic Viewpoint of Training Deep Learning Classification Systems

Deep learning systems have been reported to achieve state-of-the-art per...
research
03/06/2018

Deep Information Networks

We describe a novel classifier with a tree structure, designed using inf...
research
07/03/2023

Normalized mutual information is a biased measure for classification and community detection

Normalized mutual information is widely used as a similarity measure for...
research
03/09/2019

Deep Learning-Based Constellation Optimization for Physical Network Coding in Two-Way Relay Networks

This paper studies a new application of deep learning (DL) for optimizin...
research
03/18/2020

MINT: Deep Network Compression via Mutual Information-based Neuron Trimming

Most approaches to deep neural network compression via pruning either ev...
research
01/16/2021

DeepMI: A Mutual Information Based Framework For Unsupervised Deep Learning of Tasks

In this work, we propose an information theory based framework DeepMI to...

Please sign up or login with your details

Forgot password? Click here to reset