Unified Regularity Measures for Sample-wise Learning and Generalization

08/09/2021
by   Chi Zhang, et al.
1

Fundamental machine learning theory shows that different samples contribute unequally both in learning and testing processes. Contemporary studies on DNN imply that such sample di?erence is rooted on the distribution of intrinsic pattern information, namely sample regularity. Motivated by the recent discovery on network memorization and generalization, we proposed a pair of sample regularity measures for both processes with a formulation-consistent representation. Specifically, cumulative binary training/generalizing loss (CBTL/CBGL), the cumulative number of correct classi?cations of the training/testing sample within training stage, is proposed to quantize the stability in memorization-generalization process; while forgetting/mal-generalizing events, i.e., the mis-classification of previously learned or generalized sample, are utilized to represent the uncertainty of sample regularity with respect to optimization dynamics. Experiments validated the effectiveness and robustness of the proposed approaches for mini-batch SGD optimization. Further applications on training/testing sample selection show the proposed measures sharing the uni?ed computing procedure could benefit for both tasks.

READ FULL TEXT

page 4

page 5

page 10

page 12

page 15

research
03/03/2022

BatchFormer: Learning to Explore Sample Relationships for Robust Representation Learning

Despite the success of deep neural networks, there are still many challe...
research
04/21/2022

Out-of-distribution generalization for learning quantum dynamics

Generalization bounds are a critical tool to assess the training data re...
research
05/28/2019

Understanding the Behaviour of the Empirical Cross-Entropy Beyond the Training Distribution

Machine learning theory has mostly focused on generalization to samples ...
research
09/23/2020

Residual Embedding Similarity-Based Network Selection for Predicting Brain Network Evolution Trajectory from a Single Observation

While existing predictive frameworks are able to handle Euclidean struct...
research
06/14/2023

Distribution Shift Inversion for Out-of-Distribution Prediction

Machine learning society has witnessed the emergence of a myriad of Out-...
research
10/23/2021

In Search of Probeable Generalization Measures

Understanding the generalization behaviour of deep neural networks is a ...
research
10/30/2019

Assessment of Multiple-Biomarker Classifiers: fundamental principles and a proposed strategy

The multiple-biomarker classifier problem and its assessment are reviewe...

Please sign up or login with your details

Forgot password? Click here to reset