A Framework of Learning Through Empirical Gain Maximization

09/29/2020
by   Yunlong Feng, et al.
0

We develop in this paper a framework of empirical gain maximization (EGM) to address the robust regression problem where heavy-tailed noise or outliers may present in the response variable. The idea of EGM is to approximate the density function of the noise distribution instead of approximating the truth function directly as usual. Unlike the classical maximum likelihood estimation that encourages equal importance of all observations and could be problematic in the presence of abnormal observations, EGM schemes can be interpreted from a minimum distance estimation viewpoint and allow the ignorance of those observations. Furthermore, it is shown that several well-known robust nonconvex regression paradigms, such as Tukey regression and truncated least square regression, can be reformulated into this new framework. We then develop a learning theory for EGM, by means of which a unified analysis can be conducted for these well-established but not fully-understood regression approaches. Resulting from the new framework, a novel interpretation of existing bounded nonconvex loss functions can be concluded. Within this new framework, the two seemingly irrelevant terminologies, the well-known Tukey's biweight loss for robust regression and the triweight kernel for nonparametric smoothing, are closely related. More precisely, it is shown that the Tukey's biweight loss can be derived from the triweight kernel. Similarly, other frequently employed bounded nonconvex loss functions in machine learning such as the truncated square loss, the Geman-McClure loss, and the exponential squared loss can also be reformulated from certain smoothing kernels in statistics. In addition, the new framework enables us to devise new bounded nonconvex loss functions for robust learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/21/2018

Learning with Non-Convex Truncated Losses by SGD

Learning with a convex loss function has been a dominating paradigm for...
research
09/23/2016

Changepoint Detection in the Presence of Outliers

Many traditional methods for identifying changepoints can struggle in th...
research
06/19/2020

New Insights into Learning with Correntropy Based Regression

Stemming from information-theoretic learning, the correntropy criterion ...
research
12/20/2019

Robust Estimation and Variable Selection for the Accelerated Failure Time Model

This paper considers robust modeling of the survival time for cancer pat...
research
04/21/2021

Robust Kernel-based Distribution Regression

Regularization schemes for regression have been widely studied in learni...
research
02/16/2021

From Majorization to Interpolation: Distributionally Robust Learning using Kernel Smoothing

We study the function approximation aspect of distributionally robust op...
research
03/27/2019

Iteratively reweighted least squares for robust regression via SVM and ELM

The measure of most robust machine learning methods is reweighted. To ov...

Please sign up or login with your details

Forgot password? Click here to reset