How Entropic Regression Beats the Outliers Problem in Nonlinear System Identification

05/16/2019
by   Abd AlRahman R. AlMomani, et al.
0

System identification (SID) is central in science and engineering applications whereby a general model form is assumed, but active terms and parameters must be inferred from observations. Most methods for SID rely on optimizing some metric-based cost function that describes how a model fits observational data. A commonly used cost function employs a Euclidean metric and leads to a least squares estimate, while recently it becomes popular to also account for model sparsity such as in compressed sensing and Lasso. While the effectiveness of these methods has been demonstrated in previous studies including in cases where outliers exist, it remains unclear whether SID can be accomplished under more realistic scenarios where each observation is subject to non-negligible noise and sometimes contaminated by large noise and outliers. We show that existing sparsity-focused methods such as compressive sensing, when used applied in such scenarios, can result in "over sparse" solutions that are brittle to outliers. In fact, metric-based methods are prone to outliers because outliers by nature have an unproportionally large influence. To mitigate such issues of large noise and outliers, we develop an Entropic Regression approach for nonlinear SID, whereby true model structures are identified based on relevance in reducing information flow uncertainty, not necessarily (just) sparsity. The use of information-theoretic measures as opposed to a metric-based cost function has a unique advantage, due to the asymptotic equipartition property of probability distributions, that outliers and other low-occurrence events are conveniently and intrinsically de-emphasized.

READ FULL TEXT
research
06/25/2023

Evolution of K-means solution landscapes with the addition of dataset outliers and a robust clustering comparison measure for their analysis

The K-means algorithm remains one of the most widely-used clustering met...
research
10/10/2016

Robust Bayesian Compressed sensing

We consider the problem of robust compressed sensing whose objective is ...
research
03/25/2012

Greedy Sparsity-Constrained Optimization

Sparsity-constrained optimization has wide applicability in machine lear...
research
01/12/2015

Combined modeling of sparse and dense noise for improvement of Relevance Vector Machine

Using a Bayesian approach, we consider the problem of recovering sparse ...
research
01/29/2019

A Robust Time Series Model with Outliers and Missing Entries

This paper studies the problem of robustly learning the correlation func...
research
04/04/2011

Robust Nonparametric Regression via Sparsity Control with Application to Load Curve Data Cleansing

Nonparametric methods are widely applicable to statistical inference pro...
research
02/24/2017

Characterizing Classes of Potential Outliers through Traffic Data Set Data Signature 2D nMDS Projection

This paper presents a formal method for characterizing the potential out...

Please sign up or login with your details

Forgot password? Click here to reset