Engression: Extrapolation for Nonlinear Regression?

07/03/2023
by   Xinwei Shen, et al.
0

Extrapolation is crucial in many statistical and machine learning applications, as it is common to encounter test data outside the training support. However, extrapolation is a considerable challenge for nonlinear models. Conventional models typically struggle in this regard: while tree ensembles provide a constant prediction beyond the support, neural network predictions tend to become uncontrollable. This work aims at providing a nonlinear regression methodology whose reliability does not break down immediately at the boundary of the training support. Our primary contribution is a new method called `engression' which, at its core, is a distributional regression technique for pre-additive noise models, where the noise is added to the covariates before applying a nonlinear transformation. Our experimental results indicate that this model is typically suitable for many real data sets. We show that engression can successfully perform extrapolation under some assumptions such as a strictly monotone function class, whereas traditional regression approaches such as least-squares regression and quantile regression fall short under the same assumptions. We establish the advantages of engression over existing approaches in terms of extrapolation, showing that engression consistently provides a meaningful improvement. Our empirical results, from both simulated and real data, validate these findings, highlighting the effectiveness of the engression method. The software implementations of engression are available in both R and Python.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/31/2018

On the Unbiased Asymptotic Normality of Quantile Regression with Fixed Effects

Nonlinear panel data models with fixed individual effects provide an imp...
research
08/26/2022

High-dimensional sparse vine copula regression with application to genomic prediction

High-dimensional data sets are often available in genome-enabled predict...
research
06/11/2021

Neural Networks for Partially Linear Quantile Regression

Deep learning has enjoyed tremendous success in a variety of application...
research
06/27/2023

Wasserstein Generative Regression

In this paper, we propose a new and unified approach for nonparametric r...
research
05/06/2023

Twin support vector quantile regression

We propose a twin support vector quantile regression (TSVQR) to capture ...
research
11/21/2022

First Steps Toward Understanding the Extrapolation of Nonlinear Models to Unseen Domains

Real-world machine learning applications often involve deploying neural ...
research
05/12/2023

Expertise-based Weighting for Regression Models with Noisy Labels

Regression methods assume that accurate labels are available for trainin...

Please sign up or login with your details

Forgot password? Click here to reset