DeepAI AI Chat
Log In Sign Up

On a Projection Estimator of the Regression Function Derivative

by   Fabienne Comte, et al.

In this paper, we study the estimation of the derivative of a regression function in a standard univariate regression model. The estimators are defined either by derivating nonparametric least-squares estimators of the regression function or by estimating the projection of the derivative. We prove two simple risk bounds allowing to compare our estimators. More elaborate bounds under a stability assumption are then provided. Bases and spaces on which we can illustrate our assumptions and first results are both of compact or non compact type, and we discuss the rates reached by our estimators. They turn out to be optimal in the compact case. Lastly, we propose a model selection procedure and prove the associated risk bound. To consider bases with a non compact support makes the problem difficult.


page 1

page 2

page 3

page 4


Nonparametric Estimation for I.I.D. Paths of Fractional SDE

This paper deals with nonparametric projection estimators of the drift f...

Kernel Selection in Nonparametric Regression

In the regression model Y = b(X) +ε, where X has a density f, this paper...

Robust estimators in a generalized partly linear regression model under monotony constraints

In this paper, we consider the situation in which the observations follo...

Regression in Nonstandard Spaces with Fréchet and Geodesic Approaches

One approach to tackle regression in nonstandard spaces is Fréchet regre...

Variable bandwidth kernel regression estimation

In this paper we propose a variable bandwidth kernel regression estimato...

Distribution-free properties of isotonic regression

It is well known that the isotonic least squares estimator is characteri...