Provable More Data Hurt in High Dimensional Least Squares Estimator

08/14/2020 ∙ by Zeng Li, et al. ∙ 5

This paper investigates the finite-sample prediction risk of the high-dimensional least squares estimator. We derive the central limit theorem for the prediction risk when both the sample size and the number of features tend to infinity. Furthermore, the finite-sample distribution and the confidence interval of the prediction risk are provided. Our theoretical results demonstrate the sample-wise nonmonotonicity of the prediction risk and confirm "more data hurt" phenomenon.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 2

page 19

page 20

page 21

page 22

page 23

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.