Toward a Better Understanding of Leaderboard

10/12/2015
by   Wenjie Zheng, et al.
0

The leaderboard in machine learning competitions is a tool to show the performance of various participants and to compare them. However, the leaderboard quickly becomes no longer accurate, due to hack or overfitting. This article gives two pieces of advice to prevent easy hack or overfitting. By following these advice, we reach the conclusion that something like the Ladder leaderboard introduced in [blum2015ladder] is inevitable. With this understanding, we naturally simplify Ladder by eliminating its redundant computation and explain how to choose the parameter and interpret it. We also prove that the sample complexity is cubic to the desired precision of the leaderboard.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/07/2022

Machine Learning Students Overfit to Overfitting

Overfitting and generalization is an important concept in Machine Learni...
research
02/20/2018

Sample Complexity of Stochastic Variance-Reduced Cubic Regularization for Nonconvex Optimization

The popular cubic regularization (CR) method converges with first- and s...
research
11/29/2018

Sample Efficient Stochastic Variance-Reduced Cubic Regularization Method

We propose a sample efficient stochastic variance-reduced cubic regulari...
research
04/08/2019

Feature Learning Viewpoint of AdaBoost and a New Algorithm

The AdaBoost algorithm has the superiority of resisting overfitting. Und...
research
04/30/2015

Model Selection and Overfitting in Genetic Programming: Empirical Study [Extended Version]

Genetic Programming has been very successful in solving a large area of ...
research
05/09/2023

Testing for Overfitting

High complexity models are notorious in machine learning for overfitting...

Please sign up or login with your details

Forgot password? Click here to reset