Using J-K fold Cross Validation to Reduce Variance When Tuning NLP Models

06/19/2018
by   Henry B. Moss, et al.
2

K-fold cross validation (CV) is a popular method for estimating the true performance of machine learning models, allowing model selection and parameter tuning. However, the very process of CV requires random partitioning of the data and so our performance estimates are in fact stochastic, with variability that can be substantial for natural language processing tasks. We demonstrate that these unstable estimates cannot be relied upon for effective parameter tuning. The resulting tuned parameters are highly sensitive to how our data is partitioned, meaning that we often select sub-optimal parameter choices and have serious reproducibility issues. Instead, we propose to use the less variable J-K-fold CV, in which J independent K-fold cross validations are used to assess performance. Our main contributions are extending J-K-fold CV from performance estimation to parameter tuning and investigating how to choose J and K. We argue that variability is more important than bias for effective tuning and so advocate lower choices of K than are typically seen in the NLP literature, instead use the saved computation to increase J. To demonstrate the generality of our recommendations we investigate a wide range of case-studies: sentiment classification (both general and target-specific), part-of-speech tagging and document classification.

READ FULL TEXT
research
05/27/2014

Futility Analysis in the Cross-Validation of Machine Learning Models

Many machine learning models have important structural tuning parameters...
research
12/08/2012

An Empirical Comparison of V-fold Penalisation and Cross Validation for Model Selection in Distribution-Free Regression

Model selection is a crucial issue in machine-learning and a wide variet...
research
03/03/2021

Machine Learning using Stata/Python

We present two related Stata modules, r_ml_stata and c_ml_stata, for fit...
research
10/30/2019

Find what you are looking for: A data-driven covariance matrix estimation

The global minimum-variance portfolio is a typical choice for investors ...
research
02/05/2008

V-fold cross-validation improved: V-fold penalization

We study the efficiency of V-fold cross-validation (VFCV) for model sele...
research
07/31/2019

A Leisurely Look at Versions and Variants of the Cross Validation Estimator

Many versions of cross-validation (CV) exist in the literature; and each...
research
08/24/2019

EPP: interpretable score of model predictive power

The most important part of model selection and hyperparameter tuning is ...

Please sign up or login with your details

Forgot password? Click here to reset