The performance of a learning algorithm is usually measured in terms of prediction error. It is important to choose an appropriate estimator of the prediction error. This paper analyzes the statistical properties of the K-fold cross-validation prediction error estimator. It investigates how to compare two algorithms statistically. It also analyzes the sensitivity to the changes in the training/test set. Our main contribution is to experimentally study the statistical property of repeated cross-validation to stabilize the prediction error estimation, and thus to reduce the variance of the prediction error estimator. Our simulation results provide an empirical evidence to this conclusion. The experimental study has been performed on PAL dataset for age estimation task. © 2012 Springer-Verlag.
CITATION STYLE
Chen, C., Wang, Y., Chang, Y., & Ricanek, K. (2012). Sensitivity analysis with cross-validation for feature selection and manifold learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7367 LNCS, pp. 458–467). https://doi.org/10.1007/978-3-642-31346-2_52
Mendeley helps you to discover research relevant for your work.