Nearest neighbor classification in infinite dimension

69Citations
Citations of this article
18Readers
Mendeley users who have this article in their library.

Abstract

Let X be a random element in a metric space (F, d), and let y be a random variable with value 0 or 1. Y is called the class, or the label, of X. Let (Xi, Yi)1≤i≤n be an observed i.i.d. sample having the same law as (X, Y). The problem of classification is to predict the label of a new random element X. The k-nearest neighbor classifier is the simple following rule: look at the k nearest neighbors of X in the trial sample and choose 0 or 1 for its label according to the majority vote. When (F, d) = (ℝ, ||.||), Stone (1977) proved in 1977 the universal consistency of this classifier: its probability of error converges to the Bayes error, whatever the distribution of (X, Y). We show in this paper that this result is no longer valid in general metric spaces. However, if (F, d) is separable and if some regularity condition is assumed, then the k-nearest neighbor classifier is weakly consistent. © EDP Sciences, SMAI 2006.

Cite

CITATION STYLE

APA

Cérou, F., & Guyader, A. (2006). Nearest neighbor classification in infinite dimension. ESAIM - Probability and Statistics, 10, 340–355. https://doi.org/10.1051/ps:2006014

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free