Abstract:
Let (
X,
θ) be a random vector,
X∈
Rd,
θ∈
R~1, (
Xi,
θi) be iid random samples of (
X,
θ),
i=1,…,
n, and
Ln be the conditional risk in the nearest neighbor (
NN) predic- tion under square loss. The estimate of
Ln is defined as \hatL_n=\frac1n \sum_j=1^n\left(\theta_j-\theta_n j\right)^2, where
θnj denotes the
NN prediction of
θj, based on the training sample (
X1,
θ1), …, (
Xj-1,
θj-1), (
Xj+1,
θj+1), …, (
Xn,
θn) and the observed
Xj. Ammusing only the second moment of
θ is finite and some other conditions, oonvergenee in series form and in mean is discussed.