Information Technology Reference
In-Depth Information
- Initialize latent variables X =( x 1 ,... x N ) (often employing spectral methods like
LLE), and
- minimize E ( X ) w.r.t. DSRE employing an optimization scheme.
A typical example is unsupervised kernel regression, analyzed by Klanke and Rit-
ter [16], but further methods can also be employed. Klanke and Ritter [16] introduced
an optimization scheme based on LLE, PCA, and leave-one-out cross-validation (LOO-
CV) for UKR. Various extension of UKR have been proposed, e.g., a feature space vari-
ant (employing Mercer kernels), and the use of landmark points to reduce the
non-parametric training set. Carreira-Perpinan and Lu [5] argue that training of non-
parametric unsupervised regression approaches is quite expensive, i.e.,
( N 3 ) in
O
( N 2 ) in memory. Parametric methods can accelerate learning, e.g., un-
supervised regression based on radial basis function networks (RBFs) [26], Gaussian
processes [19], and neural networks [27].
time, and
O
2.2
Nearest Neighbor Regression
In the following, we give a short introduction to K-nearest neighbor regression that
is basis of the UNN approach. KNN is a technique with a long tradition. It was first
mentioned by Fix and Hodges [8] in the fifties in an unpublished US Air Force School of
Aviation Medicine report as non-parametric classification technique. Cover and Hart [3]
investigated the approach experimentally in the sixties. Interesting properties have been
found, e.g., that for K =1 ,and N
, KNN is bound by the Bayes error rate.
The problem in regression is to predict output values y IR
→∞
d of given input values
q based on sets of N input-output examples (( x 1 , y 1 ) ,..., ( x N , y N )) . The goal
is to learn a function f : x y known as regression function. For a novel pattern x
KNN regression computes the mean of the function values of its K-nearest neighbors:
x IR
f KNN ( x ):= 1
K
y i
(2)
i∈N K ( x )
N K ( x ) containing the indices of the K -nearest neighbors of x .Theidea
of KNN is based on the assumption of locality in data space: In local neighborhoods
of a pattern x near patterns are expected to have similar labels y . Consequently, for
an unknown x the label must be similar to the labels of the closest patterns, which
is modeled by the average of the output value of the K nearest samples. KNN has
been proven well in various applications, e.g., in the detection of quasars based on
spectroscopic data [10].
with set
3
Unsupervised Nearest Neighbor Regression
In this section we introduce an iterative strategy for UNN regression [17,18] that is
based on minimization of the data space reconstruction error (DSRE).
Search WWH ::




Custom Search