Information Technology Reference
In-Depth Information
Fig. 3.3. Percentage of explained variance
1995] for analyzing and reducing the dimensions of nonlinear distributions. It
may be viewed as a nonlinear extension of principal component analysis. CCA
uses a more local criterion than PCA, which allows it to keep the local topol-
ogy of the distribution of input points. An analysis of this method, together
with examples of applications may be found in [Herault 1993; Vigneron 1997].
Figure 3.4 shows CCA applied to dimension reduction of nonlinear data
structures: on the left, the left-hand part shows a set of points defined in
R
2 . The dimension
reduction may therefore be seen as a “nonlinear” projection that retains the
proximity of points and therefore the local topology of the distribution.
In closed structures, such as a sphere or a cylinder, dimensionality re-
duction will inevitably result in some local distortion, as shown in Fig. 3.5,
which shows an example of the projection of a sphere on the plane. The main
principle of CCA is the gradual control of local distortion, during training.
Since the main goal of CCA is a dimensionality reduction that preserves the
local topology, it is ideally suited for the representation of nonlinear varieties.
A variety in
3 , and the right-hand part shows a representation in
R
R p may be defined roughly as a set of points, the local dimensions
of which are smaller than p . The envelope of a sphere defined in
3 is an
example: the dimension of the variety is 2. More strictly, a variety of dimension
q in R q is a sub-set of R n obtained by applying a function defined by R q in R q .
R
Search WWH ::




Custom Search