Digital Signal Processing Reference
In-Depth Information
TABLE 11.1
Performance of the LBG, SOM, and MMSOM
in Quantizing Several Frames of Color Image
Sequence “Trevor White”
a
Recall MSE
Frame No.
LBG
SOM
MMSOM
10
289.65
313.46
240.32
50
283.27
298.84
222.38
100
295.43
312.01
242.52
120
285.27
308.36
226.46
129
282.22
307.14
224.05
a
When the learning procedure was applied on the
first frame in the presence of mixed additive Gaussian
and impulsive noise, Codebook size
=
32.
after the completion of the first layer, clusters the documents of the train-
ing corpus into classes that contain relevant documents with respect to their
context. Accordingly, the WEBSOM architecture is a prominent candidate for
document organization and retrieval. The MMSOM has replaced the SOM in
the aforementioned architecture.
41
To assess the quality of the document or-
ganization (i.e., segmentation) that results by the SOM and the MMSOM, the
average recall-precision curve
59
has been derived for a test (unseen) document
query of known categorization. Figure 11.9 demonstrates that the MMSOM
improves the precision rate for a range of recall rates. The frequencies of bi-
grams can be sorted into descending order and signed ranks can be assigned
to them. A novel metric that incorporates these signed ranks has been pro-
posed for evaluating the context similarity between document pairs. A novel
SOM variant, the
Wilcoxon SOM
,isbuilt based on this metric.
46
Figure 11.10
depicts the average recall-precision curve for the SOM and the Wilcoxon SOM
for the “ACQ” category of the Reuters-21578 corpus.
60
11.7
A Class of Split-Merge Self-Organizing Maps
One common feature of both SOM and LBG algorithms is that they rely on
the assumption that the number of output neurons or the size of codebook
N
is known in advance or is preset to a desired value, usually a power of 2. Al-
though several VQ design techniques, such as the pairwise nearest neighbor
algorithm
39
,
61
or the tree-structured VQ,
62
employ splitting criteria that might
be useful in determining how many output neurons or codevectors ade-
quately approximate the training set provided in the input of VQ, the avail-
ability of splitting criteria is not sufficient on its own. Because a quantization
in terms of
N
1 output neurons always is expected to yield a lower MSE
than a quantization in terms of
N
output neurons,
63
the issue of deciding what
+
Search WWH ::
Custom Search