Information Technology Reference
In-Depth Information
Event
attenti
binding
dyslexi
network to transcend rote memorization of the text it-
self, and produce representations that will be effective
for processing novel text items.
attention
1
0.415
0.090
binding
0.415
1
0.118
dyslexia
0.090
0.118
1
View the GetWordRF representations for several
other units to get a sense of the general patterns across
units.
Although there clearly is sensible semantic structure
at a local level within the individual-unit representa-
tions, it should also be clear that there is no single, co-
herent theme relating all of the words represented by
a given unit. Thus, individual units participate in rep-
resenting many different clusters of semantic structure,
and it is only in the aggregate patterns of activity across
many units that more coherent representations emerge.
This network thus provides an excellent example of dis-
tributed representation.
, !
Table 10.12: Cosine matrix of the hidden representations for
the words “attention,”“binding,” and “dyslexia”.
By turning selwts off, we ensure that the previ-
ously selected units from “attention” will remain so.
Now, you should be able to see that in several cases, the
strong weights for “binding” overlap with those of the
selected units from “attention.” Thus, these overlapping
units represent both of these concepts, as one might ex-
pect given that they have a clear semantic relationship.
Now let's see what a more unrelated word looks like.
Do SelWord again, and enter “dyslexia” as the
word, and keep the other buttons as they are.
You should now observe considerably less overlap
among the strong weights from “dyslexia” and the se-
lected units from “attention,” which is appropriate given
that these words are not as closely related.
The SelWord process can be performed automati-
cally for two different words by using the PairProbe
button. The first word's weights will be selected, and
the second word's weights will be viewed as color val-
ues.
Distributed Representations via Sending Weights
A more appropriate technique for exploring the nature
of the distributed semantic representations is to look at
the sending weights from a given word to the hidden
layer. This will show the pattern of hidden units that
represent that word, and by doing this for multiple dif-
ferent words, we can get a sense of how many hidden
units the words have in common. In other words, we
can look at distributed pattern overlap as a measure of
the similarity structure (as discussed in chapter 3).
Do some PairProbe s for other word pairs that you
expect to be semantically related in this textbook, and
pairs that you expect to be unrelated.
Press SelWord on the sem_ctrl control panel.
Enter “attention” as the word, and click both the view
and selwts buttons on.
This will allow us to view the weights for the input
unit corresponding to the word “attention,” and it will
select all of the hidden units which have a weight value
greater than the rf_thresh of .5.
, !
Summarizing Similarity with Cosines
Instead of just eyeballing the pattern overlap, we can
compute a numerical measure of similarity using nor-
malized inner products or cosines between pairs of
sending weight patterns (see equation 10.1 from the
dyslexia model in section 10.3). The WordMatrix
button does this computation automatically for a list of
words (separated by spaces), producing a matrix of all
pairwise cosines, and a cluster plot of this distance ma-
trix.
Youhavetoselect s.wt in the network to actually
view the sending weights.
You should see a pattern of hidden units that have
strong weights from this word unit, with the strongest
units selected (indicated by the white dashed line sur-
rounding the unit). By itself, this pattern is not very
meaningful. However, we can now compare the pattern
with that of another, related word.
, !
Do WordMatrix , and enter “attention binding
dyslexia” as the words.
Do SelWord again, but enter in “binding” as the
word this time, and click the selwts button off.
, !
, !
Search WWH ::




Custom Search