Information Technology Reference
In-Depth Information
Hidden
Christo=Penny
Andy=Christi
Marge=Art
Vicky=James
Jenn=Chuck
Colin
Charlot
Rob=Maria
Pierro=Francy
Gina=Emilio
Angela=Tomaso
Lucia=Marco
Alf
Sophia
Figure 6.7:
The family tree structure learned in the
family
trees
task. There are two isomorphic families, one English
and one Italian. The
=
symbol indicates marriage.
Agent_Code
Relation_Code
Patient_Code
6.4.1
Exploration of a Deep Network
Relation
Agent
Patient
Now, let's explore the case of learning in a deep network
using the same family trees task as O'Reilly (1996b)
and Hinton (1986). The structure of the environment is
shown in figure 6.7. The network is trained to produce
the correct name in response to questions like “Rob is
married to whom?” These questions are presented by
activating one of 24 name units in an
agent
input layer
(e.g., “Rob”), in conjunction with one of 12 units in a
relation
input layer (e.g., “Married”), and training the
network to produce the correct unit activation over the
patient
output layer.
Figure 6.8:
The family tree network, with intermediate
Code
hidden layers that re-represent the input/output patterns.
This will bring up a window displaying the first ten
training events, which should help you to understand
how the task is presented to the network.
Go ahead and scroll down the events list, and see
the names of the different events presented to the net-
work (you can click on any that look particularly interest-
ing to see how they are represented).
Now, let's see how this works with the network itself.
,
!
Open
project
family_trees.proj.gz
in
chapter_6
.
First, notice that the network (figure 6.8) has
Agent
and
Relation
input layers, and a
Patient
output
layer all at the bottom of the network. These layers
have
localist
representations of the 24 different peo-
ple and 12 different relationships, which means that
there is no “overt” similarity in these input patterns
between any of the people. Thus, the
Agent_Code
,
Relation_Code
,and
Patient_Code
hidden lay-
ers provide a means for the network to re-represent
these localist representations as richer distributed pat-
terns that should facilitate the learning of the mapping
by emphasizing relevant distinctions and deemphasiz-
ing irrelevant ones. The central
Hidden
layer is re-
sponsible for performing the mapping between these re-
coded representations to produce the correct answers.
,
!
Press the
Step
button in the control panel.
The activations in the network display reflect the mi-
nus phase state for the first training event (selected at
random from the list of all training events).
,
!
Press
Step
again to see the plus phase activations.
The default network is using a combination of
Hebbian and GeneRec error-driven learning, with the
amount of Hebbian learning set to .01 as reflected by
the
lrn.hebb
parameter in the control panel. Let's
see how long it takes this network to learn the task.
,
!
Open up a graph log to monitor training by press-
ing
View
,
TRAIN_GRAPH_LOG
, turn the
Display
of the
network off, and press
Run
to allow the network to train
on all the events rapidly.
As the network trains, the graph log displays the er-
ror count statistic for training (in red) and the average
number of network settling cycles (in orange).
,
!
Press
View
and
select
EVENTS
on
the
family_trees_ctrl
control panel.
,
!
Search WWH ::
Custom Search