Information Technology Reference
In-Depth Information
SSE-Vs-Epoch with
Validation data set
Early Stopping
SSE-Vs-Epoch
with training data set
0
Epoch
Figure 3.16.
Early stopping of training
But still, the dilemma remains: in order to stop the training process, how do we
realize that the network has learnt all the required knowledge from the training data
and has reached its maximum generalization? Then, from learning theory we know
that after reaching the point of maximum generalization, the network - although
learning more and more from the training set - will start impairing the related test
set performance (Figure 3.16) due to its overtraining (Vapnik, 1995). To prevent
this, the method of
early stopping with cross-validation
has been suggested by
Prechelt (1998).
Cross-validation
is a traditional statistical procedure for random partitioning of
collected data into a
training set
and a
test set
, and for further partitioning of the
training set into the
estimation set
and the
validation set
. It is obvious that, if only
a restricted data set is available, the partition of the entire set reduces the size of the
training set. This, again, makes the location of the early stopping point difficult.
For managing this problem, a predicate or a
stopping criterion
should be found
that can indicate when to stop the training.
Prechelt (1998), using the error function (or the objective function)
E
, training
error
E
tr
(as the average error per example across the training set), and the test and
validation errors
E
t
and
E
v
respectively, has defined three possible stopping
criteria:
x Stop as soon as the generalization loss exceeds a threshold value
İ
,
i.e
.
when
g
( )
t
!
İ
, where the error function
g
( )
t
is based on the lowest
loss
loss
validation set error
E
and the validation error
E
.
opt
x Stop as soon as the quotient
g t
Pt
()
()
loss
!
İ
,
tr
P
tr
(
t
)
where
is the
training progress
defined by
Search WWH ::
Custom Search