Next: Termination Conditions for Backprop
Up: Neural Network Learning
Previous: Backpropagation Algorithm
- the error
in the delta rule is replaced by
- for output unit
it is the familiar
from
the delta rule multiplied by
which is derivative of
the sigmoid squashing function - for hidden unit
the derivative component is the same but
there is no target value directly available so you sum the error
terms
for each output unit influenced by
weighting each of the
by the weight,
, from the hidden
unit
to the output unit
. - This weight characterizes the degree to which each hidden unit
is responsible for the error in output unit
.
Patricia Riddle
Fri May 15 13:00:36 NZST 1998