Next: Error Plots
Up: Neural Network Learning
Previous: Backprop in Action
- train until the E on the training examples falls below some level
- why does overfitting tend to occur in later iterations?
- Initially weights are set to small random numbers, as training
proceeds weights change to reduce error over the training data &
complexity of the decision surface increases, given enough iterations
can overfit
- Weight decay - decrease each weight by a small factor on each
iteration - intuition keep weight values small to bias against complex
decision surfaces - do complex decision surfaces need to have high weights???
- stop when you reach the lowest error on the validation set
- keep current ANN weights and the best-performing weights thus
far measured by error over the validation set
- training is terminated once the current weights reach a
significantly higher error over the validation set
- Care must be taken to avoid stopping too soon!!
- If data is too small can do k-fold cv (remember to use this just
to determine the number of iterations!) then train over whole dataset
(same in decision trees)
Patricia Riddle
Fri May 15 13:00:36 NZST 1998