Back Propagation
Back Propagation
Networks
Introduction to
Backpropagation
- In 1969 a method for learning in multi-layer network,
Backpropagation,
Backpropagation was invented by Bryson and Ho.
1. Computes the error term for the output units using the
observed error.
weights
Error as function of
weights in
multidimensional space
Compute
deltas Gradient
• Trying to make error decrease the fastest
• Compute:
• GradE = [dE/dw1, dE/dw2, . . ., dE/dwn]
• Change i-th weight by
• deltawi = -alpha * dE/dwi
Derivatives of error for weights
• We need a derivative!
• Activation function must be continuous,
continuous
differentiable, non-decreasing, and easy to
compute
Can’t use LTU
• To effectively assign credit / blame to units
in hidden layers, we want to look at the
first derivative of the activation function
Compute deltas
How do we pick α ?
1. Tuning set, or
2. Cross validation, or
In the restaurant
problem NN was
worse than the
decision tree
See next
slide for
explanation
Visualization of
Backpropagation
learning
Calculate difference to
desired output
Compute the
error in output
Update weights
to output layer
Compute error in
each hidden layer
Update weights in
each hidden layer
Return learned network
Examples and
Applications
of ANN
Neural Network in Practice
• Results
– 95% accuracy on the training data
– 78% accuracy on the test set
Other Examples
• Neurogammon (Tesauro & Sejnowski, 1989)
– Backgammon learning program
– 4 hidden units
• Interactive
– activation propagates forward & backwards
– propagation continues until equilibrium is reached in
the network
– We do not discuss these networks here, complex
training. May be unstable.
Ways of learning with an ANN
• Add nodes & connections
• Subtract nodes & connections
• Modify connection weights
– current focus
– can simulate first two
• I/O pairs:
– given the inputs, what should the output be?
[“typical” learning problem]
More Neural Network
Applications
- May provide a model for massive parallel computation.
ex2. Nestor:
- Uses Nestor Learning System (NLS).
- Several multi-layered feed-forward neural networks.
- Intel has made such a chip - NE1000 in VLSI technology.
Ex1. Software tool - Enterprise Miner
•Software exists.
Problems with using ANNs
1. Insufficiently characterized development
process compared with conventional software
– What are the steps to create a neural network?
• Repeatability
– Relevant information must be captured in problem
specification and combinations of parameters
Problem 3 - Quality Assurance
• Specification of generic neural network software
(models and learning)
• Prototype of specification
Eric Wong
Eddy Li
Martin Ho
Kitty Wong