Back Propagation Back Propagation Network Network Network Network
Back Propagation Back Propagation Network Network Network Network
NETWORK
Back propagation network (BPN)
Network associated with back propagation learning
algorithm (BPLA).
BPLA is one of the most important development in neural
network.
BPLA is applied to multilayer feed forward networks
consisting of processing elements with continuous
differentiable activation function.
BPN is used to classify the input patterns correctly.
Basics of gradient descent method is used in weight update
algorithm.
Back propagation network (BPN)
Error will be propagated back to the hidden unit.
Aims to achieve a balance between the network’s ability to
respond and its ability to give reasonable responses to the
input that is similar but not identical to the training input.
Training stages in BPN network:
Output generation of the network for the input pattern
Updation of weights.
Architecture
BPN is a multi layer feed-forward neural network consisting
of
Input layer
Output layer
Differentiability
Nondecreasing monotony
Training patterns
Incremental approach for updation of weights.
Weights are being changed immediately after a single training pattern
is presented.
Learning rate
Updation rule
Larger value
speed up the convergence but might result in overshooting
Range: 10-3 to 10
Factors affecting the BPN
Momentum
Very efficient and commonly used method that allows a larger
learning rate without oscillations is adding a momentum factor to the
normal weight updation method.
Denoted as
Over-fitting or Over-training:
Network learns well but does not generalize well if there are many
trainable parameters for the given amount of training data is available,
Training data should cover the entire expected input space, and while
training, training – vector pairs should be selected randomly from the
set.
If proper value of T is selected such that T/L >> 1, then the network
can able to discriminate pattern classes using fine piecewise
hyperplane partitioning.
Factors affecting the BPN
Number of hidden layer nodes
If there is more than one hidden layer in a BPN, then calculations
performed for a single layer are repeated for other layers and are
summed up at the end.
Example:
If the network does not converge to a solution, it may need more hidden
nodes.
Also, if the network converges, the user may try a very few hidden nodes
and then settle finally on a size based on overall system performance.
Example
Input pattern [0, 1] . Target output: 1 Learning rate,
Example (Contd..,)
Initial weights:
[ v11 v21 v01] =[0.6 -0.1 0.3]
For z2 layer:
Example (Contd..,)
Applying activation function:
Now,
Example (Contd..,)
Therefore,
Here m = 1 and j = 1 to 2.
Therefore,
Example (Contd..,)
Now,
Example (Contd..,)
Now,
Example (Contd..,)
Calculate change un weights between the input and hidden
layer
Example (Contd..,)
The final weights are calculated as