Professional Documents
Culture Documents
SESSION 3
Topics that will be covered in this Session
• Learning Algorithms
• The Basic Principle of ANN Learning
• Supervised Learning
• Hebb Rule
• Perceptron Learning Rule
• Delta Rule
• Extended Delta Rule
Learning Algorithms
• ANN is characterized by three entities:
• Its architecture
• Activation function
• Learning technique
• Learning refers to the process of finding the appropriate set of weights of the
interconnections so that the ANN attains the ability to perform the designated task.
• This process is called Training the ANN
• How to find the appropriate set of weights so that the ANN is able to solve a given
problem?
• To start with a set of weights and then gradually modify them to arrive at the final
weights
The Basic Principle of ANN Learning
• The ANN starts with an initial distribution of interconnection weights and then
goes on adjusting the weights iteratively until some predefined stopping
criterion is satisfied
• The weight of a certain interconnection path at the iteration is
• The weight at iteration, is obtained by
Input Arguments
Big Small Red Green
SIZE COLOR
BIG RED
SMALL RED Size & Color
BIG GREEN
SMALL GREEN
Small Small
Big & Big & &
Red &
Green Green
• Clustering Red
Supervised Learning
Supervised Learning
• A neural network is trained with the help of a set of patterns known as the training
vectors
• The output of these vectors might be, or might not be, known beforehand
• When these are known and that knowledge is employed in the training process, the
training is termed as supervised learning
• Otherwise, the learning is said to be unsupervised
• Some popular supervised learning methods are
• Perceptron Learning
• Delta Learning
• Least-Mean-Square (LMS) Learning
• Correlation Learning
• Outstar Learning
Linearly Separable data
1. Hebb Rule
• It is one of the earliest learning rules for ANNs
• Weight adjustment is computed as
where and are the vectors corresponding to interconnection weights and inputs
and
3. Delta / LMS (Least Mean Square) / Widrow-Hoff Rule
• In LMS learning, the identify function is used as the activation function during the
training phase
• The learning rule minimizes mean square error between the activation and the
target value
• The output of LMS is in binary form
4. Extended Delta Rule
• The Extended Delta Rule removes the restriction of the output activation function
being the identity function only
• Any differentiable function can be used for this purpose
• Here the weight adjustment is computed as
where g(.) is the output activation function and g’(.) is its first derivative