Professional Documents
Culture Documents
EIE520 Module1
EIE520 Module1
.
Module 1
Introduction to Artificial Neural Network
y0
y1
y2
...
yn
Co eig
on
nn ht
we
w
ec s
igh
ts
tio
n
Input Signals
Output Signals
Output Signals
Input Signals
1 Soma Neuron
2 Dendrite Input
3 Axon Output
4 Synapse Weight
1.3.2 Differences Between the Biological and
Artificial Neural Network
1. Parameters: Human brains have about 10million
times synapses than ANN, ResNET, which is one
of the deepest ANN architecture has 60 million
synapses.
2. Topology: Human brains have no layers whereas
ANN could have several layer.
3. Operation Mode: The human brain works
asynchronously while ANN works
synchronously.
4. Learning Algorithm: ANN uses gradient descent
learning(and other learning algorithms) while the
how the human brain learn is not yet known.
5. Power consumption: Human brain uses very
little power compared to ANN.
6. Learning Phases: Human brain never stops
learning as long as it is living, while ANN first trains
then tests.
1.4 History of ANN
The history of ANN began nearly simultaneously
with that of programmable electronic computers
and it is divided into five timelines.
i) The beginning of ANN(1940s).
ii) The first golden age (1950s and 1960s).
iii) The quiet years (1970s).
iv) The renaissance or renewed enthusiasm era
(1980s and 1990s).
v) The modern era (2000s – Till Date).
?? ??
Figure 1.5: Some Founding Fathers of ANN (From left to right and top to bottom: John
von Neumann, Donald O. Hebb, Marvin Minsky, Bernard Widrow, Seymour Papert,
Teuvo Kohonen, John Hopfield, Yann LeCun, Geoffrey Hinton, Yoshua Bengio).
1.4.1 The beginning of ANN (1940s)
(a) McCulloch-Pitts Neurons:
The first artificial neuron was designed by Warren
McCulloch and Walter Pitts [McCulloch & Pitts, 1943].
The neurons are used most widely as logic circuits. Their
subsequent work [Pitts & McCulloch, 1943] addressed the
translational and rotational invariant pattern recognition.
(b)Hebb Learning:
Donald Hebb, a psychologist at McGill University
developed the first learning law for ANN in 1949.
Rochester,Holland, Haibt & Duda (1956) refined Hebb’s
work to allow computer simulations for ANN. McClelland
and Rumelhart further expanded the work of Hebb in
1988.
1.4.2 The first golden age (1950s and 1960s)
ANNs are often seen today as an alternative to (or
complement of) traditional computing, it should be noted
that John von Neumman was keenly interested in
modeling the brain[von Neumman, 1958].
(a) Perceptrons:
Frank Rosenblatt(1958,1959,1960) with several other
researchers[Block, 1962, Minsky & Papert 1969]
introduced and developed a large class of ANN called
Perceptrons.
The Perceptron learning rule uses an iterative weight
adjustment that is more powerful than the Hebb rule.
However, the limitations of what the Perceptron can learn
was demonstrated by Minsky and Papert (1969).
(b) Adaline:
Bernard Widrow and his student Marcian Hoff developed a
learning rule tagged Least Mean Squares(LMS) or delta rule that
is closely related to the perceptron learning rule[Widrow & Hoff,
1960].
The similarity of models developed in psychology by Rosenblatt
to those developed in electrical engineering by Widrow and Hoff
is an evidence of the interdisciplinary nature of ANN.
The Widrow-Hoff learning rule for a single layer network is the
precursor of the backpropagation rule for multilayer networks.
ADALINE means Adaptive Linear Neurons. It has been applied
for adaptive antenna system[Widrow, Mantey, Griffiths & Goode,
1967], rotation invariant pattern recognition, and a variety of
control systems such as broom balancing and backing up a truck.
MADALINES are multilayer extensions of ADALINES[Widrow
& Hoff, 1960; Widrow and Lehr,1990]
1.4.3 The quiet years (1970s)
(a) Kohonen:
Teuvo Kohonen’s early work at Helsinki University of
Technology dealt with associative memory neural
networks[Kohonen,1972].
He later worked on self-organising feature maps that use a
topological structure for the cluster units.
These networks have been applied to speech
recognition[Kohonen, Torkkola, Shozakai, Kangas &
Venta 1987; Kohonen, 1988],the “Traveling Salesman
Problem”[Angeniol, Vaubiis & Le Texier, 1988] and
composition of music [Kohonen, 1989b].
(b) Anderson:
James Anderson started his research in ANN with
associative memory nets at Brown University[Anderson
1968, 1972]. He developed it into the “Brain-in-a-
Box”[Anderson, Silverstein, Ritz, & Jones, 1977].
Application of this include medical diagnosis and learning
of multiplication table.
(c) Grossberg:
The works of Stephen Grossberg at the center for Adaptive
Systems at Boston University (although very mathematical
and biological) are widely known in ANN
community[Grossberg, 1976,1980, 1982, 1987, 1988].
(d) Carpenter:
Stephen Grossberg and Gail Carpenter developed a theory
of self-organizing neural networks called Adaptive
Resonance Theory for binary input patterns (ART1) and
for continuously valued inputs (ART2).
(a) Backpropagation:
The failure of single-layer Perceptron to be able to solve
some simple problems such as the XOR function and the
lack of a general method for training a multilayer networks
led to the “quiet years” of the 1970s.
Werbos (1974) developed a method for propagating errors
at the output units back to the hidden units but it did not
gain wide publicity.
(c) Neocognitron:
Kunshiko Fukushima developed an earlier self-organizing
network called the cognitron for character
recognition[Fukushima, 1975]. Cognitron failed to recognize
positions or rotation distorted characters. Neocognitron was
created to correct this defficiency[Fukushima, 1988; Fukushima,
Mijake & Ito, 1983].
(d) Boltzman machine:
Some researchers developed nondeterministic ANN in
which the weights or activations are changed on the basis of
a probability density function [Kirkpatrick, Gelatt &
Vecchi,1983].