Professional Documents
Culture Documents
31 34 Artificial Neural Network
31 34 Artificial Neural Network
Guided By Submitted By
WADHWAN CITY
DIST : SURENDRANAGAR
CERTIFICATE
ACKNOWLEDGEMENT
INDEX
Content Page
No.
1. Introduction 1
2. Definition 2
3. Structure Of Human Brain 3
4. Neurons 4
5. Basics Of ANN Model 6
6. Artificial Neural Networks 7
6.1 How ANN Differs From Conventional Comp.
6.2 ANN Vs Von Neumann Comp.
7. Perceptron 10
8. Learning Laws 11
8.1 Hebb’s Rule
8.2 Hopfield Rule
8.3 The Delta Rule
8.4 The Gradient Descent Rule
8.5 Kohonen’s Learning Rule
9. Basic Structure Of ANNS 14
10. Network Architectures 15
10.1Single Layer Feed Forward ANN
10.2Multilayer Feed forward ANN
10.3Recurrent ANN
11. Learning Of ANNS 18
11.1Learning With A Teacher
11.2Learning Without A Teacher
11.3Learning Tasks
12. Control 23
13. Adaptation 24
14. Generalization 24
15. Probabilistic ANN 25
16. Advantages 26
17. Limitations 27
18. Applications 28
19. References 32
Introduction
Ever since eternity, one thing that has made human beings stand
apart from the rest of the animal kingdom is, its brain .The most intelligent
device on earth, the “Human brain” is the driving force that has given us
the ever-progressive species diving into technology and development as
each day progresses.
Due to his inquisitive nature, man tried to make machines that could
do intelligent job processing, and take decisions according to instructions
fed to it. What resulted was the machine that revolutionized the whole
world, the “Computer” (more technically speaking the Von Neumann
Computer). Even though it could perform millions of calculations every
second, display incredible graphics and 3-dimentional animations, play
audio and video but it made the same mistake every time.
Practice could not make it perfect. So the quest for making more
intelligent device continued. These researches lead to birth of more
powerful processors with high-tech equipments attached to it, super
computers with capabilities to handle more than one task at a time and
finally networks with resources sharing facilities. But still the problem of
designing machines with intelligent self-learning, loomed large in front of
mankind. Then the idea of initiating human brain stuck the designers who
started their researches one of the technologies that will change the way
computer work Artificial Neural Networks.
Definition
Sequential Parallel
Functions: Functions:
Rules Images
Concepts Pictures
Calculations Control
Functions of Brain
Neurons
Since the components and connections and their packaging under stringent
spatial constraints make the system large-scale, the role of graph theory,
algorithm, and neuroscience is pervasive.
Perceptron
Input signals are applied to the node via input connection (dendrites
in the case of the brain.) The connections have “strength” which change as
the system learns. In neural networks the strength of the connections are
referred to as weights. Weights can either excite or inhibite the
transmission of the incoming signal. Mathematically incoming signals
values are multiplied by the value of those particular weights.
At the perceptron all weighted input are summed. This sum value is
than passed to a scaling function. The selection of scaling function is part
of the neural network design.
Perceptron
Learning Laws
Many learning laws are in common use. Most of these are some sort
of variation of the best known and oldest learning laws, hebb’s rule.
Research into different learning functions continues as new ideas routine
show up in trade publication. Some researches have the modeling of
biological learning as their main objective. Others are experimenting with
adaptation of their perceptions of how nature handles learning. Either way,
man’s understanding of how neural processing actually works is very
limited. Learning is certainly more complex rhan the simplification
represented by the learning laws currently develop. A few of the major
laws are presented as examples.
Hebb’s Rule
The first, and undoubtedly the best known, learning rule were introduced
by Donald Hebb. The description appeared in his book the Organization of
behavior in 1949. His basic rule is: if a neuron receives an input from
another neuron, and if both are highly active (mathematically have the
same sign), the weight between the neurons should be strengthened.
Hopfield Law
It is similar to Hebb’s rule with the exception that it specifies the
magnitude of the strengthening or weakening. It states,” if the desired
output and the input are both active and both inactive, increment the
connection weight by the learning rate, otherwise decrement the weight by
the learning rate”.
processing element. Their rule changes the synaptic weights in the way that
minimizes the mean squared error of the network.
This rule is also referred to as windrows-Hoff Learning rule and the least
mean square (LMS) Learning Rule.
The way that the Delta Rule works is that the delta rule error in the
output layer is transformed by the derivative of the transfer function and is
then used in the previous neural layer to adjust input connection weights. In
other words, the back-propagated into previous layers one layer at a time.
The process of back-propagating the network errors continues until the first
layer is reached. The network typed called feed forward; back-propagation
derives its name from this method of computing the error term. When using
the delta rule, it is important to ensure that the input data set is well
randomized. Well-ordered or structured presentation of the training set can
lend to a network, which cannot converge to the desired accuracy. If that
happens, then network is incapable of learning the problem.
It has been shown that different learning rates for different layers of
network help the learning process converge faster. In these tests, the
learning rates for those layers close to the output were set lower than those
layers near the input. This especially important for applications where the
input data is not derived from a strong underlying model.
Hidden layer:
The in-between input and output layer the layers are knownas hidden layers
where the knowledge of past experience / training is the
Output Layer:
The topmost layer which give the final output. In this case z1 and z2 are
output neurons.
Network architectures
2). Multilayer feed forward networks: The second class of the feed
forward neural network distinguishes itself by one or more hidden
layers, whose computation nodes are correspondingly called neurons
Learning of ANNS
The property that is of primary significance for a neural network is the
ability of the network to learn from environment, and to improve its
performance through learning.
the learning process in which the teacher teaches the network by giving the
network the knowledge of environment in the form of sets of the inputs-
outputs pre-calculated examples.
As shown in figure
As shown in figure.
As indicated in figure.
Learning tasks
Pattern recognition:
Humans are good at pattern recognition. We can recognize the familiar face
of the person even though that person has aged since last encounter,
identifying a familiar person by his voice on telephone, or by smelling the
fragments comes to know the food etc.
The network is able to identify the class of that particular pattern because
of the information it has extracted from the training data. Pattern
recognition performed by neural network is statistical in nature, with the
pattern being represented by points in a multidimensional decision space.
The decision space is divided into regions, each one of which is associated
with class. The decision boundries are determined by the training process.
Control
The control of a plant is another learning task that can be done by a neural
network; by a ‘plant’ we mean a process or critical part of a system that is
to be maintained in a controlled condition. The relevance of learning to
control should not be surprising because, after all, the human brain is a
computer, the output of which as a whole system are actions. In the context
of control, the brain is living proof that it is possible to build a generalized
controller that takes full advantages of parallel distributed hardware, can
control many thousands of processes as done by the brain to control the
thousands of muscles.
Adaptation
The environment of the interest is no stationary, which means that the
statistical parameters of the information bearing generated by the
environment vary with the time. In situation of the kind, the traditional
methods of supervised may learning may prove to be inadequate because
the network is not equipped with the necessary means to track the statistical
variation of the environment in which it operates. To overcome these
shortcomings, it is desirable for a neural network to continually adapt its
free parameters to variation in the incoming signals in a real time fashion.
Thus an adaptive system responds to every distinct input as a novel one. In
other words the learning process encountered in the adaptive system never
stops, with learning going on while signal processing is being performed by
the system. This form of learning is called continuous learning or learning
on the fly.
Generalization
In back propagation learning we typically starts with a training sample and
uses the back propagation algorithm to compute the synaptic weights of a
multiplayer preceptor by loading (encoding) as many as of the training
example as possible into the network. The hope is that the neural network
so design will generalize. A network is said generalize well when the input
output mapping computed by the network is correct or nearly so for the test
data never used in creating or training the network; the term generalization
is borrowed from psychology.
data but not true for the underlining function that is to be modeled. Such a
phenomena is referred to as an over fitting or over training. When the
network is over trained it looses the ability to generalize between similar
input output pattern.
Gershenfield and weighed defined three goals for time series analysis:
forecasting, modeling, and characterization. Forecasting is predicting the
short-term evolution of the system. Modeling involves finding a description
that accurately captures the features of the long-term behavior. The goal of
characterization is to determined the fundamental properties of the system,
such as the degrees of freedom or the amount of randomness. The
traditional methods used for time series prediction are the moving average
(ma), autoregressive (ar), or the combination of the two, the ARMA model.
2) Speech Generation
One of the earliest successful applications of the back propagation
algorithm for training multiplayer feed forward networks were in a speech
generation system called NET talk, developed by Sejnowski and
Rosenberg. Net talk is a fully connected layered feed forward network with
only one hidden layer. It was trained to pronounce written English text.
Turning a written English text into speech is a difficult task, because most
phonological rules have exceptions that are context-sensitive.
Net talk is a simplest network that learns the function in several hours
using exemplars.
3) Speech Recognition
Kohonen used his self-organizing map for inverse problem to that
addressed by Net talk: speech recognition. He developed a phonetic
typewriter for the Finnish language. The phonetic typewriter takes as input
a speech as input speech and converts it into written text. Speech
recognition in general is a much harder problem that turning text into
speech. Current state-of-the-art English speech recognition systems are
based on hidden Markov Model (HMM). The HMM, which is a Markov
process; consist of a number of states, the transitions between which
depend on the occurrence of some symbol.
drove the CMU Navlab, equipped with video cameras and laser range
sensors, for 21.2 miles with an average speed of 55 mph on a relatively old
highway open to normal traffic. ALVINN was not distributed by passing
cars while it was driven autonomously. ALVINN nearly doubled the
previous distance world record for autonomous navigation.
5) Handwriting Recognition
Members of a group at AT&T Bell Laboratories have been working in
the area of neural networks for many years. One of their projects involves
the development of a neural network recognizer for handwritten digits. A
feed forward layered network with three hidden layers is used. One of the
key features in this network that reduces the number of free parameters to
enhance the probability of valid generalization by the network. Artificial
neural network is also applied for image processing.
6) In Robotics Field:
With the help of neural networks and artificial Intelligence. Intelligent
devices, which behave like human, are designed. Which are helpful to
human in performing various tasks.
Business
o Marketing
o Real Estate
Document and Form Processing
o Machine Printed Character Recognition
o Graphics Recognition
o Hand printed Character Recognition
o Cursive Handwriting Character Recognition
Food Industry
o Odor/Aroma Analysis
o Product Development
o Quality Assurance
Financial Industry
o Market Trading
o Fraud Detection
o Credit Rating
Energy Industry
o Electrical Load Forecasting
o Hydroelectric Dam Operation
o Oil and Natural Gas Company
Manufacturing
o Process Control
o Quality Control
Medical and Health Care Industry
o Image Analysis
o Drug Development
o Resource Allocation
Science and Engineering
o Chemical Engineering
o Electrical Engineering
Reference:
“Neural Network”
“Internet”: -
www.anns.mit/edu.com
www.cs.barkely.com
www.academicresources.com