Professional Documents
Culture Documents
---------------------------------------------------------------------------
1.Machine Vision
---------------------------------------------------------------------------
A. Definition
Achieving this goal may require varying the amount of light used
(intensity), the style of lighting (dome vs. ring light, for example)
and the placement of the light source relative to the part and the
optical system or camera.
Lighting options include LED lighting and strobe lights for capturing
images with fast shutter speeds.
Optical System
The optical components in a machine vision system are typically a lens
or a camera, which integrates the lens with other elements such as the
sensor. The lens selection will establish the field of view, which is
the two-dimensional area over which observations can be made.
The lens also will determine the depth of focus and the focal point,
both of which will relate to the ability to observe features on the
parts being processed by the system.
Sensors use CMOS or CCD technology to capture the light and convert
that to a set of pixels that show the presence of light in the
different areas of the original part being observed.
A discrete I/O signal may be fed to a PLC that will use that to light
a stack light, or power solenoid driven actuator to move a rejected
part out of the main production pathway.
Measurement
Counting
Decoding
Location
Measurement functions are done through the comparison of a recorded
dimension from a digital image against a standard value to establish a
tolerance or to determine if the observed value of the dimension is within
acceptable levels of tolerance as called for in the design specification
for that part.
Counting functions are used to establish whether the correct quantity of
product is in place or whether the proper number of components or elements
in a design has been produced. As an example, machine vision could be used
to determine whether a six-pack of soft drinks coming off a production line
at a bottling plant has six cans or bottles, or whether one or more is
missing. At a manufacturing facility, machine vision might be used to
inspect flanges that have been put through an automated drilling operation
to determine if the proper number of holes has been drilled into each
flange.
Decoding functions are used to decode or read one dimensional and two-
dimensional symbologies used to uniquely tag products, such as linear bar
codes, stacked symbologies, data matrix codes, QR codes, or Optical
Character Recognition (OCR) fonts. This functional capability allows for
the recording of historical data on a production process so that a record
is available of a part’s production. It can also enable automation of
product sorting and serve to validate that the correct item is coming
through the process at the correct time.
Location functions deal with establishing the position and orientation of a
part in a process. This type of capability is valuable in automated
assembly operations, as it can be used to verify that the component needed
is in the correct place and is properly aligned within allowed tolerances,
for the next step in the assembly process to occur. Machine vision systems
can also be used to identify a specific part or component by locating a
unique pattern or feature of that part, thus assuring that the item needed
is not only in the correct position but that it is the correct item and not
something else of similar appearance.
Periodic calibration of machine vision systems should take place just as
with other types of metrology equipment.
E. Specific applications
The applications for machine vision systems are widespread and cross many
types of industries. Here are a few examples of applications for machine
vision systems in different industrial settings to provide a general sense
of the diverse nature of these systems:
Automotive manufacturing
Guiding assembly or welding robots
Verifying orientation of parts
Counting the number of welds
Checking for surface defects prior to painting
Electronics
Verifying the shape and position of connector pins
Parts selection and orientation for robotic pick & place systems
Checking for solder connection issues or other conditions on
PCBs
General inspection of manufactured components such as LEDs
Semiconductors
Inspecting wafers and masks using deep ultraviolet wavelength
light (DUV) to achieve the needed high speed and high
resolution.
---------------------------------------------------------------------------
2. Natural Language Processing (NLP):
---------------------------------------------------------------------------
A. Definition:
Speech recognition
Given a sound clip of a person or people speaking, determine the
textual representation of the speech. This is the opposite of
text to speech and is one of the extremely difficult problems
colloquially termed "AI-complete".
In natural speech there are hardly any pauses between successive
words, and thus speech segmentation is a necessary subtask of
speech recognition (see below). In most spoken languages, the
sounds representing successive letters blend into each other in
a process termed coarticulation, so the conversion of the analog
signal to discrete characters can be a very difficult process.
Also, given that words in the same language are spoken by people
with different accents, the speech recognition software must be
able to recognize the wide variety of input as being identical
to each other in terms of its textual equivalent.
Speech segmentation
Given a sound clip of a person or people speaking, separate it into
words. A subtask of speech recognition and typically grouped with it.
Text-to-speech
Given a text, transform those units and produce a spoken
representation. Text-to-speech can be used to aid the visually impaired
Morphological analysis
Lemmatization
The task of removing inflectional endings only and to return the base
dictionary form of a word which is also known as a lemma. Lemmatization
is another technique for reducing words to their normalized form. But
in this case, the transformation actually uses a dictionary to map
words to their actual form.[22]
Morphological segmentation
Separate words into individual morphemes and identify the class of the
morphemes. The difficulty of this task depends greatly on the
complexity of the morphology (i.e., the structure of words) of the
language being considered. English has fairly simple morphology,
especially inflectional morphology, and thus it is often possible to
ignore this task entirely and simply model all possible forms of a word
(e.g., "open, opens, opened, opening") as separate words. In languages
such as Turkish or Meitei,[23] a highly agglutinated Indian language,
however, such an approach is not possible, as each dictionary entry has
thousands of possible word forms.
Part-of-speech tagging
Given a sentence, determine the part of speech (POS) for each word.
Many words, especially common ones, can serve as multiple parts of
speech. For example, "book" can be a noun ("the book on the table") or
verb ("to book a flight"); "set" can be a noun, verb or adjective; and
"out" can be any of at least five different parts of speech.
Stemming
The process of reducing inflected (or sometimes derived) words to a
base form (e.g., "close" will be the root for "closed", "closing",
"close", "closer" etc.). Stemming yields similar results as
lemmatization, but does so on grounds of rules, not a dictionary.
Syntactic analysis
Grammar induction[24]
Generate a formal grammar that describes a language's syntax.
Sentence breaking (also known as "sentence boundary disambiguation")
Given a chunk of text, find the sentence boundaries. Sentence
boundaries are often marked by periods or other punctuation marks, but
these same characters can serve other purposes (e.g., marking
abbreviations).
Parsing
Determine the parse tree (grammatical analysis) of a given sentence.
The grammar for natural languages is ambiguous and typical sentences
have multiple possible analyses: perhaps surprisingly, for a typical
sentence there may be thousands of potential parses (most of which will
seem completely nonsensical to a human). There are two primary types of
parsing: dependency parsing and constituency parsing. Dependency
parsing focuses on the relationships between words in a sentence
(marking things like primary objects and predicates), whereas
constituency parsing focuses on building out the parse tree using a
probabilistic context-free grammar (PCFG) (see also stochastic
grammar).
Terminology extraction
The goal of terminology extraction is to automatically extract relevant
terms from a given corpus.
Word sense disambiguation (WSD)
Many words have more than one meaning; we have to select the meaning
which makes the most sense in context. For this problem, we are
typically given a list of words and associated word senses, e.g. from a
dictionary or an online resource such as WordNet.
Entity linking
Many words - typically proper names - refer to named entities; here we
have to select the entity (a famous individual, a location, a company,
etc.) which is referred to in context.
Argument mining
The goal of argument mining is the automatic extraction and
identification of argumentative structures from natural language text
with the aid of computer programs.[26] Such argumentative structures
include the premise, conclusions, the argument scheme and the
relationship between the main and subsidiary argument, or the main and
counter-argument within discourse.[27][28]
---------------------------------------------------------------------------
3. Industry 4:
The first industrial revolution came with the advent of mechanisation,
steam power and water power.
Characters A-Z, a-z, 0-9, and other special symbols are considered as
genes
A string generated by these characters is considered as
chromosome/solution/Individual
Fitness score is the number of characters which differ from characters in
target string at a particular index. So individual having lower fitness
value is given more preference.
Why use Genetic Algorithms
MYCIN
One of the earliest expert systems based on backward chaining. It can
identify various bacteria that can cause severe infections and can
also recommend drugs based on the person’s weight.
DENDRAL
It was an artificial intelligence-based expert system used for
chemical analysis. It used a substance’s spectrographic data to
predict its molecular structure.
R1/XCON
It could select specific software to generate a computer system wished
by the user.
PXDES
It could easily determine the type and the degree of lung cancer in a
patient based on the data.
CaDet
It is a clinical support system that could identify cancer in its
early stages in patients.
DXplain
It was also a clinical support system that could suggest a variety of
diseases based on the findings of the doctor.
Components of an Expert System :
Knowledge Base –
The knowledge base represents facts and rules. It consists of
knowledge in a particular domain as well as rules to solve a problem,
procedures and intrinsic data relevant to the domain.
Inference Engine –
The function of the inference engine is to fetch the relevant
knowledge from the knowledge base, interpret it and to find a solution
relevant to the user’s problem. The inference engine acquires the
rules from its knowledge base and applies them to the known facts to
infer new facts. Inference engines can also include an explanation and
debugging abilities.
Knowledge Acquisition and Learning Module –
The function of this component is to allow the expert system to
acquire more and more knowledge from various sources and store it in
the knowledge base.
User Interface –
This module makes it possible for a non-expert user to interact with
the expert system and find a solution to the problem.
Explanation Module –
This module helps the expert system to give the user an explanation
about how the expert system reached a particular conclusion.
The Inference Engine generally uses two strategies for acquiring knowledge
from the Knowledge Base, namely –
Forward Chaining
Backward Chaining
Forward Chaining –
Forward Chaining is a strategic process used by the Expert System to answer
the questions – What will happen next. This strategy is mostly used for
managing tasks like creating a conclusion, result or effect. Example –
prediction or share market movement status.
Backward Chaining –
Backward Chaining is a storage used by the Expert System to answer the
questions – Why this has happened. This strategy is mostly used to find out
the root cause or reason behind it, considering what has already happened.
Example – diagnosis of stomach pain, blood cancer or dengue, etc.
Backward Chaining
Characteristics of an Expert System :
The term “deep” usually refers to the number of hidden layers in the
neural network. Traditional neural networks only contain 2-3 hidden
layers, while deep networks can have as many as 150.
Deep learning models are trained by using large sets of labeled data and
neural network architectures that learn features directly from the data
without the need for manual feature extraction.
Figure 1: Neural networks, which are organized in layers consisting of a
set of interconnected nodes. Networks can have tens or hundreds of hidden
layers.
One of the most popular types of deep neural networks is known as
convolutional neural networks (CNN or ConvNet). A CNN convolves learned
features with input data, and uses 2D convolutional layers, making this
architecture well suited to processing 2D data, such as images.
CNNs eliminate the need for manual feature extraction, so you do not need
to identify features used to classify images. The CNN works by extracting
features directly from images. The relevant features are not pretrained;
they are learned while the network trains on a collection of images. This
automated feature extraction makes deep learning models highly accurate for
computer vision tasks such as object classification.
Figure 2: Example of a network with many convolutional layers. Filters are
applied to each training image at different resolutions, and the output of
each convolved image serves as the input to the next layer.
CNNs learn to detect different features of an image using tens or hundreds
of hidden layers. Every hidden layer increases the complexity of the
learned image features. For example, the first hidden layer could learn how
to detect edges, and the last learns how to detect more complex shapes
specifically catered to the shape of the object we are trying to recognize.
Difference Between Machine Learning and Deep Learning?
Deep learning is a specialized form of machine learning. A machine learning
workflow starts with relevant features being manually extracted from
images. The features are then used to create a model that categorizes the
objects in the image. With a deep learning workflow, relevant features are
automatically extracted from images. In addition, deep learning performs
“end-to-end learning” – where a network is given raw data and a task to
perform, such as classification, and it learns how to do this
automatically.
Another key difference is deep learning algorithms scale with data, whereas
shallow learning converges. Shallow learning refers to machine learning
methods that plateau at a certain level of performance when you add more
examples and training data to the network.
A key advantage of deep learning networks is that they often continue to
improve as the size of your data increases.
1. Virtual Assistants
Virtual Assistants are cloud-based applications that understand natural
language voice commands and complete tasks for the user. Amazon Alexa,
Cortana, Siri, and Google Assistant are typical examples of virtual
assistants. They need internet-connected devices to work with their full
capabilities. Each time a command is fed to the assistant, they tend to
provide a better user experience based on past experiences using Deep
Learning algorithms.
2. Chatbots
Chatbots can solve customer problems in seconds. A chatbot is an AI
application to chat online via text or text-to-speech. It is capable of
communicating and performing actions similar to a human. Chatbots are used
a lot in customer interaction, marketing on social network sites, and
instant messaging the client. It delivers automated responses to user
inputs. It uses machine learning and deep learning algorithms to generate
different types of reactions.
The next important deep learning application is related to Healthcare.
3. Healthcare
Deep Learning has found its application in the Healthcare sector. Computer-
aided disease detection and computer-aided diagnosis have been possible
using Deep Learning. It is widely used for medical research, drug
discovery, and diagnosis of life-threatening diseases such as cancer and
diabetic retinopathy through the process of medical imaging.
4.
Deep learning applications are used in industries from automated driving to
medical devices.
Automated Driving: Automotive researchers are using deep learning to
automatically detect objects such as stop signs and traffic lights. In
addition, deep learning is used to detect pedestrians, which helps decrease
accidents.
Aerospace and Defense: Deep learning is used to identify objects from
satellites that locate areas of interest, and identify safe or unsafe zones
for troops.
Medical Research: Cancer researchers are using deep learning to
automatically detect cancer cells. Teams at UCLA built an advanced
microscope that yields a high-dimensional data set used to train a deep
learning application to accurately identify cancer cells.
Industrial Automation: Deep learning is helping to improve worker safety
around heavy machinery by automatically detecting when people or objects
are within an unsafe distance of machines.
Electronics: Deep learning is being used in automated hearing and speech
translation. For example, home assistance devices that respond to your
voice and know your preferences are powered by deep learning applications.