Professional Documents
Culture Documents
PROJECT REPORT
ON
Submitted by:
DIVYA SETH
(BANASTHALI VIDYAPITH)
PROJECT INTERN
TVS MOTOR
CERTIFICATE
Certified that this project report “Vison based Programming, AI in camera vision application
” is the bonafide work of “DIVYA SETH” who carried out the project work under my
supervision.
Himansu Shaw
Assistant Manager
Project Guide
TVS Motor
ACKNOWLEDGEMENT
I would like to express our special thanks of gratitude to my project mentor S.K. VIJAY, as well
as my project guide I VENKATA VISHNU VIHARI & my project guide HIMANSU SHAW
who gave me the golden opportunity to do this wonderful project on the topic “Vision-based
Programming, AI in camera vision application”, which also helped me in doing a lot of research
and I, came to know about so many new things. I’m thankful to them. Secondly, I would also
like to thank our parents who helped us a lot in finalizing this project within the limited time
frame.
PROJECT INTERN
TVS MOTOR
TABLE OF CONTENTS
TITLE
LIST OF FIGURES
1.0 INTRODUCTION
1.1 AI Machine Vision
1.2 Applications of MV
2.0 IMPLEMENTATION
2.1 Pre-Design Solution
2.2 Camera Selection Criteria
2.3 Problem Statement
3.0 METHODOLOGY
3.1 7 Layers of OSI Model
4.0 CHALLENGES
4.1 System Integration
5.0 RESULT AND DISCUSSION
5.1 Node-Red
5.2 Dashboard Result
6.0 CONCLUSION
7.0 DISCUSSION & FUTURE WORK
REFERENCES
LIST OF FIGURES
CHAPTER 1
INTRODUCTION
1.1 AI Machine Vision
The overall machine vision process includes planning the details of the requirements and
project and then creating a solution. During run-time, the process starts with imaging,
followed by automated analysis of the image and extraction of the required information.
The primary uses for machine vision are imaging-based automatic inspection and sorting and
robot guidance. The former is abbreviated as "automatic inspection".
Equipment
The components of an automatic inspection system usually include lighting, a camera or
other imager, a processor, software, and output devices.
Imaging
The imaging device (e.g. camera) can either be separate from the main image processing unit
or combined with it in which case the combination is generally called a smart camera or
smart sensor. Inclusion of the full processing function into the same enclosure as the camera
is often referred to as embedded processing. When separated, the connection may be made to
specialized intermediate hardware, a custom processing appliance, or a frame grabber within
a computer using either an analog or standardized digital interface (Camera Link,
CoaXPress). MV implementations also use digital cameras capable of direct connections
(without a frame grabber) to a computer via FireWire, USB, or Gigabit Ethernet interfaces.
While conventional (2D visible light) imaging is most commonly used in MV, alternatives
include multispectral imaging, hyperspectral imaging, imaging various infrared bands, line
scan imaging, 3D imaging of surfaces, and X-ray imaging. Key differentiations within MV
2D visible light imaging are monochromatic vs. color, frame rate, resolution, and whether or
not the imaging process is simultaneous over the entire image, making it suitable for moving
processes.
Though the vast majority of machine vision applications are solved using two-dimensional
imaging, machine vision applications utilizing 3D imaging are a growing niche within the
industry. The most commonly used method for 3D imaging is scanning-based triangulation
which utilizes the motion of the product or image during the imaging process. A laser is
projected onto the surfaces of an object. In machine vision, this is accomplished with a
scanning motion, either by moving the workpiece or by moving the camera & laser imaging
system. The line is viewed by a camera from a different angle; the deviation of the line
represents shape variations. Lines from multiple scans are assembled into a depth map or
point cloud. Stereoscopic vision is used in special cases involving unique features present in
both views of a pair of cameras. Other 3D methods used for machine vision are time of flight
and grid-based. One method is grid array-based systems using the pseudorandom structured
light system as employed by the Microsoft Kinect system circa 2012.
Image processing
After an image is acquired, it is processed. Central processing functions are generally done
by a CPU, a GPU, an FPGA, or a combination of these. Deep learning training and inference
impose higher processing performance requirements. Multiple stages of processing are
generally used in a sequence that ends up as the desired result. A typical sequence might start
with tools such as filters which modify the image, followed by extraction of objects, then
extraction (e.g. measurements, reading of codes) of data from those objects, followed by
communicating that data, or comparing it against target values to create and communicate
"pass/fail" results. Machine vision image processing methods include;
Neural net / deep learning/machine learning processing: weighted and self-training multi-
variable decision making Circa 2019 there is a large expansion of this, using deep learning
and machine learning to significantly expand machine vision capabilities.
Pattern recognition including template matching. Finding, matching, and/or counting specific
patterns. This may include the location of an object that may be rotated, partially hidden by
another object, or varying in size.
Outputs
A common output from automatic inspection systems is pass/fail decisions. These decisions
may in turn trigger mechanisms that reject failed items or sound an alarm. Other common
outputs include object position and orientation information for robot guidance systems.
Additionally, output types include numerical measurement data, data read from codes and
characters, counts and classification of objects, displays of the process or results, stored
images, alarms from automated space monitoring MV systems, and process control signals.
This also includes user interfaces, interfaces for the integration of multi-component systems,
and automated data interchange.
CHAPTER 2
IMPLEMENTATION
These two beams will interfere—hence the name, “interferometry”—with each other,
creating light and dark bands. When a portion of the specimen is in focus, the bands are
recorded onto a CCD camera, while the software reconstructs the 3-D data set. The
magnification also can be increased to measure the overall surface profile at surface
roughness resolutions.
5. Camera Technologies
Proprietary technology for high-speed 3-D reconstruction of pixel-based images, Points from
Focus (PFF) provides resolution down to 50 nanometers. Relatively speaking, this is a low-
resolution method for measuring the Z-axis, but the speed of PFF makes it an excellent
technology choice for measuring components such as diamond chips, fastener tips, surgical
screws, and a variety of other manufactured micro-components.
6. Structured illumination
A more recently developed technology, structured illumination will continue being refined
for far greater capabilities and super-resolution levels than are available today. With the
increased speeds of image capture devices, image processors (such as frame grabber boards),
and computer systems, processing massive amounts of raw images will not be the bottleneck.
Structured illumination microscopy provides resolution capability similar to WLI
technology. This technology is currently used on biological specimens because of its optical
sectioning capability, which measures focal planes deep into a specimen that would
otherwise require manual sectioning or cutting.
Sensor size
Sensor size and the optics determine the FOV. For machine vision, the most popular sensor sizes
are 1/2 and 1/3 type (also called 1/2 or 1/3 in.). The 1/2 type refers to the active image area of an
old 1/2-in. imaging tube. When selecting a lens, make sure that it is compatible with a sensor size
equal to or greater than the sensors in the camera. Pixels (or cell) size or pitch is typically
reported in microns. The pixel size is always measured from pixel center to pixel center. The
active imaging size of a sensor is therefore the pixel size times the number of active or effective
pixels.
Signal-to-noise ratio
The signal-to-noise ratio is the ratio of the maximum signal and the minimum noise the camera
can achieve within an output image. This is usually expressed in decibels (dB) in a video camera.
The higher the decibel spec, the better the contrast between signal and noise from the camera. In
machine vision, a digital image is needed, so the image is digitized either in a frame grabber or in
the camera. Since 6 dB represents a doubling of the ratio, 48 dB is equal to 8 bits, and 60 dB is
equal to 10 bits. The effective number of bits will depend on analog signal transmission losses
and the frame grabber.
Robots and human labor both are required during the execution of any Automatic or Manual
programming or development, in which the inclusion of humans brings in the human error too
likely child part detection which is missing or it could be any issue big or small. This issue
comes up with a solution of the vision detection system, & hence the development of AI
Machine Vision arrives.
CHAPTER 3
METHODOLOGY
CHALLENGES
System Integration for Machine Vision camera has a lot of challenges to face, like
1. The decision of hardware selection that is compactable to all the four components i.e.
Robot, PLC, Camera, and User Interface.
2. The robot doesn’t support any of the communication protocols which can directly
communicate with PLC & Camera……. Hence, We are in requirement of usage of
Digital I/O Communication.
3. Communication between Node-Red & PLC requires MODBUS TCP/IP.
4. Communication between Node-Red & Camera requires PROFINET.
MODBUS TCP/IP, PROFINET, ETHERNET, COAXIAL CABLE, Rs232, etc are types of
data links, and communication between each of the four components requires a different type
of datalink which is the major challenge for SYSTEM INTEGRATION and approach to OSI
MODEL.
CHAPTER 5
5.1 NODE-RED
Node-RED is a programming tool for wiring together hardware devices, APIs, and online
services in new and interesting ways. It provides a browser-based editor that makes it
easy to wire together flows using the wide range of nodes in the palette that can be
deployed to its runtime in a single click.
Node-RED provides a browser-based flow editor that makes it easy to wire together
flows using the wide range of nodes in the palette. Flows can be then deployed to the
runtime in a single click. JavaScript functions can be created within the editor using a
rich text editor. A built-in library allows you to save useful functions, templates, or flows
for re-use.
The lightweight runtime is built on Node.js, taking full advantage of its event-driven,
non-blocking model. This makes it ideal to run at the edge of the network on low-cost
hardware such as the Raspberry Pi as well as in the cloud. With over 225,000 modules in
Node's package repository, it is easy to extend the range of palette nodes to add new
capabilities.
The flows created in Node-RED are stored using JSON which can be easily imported and
exported for sharing with others. An online flow library allows you to share your best
flows with the world.
5.2 DASHBOARD RESULTS
Result 1:
Result 2:
Result 3:
CHAPTER 7
CONCLUSION
1. Did research on different Machine Vision camera systems available other than currently
used in live-project which gave me a handful of information about different cameras and
their specification and provides help with some basic day-to-day life activities too.
2. Machine vision is enabling various industries such as retail, insurance, manufacturing,
etc. to achieve greater customer delight and satisfaction. Technological advances in the
machine vision field are continuing to expand the capabilities across industrial sectors
and applications. This development is manifesting itself as greater accuracy,
repeatability, quality, and efficiencies in manufacturing.
3. Also, introduce a new species of camera, sensors, and robots.
4. With the new technology being used, manufacturing processes can also become fast and
reliable with origami robots.
5. Also, ease human intervention.
CHAPTER 8
I have studied a system that provides a web browser-based flow editor, which can be used
to create JavaScript functions. Elements of applications can be saved or shared for re-use.
The runtime is built on Node.js. The flows created in Node-RED are stored using JSON.
Since version 0.14, MQTT nodes can make properly configured TLS connections.
Results demonstrate that a design approach that simultaneously tackles geometry and
motion aspects of a robot design hand-in-hand with User-Interface, can lead to better
performance.
REFERENCES
1. https://nodered.org/docs/
2. https://nodered.org/docs/user-guide/writing-functions
3. https://nodered.org/
4. https://en.wikipedia.org/wiki/Node-RED
5. https://youtu.be/sdGqM6L9tY0
6. https://youtu.be/_R_hQSl1Lq8
7. https://robotiq.com/products/wrist-camera?
_ga=2.22285118.464791987.1623676429-1084210023.1623676429
8. https://blog.robotiq.com/ruc-challenge-update-its-been-a-challenging-
night