Professional Documents
Culture Documents
VISION AND
AUTOMATION
SOLUTIONS
FOR ENGINEERS
AND INTEGRATORS
D E S I G N
®
WORLDWIDE
SPECIAL REPORT
Machine vision:
The past, the present
and the future
Which particular individuals, companies and
organizations, technologies, products and Keystones of machine
Vision Systems Design
applications have most significantly affected PAGE 2
the adoption of machine vision and image
processing systems? What can history tell us
about the developments that have occurred in
the past 50 years and how will new technologies
be leveraged to automate every aspect of the
manual tasks now relegated to human operators?
Editor in Chief Andy Wilson takes a look back on
the past and, how, in the future, autonomous
Machine vision:
vision-based machines will be deployed in A look into the future
applications as diverse as food harvesting, PAGE 21
medical imaging and autonomous vehicles.
REPRINTED WITH REVISIONS TO FORMAT FROM VISION SYSTEMS DESIGN. COPYRIGHT 2015 BY PENNWELL CORPOR ATION
ORIGINALLY PUBLISHED SEP TEMBER 2013
Keystones of machine
Vision Systems Design
1930s with Electronic Sorting Machines (then located in New Jersey) offering food
sorters based on using specific filters and photomultiplier detectors.
While it is true that machine vision systems have only been deployed for less than
a century, some of the most significant inventions and discoveries that led to the
development of such systems date back far longer. To thoroughly chronicle this, one
could begin by highlighting the development of early Egyptian optical lens systems
dating back to 700 BC, the introduction of punched paper cards in 1801 by Joseph
Marie Jacquard that allowed a loom to weave intricate patterns automatically or
Maxwell’s 1873 unified theory of electricity and magnetism.
Along the way, it will highlight some of the people, companies and organizations
that have made such products a reality. This article will, of necessity, be a rather more
personal and opinionated piece and, as such, I welcome any additions that you feel
have been omitted.
Basler
Truesense Imaging 4%
4%
Edmund Optics
4% Cognex
30%
MVTec
4.5%
FLIR
4.5%
National
Instruments
4.5%
Pt Grey
5% FIGURE 1: When asked which
Sony
companies had made the greatest
7% Matrox impact on the machine vision over
13%
the past twenty years or more,
Keyence
9% Teledyne DALSA 30% of our readers chose Cognex.
10%
When asked what technologies and products have made the most impact on
machine vision, readers’ answers were rather more diverse (Figure 2). Interestingly,
the emergence of CMOS images sensors, smart cameras and LED lighting, all
relatively new development in the history of machine vision, were recognized as
some of the most important innovations.
Capturing images
Although descriptions of pin-hole camera date back to as early as the 5th century
BC, it was not until about 1800 that Thomas Wedgwood, the son of a famous
English potter, attempted to capture images using paper or white treated with
silver nitrate. Following this, Louis Daguerre and others demonstrated that a silver
plated copper plate exposed under iodine vapor would produce a coating of light-
Bayer filter
GigE vision 3.5%
3.5%
Vision
software CMOS imagers
3.5% 24%
High-speed
interfaces
12%
1309VSDpieChart2
sensitive silver iodide on the surface with the resultant fixed plate producing a
replica of the scene.
Developments of the mid-19th century was followed by others, notably Henry Fox
Talbot in England who showed that paper impregnated with silver chloride could be
used to capture images. While this work would lead to the development of a multi-
billion dollar photographic industry, it is interesting that, during the same time
period, others were studying methods of capturing images electronically.
In 1857, Heinrich Geissler a German physicist developed a gas discharge tube filled
with rarefied gasses that would glow when a current was applied to the two metal
electrodes at each end. Modifying this invention, Sir William Crookes discovered
that streams of electrons could be projected towards the end of such a tube using a
cathode-anode structure common in cathode ray tubes (CRTs).
Needless to say, being tube-based, such cameras were not exactly purpose built
for rugged, high-EMI susceptible applications. This was to change when, in 1969,
Willard Boyle and George E. Smith working at AT&T Bell Labs showed how
charge could be shifted along the surface of a semiconductor in what was known as
a “charge bubble device”. Although they were both later awarded Nobel prizes for
the invention of the CCD concept, it was an English physicist Michael Tompsett,
a former researcher at the English Electric Valve Company (now e2V; Chelmsford,
England; www.e2v.com), that, in 1971 while working at Bell Labs, showed how the
CCD could be used as an imaging device.
Three years later, the late Bryce Bayer, while working for Kodak, showed how by
applying a checkerboard filter of red, green, and blue to an array of pixels on an area
CCD array, color images could be captured using the device.
While the CCD transfers collected charge from each pixel during readout and
erases the image, scientists at General Electric in 1972 developed and X-Y array of
addressable photosensitive elements known as a charge injection device (CID).
Unlike the CCD, the charge collected is retained in each pixel after the image is read
and only cleared when charge is “injected” into the substrate. Using this technology,
the blooming and smearing artifacts associated with CCDs is eliminated. Cameras
based around this technology were originally offered by CIDTEC, now part of
Thermo Fischer Scientific (Waltham, MA; www.thermoscientific.com).
While the term active pixel image sensor or CMOS image sensor was not to emerge
for two decades, previous work on such devices dates back as far as 1967 when
Dr. Gene Weckler described such a device in his paper “Operation of pn junction
photodetectors in a photon flux integrating mode,” IEEE J. Solid-State Circuits
(http://bit.ly/12SnC7O).
Despite this, the success of CMOS imagers was not to become widely adopted for
the next thirty years, due in part to the variability of the CMOS manufacturing
process. Today, however, many manufacturers of active pixel image sensors widely
tout the performance of such devices as comparable to those of CCDs.
Building such devices is expensive, however and – even with the emergence
of “fabless” developers–just a handful of vendors currently offer CCD and
CMOS imagers. Of these, perhaps the best known are Aptina (San Jose, CA;
www.aptina.com), CMOSIS (Antwerp, Belgium http://cmosis.com), Sony
Electronics (Park Ridge, NJ; www.sony.com) and Truesense Imaging (Rochester,
NY; www.truesenseimaging.com), all of whom offer a variety of devices in multiple
configurations.
While the list of imager vendors may be small, however, the emergence of
such devices has spawned literally hundreds of camera companies worldwide.
While many target low-cost applications such as webcams, others such as
Basler (Ahrensburg, Germany; www.baslerweb.com), Imperx (Boca Raton, FL;
www.imperx.com) and JAI (San Jose, CA; www.jai.com) are firmly focused on the
machine vision and image processing markets often incorporating on board FPGAs
into their products.
In 1812, using a large voltaic battery, Sir Humphry Davy demonstrated that an
arch discharge would occur and in 1860 Michael Faraday, an early associate of
Davy’s, demonstrated a lamp exhausted of air that used two carbon electrodes to
produce light.
Building on these discoveries, Edison formed the Edison Electric Light Company in
1878 and demonstrated his version of an incandescent lamp just one year later. To
extend the life of such incandescent lamps, Alexander Just and Franjo Hannaman
developed and patented an electric bulb with a Tungsten filament in 1904 while
showing that lamps filled with an inert gas produce a higher luminosity than
vacuum-based tubes.
Just as the invention of the incandescent lamp predates Edison so too does the
halogen lamp. As far back as 1882, chlorine was used to stop the blackening effects
caused by the blackening of the lamp
and slow the thinning of the tungsten
filament. However, it was not until
Elmer Fridrich and Emmitt Wiley
working for General Electric in Nela
Park, Ohio patented a practical version
of the halogen lamp in 1955 that such
illumination devices became practical.
Like the invention of the incandescent lamp, the origins of the fluorescent lamp date
back to the mid 19th century when in 1857, Heinrich Geissler a German physicist
developed a gas discharge tube filled with rarefied gasses that would glow when a
current was applied to the two metal electrodes at each end. As well as leading to the
invention of commercial fluorescent lamps, this discovery would form the basis of
tube-based image capture devices in the 20th century (see “Capturing images”).
While incandescent and fluorescent lamps became widely popular in the 20th
century, it would be research in electroluminescence that would form the basis
of the introduction of solid-state LED lighting. While electroluminescence was
discovered by Henry Round working at Marconi Labs in 1907, it was pioneering
work by Oleg Losev, who in the mid-1920s, observed light emission from zinc oxide
and silicon carbide crystal rectifier diodes when a current was passed through them
(see “The life and times of the LED,” http://bit.ly/o7axVN).
Numerous papers published by Mr. Losev constitute the discovery of what is now
known as the LED. Like many other such discoveries, it would be years later before
such ideas could be commercialized. Indeed, it would not be until 1962 when,
while working at General Electric Dr. Nick Holonyak, experimenting with GaAsP
produced the world’s first practical red LED. One decade later, Dr. M. George
Craford, a former graduate student of Dr. Holonyak, invented the first yellow
LED. Further developments followed with the development of blue and phosphor-
based white LEDs.
Interface standards
The evolution of machine vision owes as much to the television and broadcast
industry as it does to the development of digital computers. As programmable
vacuum tube based computers were emerging in the early 1940s, engineers working
on the National Television System Committee (NTSC) were formulating the first
monochrome analog NTSC standard.
Adopted in 1941, this was modified in 1953 in what would become the RS-170a
standard to incorporate color while remaining compatible with the monochrome
standard. Today, RS-170 is still being used is numerous digital CCD and CMOS-
based cameras and frame grabber boards allowing 525 line images to be captured
and transferred at 30 frames/s.
Just as open computer bus architectures led to the development of both analog
and digital camera interface boards, television standards committees followed an
alternative path introducing high-definition serial digital interfaces such as SDI and
HD-SDI. Although primarily developed for broadcast equipment, these standards
are also supported by computer interface boards allowing HDTV images to be
transferred to host computers.
Apple Computer were widely adopted in the mid-1990s by many machine vision
camera companies.
Like FireWire, the Universal Serial Bus (USB) introduced in a similar time frame
by a consortium of companies including Intel and Compaq were also to become
widely adopted by both machine vision camera companies and manufacturers of
interface boards.
When first introduced, however, these standards could not support the higher
bandwidths of machine vision cameras and by their very nature were non-
deterministic and because no high-speed point to point interface formally
existed, the Automated Imaging Association (Ann Arbor, MI) formed the
Camera Link committee in the late 1990s. Led by companies such as Basler and
JAI, the well-known Camera Link standard was introduced in October 2000
(http://bit.ly/1cgEdKH).
For some, however, even the 680MByte/s data transfer rate was not sufficient to
support the high data rates demanded by (at the time) high-performance machine
vision cameras and it was Basler and others that, in 2004, by reassigning certain pins
on the Camera Link specification managed to attain a 850 MByte/s transfer rate.
Just as this point-to-point protocol was gaining hold, other technologies such
as Gigabit Ethernet were emerging to challenge the distance limitations of the
Camera Link protocol. In 2006, for example, Pleora Technologies (Kanata, Ontario,
Canada; www.pleora.com) pioneered the introduction of the GigE Vision standard,
that although primarily based on the Gigabit Ethernet standard, incorporated
numerous additions such as how camera data could be more effectively streamed,
how systems developers could control and configure devices and – perhaps more
importantly–the GenICam generic programming interface (http://bit.ly/13TDlba)
for different types of machine vision cameras.
At the same time, the limitations of the Camera Link interface were posing
problems for systems integrators because even the extended 680MByte/s interface
required multiple connectors and still could not support emerging higher speed
CMOS cameras. So it was that in 2008, a consortium of companies led by Active
Silicon (Iver, United Kingdom; www.activesilicon.com), Adimec (Eindhoven,The
Netherlands; www.adimec.com) and EqcoLogic (Brussels, Belgium;
www.eqcologic.com) introduced the CoaXPress interface standard that, as its name
implies, is a high-speed serial communications interface that allows 6.25Gbit/s
to be transferred over a single co-ax cable. To increase this speed further, multiple
channels can be used.
Under development at the same time, and primarily led by Teledyne Dalsa
(Waterloo, Ontario, Canada; www.teledynedalsa.com), the Camera Link HS
(CLHS) standard – supposedly the successor to the Camera Link standard offers
scalable bandwidths from 300 MBytes/s to 16 GBytes/s. At the time of writing,
however, more companies have endorsed the CoaXPress standard than CLHS.
Like the original Gigabit Ethernet, the USB 3.0 standard was not well-suited
to machine vision applications. Even so, Point Grey (Richmond, BC, Canada;
www.ptgrey.com) was the first to introduce a camera for this as early as 2009
(http://bit.ly/15gILiO). So it was that in January this year, the USB Vision
Technical Committee of the Automated Imaging Association announced the
introduction of the USB 3.0 Vision standard which builds on many of the advances
of GigE Vision standard including device discovery, device control, event handling,
and streaming data mechanisms, a standard that will be supported by Point Grey
and numerous others.
It was during the 1960s that many of the algorithms used in today’s machine vision
systems were developed. The pioneering work by Georges Matheron and Jean Serra
on mathematical morphology in the 1960s led to the foundation of the Center
of Mathematical Morphology at the École des Mines in Paris, France. Originally
dealing with binary images, this work was later extended to include grey-scale
images and the commonly known gradient, top-hat and watershed operators that
are used in such software packages as Amerinex Applied Imaging’s (Monroe Twp,
NJ; www.amerineximaging.com) Aphelion.
During the 1970s, researchers at SIPI developed the basic theory of image
processing and how it could be applied in image de-blurring, image coding and
feature extraction. Indeed, early work on transform coding at SIPI now forms the
basis the basis of the JPEG and MPEG standards.
Tired of the standard text images used to compare the results of such algorithms,
Dr. Alexander Sawchuk, now Leonard Silverman Chair Professor at the USC Viterbi
Stand-alone systems by companies such as Vicom and Pixar were at the same
time being challenged by modular hardware from companies such as Datacube,
the developer of the first Q-bus frame grabber for Digital Equipment Corp
(DEC) computers.
With the advent of PCs in the 1980s, board-level frame grabbers, processors and
display controllers for the open architecture ISA bus began to emerge and with it
software callable libraries for image processing.
Today, with the emergence of the PCs PCI-Express bus, off-the-shelf frame
grabbers can be used to transfer images to the host PC at very high-data rates using
a number of different interfaces (see “Interface standards”). At the same time, the
introduction of software packages from companies such as Microscan (Renton,
WA; www.microscan.com), Matrox (Dorval, Quebec, Canada; www.matrox.com),
MVTec (Munich, Germany; www.mvtec.com), Teledyne Dalsa (Waterloo,
Ontario, Canada; www.teledynedalsa.com) and Stemmer Imaging (Puchheim,
Germany; www.stemmer-imaging.de) make it increasingly easier to configure even
the most sophisticated image processing and machine vision systems.
it and it still would not have diminished the effect of the introduction of the
automobile or made their business survive. Instead we adapted and went into
different businesses, we survived, and thrived.
I live in a home larger than my ancestors, have two trucks, multiple computers,
a kitchen and garage full of gadgets, affordable clothing, plenty of food in my
fridge, and live well largely thanks to automation. I think most of us can say that
automation makes our lives better. Even our poor and unemployed live better
as a result of automation. Food is fresher and safer. Goods are more abundant
and cost less.
Mr. William Silver has been at the forefront of development in the machine vision
industry for over 30 years. In 1981, he along with fellow MIT graduate student
Marilyn Matz joined Dr. Robert J. Shillman, a former lecturer in human visual
perception at MIT to co-found a start-up company called Cognex (Natick, MA;
www.cognex.com). He was principal developer of the company’s PatMax pattern
matching technology and its normalized correlation search tool.
Mr. Bryce Bayer (1929-2012) will be long remembered for the filter that bears his
name. After obtaining an engineering physics degree in 1951, Mr. Bayer worked
as a research scientist at Eastman Kodak until his retirement in 1986. U.S. patent
3,971,065 awarded to Mr. Bayer in 1976 and titled simply “Color imaging array” is
one of the most important innovation in image processing over the last 50 years. Mr.
Bayer was awarded the Royal Photographic Society’s Progress Medal in 2009 and
the first Camera Origination and Imaging Medal from the SMPTE in 2012.
Dr. Gary Bradski is a Senior Scientist at Willow Garage (Menlo Park, CA;
www.willowgarage.com) and is a Consulting Professor in the Computer Sciences
Department of Stanford University. With 13 issued patents, Dr. Bradski is
responsible for the Open Source Computer Vision Library (OpenCV), an open
source computer vision and machine learning software library built to provide a
common infrastructure for computer vision applications in research, government
and commercial applications. Dr. Bradski also organized the vision team for
Stanley, the Stanford robot that won the DARPA Grand Challenge and founded
the Stanford Artificial Intelligence Robot (STAIR) project under the leadership of
Professor Andrew Ng.
IMAGINE A WORLD where the laborious and repetitive tasks once performed by man
are replaced with autonomous machines. A world in which crops are automatically
harvested, sorted, processed, inspected, packed and delivered without human
intervention. While this concept may appear to be the realm of science fiction, many
manufacturers already employ automated inspection systems that incorporate
OEM components such as lighting, cameras, frame grabbers, robots and machine
vision software to increase the efficiency and quality of their products.
While effective, these systems only represent a small fraction of a machine vision
market that will expand to encompass every aspect of life from harvesting crops and
minerals to delivering the products manufactured from them directly to consumers’
doors. In a future that demands increased efficiency, such machine vision systems
will be used in systems that automate entire production processes, effectively
eliminating the manpower and errors that this entails.
then prepare the product for further processing. After which many of the machines
that are in use today will further process and package these products for delivery—
once again by autonomous vehicles. In this future, customers at supermarkets will
find no checkout personal as they will be replaced by automated vision systems that
scan, weigh and check every product.
Unconstrained environments
Today’s machine vision systems are often deployed in constrained environments
where lighting can be carefully controlled. However, to design vision-guided
robotic systems for agricultural tasks such as planting, tending and harvesting crops
requires that such systems must operate in unconstrained environments where
lighting and weather conditions may vary dramatically. To do so, requires systems
that incorporate a number of different imaging techniques depending on the
application to be performed.
Currently the subject of many research projects, these systems currently divide
agricultural tasks into automated systems that plant, tender and harvest. Since each
task will be application dependent, so too will the types of sensors that are used in
each system. For crop picking, for example, it is necessary to determine the color of
each fruit to determine its ripeness. For harvesting crops such as wheat, it may only
be necessary to guide an autonomous system across the terrain.
In such systems, machine vision systems that use stereo vision systems, LIDAR,
INS and GPS systems and machine vision software will be used for path planning
mapping and classification of crops. Such systems are currently under development
at many Universities and research institutes around the world (see “Machine
Vision in Agricultural Robotics – A short overview”–http://bit.ly/19MHn9M),
by Emil Segerblad and Björn Delight of Mälardalen University (Västerås,
Sweden; www.mdh.se).
In many of these systems, 3D vision systems play a key role. In the development of
a path planning system, for example, John Reid of Deere & Company (Moline, IL,
USA; www.deere.com) has used a 22 cm baseline stereo camera from Tyzx (Menlo
Park, CA, USA; www.tyzx.com) mounted on the front of a Gator utility vehicle
from Deere with automatic steering capabilities to calculate the position of potential
obstacles and determine a path that for the vehicle’s on-board steering controller
(http://bit.ly/1bRFQAX).
Robotic guidance
While 3D systems can be used to aid the guidance of autonomous vehicles,
3D mapping can also aid crop spraying, perform yield analysis and detect crop
damage or disease. To accomplish such mapping tasks, Michael Nielsen, and his
colleagues at the Danish Technological Institute (Odense, Denmark; www.dti.dk)
used a Trimble GPS, a laser rangefinder from SICK (Minneapolis, MN, USA;
www.sick.com) a stereo camera and a tilt sensor from VectorNav Technologies
(Richardson, TX, USA; www.vectornav.com) mounted on a utility vehicle
equipped with a Halogen flood light and a custom made Xenon strobe. After
scanning rows of peach trees, 3D reconstruction of an orchard was performed by
using tilt sensor corrected GPS positions interpolated through encoder counts
(http://bit.ly/1fxp36W).
While 3D mapping can determine path trajectories, map fields and orchards, it can
also being classify fruits and plants. Indeed, this is the aim of a system developed
by Ulrich Weiss and Peter Biber at Robert Bosch (Schwieberdingen, Germany;
www.bosch.de). To demonstrate that it is possible to distinguish multiple plants
using a low resolution 3D laser sensor, an FX6 sensor from Nippon Signal (Tokyo,
Japan; www.signal.co.jp) was used to measure the distance and reflectance intensity
of the plants using an infrared pulsed laser light with a precision of 1 cm. After 3D
reconstruction, supervised learning techniques were used to identify the plants
(http://bit.ly/19ALVMM).
Automated harvesting
Once identified, robots and vision systems will be employed to automatically
harvest such crops. Such a system for harvesting strawberries, for example has
been developed by Guo Feng and his colleagues at the Shanghai Jiao Tong
University (Shanghai, China; www.sie.sjtu.edu.cn). In the design of the system, a
two-camera imaging system was mounted onto a harvesting robot designed for
strawberry picking. A 640 x 480 DXC-151A from Sony (Park Ridge, NJ, USA;
www.sony.com) mounted on the top of robot frame captures images of 8-10
strawberries. Another camera, a 640 x 480 EC-202 II from Elmo (Plainview, NY,
USA; www.elmousa.com) was installed on the end effector of the robot to image
one or two strawberries.
While the Sony camera localized the fruit, the Elmo camera captures images of
strawberries at a higher resolution. Image from both these cameras were then
captured by an FDM-PCI MULTI frame grabber from Photron (San Diego, CA,
USA; www.photron.com) interfaced to a PC. A ring-shaped fluorescent lamp
installed around the local camera provided the stable lighting required for fruit
location (http://bit.ly/16mWhmq).
Crop grading
Machine vision systems are also playing an increasingly important role in the
grading of crops once harvested. In many systems, this requires the use of
multispectral image analysis. Indeed, this is the approach taken by Olivier Kleynen,
and his colleagues at the Universitaire des Sciences Agronomiques de Gembloux
(Gembloux, Belgium; www.gembloux.ulg.ac.be) in a system to detect defects on
harvested apples.
Other methods such as the use of SWIR cameras have also been used for similar
purposes. Renfu Lu of Michigan State University (East Lansing, MI, USA:
www.msu.edu), for example, has shown how an InGaAs area array camera from
UTC Aerospace Systems (Princeton, NJ, USA; www.sensorsinc.com) covering
the spectral range from 900-1700nm mounted to an imaging spectrograph
from Specim (Oulu, Finland; www.specim.fi) can detect bruises on apples
(http://bit.ly/zF8jlN).
While crop picking robots have yet to be fully realized, those used for grading and
sorting fruit are no longer the realm of scientific research papers. Indeed, numerous
systems now grade and sort products ranging from potatoes, dates, carrots, and
oranges. While many of these systems use visible light-based products for these
tasks, others incorporate multi-spectral image analysis.
Multispectral image analysis is also being used to perform sorting tasks. At Insort
(Feldbach, Austria; www.insort.at), for example, a multispectral camera from
EVK DI Kerschhaggl, Raaba, Austria; www.evk.co.at) has been used in a system
to sort potatoes while Odenberg (Dublin, Ireland; www.odenberg.com) has
developed a system that can sort fruit and vegetables using an NIR spectrometer
(http://bit.ly/zqqr38).
Unmanned vehicles
After packing, of course, such goods must be transported to their final destination.
Today, manned vehicles are used to perform this task. In future, however, such
tasks will be relegated to autonomous vehicles. At present, says Professor William
Covington of the University of Washington School of Law (Seattle, WA, USA;
www.law.washington.edu) fully independent vehicles that operate without
instructions from a server based on updated map data remain in the research
prototype stage.
By 2020, however, Volvo expects accident-free cars and “road trains” guided by a
lead vehicle to become available. Other automobile manufacturers such as GM,
Audi, Nissan and BMW all expect fully autonomous, driverless cars to become
available during this time frame. (http://bit.ly/1ibdWxv). These will be equipped
with sensors such as radar, lidar, cameras, IR and GPS systems to perform this task.
Automated retail
While today’s supermarkets, rely heavily on traditional bar-code readers to price
individual objects, future check-out systems will employ sophisticated scanning,
weighing and pattern recognition to relive human operators of such tasks.
FIGURE 3: Wincor Nixdorf has developed a fully automated system known as the Scan
Portal that the company claims is the world’s first practicable fully automatic scanning
system.
Companies mentioned
VisionSystems
Subscribe
VISION AND AUTOMATION SOLUTIONS
D E S I G N
®
WORLDWIDE
Industrial
Camera
DIRECTORY
now at www.vision-systems.com/
MAGAZINE subscriptions.html