Professional Documents
Culture Documents
Image Processing Group 17
Image Processing Group 17
Concepts Overview
We will explore the fundamentals, historical context, techniques, applications, challenges, and
future prospects of multimedia image processing.
Introduction and Fundamentals
What's an image?
A simple image model
Fundamental steps in image processing
Elements of digital image processing systems
Applications of image processing
Benefits of image processing
Image
An image refers to a 2D light intensity function f(x,y), where (x,y) denote spatial coordinates
and the value of f at any point (x,y) is proportional to the brightness or gray levels of the image
at that point.
Before we jump into image processing, we need to first understand what exactly constitutes an
image. An image is represented by its dimensions (height and width) based on the number of
pixels. For example, if the dimensions of an image are 500 x 400 (width x height), the total
number of pixels in the image is 200000.
This pixel is a point on the image that takes on a specific shade, opacity or color. It is usually
represented in one of the following:
Analog Image:
- An analog image is a continuous representation of visual information.
- It's captured directly from the real world using devices like cameras and scanners.
- Analog images have an infinite range of colors and tones.
- The information in an analog image is stored in a continuous manner, like a film negative.
- Analog images are susceptible to degradation and noise due to external factors.
Digital Image:
• A digital image is an image f(x,y) that has been discretized both in spatial coordinates and
brightness.
That is a discrete representation of visual information.
• The elements of such a digital array are called image elements or pixels.
Image processing requires fixed sequences of operations that are performed at each pixel of an
image. The image processor performs the first sequence of operations on the image, pixel by
pixel. Once this is fully done, it will begin to perform the second operation, and so on. The
output value of these operations can be computed at any pixel of the image.
- It's obtained by converting analog information into digital data using devices like digital
cameras and scanners.
- Digital images have a finite range of colors and tones, determined by the bit depth.
- The information in a digital image is stored in discrete values, usually as pixels with specific
color or intensity values.
- Digital images are more resistant to degradation and noise, as they can be stored and
transmitted in a more controlled manner.
Emergence of Television and Photography
Early television systems and photographic techniques laid the foundation for modern image
processing. Techniques like darkroom manipulation were precursors to digital image editing.
Example: George Carey's 1858 photograph "First Photograph Ever Taken" marked the early
use of image manipulation to remove people from the scene.
Example: Sony's Mavica (1981) was one of the first digital cameras, capturing images on a
floppy disk.
Russell Kirsch's First Digital Image
Russell Kirsch's 1957 176x176 pixel digital image of his son paved the way for digital imaging.
This marked the beginning of pixel-based digital representations.
Image Processing
Multimedia Image Processing is the field that encompasses the manipulation, analysis, and
enhancement of visual data, ranging from photographs and videos to animations and virtual
reality content. It is driven by the goal of improving visual quality, extracting meaningful
information, and optimizing multimedia experiences.
There are basically 2 main forms of image processing. That is Analogue image processing and
Digital image processing.
The key difference between analog and digital images lies in their continuous vs. discrete nature
and the methods used for processing. Analog image processing involves direct physical
manipulation of continuous images, while digital image processing involves algorithmic
manipulation of discrete digital images using computers and software.
Image Acquisition
Image acquisition is the first step in image processing. This step is also known as preprocessing
in image processing. It involves retrieving the image from a source, usually a hardware-based
source.
Image Enhancement
Image enhancement is the process of bringing out and highlighting certain features of interest in
an image that has been obscured. This can involve changing the brightness, contrast, etc.
Image Restoration
Image restoration is the process of improving the appearance of an image. However, unlike
image enhancement, image restoration is done using certain mathematical or probabilistic
models.
Compression
Compression is a process used to reduce the storage required to save an image or the bandwidth
required to transmit it. This is done particularly when the image is for use on the Internet.
Morphological Processing
Morphological processing is a set of processing operations for morphing images based on their
shapes.
Segmentation
Segmentation is one of the most difficult steps of image processing. It involves partitioning an
image into its constituent parts or objects.
Recognition
Recognition assigns a label to an object based on its description.
Computer
A general-purpose computer, which may be anything from a PC to a supercomputer, is used in
an image processing system. Sometimes, specifically built computers are utilized in specialized
applications to reach a specified degree of performance.
Massive Storing
In applications involving image processing, the skill is essential. The three main types of digital
storage for image processing applications are as follows: Three types of storage exist (1) short-
term storage, (2) online storage for quick recall (3) archive storage, which is characterized by
rare access.
Camera Sensors
It alludes to perception. The image sensor's primary function is to collect incoming light,
transform it into an electrical signal, measure that signal, and then output it to supporting
electronics. It consists of a two-dimensional array of light-sensitive components that convert
photons into electrons. Images are captured by equipment like digital cameras using image
sensors like CCD and CMOS. Two components are often needed on image sensors to collect
digital pictures. The first is an actual tool (sensor) that can detect the energy emitted by the
object we want to turn into an image. The second is a digitizer, which transforms a physical
sensing device's output into digital form.
Image Display
The pictures are shown.
Software
The image processing software comprises specialized modules that carry out particular functions.
Hardcopy Equipment
Laser printers, film cameras, heat-sensitive equipment, inkjet printers, and digital equipment like
optical and CDROM discs are just a few examples of the instruments used to record pictures.
Networking
To send visual data through a networked computer, it is a necessary component. The most
important factor in picture transmission is bandwidth since image processing applications require
vast amounts of data.
Blind Deconvolution in Image Processing
Blind image deconvolution is the challenge of recovering a clear picture from a blurry and noisy
one without knowing precisely how the image was blurred, as obtained by an ideal pinhole
camera. The unidentified blurring operation might be brought on by defocus, camera movement,
scene motion, or other optical defects. A trade-off between exposure duration and aperture
setting is necessary for proper photography exposure. The photographer might utilize a big
aperture or a lengthy exposure period when the lighting is bad. The first option produces motion
blur when the camera moves concerning objects in the scene while the exposure takes place.
When using the second option, things farther from the focal plane become out-of-focus blurred.
This may lead to blind deconvolution in image processing.
Importance of Phase in Image Processing
Phase, in a nutshell, contains information about the positions of features. Phase-only and
magnitude-only photos cannot be combined to produce the original. To obtain the original,
multiply them in the Fourier domain and reverse the transformation. A repeated waveform's
phase describes the position or timing of a particular point within a wave cycle. Instead of the
actual absolute phases of the signals, the phase difference between waves usually matters.
Basics of color
(a) Light and spectra
• Color is the perceptual result of light in the
visible region of the spectrum, having in the
region of 400nm to 700nm, incident upon the
retina.
• Visible Light is a form of electromagnetic
energy consisting of a spectrum of frequencies
having wavelengths range from about 400nm for
violet light to about 700nm for red light.
• Most light we see is a combination of many
wavelengths.
(b) Primaries
• Any color can be matched by proper proportions
of three component colors called primaries.
• The most common primaries are red, blue and
green.
U = 0.565(B-Y)
V = 0.713(R-Y)
• Interest in digital image processing methods stems from 2 principal application areas:
(1) improvement of pictorial information for human interpretation, and
(2) (2) processing of scene data for autonomous machine perception. • In the second application
area, interest focuses on procedures for extracting from an image information in a form
suitable for computer processing.
(3) • Examples include automatic character recognition,
(4) industrial machine vision for product assembly and inspection,
(5) military recognizance,
(6) automatic processing of fingerprints etc.
Example: In medical imaging, reducing noise in X-rays enhances the visibility of subtle
abnormalities.
Example: Astronomical images often require noise reduction to unveil distant celestial objects.
Enhancement: Contrast Adjustment
Contrast enhancement techniques modify pixel intensities to improve visual perception of
details.
Example: Enhancing satellite images helps detect changes in land use or deforestation.
Example: JPEG is commonly used for photographs on websites due to its balance between
quality and size.
Example: Fourier analysis helps identify specific frequencies in medical EEG signals.
Example: Self-driving cars use object detection to identify pedestrians and other vehicles.
Example: Lossy compression is suitable for web images, while lossless compression is
preferred for medical images.
Pattern recognition measures the various patterns around the objects in the image
Haar Cascades: Face Detection
Haar cascades are used for real-time face detection by identifying features like eyes and noses.
Left - normal traffic image | Right - a VIPS image with detection zones
Besides this, it can auto record the license plate of the vehicle, distinguish the type of vehicle,
monitor the speed of the driver on the highway and lots more.
Image Reconstruction
Image processing can be used to recover and fill in the missing or corrupt parts of an image. This
involves using image processing systems that have been trained extensively with existing photo
datasets to create newer versions of old and damaged photos.
Fig: Reconstructing damaged images using image processing
Face Detection
One of the most common applications of image processing that we use today is face detection. It
follows deep learning algorithms where the machine is first trained with the specific features of
human faces, such as the shape of the face, the distance between the eyes, etc. After teaching the
machine these human face features, it will start to accept all objects in an image that resemble a
human face. Face detection is a vital tool used in security, biometrics and even filters available
on most social media apps these days.
The digital image can be made available in any desired format (improved image, X-
Ray, photo negative, etc)
It helps to improve images for human interpretation
Information can be processed and extracted from images for machine interpretation
The pixels in the image can be manipulated to any desired density and contrast
Images can be stored and retrieved easily
It allows for easy electronic transmission of images to third-party providers
Example: Autonomous vehicles use real-time image processing for safe navigation.
Example: GANs are used for generating realistic images, like in the "This Person Does Not
Exist" project.
Acknowledgments
Special thanks to the pioneers in the field of multimedia image processing and the researchers
who continue to advance this fascinating discipline.
This concludes the concept on multimedia image processing. We've delved into the foundational
concepts, historical developments, diverse applications, challenges, and exciting future directions
of this dynamic field. As technology continues to evolve, the role of image processing in
reshaping how we perceive and interact with visual information remains ever more critical.
Reference:
1. Gonzalez, R.C., & Woods, R.E. (2018). "Digital Image Processing". Pearson Education.
2. Szeliski, R. (2010). "Computer Vision: Algorithms and Applications". Springer.
3. Pratt, W.K. (2001). "Digital Image Processing". John Wiley & Sons.
4. OpenCV Documentation: [https://docs.opencv.org/](https://docs.opencv.org/)
5. IEEE Xplore: [https://ieeexplore.ieee.org/](https://ieeexplore.ieee.org/)
6. Goodfellow, I., Bengio, Y., & Courville, A. (2016). "Deep Learning". MIT Press.
7. Coursera - Digital Image Processing: [https://www.coursera.org/](https://www.coursera.org/)
8. Watt, A., & Policarpo, F. (2016). "3D Computer Graphics". Addison-Wesley Professional.
9. Vincent, L., & Soille, P. (1991). "Watersheds in Digital Spaces: An Efficient Algorithm Based
on Immersion Simulations". IEEE Transactions on Pattern Analysis and Machine Intelligence.
10. Kirsch, R.A. (1957). "A Program for Aligning Pictures on a Digital Computer". Journal of
the ACM.
11. Pratt, W.K. (2001). "Digital Image Processing". John Wiley & Sons.
12. Szeliski, R. (2010). "Computer Vision: Algorithms and Applications". Springer.
13. Gonzalez, R.C., & Woods, R.E. (2018). "Digital Image Processing". Pearson Education.
14. Farid, H. (2020). "Photo Forensics". MIT Press.
15. Canny, J. (1986). "A Computational Approach to Edge Detection". IEEE Transactions on
Pattern Analysis and Machine Intelligence.
16. Viola, P., & Jones, M. (2001). "Rapid Object Detection using a Boosted Cascade of Simple
Features". CVPR.
17. LeCun, Y., Bengio, Y., & Hinton, G. (2015). "Deep Learning". Nature.
18. Pitas, I. (2000). "Digital Image Processing Algorithms and Applications". John Wiley &
Sons.