You are on page 1of 91

Digital Image Processing

By:

Dr. Ibrahim Ahmed Ali Babikir

Al Neelain University
Faculty of Petroleum and Minerals

Fifth Year

2013-2014

Dr. Ibrahim Babikir & Dr. Sami El Khidir 1


DIGITAL IMAGE
 A digital image is a two dimensional regular grid array of squared units
(pixel), where each unit square (pixel) is assigned a number, which is
related to some parameter e.g. reflectance, emittance, temperature,
topographic elevation or any other computed parameter (Gibson, 2000).

 Each pixel in the digital image is characterized by:


- its dimensions (x, y),
- a digital number (DN), which is the Z value
- and location within the total area of the image or within the globe

 The dimension of the pixel defines its spatial resolution.

 Digital imagery is also termed raster data.


 The digital numbers in digital imagery are generally coded in the 8-bit

binary range corresponding to 0-255.

 The range 0-255 represents the full grey scale whereas each level within

the range defines a separate grey tone.

 The zero value defines the pure black tone whereas the 255 defines the

white tone.
Architecture of a digital image
Raster array and accompanying digital numbers (DN) values for a single band image.
Dark pixels have low DN values while bright pixels have high DN values
Image Resolution
• Digital images are usually described by their resolution.
• Resolution is a broad term commonly used to describe the number of
pixels that can be displayed on a display device or the area on the
ground that a pixel represents in an image file.
Four major types of resolution are always considered:
 Spectral resolution: describing the specific wavelength intervals that a sensor
can record;
 Spatial resolution: describes the area on the ground represented by each
pixel;
 Radiometric resolution: indicates the number of possible data files values in
each band
 Temporal resolution: indicating how often a sensor obtains imagery of
particular area
The spectral resolution of a sensor system is the number and width of spectral bands in
the sensing device. The simplest form of spectral resolution is a sensor with one band
only, which senses visible light. An image from this sensor would be similar in
appearance to a black and white photograph from an aircraft. A sensor with three
spectral bands in the visible region of the EM spectrum would collect similar information
to that of the human vision system. The Landsat TM sensor has seven spectral bands
located in the visible and near to mid infrared parts of the spectrum.

The spatial resolution (ground resolution) is the ground area imaged for the instantaneous
field of view (IFOV) of the sensing device. Spatial resolution may also be described as the
ground surface area that forms one pixel in the satellite image.

The radiometric resolution of an imaging system describes its ability to


discriminate very slight differences in energy. The finer the radiometric
resolution of a sensor, the more sensitive it is to detecting small differences in
reflected or emitted energy.
The temporal resolution is a measure of the repeat cycle or frequency with which a sensor
revisits the same part of the Earth’s surface. The frequency will vary from several times
per day, for a typical weather satellite, to 8—20 times a year for a moderate ground
resolution satellite, such as Landsat TM. The frequency characteristics will be
determined by the design of the satellite sensor and its orbit pattern.
Digital Image Processing & Analysis

Interpretation and analysis of remotely sensed data involves the identification

and/or measurement of various targets in an image in order to extract useful

information about them.

Currently, most remotely sensed data is represented in digital form as an array of


pixels where the actual brightness value (Bv) of a pixel (digital number DN) is a
representation of the pixels reflected radiation. When remote sensing data are
available in digital format, digital processing and analysis is performed using
various computer software.
Digital Image Processing
• Purpose
To enhance the appearance of the image and to improve its visual
interpretability by increasing the apparent distinction between
the features that exist within the image.
• Why
To amplify the differences among the features in the digital image
and to create a new look of the image to make its features easily
recognizable and to enhance the interpretability of the image.
• How
Digital image processing is performed using specialized
computer-based software. It often involves the manipulation
mathematically complex procedures and techniques. Choosing the
appropriates enhancement technique(s) depends largely on the
objective of the interpreter and his experience and personal
preference.
• Digital processing performed usually to enhance data as a prelude
to visual interpretation.

• Digital processing and analysis procedures and techniques are


often carried out to automatically identify targets and digitally
extract information (occasionally) without any manual
intervention by a human interpreter.

• The key to visual interpretation targets recognition and


information extraction in an image is the abilty of the interpreter
to utilize the of the visual image elements tone, shape, size,
pattern, texture, shadow, and association.
What do we need to facilitate DIP?

• In order to process remotely sensed imagery digitally, the data must be


available in a readable digital format suitable for storage on a computer.

• The other requirement for digital image processing is a computer system,


sometimes referred to as an image analysis system, with the appropriate
hardware and software to process the data.
Image file formats
 BSQ (Band Sequential Format):
 each line of the data followed immediately by the next line in the same spectral band. This

format is optimal for spatial (X, Y) access of any part of a single spectral band. Good for
multispectral images

 BIP (Band Interleaved by Pixel Format):


 the first pixel for all bands in sequential order, followed by the second pixel for all bands,

followed by the third pixel for all bands, etc., interleaved up to the number of pixels. This
format provides optimum performance for spectral (Z) access of the image data. Good for
hyperspectral images

 BIL (Band Interleaved by Line Format):


 the first line of the first band followed by the first line of the second band, followed by the first

line of the third band, interleaved up to the number of bands. Subsequent lines for each band
are interleaved in similar fashion. This format provides a compromise in performance between
spatial and spectral processing and is the recommended file format for most ENVI processing
tasks. Good for images with 20-60 bands
Band 2 Band 3 Band 4
20 50 50 90 90 120 150 100 120 103 210 250 250 190 245
76 66 55 45 120 176 166 155 85 150 156 166 155 415 220
80 80 60 70 150 85 80 70 77 135 180 180 160 170 200
100 93 97 101 105 103 90 70 120 133 200 0 123 222 215

Matrix notation for band 2


1,1,2 2,1,2 3,1,2 4,1,2 5,1,2
1,2,2 2,2,2 3,2,2 4,2,2 5,2,2
1,3,2 2,3,2 3,3,2 4,3,2 5,3,2
1,4,2 2,4,2 3,4,2 4,4,2 5,4,2

10 15 17 20 21 20 50 50 90 90 120 150 100 120 103 210 250 250 190 245
15 16 18 21 23 76 66 55 45 120 176 166 155 85 150 156 166 155 415 220
17 18 20 22 22 80 80 60 70 150 85 80 70 77 135 180 180 160 170 200
BIL
18 20 22 24 25 100 93 97 101 105 103 90 70 120 133 200 0 123 222 215

10 15 17 20 21 15 16 18 21 23 17 18 20 22 22 18 20 22 24 25
20 50 50 90 90 76 66 55 45 120 80 80 60 70 150 100 93 97 101 105
120 150 100 120 103 176 166 155 85 150 85 80 70 77 135 103 90 70 120 133
BSQ
210 250 250 190 245 156 166 155 415 220 180 180 160 170 200 200 0 123 222 215

10 20 120 210 15 50 150 250 17 50 100 250 20 90 120 190 21 90 103 245

BIP
15 76 176 156 16 66 166 166 18 55 155 155 21 45 85 415 23 120 150 220
17 80 85 180 18 80 80 180 20 60 70 160 22 70 77 170 22 150 135 200
18 100 103 200 20 93 90 0 22 97 70 123 24 101 120 222 25 105 133 215
 Band sequential (BSQ) format stores information
for the image one band at a time. In other words,
data for all pixels for band 1 is stored first, then
data for all pixels for band 2, and so on.
 Value=image(c, r, b)
 Band interleaved by pixel (BIP) data is similar to
BIL data, except that the data for each pixel is
written band by band. For example, with the same
three-band image, the data for bands 1, 2 and 3 are
written for the first pixel in column 1; the data for
bands 1, 2 and 3 are written for the first pixel in
column 2; and so on.
 Value=image(b, c, r)
 Band interleaved by line (BIL) data stores pixel
information band by band for each line, or row, of
the image. For example, given a three-band image,
all three bands of data are written for row 1, all
three bands of data are written for row 2, and so
on, until the total number of rows in the image is
reached.
 Value=image(c, b, r)
 Most of the image processing functions available in image analysis

systems can be categorized into the following four categories:

 Pre-processing
 Image Enhancement
 Image Transformation
 Image Classification and Analysis
Pre-processing involve operations that are essentially done prior to other image processing

techniques and includes image rectification and image registration. Image rectification

includes the radiometric and geometric correction of the satellite raw data. Image

registration is the superimposition of different images over one another, or onto a

standard map projection (List, 1993).

Image enhancement involves techniques for the creation of new modified images that contain

more information to ease the visual interpretation of the overall image &/or certain features.

Image transformation usually deals with the processing of multiple band images of the same

area together to generate a computed (transformed) image.

Image classification involves the categorization of the pixels of a scene into various thematic

groups, based on spectral response characteristics (Gupta, 2003).


Image Statistics

The extraction and calculation of fundamental univariate and multivariate

statistics of the multispectral satellite data is the initial step towards digital image

processing and analysis.

The statistical parameters usually calculated during the different phases of digital

image processing include: the maximum and minimum values of digital numbers

(DN) “Brightness value BV” for each band of the whole satellite data set, the mean,

the standard deviation, between band variance –covariance matrix, the

correlation matrix and the frequency of brightness values in each band. Some

valuable information is always deduced from these statistics.

For more information please refer to the image statistics handout


Univariate descriptive image statistics

 The mode is the value that occurs


most frequently in a distribution
and is usually the highest point on
the curve (histogram). It is
common, however, to encounter
more than one mode in a remote
sensing dataset.
 The median is the value midway
in the frequency distribution.
One-half of the area below the
distribution curve is to the right of
the median, and one-half is to the
left
 The mean is the arithmetic
average and is defined as the sum
of all brightness value n
observations divided by the
number of observations.  BV ik
k  i 1
n
 Min n

 BV  k 
2
 Max ik
 Variance vark  i 1

 Standard deviation n 1
 Coefficient of

sk   k  vark
variation (CV)
 Skewness
 Kurtosis
 Moment
k
CV 
k
Multivariate Image Statistics
 Remote sensing research is often concerned with the
measurement of how much radiant flux is reflected or emitted
from an object in more than one band. It is useful to compute
multivariate statistical measures such as covariance and
correlation among the several bands to determine how the
measurements covary. Variance–covariance and correlation
matrices are used in remote sensing principal components
analysis (PCA), feature selection, classification and
accuracy assessment.
Covariance
 The different remote-sensing-derived spectral measurements for each pixel
often change together in some predictable fashion. If there is no relationship
between the brightness value in one band and that of another for a given
pixel, the values are mutually independent; that is, an increase or decrease in
one band’s brightness value is not accompanied by a predictable change in
another band’s brightness value. Because spectral measurements of
individual pixels may not be independent, some measure of their mutual
interaction is needed. This measure, called the covariance, is the joint
variation of two variables about their common mean.

n n

 BV  BV SPkl
cov kl 
n ik il
SPkl   BVik BVil   i 1 i 1

i 1 n n 1
Correlation
To estimate the degree of interrelation between variables in a manner not influenced by
measurement units, the correlation coefficient, is commonly used. The correlation
between two bands of remotely sensed data, rkl, is the ratio of their covariance (covkl)
to the product of their standard deviations (sksl); thus:

cov kl
rkl 
sk sl
If we square the correlation coefficient (rkl), we obtain the sample coefficient of
determination (r2), which expresses the proportion of the total variation in the values
of “band l” that can be accounted for or explained by a linear relationship with the
values of the random variable “band k.” Thus a correlation coefficient (rkl) of 0.70
results in an r2 value of 0.49, meaning that 49% of the total variation of the values of
“band l” in the sample is accounted for by a linear relationship with values of “band
Pixel Band 1
(green)
Band 2
(red)
Band 3
(ni)
Band 4
(ni) example
(1,1) 130 57 180 205
(1,2) 165 35 215 255
(1,3) 100 25 135 195
SP12  (31,860) 
675232
(1,4) 135 50 200 220
5
(1,5) 145 65 205 235
540
cov12   135
4
Band 1 (Band 1 x Band 2) Band 2
130 7,410 57

165 5,775 35

100 2,500 25

135 6,750 50

145 9,425 65

675 31,860 232


Band 1 Band 2 Band 3 Band 4
Mean (k) 135 46.40 187 222
Variance (vark) 562.50 264.80 1007 570
(sk) 23.71 16.27 31.4 23.87
(mink) 100 25 135 195
(maxk) 165 65 215 255
Range (BVr) 65 40 80 60
Univariate statistics

Band 1 Band 2 Band 3 Band 4 Band Band Band 3 Band


1 2 4
Band 1 562.25 - - -
Band 1 - - - -
Band 2 135 264.80 - -
Band 2 0.35 - - -
Band 3 718.75 275.25 1007.50 -
Band 4 537.50 64 663.75 570 Band 3 0.95 0.53 covariance
- -
Band 4 0.94 0.16 0.87 -

Covariance Correlation coefficient


Mean

Variance

Multi-variate
Statistics -
Sum of
Products
Multi-variate Statistics - Sum
of Products

Multi-variate Statistics -
Sum of Squares

Multi-variate Statistics -
Sum of Covariance

Multi-variate Statistics –
Correlation coefficient
Image distortions are often classified into: geometric distortions
and radiometric distortions. Errors that occur during Satellite
data acquisition are of two types:

Internal errors created by the sensor. They are characteristically


systematic (predictable) and constant (Jensen, 1996). Their determination
is possible by prelaunch or in-flight calibration measurements.

External errors might occur due to perturbations and the modulation of


atmospheric and scene characteristics (Jensen, 1996).
Image Rectification:

This processing category includes those operations that are normally required
prior to the main data analysis and information extraction, its utilization is
necessary to restore the image to resemble the scene on the terrain by
compensating for image errors, noise, and geometric distortions introduced
during the scanning, recording and play back operations (Sabins, 1999) and are
generally grouped into radiometric or geometric corrections:

• Geometric corrections include correcting for geometric distortions due to


sensor-Earth geometry variations, and/or conversion of the data to real world
coordinates (e.g. latitude and longitude) on the Earth's surface.

• Radiometric corrections include correcting the data for sensor irregularities


and unwanted sensor or atmospheric noise, and converting the data so they
accurately represent the reflected or emitted radiation measured by the
sensor.
Image Rectification … … … … … … … .. for more clarification

Rectification is the process of transforming the data from one grid system into

another grid system using a geometric transformation (ERDAS IMAGINE, 2003).

In other words rectification is a procedure that leads to the adjustment of the grid of

image pixels onto a known projection and datum.

The main objective of rectification is the creation of a faithful representation of the

satellite Scene in terms of position and radiance. Rectification procedures are usually

carried out whenever the Satellite data are unprojected, needs to be reprojected, or

when geometric corrections are necessary.


Image Rectification … … … … … … … .. for more clarification

There are two methods of rectifying the Satellite imageries:

• the first by registering the data to another image that has been projected, and

is known as image-to-image registration or relative registration.

• the second is by assigning coordinates to the unprojected image to a hard or

digital georeferenced map. The first method which the second one is named

image-to-map registration or absolute registration.


Geometric Corrections

The corrections referred to here are the ones related to registering an image to a
specified geographic projection, or another image.
The geometric registration process involves
identifying image coordinates of several
clearly points, called ground control points
(or GCPs), in the distorted image (A - A1 to
A4), and matching them to their true
positions in ground coordinates (e.g.
latitude, longitude). The true ground
coordinates are typically measured from a
map (B - B1 to B4), in paper or digital format.
This is image-to-map registration.
Geometric Corrections ……….cont.

Once several well-distributed GCP pairs have been


identified, the coordinate information is processed
by the computer to determine the proper
transformation equations to apply to the original
image coordinates to map them into their new
ground coordinates.

Geometric registration may also be performed by


registering one (or more) images to another
image, instead of to geographic coordinates. This
is called image-to-image registration and is often

done prior to performing various image


transformation procedures.
Haze Correction

It is well known that the shorter wavelengths of the Ultra violet and the blue light suffer

scattering when passing through the atmosphere. This process is termed selective

scattering. Selective Scattering is caused by fumes and gasses such as nitrogen, Oxygen and

Carbon dioxide (and it causes haze and reduces their contrast ratio of images). The major

objectives of remote sensing are to record the amount of energy reflected from every

particular part of the earth's surface. Scene the selective scattering of EMR by the

atmosphere leads to an additive portion to the recorded radiation reflected from the

surface of the earth, correction for this portion is essential. This correction is termed haze

correction.
Haze Correction ………….cont.

Unfortunately, the accurate estimation of the added value needs precise data about the
atmospheric conditions e.g. temperature, relative humidity, atmospheric pressure and
visibility, which is always nearly impossible to be available for extendable areas.

Haze Correction is based on the assumption that, in each band in a scene under
investigation, there should be some pixels at or close to zero brightness value but that
atmospheric effect, has added a constant value to each pixel in each band (Richards & Jia,
1999). Completely dark pixels might correspond to area of deep, clear, open water bodies or
deep shadows (Gupta, 2003).

Such pixels record non-zero value. This value is usually reflected by the minimum DN in
each band. In this study the dark-pixel subtraction or the histogram minimum method
described by Crane (1971). Chavez et.al. (1977) had been adopted for the haze removal.

Thorough examination of the histograms of each band reveals a progressive offset increase
of the plots from band 7 through to band 1. The amount of this offset account for the path
radiance that is then subtracted from each pixel in each corresponding band.
BANDS DISPLAY

It is well known that the human eye can discriminate more color hues than grey
levels or tones. Colored images are more informative than black and white
images. It is also known that white color can result from the mixing of equal
amount of Red (R), Green (G) and Blue (B), where as other colors can be
obtained by mixing different percentage of R, G & B. This process of color
mixing is known as additive color mixing. To display a colored Satellite image
only three bands are needed. These bands should be of the same resolution and
registered to the same geographic system. Each of the three bands is displayed
in one of the three major color channels, Red, Green or Blue. The resultant is a
color composite image. RGB color composite images are simple and effective as
the mixing of the three primary colors (red, green & blue) can produce a wide
range of colors (Harris et. al, 1999).
BANDS DISPLAY
A natural color composite image result from red, green and blue wavelengths
bands when represented in R, G, B respectively. Other band combinations
produce false-color images. There is a simple rule for band combination is to
render the most informative band for a particular purpose in red, the next in
green and the least informative in blue (Drury, 1993).

Optimal geologic information depiction in a color composite image relies upon


the selection of three channels that are individually informative and collectively
minimally redundant (Crippen, 1989).

The six reflected bands of Landsat 7 (Bands 1, 2, 3, 4, 5 & 7) give the chance for
20 combinations, for each of which there are six RGB order (options).

E N O R M O U S C O N F U S I O N !!!!!!
Landsat7 band3 set in the Red Channel

Landsat7 band2 set in the Green Channel

Landsat7 band1set in the Blue Channel

TRUE COLOR COMPOSITE True color composite of bands3, 2, & 1 set in


R, G, B, channels respectively
Landsat7 band4 set in the Red Channel

Landsat7 band3 set in the Green Channel

Landsat7 band2 set in the Blue Channel

False color composite of bands4, 3, & 2 set in


FALSE COLOR COMPOSITE R, G, B, channels respectively
Landsat7 band7 set in the Red Channel

Landsat7 band4 set in the Green Channel

Landsat7 band1set in the Blue Channel

False color composite of bands7, 4, & 1 set in


FALSE COLOR COMPOSITE R, G, B, channels respectively
True color composite of bands3, 2, & 1 set in False color composite of bands4, 3, & 2 set in
R, G, B, channels respectively R, G, B, channels respectively

False color composite of bands7, 4, & 1 set in


R, G, B, channels respectively
Image Enhancement
The main purpose of image enhancement, is mainly to improve the
appearance of the imagery to assist the visual interpretation and improve the
process of information extraction. Examples of enhancement functions include
contrast stretching to increase the tonal distinction between various features in
a scene, and spatial filtering to enhance (or suppress) specific spatial patterns in
an image.
Image Enhancement … … … … … … … cont.

Image enhancement is the process of making an image more interpretable for a


particular application (Faust, 1989). It is obvious that image enhancement should be
carried out after image correction to avoid the enhancement of the different image
distortions and noises.

Image enhancement involves the modification of the original set of digital numbers.
Point operations modify the digital numbers of each pixel of the image under
enhancement, while local operations modify the digital numbers in the context of
the surrounding ones (Jehsen, 1996).

Enhancing certain features of an image may occur at the expense of other features
which may become relatively subdued (Gupta, 2003).
Image Enhancement … … … … … … … cont.

 Enhancements are used to make the image easier for visual interpretation and
understandable.

 Raw image data are usually dark and occupies only a limited portion of the data
dynamic range (0 – 255 brightness levels in the case of 8-bit data).

 Contrast enhancement involves changing the original values (DN or BV) so that
more of the available range is used, thereby increasing the contrast between targets
and their backgrounds. The key to understanding contrast enhancements is to
understand the concept of an image histogram.
Image Histogram

An image histogram is a graphical representation of the brightness values that


construct an image. The brightness values (i.e. 0-255) are displayed along the x-
axis of the graph. The frequency of occurrence of each of these values in the
image is shown on the y-axis.

By changing the range of digital values in an image, we can apply various


enhancements to the data.

There are different techniques and methods of enhancing contrast and detail in an
image.
Image Enhancement … … … … … … … cont.

The raw satellite data are usually dim and lack contrast, because natural

features have a low range of reflectance in a specific wave band (Gibson &

Power, 2000). The grey-scale contains 256 grey levels.

Digital numbers in satellite data do not encompass the entire dynamic range of

the grey-scale (0 – 255) and are often compressed into a small part of the

available range.

To view satellite images, the full grey-scale must be utilized by "stretching" the

digital numbers in the raw data. Data stretching is either linear or nonlinear.
Linear contrast stretching rescales the digital numbers range of the raw image to
fill the whole dynamic range (0 – 255) by assigning the 0 value to the digital number
with the minimum value and a value of 255 to the digital number with maximum value
and stretches the middle digital numbers accordingly.
This type of contrast stretching is largely dependent upon the statistics of the scene.
Mainly the mean and the standard deviation, in addition to the minimum and the
maximum values control the calculation. The output DN values are calculated by the
equation:
Image Enhancement … … … … … … … cont.

 Linear Contrast Stretching: this is the simplest form of enhancement. It


involves identifying the minimum and maximum values in the image
histogram and assigning these to the minimum and maximum values of
the dynamic range through a simple linear transformation.

 This form of enhancement works well if the histogram values are uniformly
distributed.
Non-Linear Contrast Enhancement
In these methods, the input and output data values follow a non-linear
transformation. The general form of the non-linear contrast
enhancement is defined by y = f (x), where x is the input data value and
y is the output data value. The non-linear contrast enhancement
techniques have been found to be useful for enhancing the color contrast
between the nearly classes and subclasses of a main class.
A type of non linear contrast stretch involves scaling the input data
logarithmically. This enhancement has greatest impact on the brightness
values found in the darker part of histogram. It could be reversed to
enhance values in brighter part of histogram by scaling the input data
using an inverse log function.
Non-linear contrast stretches is used when certain parts of the histogram needs
to be stretched more preferably than the rest of it. Logarithmic and exponential
contrast stretches in addition to the Gaussian and histogram equalization contrast
stretch represent the most popular types of the non – linear contrast stretches.

Histogram Equalization: this stretch assigns more display values (range) to the
frequently occurring portions of the histogram. In this way, the detail in these areas
will be better enhanced relative to those areas of the original histogram where
values occur less frequently.

Adaptive Enhancement: (can have other names): in this situation it may be desirable
to enhance the contrast in only a specific portion(s) of the image histogram. Each
portion can then be stretched independently from other portions.

Other forms of enhancement exit. For example applying a particular equation to the
brightness values such as square root etc … or an analysis can develop his own
equations.
Linear Contrast Stretch (source Lillesand and Kiefer, 1993)
Image Enhancement … … … … … … … cont.

Spatial Frequency
 The number of changes of brightness values per unit distance for any part of
an image.
 Defines the “roughness” of the brightness variations occurring in an image.
 Spatial frequency is the analog of the frequency of a wave in time.
Spatial Frequency in an Image

• High Frequency
 Occurs often at the feature-to-feature boundary.
 Enhancing the high spatial frequency of an image helps distinguishing the
different features and their boundaries.

• Low Frequency
 Occurs in homogenous areas, where no small features variations are seen.
 Enhancing the low frequency eliminates the unwanted noise from image
details.
Image Enhancement … … … … … … … cont.

Spatial Filtering
These operations that involve enhancing
rough textured features e.g.. edges.

The process usually involves a moving


window (or filter) with a specific size that
passes through an image to identify abrupt
changes.

A low pass filter leads to the supressing


of small details while highlighting larger
details. This process leads to the
smoothing of the image.

A high pass filter leads to the


enhancement of areas with abrupt changes
such as edges. The process leads to
sharpening of the image details.
Low frequency (smooth) is represented by the two major classes; the soil and the vegetation
High frequency (rough) is represented by the details in the two major classes:
- healthy vegetated fields, non vegetated fields, canals,… etc.
- Soils, drainage, tracks, …etc.
Image Enhancement … … … … … … … cont.

Spatial Convolution Filtering


Defining Filters
Low-Frequency Filtering
Low-pass filter is used to remove noise and unwanted
details from the image
High-Frequency Filtering
Application of High-Pass Filters
Edge Enhancement and Edge Detecting
Directional Edge Detectors
Vertical vs. Horizontal Edge Detectors
Image transformations
These are the operations which are normally applied only to a single
channel of data at a time, image transformations usually involve
combined processing of data from multiple spectral bands.

Arithmetic operations (i.e. subtraction, addition, multiplication,


division) are performed to combine and transform the original bands into
"new" images which better display or highlight certain features in the
scene.

Some of these operations including various methods of spectral or band


ratioing, and a procedure called principal components analysis (PCA)
which is used to more efficiently represent the information in
multichannel imagery.
Image Transformations ………………….cont.

Basic image transformations apply simple arithmetic


operations to the image data.

For example, image subtraction is often used to


identify changes that have occurred between images
collected on different dates. Typically, two images
which have been geometrically registered, are used
with the pixel (brightness) values in one image 1 being
The mathematical expression
subtracted from the pixel values in the other 2. of the ratio function is
Image ratioing is another useful transformation
BVi,j,r = BVi,j,k/BVi,j.l
usually applied to images. Ratio images are particularly
useful in remote sensing and highlight subtle features
In the spectral response of many land cover types.
Band ratioing
Sometimes differences in brightness values from identical surface materials are
caused by topographic slope and aspect, shadows, or seasonal changes in
sunlight illumination angle and intensity. These conditions may hamper the
ability of an interpreter or classification algorithm to identify correctly surface
materials or land use in a remotely sensed image. Fortunately, ratio
transformations of the remotely sensed data can, in certain instances, be applied
to reduce the effects of such environmental conditions. In addition to
minimizing the effects of environmental factors, ratios may also provide unique
information not available in any single band that is useful for discriminating
between soils and vegetation.

The mathematical expression of the ratio function is

BV = BV /BV
i,j,r i,j,k i,j.l
Ratio function is: BVi,j,r = BV /BV
i,j,k i,j.l

where BVi,j,r is the output ratio value for the pixel at row, i, column j; BVi,j,k is the
brightness value at the same location in band k, and BVi,j,l is the brightness value
in band l. Unfortunately, the computation is not always simple since BVi,j = 0 is
possible. However, there are alternatives. For example, the mathematical domain
of the function is 1/255 to 255 (i.e., the range of the ratio function includes all
values beginning at 1/255, passing through 0 and ending at 255). The way to
overcome this problem is simply to give any BVi,j with a value of 0 the value of 1.
Ratio images can be meaningfully interpreted because they can be directly related
to the spectral properties of materials. Ratioing can be thought of as a method of
enhancing minor differences between materials by defining the slope of spectral
curve between two bands. We must understand that dissimilar materials having
similar spectral slopes but different albedos, which are easily separable on a
standard image, may become inseparable on ratio images.
The figure below shows a situation where Deciduous and Coniferous Vegetation
crops out on both the sunlit and shadowed sides of a ridge. In the individual
bands the reflectance values are lower in the shadowed area and it would be
difficult to match this outcrop with the sunlit outcrop.
The ratio values, however, are nearly identical in the shadowed and sunlit areas
and the sandstone outcrops would have similar signatures on ratio images. This
removal of illumination differences also eliminates the dependence of
topography on ratio images.
The difference between deciduous and coniferous trees is that Conifer
trees have those prickly, thorny leaves which don't really fall off, while
deciduous trees shed their leaves in the fall, and the leaves are mainly
flat, wide, and thin. Most softwoods are coniferous while hardwoods are
deciduous. Coniferous trees stay green all year round, but deciduous
change leaf color.
Image Histogram

Image processing software can illustrate the distribution of the BV (DN) values within a

satellite image. The distribution of the BVs is displayed as a histogram diagram; where the

spread of the BV values (from the MinBV to the MaxBV in the data set of the image) is shown

in the X-axis; while the frequency or how many pixels in the image each BV value has is

represented in the Y-axis of the chart.

Histogram enables the digital image processor (analyst) to quickly access the type of

distribution maintained by the data (normal, bimodal or skewed). Histograms are useful

when images are enhanced.


Lookup Tables (LUT)

A lookup table (LUT) graphs the intensity of the input pixel value relative to the

output BV observed on the scene. The curve does not provide information about

the frequency of BVs, instead it provides information regarding the range

associated with the brightness levels.

An image enhancement can be modeled on a lookup table to better evaluate the

relationship between the unaltered raw data and the adjusted display data.
Scatter Plot Diagrams

DIP software can generate scatter plots that show the correlation between the

different bands of the same scene. The diagram graph the BVs of one band relative

to those of another band of the same scene. Bands that are highly correlated will

produce plots with linear relationship and slight deviation from the line. On the

contrary, bands that are not well correlated will lack linear relationship and the

distribution of the BVs of both band will distribute randomly in the plot.

Scatter plot diagrams allow for a quick assessment of the usefulness of particular

band combinations.
Image Classification and Analysis
Image classification and analysis operations are used to digitally identify and
classify pixels in the data. It is the categorization of all the pixels in the scene or
subset of a scene (within all the bands or a set of selected bands) into classes
(landcover/use classes, lithologies, types of vegetations…etc).
Image Classification is usually performed on multi-channel data sets (A) and
this process assigns each pixel in an image to a particular class, land cover or
theme (B) based on a certain set of statistical characteristics of the pixel
brightness values.
Image classification uses the spectral information represented by the BVs (DNs)
in one or more spectral bands, and attempts to classify each individual pixel based
on this spectral information. This type of classification is termed spectral
pattern recognition.

Digital conversion of image data to


information
Image Classification ……………………cont.

The resulting classified image is comprised of


a mosaic of pixels, each of which belong to a
particular theme, and is essentially a
thematic "map" of the original image.

Classification algorithms can be classified to two main types. Supervised


and Unsupervised Classification. In the supervised classification
method the analyst have a good knowledge of the image classes, selects
homogenous pixels of the same type (training areas or Regions of
interrest) that represent each class, and then use that information to
classify the image. The selection of the training areas is based on prior
knowledge and familiarity with the area and image. This is why we call it
supervised classification.
Image Classification ……………………cont.
Supervised vs. Unsupervised Classification
Use Unsupervised classification to cluster pixels in a dataset based on statistics only,
without any user-defined training classes. The unsupervised classification techniques
available are ISODATA and K-Means (ENVI4.8 HELP file)

ISODATA unsupervised classification calculates class means evenly distributed in the data
space then iteratively clusters the remaining pixels using minimum distance techniques.
Each iteration recalculates means and reclassifies pixels with respect to the new means.
Iterative class splitting, merging, and deleting is done based on input threshold
parameters. All pixels are classified to the nearest class unless a standard deviation or
distance threshold is specified, in which case some pixels may be unclassified if they do not
meet the selected criteria. This process continues until the number of pixels in each class
changes by less than the selected pixel change threshold or the maximum number of
iterations is reached.
ISODATA Method
ISODATA First Iteration
Unsupervised Cluster Busting
K-Means unsupervised classification calculates initial class means evenly distributed in the
data space then iteratively clusters the pixels into the nearest class using a minimum
distance technique. Each iteration recalculates class means and reclassifies pixels with
respect to the new means. All pixels are classified to the nearest class unless a standard
deviation or distance threshold is specified, in which case some pixels may be unclassified
if they do not meet the selected criteria. This process continues until the number of pixels
in each class changes by less than the selected pixel change threshold or the maximum
number of iterations is reached (ENVI 4.8 Help file).
Parallelepiped Classification
Minimum Distance to Means
Classification
Nearest Neighbor Classification
Maximum Likelihood
Classification
Image Classification ……………………cont.

The numerical information in all spectral bands for the pixels in the
training areas are used to "train" the computer to recognize spectrally
similar areas for each class. This is how the classification is achieved.

Similarity between any random pixel to the training areas is based on the
statistical information, such as mean, standard deviation, variance and
covariance. A maximum likelihood algorithm is widely used to calculate
the probability of a pixel belonging to a particular class. The probabilities
are calculated based on the above statistical information.

The set of the statistical information for each class is called “spectral
signature of the class”.
THANK YOU

You might also like