# August 2012 Master of Computer Application (MCA) – Semester 6 MC0086 – Digital Image Processing– 4 Credits (Book ID: B1007) Assignment

Set – 1 (60 Marks) Answer all Questions Each Question carries TEN Marks
1. Explain the process of formation of image in human eye. Answer
Image formation in the Eye The principal difference between the lens of the eye and an

ordinary optical lens is that the former is flexible. As illustrated in Fig. 2.1, the radius of curvature of the anterior surface of the lens is greater than the radius of its posterior surface. The shape of the lens is controlled by tension in the fibers of the ciliary body. To focus on distant objects, the controlling muscles cause the lens to be relatively flattened. Similarly, these muscles allow the lens to become thicker in order to focus on objects near the eye. The distance between the center of the lens and the retina (called the focal length) varies from approximately 17 mm to about 14 mm, as the refractive power of the lens increases from its minimum to its maximum. When the eye focuses on an object farther away than about 3 m, the lens exhibits its lowest refractive power. When the eye focuses on a nearby object, the lens is most strongly refractive. This information makes it easy to calculate the size of the retinal image of any object.

Figure 2.2: Graphical representation of the eye looking at a palm tree. Point c is the optical center of the lens.

In Fig. 2.2, for example, the observer is looking at a tree 15 m high at a distance of 100 m. If h is the height in mm of that object in the retinal image, the geometry of Fig. 2.2 yields 15/100=h/17 or h=2.55 mm. The retinal image is reflected primarily in the area of the fovea. Perception then takes place by the relative excitation of light receptors, which transform radiant energy into electrical impulses that are ultimately decoded by the brain.

Brightness Adaptation and Discrimination Since digital images are displayed as a discrete set of intensities, the eye’s ability to discriminate between different intensity levels is an important consideration in presenting image-processing results. The range of light intensity levels to which the human visual system can adapt is of the order of 1010 – from the scotopic threshold to the glare limit. Experimental evidence indicates that subjective brightness (intensity as perceived by the human visual system) is a logarithmic function of the light intensity incident on the eye. A plot of light intensity versus subjective brightness, illustrating this characteristic is shown in Fig. 2.3.

Figure 2.3: Range of subjective brightness sensations showing a particular adaptation level. The long solid curve represents the range of intensities to which the visual system can adapt. In photopic vision alone, the range is about 106. The transition from scotopic to photopic vision is gradual over the approximate range from 0.001 to 0.1 millilambert (–3 to –1 mL in the log scale), as the double branches of the adaptation curve in this range show.

2. Explain the process of formation of image in human eye. Answer
Image formation in the Eye The principal difference between the lens of the eye and an

ordinary optical lens is that the former is flexible. As illustrated in Fig. 2.1, the radius of curvature of the anterior surface of the lens is greater than the radius of its posterior surface. The shape of the lens is controlled by tension in the fibers of the ciliary body. To focus on distant objects, the controlling muscles cause the lens to be relatively flattened. Similarly, these muscles allow the lens to become thicker in order to focus on objects near the eye. The distance between the center of the lens and the retina (called the focal length) varies from approximately 17 mm to about 14 mm, as the refractive power of the lens increases from its minimum to its maximum. When the eye focuses on an object farther away than about 3 m, the lens exhibits its lowest refractive power. When the eye focuses on a nearby object, the lens is most strongly refractive. This information makes it easy to calculate the size of the retinal image of any object .

Figure 2.2: Graphical representation of the eye looking at a palm tree. Point c is the optical center of the lens.

In Fig. 2.2, for example, the observer is looking at a tree 15 m high at a distance of 100 m. If h is the height in mm of that object in the retinal image, the geometry of Fig. 2.2 yields 15/100=h/17 or h=2.55 mm. The retinal image is reflected primarily in the area of the fovea. Perception then takes place by the relative excitation of light receptors, which transform radiant energy into electrical impulses that are ultimately decoded by the brain. Brightness Adaptation and Discrimination Since digital images are displayed as a discrete set of intensities, the eye’s ability to discriminate between different intensity levels is an important consideration in presenting image-processing results. The range of light intensity levels to which the human visual system can adapt is of the order of 1010 – from the scotopic threshold to the glare limit. Experimental evidence indicates that subjective brightness (intensity as perceived by the human visual system) is a logarithmic function of the light intensity incident on the eye. A plot of light intensity versus subjective brightness, illustrating this characteristic is shown in Fig. 2.3.

Figure 2.3: Range of subjective brightness sensations showing a particular adaptation level. The long solid curve represents the range of intensities to which the visual system can adapt. In photopic vision alone, the range is about 106. The transition from scotopic to photopic vision is gradual over the approximate range from 0.001 to 0.1 millilambert (–3 to –1 mL in the log scale), as the double branches of the adaptation curve in this range show.

3. Explain different linear methods for noise cleaning? Noise Cleaning An image may be subject to noise and interference from several sources, including electrical sensor noise, photographic grain noise and channel errors. Image noise arising from a noisy sensor or channel transmission errors usually appears as discrete isolated pixel variations that are not spatially correlated. Pixels that are in error often appear visually to be markedly different from their neighbors. Linear Noise Cleaning Noise added to an image generally has a higher-spatialfrequency spectrum than the normal image components because of its spatial decorrelatedness. Hence, simple low-pass filtering can be effective for noise cleaning. We will now disucss convolution method of noise cleaning. A spatially filtered output image G(j,k) can be formed by discrete convolution of an input image F(m,n) with a L * L impulse response array H(j,k) according to the relation G(j,k)= ΣΣ F(m,n) H(m+j+C, n+k+C) where C=(L+1)/2 …… [Eq 4.8] For noise cleaning, H should be of low-pass form, with all positive elements. Several common pixel impulse response arrays of lowpass form are used and two such forms are given below.

These arrays, called noise cleaning masks, are normalized to unit weighting so that the noise-cleaning process does not introduce an amplitude bias in the processed image. Another linear noise cleaning technique Homomorphic Filtering. Homomorphic filtering (16) is a useful technique for image enhancement when an image is subject to multiplicative noise or interference. Fig. 4.9 describes the process.

Figure 4.9: Homomorphic Filtering The input image F(j,k) is assumed to be modeled as the product of a noise-free image S(j,k) and an illumination interference array I(j,k). Thus, F(j,k) = S(j,k) I(j,k) Taking the logarithm yields the additive linear result log{F(j, k)} = log{I(j, k)} + log{S(j, k) Conventional linear filtering techniques can now be applied to reduce the log interference component. Exponentiation after filtering completes the enhancement process.

4. Which are the two quantitative approaches used for the evaluation of image features? Explain.

There are two quantitative approaches to the evaluation of image features: prototype performance and figure of merit. In the prototype performance approach for image classification, a prototype image with regions (segments) that have been independently categorized is classified by a classification procedure using various image features to be evaluated. The classification error is then measured for each feature set. The best set of features is, of course, that which results in the least classification error. The prototype performance approach for image segmentation is similar in nature. A prototype image with independently identified regions is segmented by a segmentation procedure using a test set of features. Then, the detected segments are compared to the known segments, and the segmentation error is evaluated. The problems associated with the prototype performance methods of feature evaluation are the integrity of the prototype data and the fact that the performance indication is dependent not only on the quality of the features but also on the classification or segmentation ability of the classifier or segmenter. The figure-of-merit approach to feature evaluation involves the establishment of some functional distance measurements between sets of image features such that a large distance implies a low classification error, and vice versa. Faugeras and Pratt have utilized the Bhattacharyya distance figure-of-merit for texture feature evaluation. The method should be extensible for other features as well. The Bhattacharyya distance (B-distance for simplicity) is a scalar function of the probability densities of features of a pair of classes defined as

where x denotes a vector containing individual image feature measurements with conditional density p (x | S1).

5. Explain with diagram Digital image restoration model. In order to effectively design a digital image restoration system, it is necessary quantitatively to characterize the image degradation effects of the physical imaging system, the image digitizer and the image display. Basically, the procedure is to model the image degradation effects and then perform operations to undo the model to obtain a restored image. It should be emphasized that accurate image modeling is often the key to effective image restoration. There are two basic approaches to the modeling of image degradation effects: a priori modeling and a posteriori modeling. In the former case, measurements are made on the physical imaging system, digitizer and display to determine their response to an arbitrary image field. In some instances, it will be possible to model the system response deterministically, while in other situations it will only be possible to determine the system response in a stochastic sense. The posteriori modeling approach is to develop the model for the image degradations based on measurements of a particular image to be restored.

Figure 5.1: Digital image restoration model. Basically, these two approaches differ only in the manner in which information is gathered to describe the character of the image degradation. Fig. 5.1 shows a general model of a digital imaging system and restoration process. In the model, a continuous image light distribution C(x,y,t,) dependent on spatial coordinates (x, y), time (t) and spectral wavelength () is assumed to exist as the driving force of a physical imaging system subject to point and spatial degradation effects and corrupted by deterministic and stochastic disturbances. Potential degradations include diffraction in the optical system, sensor nonlinearities, optical system aberrations, film nonlinearities,

atmospheric turbulence effects, image motion blur and geometric distortion. Noise disturbances may be caused by electronic imaging sensors or film granularity. In this model, the physical imaging system produces a set of output image fields FO (i)( x ,y ,t j ) at time instant t j described by the general relation.

where OP { . } represents a general operator that is dependent on the space coordinates (x, y), the time history (t), the wavelength () and the amplitude of the light distribution (C). For a monochrome imaging system, there will only be a single output field, while for a natural color imaging system, FO (i)( x ,y ,t j ) may denote the red, green and blue tristimulus bands for i = 1, 2, 3, respectively. Multispectral imagery will also involve several output bands of data. In the general model of Fig. 5.1 each observed image field FO (i)( x ,y ,t j ) is digitized, to produce an array of image samples E S (i) ( m1 , m2 , t j ) at each time instant t j. The output samples of the digitizer are related to the input observed field by

6. Describe the basic concepts in Sampling and Quantization; and Linear and Non – Linear Operations of Image Sampling and Quantization. Answer - Basic Concepts in Sampling and Quantization To create a digital image, we need to convert the continuous sensed data into digital form. This involves two processes: sampling and quantization. The basic idea behind sampling and quantization is illustrated in Fig. 3.1. Fig. 3.1 (a) shows a continuous image, f (x, y), that we want to convert to digital form. An image may be continuous with respect to the x – and y-coordinates, and also in amplitude. To convert it to digital form, we have to sample the function in both coordinates and in amplitude. Digitizing the coordinate values is called sampling. Digitizing the amplitude values is called quantization.

Figure 3.1 generating a digital image. (a) Continuous image. (b) A scan line from A to B in the continuous image, used to illustrate the concepts of sampling and quantization. (c) Sampling and quantization (d) Digital scan line. The one dimensional function shown in the fig. 3.1 (b) is a plot of amplitude (gray level) values of the continuous image along the line segment AB in fig. 3.1 (a). The random variations are due to image noise. To sample this function, we take equally spaced samples along line AB, as shown in fig. 3.1(c). The location of each sample is given by

a vertical tick mark in the bottom part of the figure. The samples are shown as small white squares superimposed on the function. The set of these discrete locations gives the sampled function. However, the values of the samples still span (vertically) a continuous range of gray-level values. In order to from a digital function, the gray-level values also must be converted (quantized) into discrete quantities. The right side of Fig. 3.1(c) shows the gray-level scale divided into discrete levels, ranging form black to white. The vertical tick marks indicate the specific value assigned to each of the eight gray levels. The continuous gray levels are quantized simply by assigning one of the eight discrete gray levels to each sample. The assignment is made depending on the vertical proximity of a sample to a vertical tick mark. The digital samples resulting from both sampling and quantization are shown in Fig. 3.1 (d). Starting at the top of the image and carrying out this procedure line by line produces a two-dimensional digital image. Representation of Digital Images The result of sampling and quantization is a matrix of real numbers. We will use two principal ways in this book to represent digital images. Assume that an image f(x, y) is sampled so that the resulting digital image has M rows and N columns. The values of the coordinates (x, y) now become discrete quantities. For notational clarity and convenience, we shall use integer values for these discrete coordinates. Thus, the values of the coordinates at the origin are (x, y)=(0, 0).The next coordinate values along the first row of the image are represented as (x, y)=(0, 1). It is important to keep in mind that the notation (0, 1) is used to signify the second sample along the first row. It does not mean that these are the actual values of physical coordinates when the image was sampled. Fig. 3.3 shows the coordinate convention used to represent digital images. Spatial and Gray-Level Resolution Sampling is the principal factor determining the spatial resolution of an image. Basically, spatial resolution is the smallest discernible detail in an image. Suppose that we construct a chart with vertical lines of width W and with the space between the lines also having width W. A line pair consists of one such line and its adjacent space. Thus, the width of a line pair is 2W, and there are 1/2W line pairs per unit distance. A widely used definition of resolution is simply the smallest number of discernible line pairs per unit distance; for example, 100 line pairs per millimeter. Gray-level resolution similarly refers to the smallest discernible change in gray level, but, measuring discernible changes in gray level is a highly subjective process. We have considerable discretion regarding the number of samples used to generate a digital image, but this is not true for the number of gray levels. Due to hardware considerations, the number of gray levels is usually an integer power of 2, as mentioned in the previous section. The most common number is 8 bits, with 16 bits being used in some applications where enhancement of specific gray-level ranges is necessary. Sometimes we find systems that can digitize the gray levels of an image with 10 or 12 bits of accuracy, but these are exceptions rather than the rule. When an actual measure of physical resolution relating pixels and the level of detail they resolve in the original scene are not necessary, it is not uncommon to refer to an Llevel digital image of size M*N as having a Digital Image Processing Unit 3 Sikkim Manipal University Page No. 45

spatial resolution of M*N pixels and a gray-level resolution of L levels. We will use this terminology from time to time in subsequent discussions, making a reference to actual resolvable detail only when necessary for clarity. Fig. 3.4 shows an image of size 1024*1024 pixels whose gray levels are represented by 8 bits. The other images shown in Fig. 3.4 are the results of subsampling the 1024*1024 image. The subsampling was accomplished by deleting the appropriate number of rows and columns from the original image. For example, the 512*512 image was obtained by deleting every other row and column from the 1024*1024 image. The 256*256 image was generated by deleting every other row and column in the 512*512 image, and so on. The number of allowed gray levels was kept at 256. These images show the dimensional proportions between various sampling densities, but their size differences make it difficult to see the effects resulting from a reduction in the number of samples. The simplest way to compare these effects is to bring all the subsampled images up to size 1024*1024 by row and column pixel replication. The results are shown in Figs. 3.5(b) through (f). Fig. 3.4(a) is the same 1024*1024, 256-level image shown in Fig. 3.4; it is repeated to facilitate comparisons.

Figure 3.4 : A 1024*1024, 8-bit image subsampled down to size 32*32 pixels. The number of allowable gray levels was kept at 256. Compare Fig. 3.5(a) with the 512*512 image in Fig. 3.5(b) and note that it is virtually impossible to tell these two images apart. The level of detail lost is Digital Image Processing Unit 3 Sikkim Manipal University Page No. 46 simply too fine to be seen on the printed page at the scale in which these images are shown. Next, the 256*256 image in Fig. 3.5(c) shows a very slight fine checkerboard pattern in the borders between flower petals and the black background. A slightly more pronounced graininess throughout the image also is beginning to appear. These effects are much more visible in the 128*128 image in Fig. 3.5(d), and they become pronounced in the 64*64 and 32*32 images in Figs. 3.5(e) and (f), respectively.

Figure 3.5: (a) 1024*1024, 8-bit image. (b) 512*512 image resampled into 1024*1024 pixels by row and column duplication. (c) through (f) 256*256, 128*128, 64*64, and 32*32 images resampled into 1024*1024 pixels. 3.2.3 Aliasing and Moiré Patterns Functions whose area under the curve is finite can be represented in terms of sines and cosines of various frequencies. The sine/cosine component with the highest frequency determines the highest “frequency content” of the function. Suppose that this highest frequency is finite and that the function is of unlimited duration (these functions are called band-limited functions), then, the Shannon sampling theorem tells us that, if the function is sampled at a rate equal to or greater than twice its highest frequency, it is possible to recover completely the original function from its samples. If the function is undersampled, then a phenomenon called aliasing corrupts the sampled image. The corruption is in the form of additional frequency components being introduced into the sampled function. These are called aliased frequencies. Note that the sampling rate in images is the number of samples taken (in both spatial directions) per unit distance. As it turns out, except for a special case discussed in the following paragraph, it is impossible to satisfy the sampling theorem in practice. We can only work with sampled data that are finite in duration. We can model the process of converting a function of unlimited duration into a

function of finite duration simply by multiplying the unlimited function by a “gating function” that is valued 1 for some interval and 0 elsewhere. Unfortunately, this function itself has frequency components that extend to infinity. Thus, the very act of limiting the duration of a band-limited function causes it to cease being band limited, which causes it to violate the key condition of the sampling theorem. The principal approach for reducing the aliasing effects on an image is to reduce its high-frequency components by blurring the image prior to sampling. However, aliasing is always present in a sampled image. The effect of aliased frequencies can be seen under the right conditions in the form of so called Moiré patterns. There is one special case of significant importance in which a function of infinite duration can be sampled over a finite interval without violating the sampling theorem. When a function is periodic, it may be sampled at a rate equal to or exceeding twice its highest frequency, and it is possible to recover the function from its samples provided that the sampling captures exactly an integer number of periods of the function. This special case allows us to illustrate vividly the Moiré effect. Fig. 3.6 shows two identical periodic patterns of equally spaced vertical bars, rotated in opposite directions and then superimposed on each other by multiplying the two images. A Moiré pattern, caused by a breakup of the periodicity, is seen in Fig. 3.6 as a 2-D sinusoidal (aliased) waveform(which looks like a corrugated tin roof) running in a vertical direction. A similar pattern can appear when images are digitized (e.g., scanned) from a printed page, which consists of periodic ink dots.

Figure 3.6: Illustration of the Moiré pattern effect. Zooming and Shrinking Digital Images We conclude the treatment of sampling and quantization with a brief discussion on how to zoom and shrink a digital image. This topic is related to image sampling and quantization because zooming may be viewed as oversampling, while shrinking may be viewed as undersampling. The key difference between these two operations and sampling and quantizing an original continuous image is that zooming and shrinking are applied to a digital image. Zooming requires two steps: the creation of new pixel

locations, and the assignment of gray levels to those new locations. Let us start with a simple example. Suppose that we have an image of size 500*500 pixels and we want to enlarge it 1.5 times to 750*750 pixels. Conceptually, one of the easiest ways to visualize zooming is laying an imaginary 750*750 grid over the original image. Obviously, the spacing in the grid would be less than one pixel because we are fitting it over a smaller image. In order to perform gray-level assignment for any point in the overlay, we look for the closest pixel in the original image and assign its gray level to the new pixel in the grid. When we are done with all points in the overlay grid, we simply expand it to the original specified size to obtain the zoomed image. This method of gray-level assignment is called nearest neighbor interpolation. Linear and Nonlinear Operations A function or an operator H is linear if: H (a f + b g) = a H (f) + b H (g) Where f and g are images, and a and b are constants. The result of applying a linear operator to the sum of two images is identical to applying the operator to the images individually, multiplying the results by the appropriate constants, and then adding those results. For example, an operator whose function is to compute the sum of K images is a linear operator. An operator that computes the absolute value of the difference of two images is not. An operation is non-linear if the equation above is not satisfied.

August 2012 Master of Computer Application (MCA) – Semester 6 MC0086 – Digital Image Processing– 4 Credits (Book ID: B1007) Assignment Set – 2 (60 Marks)
Answer all Questions Each Question carries TEN Marks

1. Describe the following with respect to Digital Images: a. Elements of Visual Perception b. Light and the Electromagnetic Spectrum c. Image Sensing and Acquisition Answer a. Elements of Visual Perception In addition to foundation of mathematical and probabilistic formulations on which digital image processing is based, human intuition and analysis play a central role which depends on subjective, visual judgments. Hence, developing a basic understanding of human visual perception is the fundamental step. Our interest lies in the mechanics and parameters related to how images are formed in the eye and in learning the physical limitations of human vision in terms of factors that are also used in our work with digital images. Thus, factors such as how human and electronic imaging compare in terms of resolution and ability to adapt to changes in illumination are important from a practical point of view. b. Light and the Electromagnetic Spectrum In 1666, Sir Isaac Newton discovered that when a beam of sunlight is passed through a glass prism, the emerging beam of light is not white but consists instead of a continuous spectrum of colors ranging from violet at one end to red at the other. As shown in Fig. 2.8, the range of colors we perceive in visible light represents a very small portion of the electromagnetic spectrum. On one end of the spectrum are radio waves with

wavelengths billions of times longer than those of visible light. At the other end of the spectrum are gamma rays with wavelengths millions of times smaller than those of visible light.

The electromagnetic spectrum can be expressed in terms of wavelength, frequency, or energy. Wavelength (λ) and frequency (ν) are related by the expression λ = c / ν (2.3-1) where c is the speed of light (2.998*108 m/s).The energy of the various components of the electromagnetic spectrum is given by the expression E =h ν (2.3-2) where h is Planck’s constant. The units of wavelength are meters, with the terms microns (denoted μm and equal to 10–6 m) and nanometers (10–9 m) being used frequently. Frequency is measured in Hertz (Hz), with one Hertz being equal to one cycle of a sinusoidal wave per second. Electromagnetic waves can be visualized as propagating sinusoidal waves with wavelength l as shown Fig. 2.9, or they can be thought of as a stream of massless particles, each traveling in a wavelike pattern and moving at the speed of light. Each massless particle contains a certain amount (or bundle) of energy. Each bundle of energy is called a photon. We see from Eq. (2.3-2) that energy is proportional to frequency, so the higher-frequency (shorter wavelength) electromagnetic phenomena carry more energy per photon. Thus, radio waves have photons with low energies; microwaves have more energy than radio waves, infrared still more, then visible, ultraviolet, X-rays, and finally gamma rays, the most energetic of all. This is the reason that gamma rays are so dangerous to living organisms.

Light is a particular type of electromagnetic radiation that can be seen and sensed by the human eye. The visible band of the electromagnetic spectrum spans the range from approximately 0.43 μm (violet) to about 0.79 μm (red).For convenience, the color spectrum is divided into six broad regions: violet, blue, green, yellow, orange, and red. No color (or other component of the electromagnetic spectrum) ends abruptly, but rather each range blends smoothly into the next. The colors that humans perceive in an object are determined by the nature of the light reflected from the object. A body that reflects light and is relatively balanced in all visible wavelengths appears white to the observer. However, a body that favors reflectance in a limited range of the visible spectrum exhibits some shades of color. For example, green objects reflect light with wavelengths primarily in the 500 to 570 nm range while absorbing most of the energy at other wavelengths.

from each sensor by digitizing its response. In this section, we look at the principal modalities for image sensing and generation.

Image Acquisition using a Single Sensor Fig. 2.10 shows the components of a single sensor. The most common sensor of this type is the photodiode, which is constructed of silicon materials and whose output voltage waveform is proportional to light. The use of a filter in front of a sensor improves selectivity. For example, a green (pass) filter in front of a light sensor favors light in the green band of the color spectrum. As a consequence, the sensor output will be stronger for green light than for other components in the visible spectrum.

Figure 2.11: Combining a single sensor with motion to generate a 2-D image. In order to generate a 2-D image using a single sensor, there have to be relative displacements in both the x- and y-directions between the sensor and the area to be imaged. Figure 2.11 shows an arrangement used in high-precision scanning, where a film negative is mounted onto a drum whose mechanical rotation provides displacement

in one dimension. The single sensor is mounted on a lead screw that provides motion in the perpendicular direction. Since mechanical motion can be controlled with high precision, this method is an inexpensive (but slow) way to obtain high-resolution images. Other similar mechanical arrangements use a flat bed, with the sensor moving in two linear directions. These types of mechanical digitizers sometimes are referred to as microdensitometers. Another example of imaging with a single sensor places a laser source coincident with the sensor. Moving mirrors are used to control the outgoing beam in a scanning pattern and to direct the reflected laser signal onto the sensor. Image Acquisition using Sensor Strips A geometry that is used much more frequently than single sensors consists of an in-line arrangement of sensors in the form of a sensor strip. The strip provides imaging elements in one direction. Motion perpendicular to the strip provides imaging in the other direction, as shown in Fig. 2.12.

Figure 2.12: Image acquisition using a linear sensor strip. This is the type of arrangement used in most flatbed scanners. Sensing devices with 4000 or more in-line sensors are possible. In-line sensors are used routinely in airborne imaging applications, in which the imaging system is mounted on an aircraft that flies at a constant altitude and speed over the geographical area to be imaged. Onedimensional imaging sensor strips that respond to various bands of the electromagnetic spectrum are mounted perpendicular to the direction of flight. The imaging strip gives one line of an image at a time, and the motion of the strip completes the other dimension of a two-dimensional image. Lenses or other focusing schemes are used to project the area to be scanned onto the sensors.

2. Describe various texture features of historical and practical importance. Texture Features There is no commonly accepted quantitative definition of visual texture. As a consequence, researchers seeking a quantitative texture measure have been forced to search intuitively for texture features, and then attempt to evaluate their performance. The following subsections describe several texture features of historical and practical importance. A.Fourier Spectra Methods Several studies have considered textural analysis based on the Fourier spectrum of an image region, as discussed in Section 7.3. Because the degree of texture coarseness is proportional to its spatial period, a region of coarse texture should have its Fourier spectral energy concentrated at low spatial frequencies. Conversely, regions of fine texture should exhibit a concentration of spectral energy at high spatial frequencies. Although this correspondence exists to some degree, difficulties often arise because of spatial changes in the period and phase of texture pattern repetitions. Experiments have shown that there is considerable spectral overlap of regions of distinctly different natural texture, such as urban, rural and woodland regions extracted from aerial photographs. On the other hand, Fourier spectral analysis has proved successful in the detection and classification of coal miner‟s black lung disease, which appears as diffuse textural deviations from the norm. B.Edge Detection Methods Rosenfeld and Troy have proposed a measure of the number of edges in a neighborhood as a textural measure. As a first step in their process, an edge map array E(j, k) is produced by some edge detector such that E(j, k) = 1 for a detected edge and E(j, k) = 0 otherwise. Usually, the detection threshold is set lower than the normal setting for the isolation of boundary points. This texture measure is defined as

Where W = 2w + 1 is the dimension of the observation window. A variation of this approach is to substitute the edge gradient G(j, k) for the edge map array in Eq.6.

C.Autocorrelation Methods The autocorrelation function has been suggested as the basis of a texture measure. Although it has been demonstrated in the preceding section that it is possible to generate visually different stochastic fields with the same autocorrelation function, this does not necessarily rule out the utility of an autocorrelation feature set for natural images. The autocorrelation function is defined as

for computation over a W X W window with - T m, n T pixel lags. Presumably, a region of coarse texture will exhibit a higher correlation for a fixed shift than will a region of fine texture. Thus, texture coarseness should be proportional to the spread of the autocorrelation function. Faugeras and Pratt have proposed the following set of autocorrelation spread measures:

3. Explain different methods for wavelet transform in two dimensions. Wavelet transforms in two dimensions For many applications, in particular for image processing and image compression, we need to generalize wavelets transforms to two dimensions. First, we will consider how to perform a wavelet transform of the pixel values in a two-dimensional image. Then the scaling functions and wavelets that form a two dimensional wavelet basis. We will use the Haar basis as a simple example, but it will apply to other bases as well. There are two ways we can generalize the one-dimensional wavelet transform to two dimensions, standard and non-standard decomposition (since a multi-dimensional wavelet transform is frequently referred to in the literature as a wavelet decomposition). Standard decomposition To obtain the standard decomposition of an image, we first apply the one-dimensional wavelet transform to each row of pixel values. This operation gives us an average value along with detail coefficients for each row. Next, we treat these transformed rows as if they were themselves an image, and apply the one-dimensional transform to each column. The resulting values are all detail coefficients except for a single overall average coefficient. Non-Standard decomposition The second type of two-dimensional wavelet transform, called the non-standard decomposition, alternates between operations on rows and columns. First, we perform one step of horizontal pairwise averaging and differencing on the pixel values in each row of the image. Next, we apply vertical pairwise averaging and differencing to each column of the result. To complete the transformation, we repeat this process recursively on the quadrant containing averages in both directions.

4. Discuss Orthogonal Gradient Generation for first order derivative edge detection. First-Order Derivative Edge Detection There are two fundamental methods for generating first-order derivative edge gradients. One method involves generation of gradients in two orthogonal directions in an image; the second utilizes a set of directional derivatives. We will be discussing the first method. Orthogonal Gradient Generation An edge in a continuous domain edge segment F(x,y) can be detected by forming the continuous one-dimensional gradient G(x,y) along a line normal to the edge slope, which is at an angle Θ with respect to the horizontal axis. If the gradient is sufficiently large (i.e., above some threshold value), an edge is deemed present. The gradient along the line normal to the edge slope can be computed in terms of the derivatives along orthogonal axes according to the following

…………(8.3a) Fig. 8.5 describes the generation of an edge gradient in the discrete domain in terms of a row gradient and a column gradient. The spatial gradient amplitude is given by

Figure 8.5: Orthogonal gradient generation For computational efficiency, the gradient amplitude is sometimes approximated by the magnitude combination

The orientation of the spatial gradient with respect to the row axis is

The remaining issue for discrete domain orthogonal gradient generation is to choose a good discrete approximation to the continuous differentials of Eq. 8.3a. The simplest method of discrete gradient generation is to form the running difference of pixels along rows and columns of the image. The row gradient is defined as

Diagonal edge gradients can be obtained by forming running differences of diagonal pairs of pixels. This is the basis of the Roberts cross-difference operator, which is defined in magnitude form as

and in square-root form as

Prewitt has introduced a pixel edge gradient operator described by the pixel numbering convention of Fig. 8.6. The Prewitt operator square root edge gradient is defined as

where K = 1. In this formulation, the row and column gradients are normalized to provide unit-gain positive and negative weighted averages about a separated edge position.

Figure 8.6: Numbering convention for 3 × 3 edge detection operators. The Sobel operator edge detector differs from the Prewitt edge detector in that the values of the north, south, east and west pixels are doubled (i.e., K = 2). The motivation for this weighting is to give equal importance to each pixel in terms of its contribution to the spatial gradient. Fig. 8.7 shows examples of the Prewitt and Sobel gradients of the peppers image.

(a)Original Image

The row and column gradients for all the edge detectors mentioned previously in this subsection involve a linear combination of pixels within a small neighborhood. Consequently, the row and column gradients can be computed by the convolution relationships.

where HR(j,k) and HC (j,k) are 3 * 3 row and column impulse response arrays, respectively, as defined in Fig. 8.7.

Figure 8.7: Impulse response arrays for 3 × 3 orthogonal differential gradient edge operators.

5. Explain about the Region Splitting and merging with example. Region Splitting and Merging Sub-divide an image into a set of disjoint regions and then merge and/or split the regions in an attempt to satisfy the conditions stated in section 10.3.1. Let R represent the entire image and select predicate P. One approach for segmenting R is to subdivide it successively into smaller and smaller quadrant regions so that, for ant region, P() = TRUE. We start with the entire region. If then the image is divided into quadrants. If P is FALSE for any quadrant, we subdivide that quadrant into sub quadrants, and so on. This particular splitting technique has a convenient representation in the form of a so called quad tree (that is, a tree in which nodes have exactly four descendants), as shown in Fig. (10.3.3) 10.4. The root of the tree corresponds to the entire image and that each node corresponds to a subdivision. In this case, only was sub divided further.

6. Explain the concepts of point and line detection. Line and Spot Detection A line in an image could be considered to be composed of parallel, closely spaced edges. Similarly, a spot could be considered to be a closed contour of edges. This method of line and spot detection involves the application of scene analysis techniques to spatially relate the constituent edges of the lines and spots. We will consider only small-scale models of lines and edges and to apply the detection methodology developed previously for edges. For the unit-width line models, line detection can be accomplished by threshold detecting a line gradient

where Hm(j, k) is a 3 × 3 line detector impulse response array corresponding to a specific line orientation. Unit-width step spots can be detected by thresholding a spot gradient

where H(j,k) is an impulse response array chosen to accentuate the gradient of a unitwidth spot. Cook and Rosenfeld (43) and Zucker et al. (44) have suggested several algorithms for detection of large spots. In one algorithm, an image is first smoothed with a W × W low-pass filter impulse response array. Then the value of each point in the averaged image is compared to the average value of its north, south, east and west neighbors spaced W pixels away. A spot is marked if the difference is sufficiently large. A similar approach involves formation of the difference of the average pixel amplitude in a W × W window and the average amplitude in a surrounding ring region of width W.