Professional Documents
Culture Documents
INTRODUCTION
1.1. BASICS OF IMAGE PROCESSING
An image is a two-dimensional picture, which has a similar appearance to
some subject usually a physical object or a person. Image is a two-dimensional, such
as a photograph, screen display, and as well as a three-dimensional, such as a statue.
They may be captured by optical devices such as cameras, mirrors, lenses, telescopes,
microscopes, etc. and natural objects and phenomena, such as the human eye or water
surfaces.
1.1.1. Image
In simple words an image is a representation of a real scene, either in black
and white or in color. Technically an image may be defined as a two dimensional
function f(x,y) where x and y are spatial (plane) coordinates (x,y) is called the
intensity or gray level of the image at that point.
1.1.2. Types of Images
There are several standard types of images. They are
1) Binary
2) Gray Scale
3) Color
Binary Image
Binary images are the simplest type of images and can take on two values,
typically black and white, or 0 and 1. A binary image is referred to as a 1-bit image
because it takes only 1 binary digit to represent each pixel. These types of images are
frequently used in applications where the only information required is general shape
or outline.
Binary images are often created from the gray-scale images via a threshold
operation, where every pixel above the threshold value is turned white (1), and those
below it turned black (0). In fig. 1.1 a binary image is shown.
1
Typical color images are represented as red, green, and blue (RGB images).
Using the 8-bits monochrome standard as a model, the corresponding color image
would have 24-bits/pixels (8-bits for each of the three color bands red, green, and
blue). Fig. 1.3 shows the RGB image.
IMAGE
OUTPUT
(IMAGE)
PROCESSING
(IMAGE/DATA)
1.2.5. Neighbourhood
A pixel P at coordinates (x,y) has four horizontal and vertical neighbours
whose coordinates are given by (x+1, y), (x-1, y), (x, y+1), (x, y-1).This set of pixels
called the 4-neighbors of P is denoted by N4(P).Each pixel is a unit distance from
(x,y), and some of the neighbours of P lie outside the digital image if (x,y) is on the
border of the image. The four diagonal neighbours of P have coordinates (x+1, y+1),
(x+1, y-1), (x-1, y+1), (x-1, y-1)are denoted by Ng (P).
NW
SW
NE
SE
1.3. Background
The term inpainting comes from art restoration, where it is also called retouching. The term inpainting is borrowed from paper art, where restoration artists are
tasked with restoring faded and damaged paintings. Medieval artwork started to be
restored as early as the Renaissance, the motives being often as much to bring
medieval pictures up to date" as to fill-in any gaps. The need to retouch the image in
an unobtrusive way extended naturally from paintings to photography and film. The
purposes remained the same: to revert deterioration (e.g. scratches and dust spots), or
to add or remove elements. In art however, the major concern is to hide the damage in
whichever way complements the existing pigments and image the best, rather than
repaint the damage parts of the painting since erasing paintings is generally not an
option
Image retouching ranges from the restoration of paintings to scratched
photographs or films to the removal or replacement of arbitrary objects in images.
6
1.4.
Digital Inpainting
The idea of object removal is to remove objects from digital photographs and
fill the hole with the information extracted from the surrounding area. In this work,
Super-resolution based Inpainting is used for object removal in Digital photographs.
Inpainting vs Denoising
Image inpainting is different than image denoising. Image inpainting is an
iterative method for repairing damaged pictures or removing unnecessary elements
from pictures. Classical image denoising algorithms don't apply to image inpainting.
In common image enhancement applications, the pixels contain both information
about real data and the noise, while in image inpainting, there is no significant
information in the region to be inpainted. The information is mainly in the regions
surrounding the areas to be inpainted. Another difference lies within the size of the
data to be processed, the region of missing data in inpainting is usually large like long
cracks in photographs, superimposed large fonts, and so on.
Image processing
f
Fig 1.8: Linear Transformation through an Image Processor f
and i is the transformed image (i.e., the digitally inpainted image). The image
processor can be looked upon as a function f as follows:
f : io i i.e.
f (io ) i
1.7.
Overview
Although tremendous progress has been made in the past years on inpainting,
difficulties exist when the hole or the area of object to be removed is very large and
the computational time required in general is high. These two problems are addressed
by considering a two-step or hierarchical approach in which inpainting is performed
on a input image and a super resolution algorithm is used to construct a high
resolution (HR) image. Super-resolution (SR) imaging aims to overcome or
compensate the limitation or shortcomings of the image acquisition device/system
and/or possibly ill-posed acquisition conditions to produce a higher-resolution image
based on a set of images that were acquired from the same scene. With rapid progress
in image processing for visual communications and scene understanding, there is a
strong demand for providing the viewer with high-resolution imaging not only for
providing better visualization (fidelity issue) but also for extracting additional
information details (recognition issue). A HR image makes it easy to achieve a better
classification of regions in a multi-spectral remote sensing image or to assist
radiologist for making diagnosis based on a medical imagery.
1.8.
Organization of Thesis
Remaining portion of this work covers the following: Chapter 2 describes
1.9.
Resources Required
10
Chapter-2
LITERATURE SURVEY
2.1. Types of Inpainting
Mainly three groups of image inpainting algorithms can be found. The first
one to be noted is structural inpainting, the second one is texture inpainting and the
last one is a combination of these two techniques. All these inpainting methods have
one thing in common i.e. they use the information of the known or undestroyed image
areas in order to fill the gap.
Structural inpainting
Structural inpainting uses geometric approaches for filling in the missing information
in the region which should be inpainted. These algorithms focus on the consistency of
the geometric structure.
Textural inpainting
Like everything else the structural inpainting methods have both, advantages and
disadvantages. The main problem is that all the structural inpainting methods are not
able to restore texture. Texture has a repetitive pattern which means that a missing
portion cannot be restored by continuing the level lines into the gap.
Combined Structural and Textural inpainting
Combined structural and textural inpainting approaches simultaneously try to perform
texture and structure filling in regions of missing image information. Most parts of an
image consist of texture and structure. The boundaries between image regions
accumulate structural information which is a complex phenomenon. This is the result
when blending different textures together. That is why, the state of the art inpainting
method attempts to combine structural and textural inpainting.
into the missing region at the pixel level. Basically these algorithms are based on
theory of variational method and Partial Differential equation (PDE). The diffusionbased Inpainting algorithm produces superb results or filling the non-textured or
relatively smaller missing region. The drawback of the diffusion process is it
introduces some blur, which becomes noticeable when filling larger regions. All the
PDE based in painting models are more suitable for completing small, non-textured
target region.
2.2.2.Texture Synthesis Based Inpainting
Texture synthesis based algorithms are one of the earliest methods of image
Inpainting. And these algorithms are used to complete the missing regions using
similar neighbourhoods of the damaged pixels. The texture synthesis algorithms
synthesize the new image pixels from an initial seed. And then strives to preserve the
local structure of the image . All the earlier Inpainting techniques utilized these
methods to fill the missing region by sampling and copying pixels from the
neighbouring area. For e. g, Markov Random Field (MRF) is used to model the local
distribution of the pixel. And new texture is synthesized by querying existing texture
and finding all similar neighbourhoods. Their differences exist mainly in how
continuity is maintained between existing pixels and Inpainting hole. The main
objective of texture synthesis based inpainting is to generate texture patterns, which is
similar to a given sample pattern, in such a way that the reproduced texture retains the
statistical properties of its root texture .
2.2.3. PDE based Inpainting
This algorithm is the iterative algorithm. The main idea behind this algorithm is to
continue geometric and photometric information that arrives at the border of the
occluded area into area itself . This is done by propagating the information in the
direction of minimal change using isophote lines. This algorithm will produce good
results if missed regions are small one. But when the missed regions are large this
algorithm will take so long time and it will not produce good results. Then inspired by
this work proposed the Total Variational (TV) Inpainting model. This model uses
Euler-Lagrange equation and anisotropic diffusion based on the strength of the
isophotes. This model performs reasonably well for small regions and noise removal
applications. But the drawback of this method is that this method neither connects
broken edges nor greats texture patterns. These algorithms were focused on
12
maintaining the structure of the Inpainting area. And hence these algorithms produce
blurred resulting image. Another drawback of these algorithms is that the large
textured regions are not well reproduced.
2.3. SUPER-RESOLUTION
Super-resolution (SR) imaging aims to overcome or compensate the limitation
or shortcomings of the image acquisition device/system and/or possibly ill-posed
acquisition conditions to produce a higher-resolution image based on a set of images
that were acquired from the same scene. With rapid development and deployment of
image processing for visual communications and scene understanding, there is a
strong demand for providing the viewer with high-resolution imaging not only for
providing better visualization (fidelity issue) but also for extracting additional
information details (recognition issue). For examples, a high resolution image is
beneficial to achieve a better classification of regions in a multi-spectral remote
sensing image or to assist radiologist for making diagnosis based on a medical
imagery. In video surveillance systems, higher-resolution video frames are always
welcomed for more accurately identifying the objects and persons of interest.
The most direct approach on obtaining higher-resolution images is to improve
the image acquisition device (e.g., digital camera) by reducing the pixel size on the
sensor (e.g., charge-coupled device). However, there is a limitation in reducing the
sensors pixel size in the sensor technology. When the sensors pixel size becomes too
13
small, the captured image quality will be inevitably degraded. This is due to the fact
that the noise power remains roughly the same, while the signal power decreases
proportional to the sensors pixel size reduction. Furthermore, higher cost is required
to increase the chip size. Owing to the above-mentioned, the SR image processing
becomes a promising alternative. The SR imaging research has grown very rapidly,
after it was first addressed by Tsai and Huang in 1984. In view of this, the purpose of
this paper is to provide a comprehensive and updated survey for the SR research
literature and contribute several inspirations for future SR research.
To understand the SR imaging, several fundamental concepts are required to
be clarified. First, it is important to note that an images resolution is fundamentally
different from its physical size. In our context, the objective of SR imaging is to
produce an image with a clearer content from its low-resolution counterpart rather
than simply achieving a larger size of image In other words, the main goal and the
first priority of super-resolution imaging is to fuse the contents of multiple input
images in order to produce one output image containing with more clear and detailed
contents. The physical size of the output image (in terms of total number of pixels)
could be the same as any one of the input images or subject to further enlargement
using an image interpolation method. Second, in our context, the term resolution of
super-resolution is referred to the spatial resolution of the image, not the temporal
resolution of the image sequence. The latter is commonly expressed in terms of the
number of frames captured per second (i.e., frame rate). Third, it is worthwhile to note
that the term super-resolution has been used in other research areas as well. For
example, in the field of optics, super-resolution refers to a set of restoration
procedures that seek to recover the information beyond the diffraction limit. In
another example on the scanning antenna research, the super-resolution technique is
exploited to resolve two closely spaced targets when a one-dimensional stepped
scanning antenna is used.
The limitation of SR computation mainly comes from the following factors
interpolation error, quantization error, motion estimation error and optical-blur. Baker
and Kanade showed that, for a sufficiently large resolution enhancement factor, any
smoothness prior image model will result in reconstructions with very little highfrequency content. Lin et al. studied a numerical perturbation model of
reconstruction-based SR algorithms for the case of translational motion and the
learning-based SR algorithms. Robinson and Milanfar analysed this issue using
14
that is lost or degraded during the image acquisition process. This is motivated by that
the wavelet transform provides a powerful and efficient multi-scale representation of
the image for recovering the high-frequency information. These approaches typically
treat the observed low-resolution images as the low-pass filtered sub bands of the
unknown wavelet-transformed high-resolution image. The aim is to estimate the finer
scale sub band coefficients, followed by applying the inverse wavelet transform, to
produce the high-resolution image. To be more specific, take the 2 2 SR
computation as an example. The low-resolution images are viewed as the
representation of wavelet coefficients after several levels (say, N levels) of
decomposition. Then, the high-resolution image can be produced by estimating the (N
+ 1)th scale wavelet coefficients, followed by applying the inverse wavelet
decomposition. In Ei-Khamy et al. proposed to first register multiple low-resolution
images in the wavelet domain, then fuse the registered low-resolution wavelet
coefficients to obtain a single image, followed by performing interpolation to get a
higher-resolution image. Ji and Fermuller proposed a robust wavelet SR approach to
handle the error incurred in both the registration computation and the blur
identification computation.
algorithm, where each unknown pixel is assigned with an intensity value that is same
as its neighbouring pixels. But this method tends to produce images with a blocky
appearance. Ur and Gross performed a non-uniform interpolation of a set of spatially
shifted low-resolution images by utilizing the generalized multichannel sampling
theorem. The advantage of this approach is that it has low computational load, which
is thus quite suitable for real-time applications. However, the optimality of the entire
reconstruction process is not guaranteed, since the interpolation errors are not taken
into account. Bose and Ahuja
estimate the intensity value at each pixel position of the high-resolution image via a
polynomial approximation using the pixels in a defined neighbourhood of the pixel
position under consideration. Furthermore, the coefficients and the order of the
polynomial approximation are adaptively adjusted for each pixel position. Three steps
that are conducted iteratively. Irani and Peleg proposed an iterative backprojection
(IBP) algorithm, where the high-resolution image is estimated by iteratively
projecting the difference between the observed low-resolution images and the
simulated low resolution images. However, this method might not yield unique
solution due to the ill-posed nature of the SR problem. A projection onto convex sets
(POCS) was proposed by Patti and Tekalp to develop a set-theoretic algorithm to
produce the high-resolution image that is consistent with the information arising from
the observed low-resolution images and the prior image model. These information are
associated with the constraint sets in the solution space; the intersection of these sets
represents the space of permissible solutions. By projecting an initial estimate of the
unknown high-resolution image onto these constraint sets iteratively, a fairly good
solution can be obtained. This kind of method is easy to be implemented; however, it
does not guarantee uniqueness of the solution. Furthermore, the computational cost of
this algorithm is very high.
the unknown. Then, the unknown high-resolution image can be estimated via some
statistics of a probability distribution of the unknown high-resolution image, which is
established by applying Bayesian inference to exploit the information provided by
both the observed low-resolution images and the prior knowledge of the unknown
high-resolution image numerous regularization based SR algorithms have been
developed for addressing this issue.
18
Chapter 3
INPAINTING ALGORITHM
The task of object removal becomes complicated when area of object to be
removed or inpainted becomes large. So far a number of methods for inpainting had
been proposed.
In this method, a new framework for inpainting is proposed which is a
combination of low resolution picture inpainting and a super resolution algorithm
based on multiscale morphology.
This process of inpainting is mainly divided into two sequential steps. The
first one is a patch sampling method known as Exemplar based inpainting. The
inpainting is performed on a LR (resized) version of an input image. This is because a
low resolution picture is less contaminated by noise and is composed of main scene
structures. Also, as the picture to inpaint is smaller than the original one, the
computation time is reduced compared to the one necessary to inpaint the original
image. The inpainted version of the image obtained is given as input to the Super
Resolution algorithm. Its goal is to enhance the resolution of the image. Fig. 1 shows
the frame work of the proposed method which is as follows:
1) A low resolution picture is built from original image
2) An exemplar based inpainting algorithm is applied to remove the object.
3) Super Resolution algorithm is applied to the input image.
Original
Image
LR
image
Exemplar
based
Inpainting
Super
Resolution
using
Morphologic
Regularization
Image mask
19
Output
(Inpainted)
Image
inward as the algorithm progresses, and so we also refer to it as the fill front. The
source region, , which remains fixed throughout the algorithm, provides samples
used in the filling process.
We now focus on a single iteration of the algorithm to show how structure and
textures are adequately handled by exemplar based synthesis. Suppose that the square
template p centred at the point p (fig. 3.2b), is to be filled. The best-match
sample from the source region comes from the patch p , which is most similar to
those parts that are already filled in p . In the example in fig. 3.2b, we see that if p
lies on the continuation of an image edge, the most likely best matches will lie along
the same (or a similarly coloured) edge (e.g., q and q in fig.3.2c).
1
11
maintained the same orientation as in the source region. In this work we focus on a
patch-based filling approach (as opposed to pixel-based). Furthermore, we note that
patch based filling improves the accuracy of the propagated structures.
20
11
21
Onion peel
Desiderata
Fig.3.3. The importance of the filling order when dealing with concave
target regions. (a) A diagram showing an image and a selected target
region (inwhite). The remainder of the image is the source. (b,c,d)
Different stages in the concentric-layer filling of the target region. (d)
The onion-peel approach produces artefacts in the synthesized horizontal
structure. (b,c,d) Filling the target region by an edge-driven filling
order achieves the desired artefact-free reconstruction. (d) The final
edge-driven reconstruction, where the boundary between the two
background image regions has been reconstructed correctly.
order in which the filling process proceeds. Furthermore, we list a number of desired
properties of the ideal filling algorithm. A comparison between the standard
concentric layer filling(onion-peel) and the desired filling behaviour is illustrated in
fig. 3.3. Figures 3.3 b, c, d show the progressive filling of a concave target region via
an anti-clockwise onion-peel strategy. As it can be observed, this ordering of the filled
patches produces the horizontal boundary between the background image regions to
be unexpectedly reconstructed as a curve.
A better filling algorithm would be one that gives higher priority of synthesis
to those regions of the target area which lie on the continuation of image structures, as
shown in figs. 3.3b,c,d. Together with the property of correct propagation of linear
structures, the latter algorithm would also be more robust towards variations in the
22
calculate the priority of patches in the filling region. The priority of a patch centered
23
on p is calculated using confidence and data terms [2]. The priority is given by the
product of these two terms.
P(p) = C(p) D(p).
(1)
Where C(p) is the confidence term and D(p) is the data term[2].
C (p)
q p (
C (q)
, D (p)
I p .n p
is the area of
(2)
p , n p is a unit vector
p , and
p I , .
24
2)
each patch in the filling region are computed. Hence the filling order is determined.
The patch with highest priority p is found and it is filled with the information (data)
from the source region. In traditional inpainting techniques, pixel value information is
propagated via diffusion which leads to smoothness and blurring of image [6].
Propagation of image texture is done by direct sampling of the source region [5]. The
patch
q which
found .
Where
d ( p , q )
arg min
d ( p , q )
(3)
(source exemplar).
3) Updating confidence values: After filling the patch p with new pixel values,
the confidence term C(p) should be updated.
C( p) C( p )p p .
(4)
As filling proceeds, confidence values decay, indicating that we are less sure of the
color values of pixels near the centre of the target region.
The above three steps are repeated in an iterative manner until the
target region is filled with the data from source region. After completion of inpainting
the next step is applying super resolution algorithm on output LR image to obtain a
HR image. Algorithm 1 (Table 3.1) briefly explains the algorithmic steps for
Exemplar based inpainting. The subscript t used in Algorithm 1 indicates the current
iteration.
25
Algorithm 1
Extract the manually selected initial front 0 .
Repeat until done:
1a. Identify the fill front t . If t ;, exit.
1b. Compute priorities P(p) p t
2a. Find the patch p with the maximum priority,
i.e.,
p arg max p t P( p)
d ( p , q ) .
3. Update C(p) p p
process starts with the patch H R having the highest priority and which is
composed of known and unknown parts.
3) For the LR patch corresponding to the HR patch having the highest priority, its
best neighbour in the inpainted images of lower resolution is sought. This
search is performed in the dictionary and within a local neighbourhood. Only
the best candidate is kept. From this LR candidate, a HR patch is simply
deduced. Its pixel values are then copied into the unknown parts of the current
HR patch H R .
After the filling of the current patch, the priority value is propagated and the
aforementioned steps are iterated while there exist unknown.
and
iteration, is one of recently developed methods for such non differentiable constraint
optimization problems. Marquina and Osher were first to use the Bregman iteration
for fast SR image reconstruction with TV regularization. However, even though all
these regularization terms for SR image reconstruction lead to a stable solution, their
performance depends on optimization technique as well as regularization term. On the
other hand, based on TV regularization, Marquina and Osher obtained superior result
by employing Bregman iteration. So we envisage that even better results would be
obtained by combining Bregman iteration and a more sophisticated regularization
method that can suppress noise in LR images and ringing artifacts occurred during
capturing the details of the HR image. We propose a new regularization method based
on multiscale morphologic filters which are nonlinear in nature. Morphological
operators and filters are well-known tools that can extract structures from images.
They are used in image denoising, image segmentation and image fusion successfully.
Since proposed morphologic regularization term uses nondifferentiable max and min
operators, we develop an algorithm based on Bregman iterations and the forward
backward operator splitting using subgradients. It is seen that the results produced by
the proposed regularization are less affected by aforementioned noise evolved during
the iterative process.
3.3.1. Problem Formulation:
The observed images of a scene are usually degraded by blurring due to atmospheric
noise or inappropriate camera settings. Downsampling is also an important factor in
degradation of images. HR and LR images can be related as follows:
Yk
DFk H k X
ek ,
1, 2...., K
(5)
and k
is the index of the LR images. Assuming that H k becomes same for all k, it may be
simply denoted as H.
Yk
DFk HX
ek ,
1, 2...., K
30
(6)
Y k which
denotes
upsampled and reverse shifted kth LR image. Thus from equation (6), we can write
Yk
ek ,
R k HX
Fk 1DT Yk and ek
where Y k
RHX
(7)
Fk 1DT ek
Y k into
e.
(8)
2
2
(9)
When K< N/M, the SR image reconstruction becomes an ill posed problem, and
therefore, it becomes necessary to impose regularization to obtain a stable solution.
( X ) on
the
arg min{ ( X ) :
X
RHX
Y 2
(10)
where is a scalar constant depending on the noise variance in the LR images. The
above equation (10) represents a constrained minimization problem. Unconstrained
minimization problem is represented as follows:
31
1
X arg min RHX Y
X 2
2
2
( X )
(11)
where is the regularization parameter that controls emphasis between data error
term and regularization term.
The conventional regularization methods choose ( X ) as the high frequency
energy and minimize its pth form to ensure smoothness. In this paper ( X ) is defined
based on morphological filters that suppress noise [8]. Bright and dark noise can be
removed using morphological opening and closing.
3.3.2. Morphologic Regularization:
Let B be disk of unit size with origin at its center and sB be a disk structuring element
(SE) of size s. Then the morphologic dilation Ds ( X ) of an image X of size m n is
given by
max r( sB )(1) {xr }
max r( sB )( 2) {xr }
Ds ( X )
max
r( sB )(mn) {xr }
(12)
min r( sB )( 2 ) {xr }
Es ( X )
min
r( sB )(mn) { xr }
(13)
(14)
Cs ( X ) Es (D s ( X ))
(15)
s 1
32
where 1 is a column vector consisting of all the 1s and is the weighing coefficient.
Therefore, the SR reconstruction problem can be written as
s
X min s 1t [Cs ( X ) Os ( X )] : RHX Y
X s 1
(16)
(17)
Rn R .
(n)
X ( n 1) arg min Bp ( X , X ( n) ( X )
X
( n 1)
( n)
p
p T ( X ( n 1) )
(18)
(n)
( n 1)
( n)
(Y RHX
( n 1)
2
2
( X )
(19)
In equation (19) (Y RHX (n) ) is the error in the nth estimation which is added to Y(n)
such that RHX Y becomes zero. The unconstrained optimization (first step in
equation (19) can be explicitly solved using the following two step algorithm.
( X )
2
2
(20)
33
(21)
( X ) 1
(X U(n1) ) 0.
(X )
(22)
( X )
(X )
(23)
X (n)
Algorithm1 is used to fill these unknown pixels with their corresponding neighbor
pixels. is a predefined value chosen depending on the variance of noise in LR
images.
The regularization function ( X ) consists of opening and closing operations
which in turn uses morphologic operators dilation and erosion respectively. These
operators are non-differentiable. Therefore , we compute the subgradients of the
regularization function as in [8].
34
Algorithm 3
Initialize
Y (0) n 0, Y, X (0) FillUnknown(Y );
While RHX ( n) Y
2
2
U ( n 1) X (n) H T RT (RHX( n) Y n
( X )
( n 1)
( n 1)
X
U
( X ) X (n)
( n 1)
( n)
( n 1)
Y
Y (Y RHX
)
End
35
Chapter 4
Results
The proposed method is applied on different images and compared with the Patch
Match method [4] which is available in Adobe Photoshop.
Fig.4.1 4.9 shows the screen shots captured of MATLAB figures. In each of these
screen shots, the first row consists of the original image (input image) and the mask
image i.e., the image specifying the target region or region of object removal
respectively. The second row consists of inpainted image i.e., obtained after
implementing Algorithm 1.
Fig.4.10 shows the results obtained after applying Super resolution (Algorithm 3)
on different natural images. Column (a) gives the original images, column (b) gives
the image masks (object to be removed) and column (c) gives the SR Inpainted
images.
Fig. 4.11 shows the comparision beween the results obtained through proposed
method and the Patch Match method available in Adobe Photoshop (one of the Stateof-art Methods). Column (a) shows the original images, column (b) shows the image
obtained through Patch Match in Adobe Photoshop and (c) gives the output image
obtained through proposed method. We observe that when the object removal is done
using Patch tool in Adobe Photoshop, the region filled in the results obtained is
blurred and does not look reasonable to human eyes.
Table 4.1 shows the time taken for inpainting and super-resolution for some natural
images. Simulation is performed on a laptop with configuration as follows: Intel i5
processor and 4Gb RAM .Software used for simulation is MATLAB 7.11.0. It is
observed that inpainting process consumes more time.
Table 4.2 gives the comparision between the results obtained through Examle-based
Super Resolution(SR) and results obtained through Super Resolution using Bregman
iteration (Algorithm 3) . The comparision is done by using the quality metric Peak
signal-to-noise ratio (PSNR). PSNR1 gives the PSNR of images obtained through
Example-based SR and PSNR2 gives the PSNR of images obtained through SR using
Bregman iteration.
36
37
38
39
41
(a)
(b)
(c)
Fig.4.10. Results of the proposed method. (a) Original Picture, (b) Mask Image
(image specifying the object to be removed), and (c) Output picture after SR process
42
(a)
(b)
(c)
Fig.4.11. Comparison of the proposed method. (a) Original Picture, (b) Patch Match
in Adobe Photoshop, and (c) Proposed method
43
Picture
Resolution
Missing
Inpainting
SR
Total
Area
(sec)
(sec)
(sec)
(%)
Farm
256X256
18
88
40
128
Elephant
320X480
17
90
38
128
Tiger
480X320
28
149
38
187
Soldier
320X480
30
140
38
178
Island
316X416
20
105
37
142
City
400X300
40
40
80
Eagle
188X258
13
70
37
107
Ship
181X241
13
76
38
114
Yanam
240X320
18
24
42
63
PSNR1
PSNR2
(Example-based
SR)
Iteration)
Elephant
16.3
16.8
Farm
15.6
18.43
City
21.8
23.4
Island
14.63
15.39
Yanam
17.6
23.53
Soldier
18.1
18.2
Eagle
18.9
21.6
Ship
17.7
19.2
Tiger
17.1
17.6
Chapter 5
CONCLUSION AND FUTURE SCOPE
5.1. Conclusion
A new approach for object removal using inpainting has been successfully
implemented in this thesis. Object removal is done in a hierarchical manner by
combining inpainting and Super-resolution in a two-step approach. Proposed method
is applied on wide variety of images and the obtained results show the effectiveness of
this method compared to other methods such as Patch Match. The Super Resolution
process used in proposed method is proved to be less complex and effective compared
to previous method such as Example based Super Resolution. The proposed method
achieves the main goal of inpainting i.e., the obtained results look reasonable to the
human eyes.
5.2.Future scope
The process of inpainting has a great scope for future work. One of the
possibilities is to use data base of similar images that represents the same scene for
the purpose of texture synthesis. Inspite of using the input image itself for texture
synthesis, the best suitable matching patch can be caluculated from these images. A
different method of Super Resolution method such as classical multi-image Superresolution in conjuction with proposed inpainting process can be used in order to
obtain better image quality.
45
REFERENCES
[1]
[2]
object removal by
examplar-based image inpainting, IEEE Trans. Image Process., vol. 13, no. 9,
pp. 12001212, Sep. 2004.
[3]
[4]
[5]
[6]
[7]
[8]
Purkait, P., Chanda, B.: Super resolution image reconstruction through bregman
iteration using morphologic regularization. IEEE Transactions on Image
Processing 21, 40294039 (2012).
46
[12]
[13]
[14]
[15]
[16]
http://www.en.wikipedia.org/wiki/Inpainting
[17]
http://www.math.ucla.edu/~imagers/htmls/inp.html
47