Professional Documents
Culture Documents
fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
AbstractLow light photography suffers from blur and noise. Another prevalent problem in object motion blur is the
In this article, we propose a novel method to recover a dense sensor noise. In low light photography, the exposure time is
estimate of spatially varying blur kernel as well as a denoised increased to accumulate more light. Since the degree of blur
and deblurred image from a single noisy and object motion
blurred image. Proposed method takes advantage of the sparse scales with the object speed and the exposure time, a real-
representation of double discrete wavelet transform (DDWT)a world object motion blur is also accompanied by (moderate
generative model of image blur that simplifies the wavelet analysis to severe) noise. Though existing deblurring methods tolerate
of a blurred imageand the Bayesian perspective of modeling noise to a degree [4], [5], [6], [7], [8], [9], [10], [11], [12],
the prior distribution of the latent sharp wavelet coefficient and [13], [14], [15], [16], the prior studies exploring the limits
the likelihood function that makes the noise handling explicit. We
demonstrate the effectiveness of proposed method on moderate of deblurring in noise have been limited to the global blur
noise and severely blurred images using simulated and real case [3], [17]. Even a moderate level of noise is enough to
camera data. severely deteriorate the image deblurring methods handling
Index TermsNoisy image blind deblurring, spatially varying the spatially varying blur cases, as evidenced by a motion
blur, Gaussian scale mixture, double discrete wavelet transform, object blur estimation method in [26] failing under noise in
sparse representation, statistical model. Figure 2(c).
I. I NTRODUCTION
Image blur is a phenomenon caused by pixel sensors record-
ing light from multiple sources. If the object represented in the
image is Lambertian, such degradation can be characterized The goal of this article is to remove moderate to severe noise
by a point spread function or the blur kernel. Camera shake, and to reverse the effects of the (unknown) spatial varying blur
defocus, and object motion are three common types of blur in caused by object motion. We accomplish this task by appealing
consumer imaging. Blur casused by camera shake is spatially to the sparse representation of blurry images via the double
slowly varying (often approximated by a global blur [1], [2], discrete wavelet transform (DDWT) we proposed previously
[3], [4], [5], [6], [7], [8], [9], [10], [11], [12], [13], [14], [15], [28], [25]. DDWT is an analytical image representation that
[16], [17]), whereas the spatial variation of defocus and object significantly simplifies the wavelet-based generative modeling
motion causes sharp blur transitions because they depend on of blurred images by decoupling a blur from the signal. Adopt-
the depth or the speed of the object, respectively. As such, ing the Bayesian perspectivethat is, the prior distribution
the homographic transformations typically used to model the of the latent image wavelet coefficients are modeledand a
slow spatial variation of the camera shake (e.g. [18], [19], likelihood function that makes the noise handling explicit, we
[20]) are clearly mismatched for the defocus and the object propose (i) a maximal likelihood estimation of the blur kernel
motion blurs. This is evidenced by Figure 2(b) where the blur and (ii) a Bayesian least squares estimate of the latent sharp
kernel boundary between the foreground and the background denoised image. Specifically, we draw on the work of [31],
was missed entirely by the blind image deblurring method. [32], [21] to model the distribution of the latent sharp image
Hence detection and removal of blur stemming from object wavelet coefficient by the Gaussian scale mixture. The heavy-
motionthe subject of this paperrequires a fundamentally tailed prior density of the wavelet coefficients helps resolve the
different technique, one that have yet to reach its maturity. ambiguities caused by the noise and the blur. At the higher
With the exception of [21], [22], [23], methods designed to level, the present problem is simpler than the camera shake
handle sharp blur kernel boundaries have restricted the blur problem in the sense that the blur kernel is parameterized
shape to a line (for an object motion) [24], [25], [26], [27]; following the examples of [24], [25], [26], [27], our method
or a line or a circle (for a defocus blur) [28], [29], [30]. The approximates the blur shape by one dimensional linear motion.
parameterized blur model constrains the motion blur problem But it is more complex than the camera shake deblurring
to the point that the blur kernel boundaries can be recovered techniques (e.g. [18], [19], [20]) in the sense that dense esti-
with the help of local image statistics [26], [27], boundary mation of blur kernel tolerates sharp blur kernel boundaries;
aware regularization [24], matting [29], dictionary learning and far more robust to noise than the existing object motion
[30] and double discrete wavelet transform [28], [25]. blur detection/removal techniques.
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
(a) Blurred input image y (b) DWT coefficients wj (c) DDWT coefficients v ij (d) Intermediate blur kernel estimate (e) Final blur kernel estimate
Fig. 3: Example of DWT and DDWT coefficients using data corrupted by noise. The motion blur manifests itself as a double
edge in DDWT, where the distance between the double edges (yellow arrows in (c)) correspond to the speed of the object;
(d) is the coarse dense blur kernel estimation in (20); (e) final blur kernel estimation including graph cut described in Section
III.D.
Define: C. Gaussian Scale Mixture
The distribution of natural images in wavelet domain typ-
vij (n) :={di () ? dj (M ) ? y(M )}(n) ically exhibits non-Gaussian behavior (e.g. tails are heavier
={di () ? dj (M ) ? (h(M ) ? x(M ) + (M ))}(n)than Gaussian). Wainwright et al. proposed Gaussian scale
={di () ? h(M )} ? {dj (M ) ? x(M )}(n) mixture (GSM) [32] to model such behavior. Let u(n)
corresponds
to the latent sharp DWT coefficients within an
+ {di () ? dj (M ) ? (M )}(n) N N spatial neighborhood of n (overlapped, and vec-
={qi ? uj }(n) + ij (n), (8) torized to RN ), a random vector u RN is a Gaussian scale
mixture if it can be written as
where
u|z N (0, z), z p(z), (13)
uj (n) j
:= {d (M ) ? x(M )}(n) = u (M n) j
where z R is a positive scalar random variable distributed
qi (n) := {di () ? h(M )}(n) according to the density function p(z) and denotes the
covariance of u. The prior density p(u) is normal density
ij (n) := {di () ? dj (M ) ? (M )}(n). (9)
p(u|z) marginalized over the hidden parameter z, as follows:
Z
Owing to the fact that h(M n) is a horizontal blur, we may
p(u) = p(u|z)p(z)dz
simplify qi (n) and vij (n) as:
Z exp uT (z)1 u
n o 2
qi (n) = kk(M1 n)k n + 12 k (n) n 12 k (n)
= p(z)dz, (14)
(2)N/2 |z|0.5
uj (n+ 21 k (n))uj (n 21 k (n))
vij (n) = kk(M n)k + ij (n), (10) where the multiplier z is known as hidden variable and control
the contribution of each Gaussian in the mixture. The benefit of
where GSM is that it simplifies non-Gaussianity by marginalization
( of Gaussians over all possible values of z. Portilla et al. [31]
[0 r cos ]T
if || /4 investigates several solutions including maximum likelihood
k (n) := M1 k(M n) = .
[0 r sin ]T
otherwise estimation of z and Jeffreys prior [33] in their denoising work.
(11) In this paper, we follow the examples of Portilla et al. in [31]
Note that when M is a non-integer matrix, it requires to adopt Jefferys prior for p(z).
interpolation. This is problematic for low resolution images,
but the sacrifice is less severe for high resolution images. III. P ROPOSED W ORK
Unshearing vij (n) results in the following
The main drawback of an autocorrelation approach to the
DDWT-based blur kernel estimation in [28] is that it is
v ij (n, ) : = v (M1 n)
sensitive to noise (like most blind deblurring methods). We
uj (n+ 12 k(n))uj (n 12 k(n))
= kk(n)k + ij (n, ), investigate a Bayesian-based approach to the DDWT-based
blur kernel estimation that is more robust to noise.
(12)
where ij (n, ) is the unsheared version of ij (n). Therefore A. Noisy Blurred Image Probability Model
in the remainder of the paper, we are concerned with recover- In contrast to the discriminative modelingincluding reg-
ing uj from v ij , which is the noisy sum of the shifted versions ularization approaches that are popular in blind and non-
of uj . For simplicity, the superscripts i and j are omitted from blind deblurring ([7], [8], [9], [10], [11], [12], [13], [14],
{uj , q i , v ij , ij , uj , qi , vij , ij }. [15], [16])DDWT admits a generative modeling of blur that
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
Fig. 4: Examples of log-likelihood function with and without object motion. (a) is the 2D likelihood function over blur direction
and length when there is motion blur; (b) is the 2D likelihood function when image is sharp.
lends itself naturally to working directly with the posterior Here, we effectively choose the most likely blur length and
densities of the latent sharp images. By DDWT relation direction by searching over the vectored candidate blur ker-
in (12) and the GSM model of (13), the motion blurred nel ` parameterized by its length > 0 and its direction
image yields a pairwise jointly Gaussian distribution between [, ), aimed at estimating the latent ground truth blur
the DDWT coefficients (sheared and unsheared; recall (12)) kernel k(n) = r[sin , cos ]T parameterized by the actual
v n + 12 k(n), and v n 12 k(n), as follows:
r > 10 and
length direction [, ). The likelihood
v n + 2 `,
u(n) p k(n) = ` is computed by marginaliz-
v n + 21 k(n),
v n 12 `,
k(n), u (n + k(n))
v n 21 k(n), ing over all possible hidden variables of (16):
u (n k(n))
v n + 12 `,
1 u (n + k(n)) u(n) 0 p k(n) = `
N , , v n 12 `,
kk(n)k u (n k(n)) + u(n) 0
v n + 12 `, k(n) = `
Z
(15)
= p
v n 12 `, z(n, `)
where u(n), v(n, ) ZN denote the DWT and the DDWT R3
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
Equation (21) suggeststhat the posterior mean is the condi- v n + 12 k(n), k(n)
v n + 1 k(n),
p 1 p(z(n, k))
k(n) v n 2 k(n), z(n, k)
tional BLS estimate E u(n) 2 , = R . (25)
v n 21 k(n),
v n + 12 k(n), k(n)
z(n, k) p p(z(n, k))dz(n, k)
R3 1
marginalized
over the posterior mixture parameter v n 2 k(n), z(n, k)
v n + 1 k(n),
z(n, k) 2 , k(n) .
v n 21 k(n), We numerically integrate over z(n, k) in (21) and (25). The
Thanks to GSM, the latent local DWT vector u(n)|z(n) is final deblurred and denoised image is recovered by the inverse
conditionally normal. Assuming the noise is additive Gaussian wavelet transform of the wavelet coefficients estimated by
in (15), conditional BLS estimation is a linear minimum mean (21). The proposed algorithm for Bayesian least squares image
square error (LMMSE) estimate: deblurring is summarized in Algorithm 2.
v n + 1 k(n),
2 k(n)
E u(n) , Algorithm 2 Non-blind deblurring with numerical integration
v n 21 k(n),
z(n, k)
1
Compute multi-scale DDWT coefficients v ij via (12).
1 v n + 2 k(n), 1:
=uv|k,z (v|k,z ) , (22)
v n 21 k(n), 2: for each subband j do
3: j from DDWT vector
Reconstruct sharp DWT vector u
where ij
v using (21).
" T #
v n + 12 k(n), k(n) end for
4:
uv|k,z = E u(n) Inverse DWT to recover sharp image x .
v n 21 k(n), z(n, k) 5:
z(n)
= , (23)
kk(n)k
and
C. Speed Optimization
v|k,z
" T #
The marginalizations over the hidden variables used in (19),
v n + 12 k(n), v n + 12 k(n), k(n)
=E (21) and (25) involve a triple integration that is computation-
v n 12 k(n), v n 21 k(n), z(n, k)
ally expensive. It stems from the fact that the (conditional)
1
(z(n + k(n)) + z(n)) z(n) probability density function on z cannot be circumvented in
=
kk(n)k2 z(n) (z(n k(n)) + z(n)) (19), (21) and (25). In order to speed up the inference, we
0 propose a simplification aimed at replacing integral with fast
+ . (24)
0 searches.
The posterior mixture parameter 1) Blur Estimation by Joint Maximum Likelihood (ML) and
v n + 1 k(n), Maximum a-Posteriori (MAP) Estimation: As an alternative to
z(n, k) 2 , k(n) can be handled by
v n 21 k(n), integration in (18), we propose a joint ML estimation of k(n)
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
estimation map by graph cut [39] based on Markov random sharp DWT coefficients at each subband, and final sharp image
field (MRF) over the entire image. Such optimization process is reconstructed by taking inverse DWT of recovered DWT
searches the estimation map that minimizes following energy: coefficients.
X X
E(k) = E1 (k(n)) + E2 (k(n), k(n0 )). (31)
n n,n0
A. Simulation Study
Here E1 enforces output blur map similar to (18) or (27). We We simulated a blurred noisy camera data using 50 8-
choose the binary cost: bit images from Berkeley segmentation dataset [42] with
( various types of details (e.g. line, curve, different shapes).
0 if k(n) = k(n) The synthetic blurry and noisy images are created by applying
E1 = . (32) global translational blur of various lengths in the horizontal
1 otherwise
direction as well as additive white Gaussian noise. We quan-
On the other hand, E2 encourages discontinuities to align with tify the overall performance (blur estimation, deblurring, and
image edges. Such cost is defined by: denoising) by comparing the average peak signal-to-noise ratio
(PSNR) of recovered images. Figure 6 summarizes the PSNR
(
0 , k(n) = k(n0 )
E2 = 2 2 , (33) scores of blind deblurring performance. Overall, the proposed
ekn k / , otherwise
work clearly outperformed the alternatives when the noise
where is set to be four times of the median value of kn k2 , variances are 25, 100 and 625. For this study, our methods
and kn k2 denotes the image gradient at n [37]. were implemented to estimate the global blurthat is, we used
(20) and (27) to find the global maximizer of the likelihood
IV. E XPERIMENTS function. Among our methods, the full integration versions
as described by Algorithms 1 and 2 performed better than
We evaluate our proposed method using simulated and real
the accelerated version of Algorithms 3 and 4, though the
camera data degraded by blur and noise. Experiments are
differences were small. The performance of [4] for a low noise
implemented in MATLAB and executed on PC using Intel(R)
case (variance = 9) were comparable. PSNRs of [13] when
Core(TM) i7-2600 CPU @ 3.4GHz with 16GB memory.
noise variance = 625 are intentionally left out due to the fact
Below we describe the simulated and real-image experiments.
that the code provided by the authors caused error when the
A few representative output images are shown in Figures 8
noise variance was high.
to 10; many more examples can be found in the supplementary
We also compare the deblurring techniques of Algorithms
material.
2 and 4 to the non-blind deblurring techniques of [6], [3],
We compared our proposed method to state-of-art blind
[4], [41], [13], [17], [28]. In Figure 7, the ground truth blur
image deblurring algorithms of [6], [15], [19], [3], [4], [5],
kernels were provided. Proposed non-blind deblurring methods
[13], [16], [17], [28]. We also compared the motion blur kernel
are comparable to state-of-art methods ([41], [17]) when noise
estimation/detection to the methods in [26], [28], [25] and
variance = 100 and 625.
[22]. We used parameter values recommended by each authors.
Since there are no alternatives for handling object motion blur
in presence of noise, we acknowledge the limitation to our B. Real Data
studyexisting methods use parametric blur kernels to handle The real data experiment is conducted on raw sensor
blur kernel boundaries (typical for object motion and defocus) data captured by Nikon D90 indoor (low light) or outdoor
but without considerations for severe noise [28], [25], [26]; (sunset/sunrise in cloudy weather) with high ISO and small
use nonparametric blur kernels but do not handle sharp blur aperture size. All input images suffer from noise and uni-
kernel boundaries (typical for camera shakes) [19]; and/or form or non-uniform translational blur (in various directions)
are designed for heavy noise, but without consideration for stemming from object motion. The raw data was first treated
spatial variations of blur [17], [3]. In the real-image results with demosaicking method of [43]; then pixel neighborhood
below, the global deblurring methods of [40], [41], [17], [4], of 22 was averaged and the image was downsampled by 2 to
[3] were applied only to the blurred regions of the images, ensure that the demosaicking artifacts has minimal impact on
as determined by our MLE blur kernel estimation method in deblurring process. We trained the covariance matrix E[uuT ]
Algorithm 3 (See Section IV.B for details). of the sharp DWT coefficients in (17) for each scale and
We set GSM neighborhood N = 3 and N = 1 in subband based on 112 sharp images captured by the same
Algorithms 1, 2 and Algorithms 3, 4, respectively. The range of camera and lens, and downsampled in the same manner. We
numerical integration in Algorithms 1 and 2 is truncated within report the noise standard deviation estimated by [34] in the
z [exp(20.5), exp(3.5)]3 . We use the single scale and ver- caption of each figure. We are aware that such estimation is not
tical highpass subband (recall blur estimation is evaluated by accurate for the true image noise model since the distribution
image shearing) of the wavelet coefficients for blur estimation. of the image sensor noise is better characterized as Poisson
Specifically, we use the third level vertical subband of DWT [44], though it is still a good proxy for the level of noise
coefficients in the cases of noise variances equal to 625 in we see in each image. We begin the real data experiments
synthetic experiment (Figure 6), as well as real image example by considering an image with a global object motion and
(Figures 8(b,c)). We use the second level vertical subband for noise. We captured an image of a color picture mounted on a
the rest of the experiment. For image deblurring, we recover Yaskawa Motoman HP3 robot. Together with the motion blur
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
Fig. 6: Blind deblurring performance on images synthetically corrupted by noise and global blur.
speed and direction that can be controlled with high precision the bilateral filtering both methods takes 385 secondsone
and the relative camera coordinates, we have the ground truth can expect significant speed up from bilateral filtering itself
motion data for these images. This experiment ensures that by considering fast implementations, such as that of [45].
blind deblurring methods in [3], [17], [4], [5], [13] and a Deblurring methods in Algorithms 2 and 4 take 228 and 9.5
combination of [40] and [41] designed for global blur can seconds, respectively.
be tested without additional measures. In a more likely scenario, Figure 9(a) shows an example
Figure 8 show the results of processing uniformly blurred of moving foreground against a stationary background; and
image (more global blur images in the supplementary mate- Figure 10(a) shows a stationary foreground against moving
rial). In this example, noise is clearly visible. In Figure 8(e,f,h), background (more examples in the supplementary material).
the alternative blind deblurring methods fail at the blur kernel Subfigures (m,n) in Figures 9 and 10 show the recovered
estimation stage. We found that our previous method (based on object speed, direction, and sharp blur kernel boundaries using
autocorrelation function of DDWT) yielded satisfactory blur the proposed Algorithms 1 and 3, respectively. Contrast this
estimation, but deblurring was low quality (Figure 8(d)). Meth- to the blur kernel estimation techniques in [28], [25], [26]
ods in [17], [5], [13] yield acceptable blur kernel estimations and the blur detection method in [22] that reported promising
but deblurring results are not satisfactory (Figure 8(g,i,j)). results in low noise case. These algorithms prove less reliable
The proposed methods (implemented with (20) and (27) for identifying object motion blur in presence of more noise,
to find the global maximizer of the likelihood function) also however, as Figure 9(o-r) and Figure 10(o-r) have shown. As
yielded satisfactory blur kernel estimation and image deblur- evidenced by the output images in Figures 9(b,c) and Figures
ring resultsconsistent with our findings in Figures 6 and 10(b,c), the proposed deblurring methods in Algorithms 2
7. As was also suggested by Figures 6 and 7, the integral and 4 yield satisfactory results, largely free of instabilities
method as described by Algorithms 1 and 2 yielded better caused by noise and ringing artifacts. For the most part, the
results than the approximated method by Algorithms 3 and edge details were recovered correctly, albeit somewhat softer
4, though the latter was far faster in execution time. The in appearance. The output from our previous DDWT-based
blur kernel estimation in Algorithm 1 takes 765 seconds to method in [28] yielded considerably softer edges, as well as
process color image of size 500x500, while the accelerated the noise artifacts (Figure 9(d) and Figure 10(d)).
method in Algorithm 3 takes 393 seconds. Note however that We also compared our results to the deblurring methods
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
Fig. 7: Non-blind deblurring performance with ground truth blur kernel on images synthetically corrupted by noise and global
blur.
in [3], [40], [41], [17], [4]. Recall that these methods cannot blur: (i) recovery of sharp blur kernel transition, (ii) dense blur
handle sharp blur boundaries (and indeed, they will fail for kernel estimation, (iii) deblurring with blur kernel boundaries,
object motion blur; recall Figure 2). Thus we first isolated and (iv) robustness to noise. Most existing methods are not
the blurred regions of the image by motion matte, computed designed for sharp blur kernel boundaries and noise. Since
by convolving a binary mask of moving pixels in Figures we tend to use long exposure and high gain to overcome
9(n) and 10(n) by the estimated blur kernel (one may also noise in low light photography, deblurring techniques for blur
consider techniques such as [46] and [47]). In Figures 9(e- caused by object motion must be robust to noise in order to
h) and 10(e-h), blur kernel and the latent sharp image were be pragmatic and useful for real scenarios.
estimated by the blind image deblurring methods in [3], We conclude from Figures 9-10(m,n) (as well as the sup-
[40]+[41], [17] and [4]. In Figures 9(i-l) and 10(i-l), the latent plemental material) that our previous methods of [28] and
sharp image was estimated by non-blind image deblurring [25] as well as the techniques of [26] and [22] cannot extract
techniques of [3], [41], [17], [4] using the blur kernel estimated blur boundaries accurately, even though these methods have
in Figures 9(n) and 10(n), respectively. Despite the relatively been reported to do better when the noise is less severe. By
good performance in simulated results we reported in Figure comparison, the proposed Algorithms 1 and 3 successfully
7, the results showed sensitivities to noise (Figures 9(i, e, h), recover the object motion with sharp blur kernel transitions in
9(e, f)), susceptibility to ringing artifacts (Figures 9(j), 10(f, i, presence of noise. The proposed methods have benefited from
j, k, l)), and failure to recover edges (Figures 9(e, g, h), 10(e, the likelihood function that makes noise handling explicit.
g, h, i, k ,l)). Overall, deblurring results of [3], [4] worked DDWT also plays a critical rolethe bivariate likelihood func-
very well on some images and failed badly on others. The tion in (18) is computationally/analytically tractable thanks to
combination of matting based Algorithm 3 and [17] performed DDWT that transforms the blurred DWT coefficients into only
more consistently. two significant DDWT coefficients.
Blur kernel estimation and deblurring of images corrupted
C. Discussion by noise and blur caused by object motion pose unique
There are four main criteria for evaluating the quality of re- challenges for techniques not specifically designed for such
covering sharp image from images corrupted by object motion purpose. Majority of prior work in image deblurring are
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
designed to handle spatially uniform blur with low noise. Our V. C ONCLUSION
experiments confirm that there is no straightforward way to
repurpose existing deblurring method for the object motion
We extended the the DDWT deblurring of [28] to tolerate
problem. Specifically, by isolating the blurred object from
sensor noise and sharp blur boundaries common in object
otherwise stationary scene using motion matting, we can
motion blurs. In order to consider the effect of noise, we
effectively treat the scene as a piecewise-uniform blur with
proposed maximum likelihood blur kernel estimation and
noise. However, experiments with real image data confirm
Bayesian least squares deconvolution using Gaussian scale
that deblurring results in instabilities such as ringing artifacts
mixture modeling. We accelerate algorithm by incorporating
that are often severe, often seen near the sharp blur kernel
expectation maximization algorithm. We demonstrated that
boundaries. Noise amplification was evident in the presence
GSM modeling combined with explicit noise modeling pro-
of moderate to heavy noise, also.
vides a significant advantage in realistic image deblurring
Compared to these alternatives, the proposed deblurring
scenarios for object motion. Our future research plan includes
technique handles the blur kernel transitions better. Specif-
blur estimation by maximum a posteriori and minimum mean
ically, the proposed DDWT-GSM algorithm was the only
square estimator using prior model on blur kernel size and
method consistently and reliably outputting deblurred images
direction.
without noticeable ringing artifacts. The absence of ringing
can be explained by the fact that each DDWT coefficients
influences the recovery of only two DWT coefficients (contrast
this to the conventional deblurring approach where the recon- A PPENDIX A
struction of each pixel depends on a very large group of pixels M ODIFIED B ILATERAL FILTER
in its neighborhood). While there are instances where certain
image features were recovered by the existing methods more
convincingly than ours, our algorithm significantly improved Bilateral filter is exploited to carryout the local smoothness
stability and overall output image quality. of log-likelihood. For notational simplicity, we define
On the other hand, deblurring results with modest/severe
noise are obviously not as high quality compared to the v n + 12 `,
low noise deblurring case (obviously). GSM-based image p (V (n, `, )) := p k(n) = ` . (34)
v n 12 `,
reconstruction (such as ours) are known to yield somewhat
softer appearance, owing to the shrinkage stemming from
The bilateral filtering can be discribed by:
coring function. As suggested by Figure 7, for example, we
may consider in the future a hybridization between the blur
kernel estimation of Algorithm 1 (or 3) and existing state-of- Bn0 (n) {log p (V (n0 , `, ))}
art image deblurring techniques in the case that noise is less 0 0
P
n0 (n) W (n ) log p (V (n , `, ))
significant. = P 0
, (35)
There are further design considerations for Algorithm 1 and n0 (n) W (n )
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
z(n, `) Q(z, z p (n, `))
v n + 21 `, k(n) = `
u(n)
= arg max p p(z)
z v n 12 `, z =E log p(z) + log p
u (n + `) z
u(n)
u (n `)
1
Z
v n + 2 `,
v n + 1 `,
= arg max p(z) p , u (n + `) k(n) = `
z v n 12 `, 2
v n 1 `, , .
z = z p (n, `)
R3N u (n `) 2
k(n) = `
u(n) (39)
d u (n + `)
z
u (n `) M-step: Choose z that maximizes Q function:
Z 1
u(n)
v n + 2 `,
z p+1 (n, `) = arg max Q(z, z p (n, `)).
= arg max p(z) p u (n + `) (40)
z R3N v n 12 `, z
u (n `)
u(n) u(n)
u(n)
k(n) = `
, p u (n + `)z d u (n + `) , (37)
z Notice that p u (n + `) z is normal and each z can be
u (n `) u (n `) u (n `)
optimized separately. Hence, for m = 0, 1, 2:
where
p+1 log |z|
= arg max E log p(z)
zm
1
u(n)
v n+ 2 `, u (n + `) , k(n) = `
z 2
p 1
v n + 1 `,
n
v 2 `, u (n `)
z 2
uTm 1 um v n 12 `,
. (41)
1
u(n) k(n) = `
v n+ 2 `,
u (n + `) , k(n) = ` 2z
=p 1 (38)
z = z p (n, `)
v n 2 `, u (n `)
v n + 1 `,
is not a function of z and is zero when taking derivative with 2
v n 1 `,
respect to z. Also notice that um 2 has the conditional
k(n) = `
E-step: Compute the expected value of log-posterior
z = z p (n, `)
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
mean: Hence
v n + 1 `,
v n + 1 `,
2
uTm (z)1 um v n 1 `,
2
v n 1 `,
2
E
m
=E u m
2 2 k(n) = `
k(n) = `
z = z p (n, `)
z = z p (n, `) XX
m ]ij [1 ]ij +
[ Tm 1
m
v n + 12 `,
1
=um v|`,zp v|`,zp (42) i j
v n 12 `, = (49)
2z
and conditional covariance To solve (41), we take derivatives of each term with respect
to z:
v n + 1 `,
m
2
uTm (z)1 um v n 12 `,
v n + 1 `,
2 E
1
z 2 k(n) = `
T v n 2 `,
=E (um m )(um m )
z = z p (n, `)
k(n) = `
P P 1 Tm 1
z = z p (n, `) i j [m ]ij [ ]ij + m
= (50)
= um v|`,zp 1
v|`,z p vum |`,z .
p (43) 2z
2
1
v n + `,
2
v n 1 `,
log|z| = N
In (42) and (43), conditional cross correlation matrix take the E
2 (51)
form of following: z 2 k(n) = ` 2z
p
z = z (n, `)
v n + 1 `,
z0p (n) 2
u0 v|`,zp = [, ] (44)
1 v n 12 `,
k`k = 1
E log
(52)
z z k(n) = ` z
z p (n)
u1 v|`,zp = 1 [, 0] (45) z = z p (n, `)
k`k
z p (n) where the local spatial neighborhood has a size of N N .
u2 v|`,zp = 2 [0, ]. (46)
k`k The final solution is obtained by taking summation of (50),
(51) and (52) and set it to zero.
Notice that vum |`,zp = Tum v|`,zp , where m = 0, 1, 2, and P P [ ] [1 ] + T 1
i j 0 ij ij 0 0
N +2
P P
v|`,zp
j [1 ]ij [
1
T
]ij +1
1
1
z p+1 (n, `) =
" T # i (53)
N +2
v n + 21 `, v n + 12 `, k(n) = `
=E
v n 12 `, v n 12 `, z = z p (n, `)
P P
1 1
i j [2 ]ij [ T
]ij +2 2
N +2
z p (n + `) + z p (n) z p (n)
=
k`k2 z p (n) z p (n `) + z p (n) ACKNOWLEDGMENT
0
+ . (47) We would like to thank the authors of [6], [15], [19], [3],
0 [4], [5], [13], [16], [17], [26], [22] for providing the code that
were used in our experiments. We also thank authors of [42]
In (41), for making the image dataset available.
v n + 1 `,
2 R EFERENCES
T 1 v n 1 `,
E um um 2
[1] F. Xue and T. Blu, A novel sure-based criterion for parametric psf
k(n) = `
estimation, Image Processing, IEEE Transactions on, vol. 24, no. 2,
z = z p (n, `) pp. 595607, 2015.
[2] J. P. Oliveira, M. Figueiredo, J. M. Bioucas-Dias et al., Parametric
v n + 1 `,
2 blur estimation for blind restoration of natural images: Linear motion
v n 1 `,
T 1 and out-of-focus, Image Processing, IEEE Transactions on, vol. 23,
= E (um
m ) (um m )
2
k(n) = ` no. 1, pp. 466477, 2014.
[3] R. Molina, J. Mateos, and A. K. Katsaggelos, Blind deconvolution
z = z p (n, `)
using a variational approach to parameter, image, and blur estimation,
+ Tm 1
m Image Processing, IEEE Transactions on, vol. 15, no. 12, pp. 3715
XX 3727, 2006.
= m ]ij [1 ]ij +
[ Tm 1
m (48) [4] D. Perrone and P. Favaro, Total variation blind deconvolution: The devil
i j is in the details, in Computer Vision and Pattern Recognition (CVPR),
2014 IEEE Conference on. IEEE, 2014, pp. 29092916.
[5] T. Michaeli and M. Irani, Blind deblurring using internal patch recur-
where i, j corresponds to row and column indices, respectively. rence, in Computer VisionECCV 2014. Springer, 2014, pp. 783798.
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
[6] A. Levin, Y. Weiss, F. Durand, and W. T. Freeman, Efficient marginal [33] H. Jeffreys, An invariant form for the prior probability in estimation
likelihood optimization in blind deconvolution, in Computer Vision and problems, Proceedings of the Royal Society of London. Series A.
Pattern Recognition (CVPR), 2011 IEEE Conference on. IEEE, 2011, Mathematical and Physical Sciences, vol. 186, no. 1007, pp. 453461,
pp. 26572664. 1946.
[7] L. Lucy, Bayesian-based iterative method of image restoration, Journal [34] D. Donoho and I. Johnstone, Adapting to unknown smoothness via
of Astronomy, vol. 79, no. 745-754, 1974. wavelet shrinkage, Journal of the american statistical association,
[8] A. Levin, Y. Weiss, F. Durand, and W. Freeman, Understanding and vol. 90, no. 432, pp. 12001224, 1995.
evaluating blind deconvolution algorithms, in Computer Vision and [35] E. Polak, Optimization: algorithms and consistent approximations.
Pattern Recognition, 2009. CVPR 2009. IEEE Conference on. IEEE, Springer-Verlag New York, Inc., 1997.
2009, pp. 19641971. [36] A. P. Dempster, N. M. Laird, and D. B. Rubin, Maximum likelihood
[9] R. Fergus, B. Singh, A. Hertzmann, S. Roweis, and W. Freeman, from incomplete data via the em algorithm, Journal of the Royal
Removing camera shake from a single photograph, ACM SIGGRAPH, Statistical Society. Series B (Methodological), pp. 138, 1977.
2006. [37] A. Chakrabarti and T. Zickler, Depth and deblurring from a spectrally-
[10] J. Cai, H. Ji, and Z. Shen, Blind motion deblurring from a single image varying depth-of-field, in Computer VisionECCV 2012. Springer,
using sparse approximation, CVPR, vol. 62, pp. 291294, January 2009. 2012, pp. 648661.
[11] M. Ben-Ezra and S. Nayar, Motion deblurring using hybrid imaging, [38] A. Levin, R. Fergus, F. Durand, and W. T. Freeman, Image and depth
in Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 from a conventional camera with a coded aperture, in ACM Transactions
IEEE Computer Society Conference on, vol. 1. IEEE, 2003, pp. I657. on Graphics (TOG), vol. 26, no. 3. ACM, 2007, p. 70.
[39] Y. Boykov and V. Kolmogorov, An experimental comparison of min-
[12] S. Nayar and M. Ben-Ezra, Motion-based motion deblurring, Pattern
cut/max-flow algorithms for energy minimization in vision, Pattern
Analysis and Machine Intelligence, IEEE Transactions on, vol. 26, no. 6,
Analysis and Machine Intelligence, IEEE Transactions on, vol. 26, no. 9,
pp. 689698, 2004.
pp. 11241137, 2004.
[13] Q. Shan, J. Jia, and A. Agarwala, High-quality motion deblurring from
[40] L. Sun, S. Cho, J. Wang, and J. Hays, Edge-based blur kernel estimation
a single image, ACM Transactions on Graphics (SIGGRAPH), 2008.
using patch priors, in Computational Photography (ICCP), 2013 IEEE
[14] T. Cho, S. Paris, B. Horn, and W. Freeman, Blur kernel estimation International Conference on. IEEE, 2013, pp. 18.
using the radon transform, in Computer Vision and Pattern Recognition [41] D. Zoran and Y. Weiss, From learning models of natural image patches
(CVPR), 2011 IEEE Conference on. IEEE, 2011, pp. 241248. to whole image restoration, in Computer Vision (ICCV), 2011 IEEE
[15] S. Cho and S. Lee, Fast motion deblurring, in ACM Transactions on International Conference on. IEEE, 2011, pp. 479486.
Graphics (TOG), vol. 28, no. 5. ACM, 2009, p. 145. [42] D. Martin, C. Fowlkes, D. Tal, and J. Malik, A database of human
[16] L. Xu, S. Zheng, and J. Jia, Unnatural l0 sparse representation for segmented natural images and its application to evaluating segmentation
natural image deblurring. algorithms and measuring ecological statistics, in Proc. 8th Intl Conf.
[17] L. Z. S. Cho, D. Metaxas, S. Paris, and J. Wang, Handling noise in Computer Vision, vol. 2, July 2001, pp. 416423.
single image deblurring using directional filters. [43] E. Dubois, Frequency-domain methods for demosaicking of bayer-
[18] M. Hirsch, C. J. Schuler, S. Harmeling, and B. Scholkopf, Fast removal sampled color images, IEEE Signal Processing Letters, vol. 12, no. 12,
of non-uniform camera shake, in Computer Vision (ICCV), 2011 IEEE p. 847, 2005.
International Conference on. IEEE, 2011, pp. 463470. [44] R. C. Gonzalez, R. E. Woods, and S. L. Eddins, Digital image processing
[19] O. Whyte, J. Sivic, A. Zisserman, and J. Ponce, Non-uniform deblur- using MATLAB. Pearson Education India, 2004.
ring for shaken images, in Computer Vision and Pattern Recognition [45] C. Karam, C. Chen, and K. Hirakawa, Stochastic bilateral filter for
(CVPR), 2010 IEEE Conference on. IEEE, 2010, pp. 491498. high-dimensional images, in Image Processing (ICIP), 2015 IEEE
[20] C. Paramanand and A. N. Rajagopalan, Non-uniform motion deblurring International Conference on. IEEE, 2015, pp. 192196.
for bilayer scenes, in Computer Vision and Pattern Recognition (CVPR), [46] A. Zomet, A. Levin, S. Peleg, and Y. Weiss, Seamless image stitching
2013 IEEE Conference on. IEEE, 2013, pp. 11151122. by minimizing false edges, Image Processing, IEEE Transactions on,
[21] K. Schelten and S. Roth, Localized image blur removal through non- vol. 15, no. 4, pp. 969977, 2006.
parametric kernel estimation, in Pattern Recognition (ICPR), 2014 22nd [47] M. W. Tao, M. K. Johnson, and S. Paris, Error-tolerant image com-
International Conference on. IEEE, 2014, pp. 702707. positing, International journal of computer vision, vol. 103, no. 2, pp.
[22] J. Shi, L. Xu, and J. Jia, Discriminative blur detection features, in 178189, 2013.
Computer Vision and Pattern Recognition (CVPR), 2014 IEEE Confer-
ence on. IEEE, 2014, pp. 29652972.
[23] Q. Qian and B. K. Gunturk, Space-varying blur kernel estimation and
image deblurring, in IS&T/SPIE Electronic Imaging. International
Society for Optics and Photonics, 2014, pp. 90 230E90 230E.
[24] T. H. Kim and K. M. Lee, Segmentation-free dynamic scene deblur-
ring, in Computer Vision and Pattern Recognition (CVPR), 2014 IEEE
Conference on. IEEE, 2014, pp. 27662773.
[25] Y. Zhang and K. Hirakawa, Fast spatially varying object motion
blur estimation, in Image Processing (ICIP), 2015 IEEE International
Conference on. IEEE, 2015.
[26] A. Chakrabarti, T. Zickler, and W. T. Freeman, Analyzing spatially- Yi Zhang Yi Zhang (S12-M16) received the B.E.
varying blur, CVPR, 2010. degree in Electronic and Information Engineering
[27] A. Levin, Blind motion deblurring using image statistics, Advances in from Nanjing University of Information Science and
Neural Information Processing Systems, vol. 19, p. 841, 2007. Technology, Nanjing, Jiangsu, China, in 2008, the
[28] Y. Zhang and K. Hirakawa, Blur processing using double dis- M.S. and Ph.D. degrees in Electrical and Computer
crete wavelet transform, in Computer Vision and Pattern Recognition Engineering from University of Dayton, Dayton,
(CVPR), 2013 IEEE Conference on. IEEE, 2013, pp. 10911098. OH, in 2011 and 2015, respectively. Dr. Zhang
[29] S. Dai and Y. Wu, Removing partial blur in a single image, in is currently a research scientist at Ford Research
Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE and Innovation Center, Palo Alto, CA. His current
Conference on. IEEE, 2009, pp. 25442551. research interests include image processing, statis-
[30] F. Couzinie-Devy, J. Sun, K. Alahari, and J. Ponce, Learning to estimate tical signal processing, computer vision and deep
and remove non-uniform image blur, in Computer Vision and Pattern learning.
Recognition (CVPR), 2013 IEEE Conference on. IEEE, 2013, pp. 1075
1082.
[31] J. Portilla, V. Strela, M. Wainwright, and E. Simoncelli, Image denois-
ing using scale mixtures of gaussians in the wavelet domain, Image
Processing, IEEE Transactions on, vol. 12, no. 11, pp. 13381351, 2003.
[32] M. J. Wainwright and E. P. Simoncelli, Scale mixtures of gaussians and
the statistics of natural images. in NIPS. Citeseer, 1999, pp. 855861.
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TIP.2016.2583069, IEEE
Transactions on Image Processing
1057-7149 (c) 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.