Professional Documents
Culture Documents
A BSTRACT. The image deblurring problem consists of reconstructing images from blur and noise contaminated available
data. In this AMS Notices article, we provide an overview of some well known numerical linear algebra techniques that are
use for solving this problem. In particular, we start by carefully describing how to represent images, the process of blurring
an image and modeling different kind of added noise. Then, we present regularization methods such as Tikhonov (on the
arXiv:2201.09831v1 [math.NA] 24 Jan 2022
standard and general form), Total Variation and other variations with sparse and edge preserving properties. Additionally, we
briefly overview some of the main matrix structures for the blurring operator and finalize presenting multilevel methods that
preserve such structures. Numerical examples are used to illustrate the techniques described.
F IGURE 4. Image with assumed zero F IGURE 6. The blurred image btrue
boundary conditions is shown on the on the left with a small amount of
left and one with assumed periodic Gaussian white noise added to obtain b
boundary condition is shown on the on the right.
right. The red box represents the FOV. Two other models of noise are illustrated in Figure 7.
Under low light intensities, as encountered in astropho-
tography, the number of photons that arrive on the CCD
3. A DDING NOISE while the image is exposed may differ from the num-
Let us consider the grayscale image x true
shown on the ber expected. A Poisson distribution provides an effec-
left of Figure 5 and its blurred version b true
= Axtrue tive description of the resulting image as seen on the left.
shown on the right. If the digital image is transmitted over a communication
channel, some of the transmitted bits may be corrupted in
transmission resulting in “salt and pepper” noise demon-
strated on the right.
5. R EGULARIZATION
As an alternative to accepting xLS as our reconstructed
image, we can compute approximations to xtrue by filter-
F IGURE 11. The coefficients ing out the noise e while retaining as much information
|uT` btrue | are seen on the left while the as possible from the measured data b. This process is
known as regularization, and there are several possible
coefficients |uT` b| are on the right.
approaches we can follow.
By contrast, the coefficients |uT` b| = |uT` btrue + uT` e|
A first natural idea is to introduce a set of filtering fac-
appear on the right of Figure 11. The discrete Picard con-
tors φ` on the SVD expansion and construct a regularized
dition means that the coefficients |uT` btrue | decrease to
solution as
zero. However, the coefficients of the added white noise m
uT b
|uT` e| remain relatively constant for large i. At some point
X
xreg = φ` ` v ` ,
the noise contributed by e overwhelms the information σ`
`=1
contained in btrue . with the filter factors being φ` ≈ 0 for large values of `
Writing both vectors xLS and xtrue as a linear combi- when the noise dominates and φ` ≈ 1 for small values of
nation of the right singular vectors v` , we see that `, which are the terms in the expansion where the compo-
Xm
uT` btrue nents of both b and btrue are the closest.
true
x = v` One option is to define φ` = 1 for ` smaller than some
σ`
`=1 cut-off k < m and φ` = 0 otherwise. That is, we can sim-
and ply truncate the expansion of xLS in terms of right sin-
m m gular vectors in an attempt to minimize the contribution
X uT b X uT (btrue + e)
xLS = `
v` = `
v` from the terms |uT` e|/σ` for large `. Then, the obtained
σ` σ`
`=1 `=1 regularized solution would be
m m
X uT` btrue X uT e
` k
= v` + v` reg
X uT b`
σ` σ` x = v` .
`=1 `=1 σ`
m `=1
X uT` e
= xtrue + v` . This solution is known as the truncated SVD (TSVD).
σ`
`=1 Remember, however, that the singular values in a dis-
Because the singular values approach 0, the coefficients crete ill-posed problem approach 0 without there being
|uT` e|/σ` grow extremely large, as seen in Figure 12. a gap that would form a natural cut-off point. Instead,
Therefore, xLS includes a huge contribution from high- Tikhonov regularization chooses the filtering factors
frequency right singular vectors v` , which means that xLS σ2
is very oscillatory and not at all related to the original im- φ` = 2 ` 2
σ` + λ
age xtrue that we seek to reconstruct.
for some parameter λ whose choice will be discussed
later. For now, notice that φ` ≈ 1 when σ` λ and
φ` ≈ 0 when σ` λ. This has the effect of truncating
the singular vector expansion at the point where the sin-
gular values pass through λ but doing so more smoothly.
This leads to the regularized solution
m
X σ`
xreg = (uT b)v` ,
σ2 + λ2 `
`=1 `
8. M ATRIX S TRUCTURES
8.1. BTTB Structure. Because of the large-scale prob-
lems, it is useful to consider the structure of the matrix
F IGURE 18. The reconstructed im-
A. For instance, when considering a spatially invariant
ages xreg using the optimal regulariza-
blur and assuming that the image has zero boundary con-
tion parameter and the discretization of
ditions, the matrix A has Block-Toeplitz-Toeplitz-Block
the first derivative operator with the 2-
(BTTB) structure [12], that is, a p by p block-Toeplitz ma-
norm regularization on the left, and TV
trix with each block being a p by p Toeplitz matrix,
regularization on the right.
In Figure 18, we reconstruct the image by applying the A0 A−1 A−2 . . . A−(p−1)
discretization of the two-dimensional first derivative op- A1 A0 A−1 . . . A−(p−2)
erator for zero boundary conditions, that is, the matrix L A= 2
A A 1 A0 . . . A−(p−3) ,
takes the form of .. .. .. .. ..
. . . . .
−1 1 Ap−1 Ap−2 Ap−3 . . . A0
I ⊗ L1 −1 1 where for ` = 1, . . . , p we have
L= with L1 = .. ..
L1 ⊗ I . .
`
a a `
a `
. . . a`
−1 1 0 −1 −2 −(p−1)
(5) a` a`0 a`−1 . . . a`−(p−2)
1
where ⊗ is the Kronecker product [11] defined, for matri-
A` = 2 a `
a `
a `
. . . a`
1 0 −(p−3) .
ces B and C, by .. .. .. .. ..
. . . . .
b11 C b12 C . . . b1m C a`p−1 a`p−2 a`p−3 . . . a`0
b21 C b22 C . . . b2m C
B⊗C= .
.. .. .
Notice that to generate these matrices, we only need
.. . . the first row and column of each matrix A` , what is called
bm1 C bm2 C . . . bmm C the Toeplitz vector a` = (a`−(p−1) , . . . , a`p−1 ).
7.2. Total Variation Regularization. In many applica- 8.2. BCCB Structure. Another feasible structure arises
tions, the image to be reconstructed is known to be piece- when still considering a spatially invariant blur but assum-
wise constant with regular and sharp edges, like the one ing that the image has periodic boundary conditions. In
we are using as an example in this article. Total varia- this case, the matrix A has a Block-Circulant-Circulant-
tion (TV) regularization is a popular choice that allows Block (BCCB) structure, that is, a p by p block-circulant
the solution to preserve edges. Such regularization can be matrix with each block being a p by p circulant matrix,
8
reordering might be needed. So, the Tikhonov solution
A1 A2 A3 ... Ap
can be written as
Ap A1 A2 ... Ap−1 dc dTr
Xλ = Vc (Uc BUr ) VrT ,
T
A = Ap−1 Ap A1 ... Ap−2
, (dc dTr )2 + λ1
.. .. .. .. ..
. . . . . where 1 is a matrix of all ones, and dr and dc are the
A2 A3 A4 ... A1 diagonals of Σr and Σc , respectively.
where for ` = 1, . . . , p Notice that if Ar and Ac are Toeplitz matrices, then
` Ar ⊗ Ac has BTTB structure, and if Ar and Ac are cir-
a`2 a`3 a`p
a1 ... culant matrices, then Ar ⊗ Ac has BCCB structure.
`
ap
` a`1 a`2 ... a`p−1
`
A` = ap−1 ap
a`1 ... a`p−2
.
. .. .. .. 9. M ULTILEVEL M ETHODS
.. ..
. . . .
Because dealing with images of large size is difficult,
a`2 a`3 a`4 ... a`1
researchers keep working on finding ways to solve these
Notice that to generate these matrices, we only need large-scale inverse problems efficiently. One possible way
the first row of each matrix A` , a` = (a`1 , . . . , a`p ), so we is by developing multilevel methods. The main idea of a
do not need to store every entry of the matrix A, only the multilevel method is to define a sequence of systems of
vectors that define the matrices A` . We might never need equations decreasing in size,
to build the matrix A either to get a regularized solution.
Furthermore, BCCB matrices are diagonalizable by the A(n) x(n) = b(n) , 0 ≤ n ≤ L,
two-dimensional Discrete Fourier Transform (DFT) with where the superscript n denotes the n-th level (n = 0 be-
eigenvalues given by the Fourier transform of the first col- ing the original system), and to get an approximate solu-
umn of A. For instance, the Tikhonov solution can be tion, a correction, or some information at each level where
written as the computational cost would be smaller than solving the
conj(â) original system. At each level, the right-hand side is de-
Xλ = IDFT 2
DFT(B) , fined by
|â| + λ1
where Xλ and B are the matrices representing xλ and b(n+1) = R(n) b(n)
b, respectively (i.e., xλ = vec(Xλ ) and b = vec(B)), and the matrix by
â = mDFT(as ), conj(·) denotes the componentwise com-
plex conjugate, |â|2 is the m×m matrix whose entries are A(n+1) = R(n) A(n) P(n) ,
the squared magnitudes of the complex entries of â, 1 is where R(n) is called the restriction operator and P(n) is
a matrix of all ones, and denotes component-wise mul- the interpolation operator.
tiplication between matrices. Notice that IDFT stands for There are many ways of defining and using this se-
Inverse Discrete Fourier Transform. For more details, we quence of systems of equations to solve many different
direct the reader to [20, Chapter 5]. mathematical problems. To learn more about multilevel
methods for image deblurring applications, we recom-
8.3. Separable Blur Operator. Another special case is
mend reference [3, 6, 9, 17]
when the blurring operator A is separable, which means
that the blurring can be separated in the horizontal and
9.1. Wavelet-based approach. Here, we consider the
vertical directions as illustrated in Equation (1) that is, the
use of wavelet transforms as restriction and interpola-
matrix A can be written as
tion operators. In particular, we will work with the Haar
A = Ar ⊗ Ac . Wavelet Transform (HWT) because, as we will see, it
keeps the structures of the matrices involved. The one-
If we have the SVDs Ar = Ur Σr VrT and Ac =
dimensional HWT is the p × p orthonormal matrix W
Uc Σc VcT , we basically have the SVD of A by writing
defined by
A = Ar ⊗ Ac (6) 1 1 0 0 ... ... 0 0
0 0 1 1 ... ... 0 0
= (Ur Σr VrT ) ⊗ (Uc Σc VcT ) (7) .. .. .. .. . . . . .. ..
1 . . . . . .. .
T W1
W = √ 01 −1
= (Ur ⊗ Uc )(Σr ⊗ Σc )(Vr ⊗ Vc ) , (8) 0 0 0 ... ... 1 1
0 0 ... ... 0 0 = ,
W2
2
0. 0. 1. −1 ... ... 0 0
except that the elements in the diagonal matrix Σr ⊗ Σc
.. .. .. ... . . . . .. ..
might not be in the decreasing order, and therefore some . .. .
0 0 0 0 ... ... 1 −1
9
where W1 and W2 have dimension p/2 × p. The two-
dimensional HWT can be defined in terms of the one-
dimensional HWT by W2D = W ⊗ W ∈ Rm×m
with m = p2 . So, we define the restriction operator by
R = W1 ⊗ W1 ∈ R(m/4)×m and the interpolation oper-
ator by P = RT .
To motivate the use of HWT as a restriction operator,
on the left of Figure 19, we can see the coarser version of
xtrue defined by x(1) = Rxtrue . This is a 32 × 32 image
that still shows basically the same information as the orig-
inal 64 × 64 image with the letter H. So, reconstructing a F IGURE 20. The reconstructed coarse
coarse version could be enough for some tasks. Another images xreg of size 32 × 32, obtained
reason is that, we actually do this many times when com- using the optimal regularization param-
pressing images to save storage space in the computer. So eter and the discretization of the first de-
imagine getting the image b, and before transmitting it, rivative operator with the 2-norm regu-
we compute its coarser version b(1) = Rb and transfer larization on the left, and TV regular-
just b(1) (see right of Figure 19). The image b(1) might ization on the right.
have enough information to recover x(1) . A similar result can be shown for circulant matrices.
Corollary 1. Let C be a p × p circulant matrix and
p = 2s . Then, the 2s−1 × 2s−1 matrix C1 = W1 CW1T
is also circulant.
Let us consider the case when A = Ar ⊗Ac ∈ Rm×m .
Then, by property of the Kronecker product we have that
A(1) = (W1 ⊗ W1 )(Ar ⊗ Ac )(W1 ⊗ W1 )T
= (W1 Ar W1T ) ⊗ (W1 Ac W1T )
= A(1) (1)
r ⊗ Ac .
s−1 s−1 1 T
the 2 ×2 matrix T = W1 TW1 is also Toeplitz = (U(n) (n) (n) (n) (n) (n) T
r ⊗ Uc )(Σr ⊗ Σc )(Vr ⊗ Vc ) .
with Toeplitz vector t1 = t̃(1 : 2 : 2p − 3), where t̃ = T̃t
with T̃ being the Toeplitz matrix with Toeplitz vector with This gives us basically the SVD of A(n) , except that the
all zeros and t0 = t−2 = 1/2 and t−1 = 1. elements in the diagonal matrix Σ(n)
r ⊗ Σ(n)
c might not
10
be in the decreasing order, and therefore some reordering linear operators, and large-scale computation. The visual
would be needed. nature of the problem provides compelling motivation for
There are also other efficient methods such as com- students and allows the efficacy of various techniques to
puting the closest BCCB matrix and using it as precon- be easily assessed. The references [13, 14, 20] contain
ditioner (see [20, Chapter 4] or [14] for more details). excellent introductions to this subject. Furthermore, the
code to generate the figures appearing in this article is
available at https://github.com/ipiasu/AMS_
10. C ONCLUSIONS AND OUTLOOK
Notices_AEP.
Our intention has been to introduce readers to the prob- In addition, this field is an active area of research that
lem of image deblurring, the mathematical issues that aligns with the recent developments in machine learning
arise, and a few techniques for addressing them. As men- and convolutional neural networks. Many classical tech-
tioned in the introduction, these techniques may be ap- niques that have been traditionally used in image deblur-
plied to a wide range of related problems. Indeed, we ring can serve as a tool to speed up the computations of
outlined a number of alternative strategies, for example, the more recent methods that aim to learn models and pa-
in choosing appropriate boundary conditions or in select- rameters when training data are available or train a net-
ing the best regularization parameter, as some strategies work to classify images. More particularly, we mentioned
are better suited to a specific range of applications. earlier in the manuscript that Tikhonov regularization is
We believe this subject is accessible to both undergrad- related to ridge regression, a fundamental tool in machine
uate and graduate students and can serve as a good intro- learning. The connection to machine learning goes much
duction to inverse problems, working with ill-conditioned deeper.
11. ACKNOWLEDGMENTS
This article will appear on the AMS Notices.
R EFERENCES
[1] Harbir Antil, Zichao Wendy Di, and Ratna Khatri. Bilevel optimization, deep learning and fractional laplacian regularization with applications
in tomography. Inverse Problems, 36(6):064001, 2020.
[2] Emmanuel J Candes, Justin K Romberg, and Terence Tao. Stable signal recovery from incomplete and inaccurate measurements. Communi-
cations on Pure and Applied Mathematics: A Journal Issued by the Courant Institute of Mathematical Sciences, 59(8):1207–1223, 2006.
[3] Raymond H Chan and Ke Chen. A multilevel algorithm for simultaneously denoising and deblurring images. SIAM Journal on Scientific
Computing, 32(2):1043–1063, 2010.
[4] Julianne Chung, Matthias Chung, Silvia Gazzola, and Mirjeta Pasha. Efficient learning methods for large-scale optimal inversion design. arXiv
preprint arXiv:2110.02720, 2021.
[5] Julianne Chung and Malena I Español. Learning regularization parameters for general-form tikhonov. Inverse Problems, 33(7):074004, 2017.
[6] Marco Donatelli and Stefano Serra-Capizzano. On the regularizing power of multigrid-type algorithms. SIAM Journal on Scientific Computing,
27(6):2053–2076, 2006.
[7] Vijayan Ellappan and Vishal Chopra. Reconstruction of noisy and blurred images using blur kernel. In IOP Conference Series: Materials
Science and Engineering, volume 263, page 042024. IOP Publishing, 2017.
[8] Heinz Werner Engl. Discrepancy principles for tikhonov regularization of ill-posed problems leading to optimal convergence rates. Journal of
optimization theory and applications, 52(2):209–215, 1987.
[9] Malena I Español and Misha E Kilmer. Multilevel approach for signal restoration problems with toeplitz matrices. SIAM Journal on Scientific
Computing, 32(1):299–319, 2010.
[10] Gene H Golub, Michael Heath, and Grace Wahba. Generalized cross-validation as a method for choosing a good ridge parameter. Technomet-
rics, 21(2):215–223, 1979.
[11] Gene H Golub and Charles F Van Loan. Matrix Computations, 4th ed. Johns Hopkins University Press, Baltimore, 2013.
[12] Per Christian Hansen. Regularization tools: a matlab package for analysis and solution of discrete ill-posed problems. Numerical algorithms,
6(1):1–35, 1994.
[13] Per Christian Hansen. Discrete inverse problems: insight and algorithms. SIAM, 2010.
[14] Per Christian Hansen, James G Nagy, and Dianne P O’leary. Deblurring images: matrices, spectra, and filtering. SIAM, 2006.
[15] Per Christian Hansen and Dianne Prost O’Leary. The use of the l-curve in the regularization of discrete ill-posed problems. SIAM journal on
scientific computing, 14(6):1487–1503, 1993.
[16] Michael Lustig, David L Donoho, Juan M Santos, and John M Pauly. Compressed sensing MRI. IEEE signal processing magazine, 25(2):72–
82, 2008.
[17] Serena Morigi, Lothar Reichel, Fiorella Sgallari, and Andriy Shyshkov. Cascadic multiresolution methods for image deblurring. SIAM Journal
on Imaging Sciences, 1(1):51–74, 2008.
[18] Mirjeta Pasha, Arvind K Saibaba, Silvia Gazzola, Malena I Espanol, and Eric de Sturler. Efficient edge-preserving methods for dynamic
inverse problems. arXiv preprint arXiv:2107.05727, 2021.
11
[19] Rosemary A Renaut, Saeed Vatankhah, and Vahid E Ardestani. Hybrid and iteratively reweighted regularization by unbiased predictive risk
and weighted GCV for projected systems. 39(2):B221–B243, 2017.
[20] Curtis R Vogel. Computational methods for inverse problems. SIAM, 2002.
1 D EPARTMENT OF M ATHEMATICS , G RAND VALLEY S TATE U NIVERSITY, A LLENDALE , MI.
2 S CHOOL OF M ATHEMATICAL AND S TATISTICAL S CIENCES , A RIZONA S TATE U NIVERSITY, T EMPE , AZ.
E-mail addresses: austind@gvsu.edu, malena.espanol@asu.edu, mpasha@asu.edu
12