Professional Documents
Culture Documents
is well-posed provided
(Uniqueness) ; and
(Stability) is continuous.
exists
det
well-posed
..
.
-
--
-
('&
%$
#"
!
/
+,
*
)
5
3-
0 -
! -
-4
('&
%$
0
#"
/
1!
(Compact) diagonal operator on (Hilbert) space
6
<
8
9
8 7
@
: 8
/ A
: 8
=
:
:
/
/
/
/
/
/
?>=
Define by
6
6
8
8
8
B:
:
/
/
/
:
:
/
/
/
/
Formal (unbounded) inverse is
9C 7
;
B
C
C
:
:
:
/
/
/
:
/
/
/
:
so we have uniqueness (and existence of solutions for
certain ).
;
D
I :
:
/
/
/
:
:
:
/
/
/
GF
K H
J
Then , but
--
--
/ A
7
;
E.g., , but
D
DB
D
EL
D
D
D
:
:
/
/
/
:
:
:
/
/
/
7
.
6
D
D
D
:
:
:
/
/
/
Can replace (finite dimensional) by pseudo-inverse
5
But ...
QP
X
T
RS
such that
TZ
Y
N M
P
/T
N
RS
,
:
Hilbert spaces. To obtain regularized soln to ,
choose to fit data in least-squares sense, but penalize
solutions of large norm. Solve minimization problem
-
--
--
X -
-
('&
U
%$
#"
[
N
[
*
)
;
]
]
7
^
U
/
X
I
G
H
` _
a
is called the regularization parameter.
b
X
d d
d ed d
d ed d
d e
d ed d
d e
d e
d e
d e
d f
d f
d f
d f
d e
d e
d e
d e
h heeh h
k e k k
D
D
8
d e
d e
d e
d e
k e
/
7
;
c
U
d e
d e
d e
d e
e
e
D
d e
d e
d e
d e
8
/
/
def
def
def
def
F
I
F
GF
GF
I
G
g H
H
)
j
2
Contours for ||Ax−y|| Contours for ||Ax−(y−η)||2
2 2
1.8 1.8
1.6 1.6
1.4 1.4
1.2 1.2
1 1
y
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
x x
SAMSI Opening Workshop – p.10/33
Geometry of Tikhonov Regularization
;-
--
7
;
--
-
-
c
D
VU
U
/
F
I
G
H
N
Contours for ||Ax−y||2 + α*||x||2 Contours for ||Ax−(y+η)||2 + α*||x||2
2 2
1.8 1.8
1.6 1.6
1.4 1.4
1.2 1.2
1 1
y
y
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
x x
:
random variables.
7
-
;
7
-
7
;
7
;
l
GF
GF
v I
GF
pr GF
I
H
u = H
H
rs
rs
s
=
r
=
=
QpqJ
nJ
q
m
nq
mno
nq
tn
mn
)
Maximum a posteriori (MAP) extimator is max of posterior
pdf. Equivalently, minimize w.r.t.
7
-
;
7
-
7
;
yxw
yxw
yxw
$
l
l
.
.
F
I
v G
v H
H
nz v
p{
n =K
nz v
=
=
n
q
m
nq
First term on rhs is “fit-to-data” term; second is
“regularization” term.
c
7
; l
7 VU
-
Normal
-- l 7
; 7
and ; :
V ~
B7
|
. ^
B7 ;
- l
- ~ D l
j
; ~
D
7
--
:
:
, then prior is
. - ~
- j
X ^ . -
; -
--
~ ~ .
j BE
-
-
~
Illustrative Example
BE
~
j
SNR
/
, conditional pdf is
]
/
9
7
;
In matrix case, , diag ,
:
:
:
/
/
/
:
/
/
/
and with
:
:
/
/
/
/
/
/
:
]
=
=
=
=
=
:
:
]
]
^
^
: 0
:
=
: =
:
SAMSI Opening Workshop – p.14/33
Tikhonov Filtering
In the case of Tikhonov regularization, using the SVD
(and assuming matrix with for
]
b
=
simplicity),
;
]
]
7
N M
^
U
X
;
]
]
7
X
^
U
;
]
]
7
U
X
D
=
]
7
;
diag
X
=
=
GF
I
H
`
o
97
= ;
If , then , so
D
X
N
TZ
Y
5
]
7
E
9
;
diag as
N M
D
X
=
/
SAMSI Opening Workshop – p.15/33
Tikhonov Filtering, Continued
7
;
Plot of Tikhonov filter function shows that
o
N
o
N
Tikhonov regularization filters out singular components that
are small (relative to ) while retaining components that are
X
large.
Filter Functions for TSVD and Tikhonov Regularization
0.8
w (s2)
0.6
α
0.4
0.2
0
−5 −4 −3 −2 −1 0 1
10 10 10 10 10 10 10
s2
: X
¡!
7
=
:
= ;
N
D
/ X
=
:
Has “sharp cut-off” behavior instead of “smooth roll-off
behavior” of Tikhonov filter.
-
7
;
--
-
c
B
£
using gradient descent iteration, initial guess , and
--
E-
-
fixed step length parameter .
D
0¤@
{
{
7
;
grad
¥
c
DB
.¤
:
:
:
:
/
/
/
{
{
]
7
;
.¤
.
{
]
]
7
;
^
¤U
.¤
= ;{
]
7
;
diag
D
/
.¤
.
F
I
G
H
s
s
r
pq ¨
v
§u¦
qQpJ =
p
Landweber Filter Functions for k=10(−) and k=100(−−)
0.8
w (s )
2
0.6
k
0.4
0.2
0
−4 −3 −2 −1 0 1
10 10 10 10 10 10
SAMSI Opening Workshop – p.19/33
s2
Effect of Regularization Parameter
Illustrative Example: 1-D deconvolution with Gaussian
. ª
ǻ7
;
BE
;
kernel and discrete data
¬
©
7
P8 ;
ª7
;
noise
®
ª®
D
¯
ª
U
©
/
=
.=
T
RS
:
/
/
/
:
£
0.8
0.6
0.4
0.2
−0.2
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 SAMSI Opening Workshop – p.20/33
x axis
Graphical Representation of SVD
Singular Values Singular Vectors
Singular Values of K Singular Vector v1 v4
0
10 0.2 0.2
−2
10 0.15 0.1
−4
10 0.1 0
−6
10 0.05 −0.1
−8 0 −0.2
10 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
x axis x axis
i
σ
−10 v v
10 10 20
0.2 0.3
−12 0.2
10
0.1
0.1
−14 0
10
0
−0.1
−16
10 −0.2
−0.1
−0.3
−18
10 −0.2 −0.4
0 10 20 30 40 50 60 70 80 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
index i x axis x axis
;
7
Tikhonov regularized solution is .
°]
^
U
8
X
N
Solution error is .
N ±
8
.P 8
T
N
RS
Norm of Tikhonov Solution Error α = 0.0014384
4 1.2
3.5 1
3 0.8
2.5 0.6
||eα||
xα(t)
2 0.4
1.5 0.2
1 0
0.5 −0.2
−6 −4 −2 0 0 0.2 0.4 0.6 0.8 1
10 10 10 10
α t axis
α = 1e−06 α=1
2 1.2
1.5 1
1 0.8
0.5 0.6
xα(t)
xα(t)
0 0.4
−0.5 0.2
−1 0
−1.5 −0.2
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
t axis t axis
Y
TZ
N ²
N Y
TZ
Solution Error:
N
. N
Predictive Error:
P
. QP
N M
O
RS °
Regularized Solution:
T RS
°
T
F 7
N ²
N M QP
diag
7 RS
³ G
7 uo ¨ T
. N
Linear, Additive Noise Data Model:
H VU
´´= ^ 7
N M
O QP ;
RS = ;
. E
9
^ I T
; ;
Error Indicators
U =
QP ¶µ³
uq
G ]°
RS F NOM
u V
T H
= I
U tp
´´
NOV M
SAMSI Opening Workshop – p.23/33
Unbiased Predictive Risk Estimator
The Influence Matrix is
TZ
Y
7
;
¸
N M
O
X
:
so we can write the predictive error as
7
;
7
;
¸
¸
U
N ·
QP
X
XV
T
RS
Residual is
TZ
Y
7
;
7
;
V ;
. °
^
U
N ¹
ºP
X
X
.
T
N
RS
I
G
qQpJ H
p
»
Let denote expected value operator. Assume is
¼
P
T
RS
V ¼7
;
deterministic (or independent of ), assume , and
V
7
;
So up to const
7 ¼
X V -- .
; B
-- --
V ¼ F
Y 7
XV
;
¸ V ¼
UPRE, Continued
7
X ]
;
¸
, an unbiased estimator for
7
-- X
N · V ;
-- I
is
X
¦
7
;
.
X
--
--
Predictive error norm and solution error norm
N ·
--
--
--
--
N ¹
TZ
Y
7
;
;
/
;
trace ^
¸
X
.
This does not require prior knowledge of variance .
~
j
SAMSI Opening Workshop – p.26/33
Illustrative Example of Indicators
:
7
;
2-D image reconstruction problem, noise N ,
^
V
~
|
j
Tikhonov regularization. o-o indicates soln error norm;
7
;
indicates GCV; – indicates ; and indicates
X
.
.
.
.
predictive error norm.
U(α) is solid line; P(α) is dashed line; GCV(α) is dot−dashed line.
7
10
6
10
5
10
4
10
3
10
2
10
−6 −5 −4 −3 −2 −1 0 1
10 10 10 10 10 10 10 10
Regularization Parameter α
SAMSI Opening Workshop – p.27/33
Mathematical Summary
There exists a well-developed mathematical theory of
regularization.
There are a number of different approaches to
regularization.
optimization-based (equivalent to MAP)
filtering-based
iteration-based
There are robust schemes for choosing regularization
parameters.
 Ã
Ã
 Ã
Á97
Á 7
9
Á 7
9
;
Á ®
®
^
.Á
/ Â
.
:
:
F
I
F
I
G
ÆÅ H
H
Ä
¨
n
tJ
p
This is measured by a ccd array (digital camera), giving
data
Á 7
9
= ;
“noise”
®
U
Â
=
: =
/
For high contrast imaging (dim object near very bright
object), accurate modeling of noise is critical.
With ordinary (and even weighted) least squares, dim
object is missed.
7
;
®
D
¯
U
= WVU
Ç
=
=
:
:
/
/
/
:
Photon count for “signal”
7
| ;
Á97
= Ê;
Poisson
7É
Ç
Â
=
: =
: =
=
/
Background photon count
7
;
Poisson fixed, known
| È
È
=
/
Instrument “read noise”
7
N fixed, known
V
~
|
=
/
SAMSI Opening Workshop – p.30/33
Imaging Example, Continued
°7
Ì-
;
Log likelihood ( ) is messy
yxw
$
Ë
Ï
Ð
<
È
Ì7
;
= U
±
Ï
¨
ÓÒ
g
r
Ì7
;
°
yxw
Í
±
$
Î
.
ÀÑ
>=
>£
Light source (object) intensity is nonnegative.
ǻ7
;
Constraint .
8
°7
;
ÔÕ
b
X
:
/
Regularization parameter (strength of prior) is unknown.
SAMSI Opening Workshop – p.31/33
Applied Mathematician’s Wish List
Optimization-based regularization methods (Tikhonov,
MAP) require soln of minimization problems. Need fast,
robust, large-scale, nonlinear constrained numerical
optimization techniques.
When the parameter-to-observation map is nonlinear,
regularization functionals may be non-convex. Need
optimization methods which yield the global minimizer
(not just a local min) and are fast, robust, ....
Need indicators of reliability (e.g., confidence intervals)
for regularized solutions.
Need good priors.
Need fast, robust schemes for choosing regularization
parameters.
SAMSI Opening Workshop – p.32/33
Challenge for the Statistics Community
Can MCMC techniques provide fast, robust alternatives
to optimization-based regularization methods?