Are you sure?
This action might not be possible to undo. Are you sure you want to continue?
Testorf/1597980/Ch04
C HA P T E R 4
The RadonWigner
Transform in Analysis,
Design, and Processing
of Optical Signals
Walter D. Furlan and Genaro Saavedra
Universitat de Val` encia, Optics Department
Burjassot, Spain
4.1 Introduction
One of the main features of phase space is that its conjugate coor
dinates are noncommutative and cannot be simultaneously speciﬁed
withabsolute accuracy. As a consequence, there is nophasespace joint
distribution that can be formally interpretedas a joint probability den
sity. Indeed, most of the classic phasespace distributions, such as the
Wigner distribution function (WDF), the ambiguity function (AF), or
the complex spectrogram, have difﬁcult interpretation problems due
to the complex, or negative, values they have in general. Besides, they
maybe nonzeroeveninregions of the phase space where either the sig
nal or its Fourier transform vanishes. This is a critical issue, especially
for the characterization of nonstationary or nonperiodic signals. As
an alternative, the projections (marginals) of the phasespace distribu
tions are strictly positive, and as we will see below, they give informa
tion about the signal on both phasespace variables. These projections
can be formally associated with probability functions, avoiding all
interpretation ambiguities associated with the original phasespace
distributions. This is the case of the RadonWigner transform (RWT),
107
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
108 C h a p t e r F o u r
closely related to the projections of the WDF in phase space and also
intimately connected with AF, as will be shown.
The general structure of this chapter is as follows. In Sec. 4.2, a gen
eral overview of mathematical properties of the RWT is given, and
a summary of different optical setups for achieving it is presented.
Next, the use of this representation in the analysis of optical signals
and systems is developed in several aspects, namely, the computation
of diffraction intensities, the optical display of Fresnel patterns, the
amplitude and phase reconstruction of optical ﬁelds, and the calcula
tion of merit function in imaging systems. Finally, in Sec. 4.4, a review
of design techniques, based on the utilization of the RWT, for these
imaging systems is presented, along with some techniques for optical
signal processing.
4.2 Projections of the Wigner Distribution
Function in Phase Space: The
RadonWigner Transform (RWT)
The RWT was ﬁrst introduced for the analysis and synthesis of
frequencymodulated time signals, and it is a relatively new formal
isminoptics.
1, 2
However, it has foundseveral applications inthis ﬁeld
during the last years. Many of them, such as the analysis of diffrac
tionpatterns, the computationof merit functions of optical systems, or
the tomographic reconstruction of optical ﬁelds, are discussed in this
chapter. We start by presenting the deﬁnition and some basic proper
ties of the RWT. The optical implementation of the RWT which is the
basis for many of the applications is discussed next.
Note, as a general remark, that for the sake of simplicity most of the
formal deﬁnitions for the signals used hereafter are restricted to one
dimensional signals, i.e., functions of a single variable f (x). This is
mainly justiﬁed by the speciﬁc use of these properties that we present
in this chapter. The generalization to more than one variable is in most
cases straightforward. We will refer to the dual variables x and as
spatial andspatialfrequency coordinates, since we will deal mainlywith
signals varying on space. Of course, if the signal is a function of time
instead of space, the terms time and frequency should be applied.
4.2.1 Deﬁnition and Basic Properties
We start this section by recalling the deﬁnition of the WDF associated
with a complex function f (x), namely,
W{ f (x
), x, } = W
f
(x, )
=
+∞
−∞
f
x +
x
2
f
∗
x −
x
2
exp(−i 2x
) dx
(4.1)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 109
which also can be deﬁned in terms of the Fourier transform (FT) of
the original signal
F{ f (x), } = F() =
+∞
−∞
f (x) exp(−i 2x) dx (4.2)
as
W
f
(x, ) =
+∞
−∞
F
+
2
F
∗
−
2
exp(i 2
x) d
(4.3)
It is interesting to remember that any WDF can be inverted to recover,
up to a phase constant, the original signal or, equivalently, its Fourier
transform. The corresponding inversion formulas are
3
f (x) =
1
f
∗
(x
)
+∞
−∞
W
f
x + x
2
,
exp[i 2(x − x
)] d (4.4)
F() =
1
F
∗
(
)
+∞
−∞
W
f
x,
+
2
exp[−i 2( −
)x] dx (4.5)
Note that these equations state the uniqueness of the relationship be
tween the signal and the corresponding WDF (except for a phase con
stant). It is straightforward to deduce from these formulas that the
integration of the WDF on the spatial or spatialfrequency coordinate
leads to the modulus square of the signal or its Fourier transform,
respectively, i.e.,
 f (x)
2
=
+∞
−∞
W
f
(x, ) d (4.6)
F()
2
=
+∞
−∞
W
f
(x, ) dx (4.7)
These integrals, or marginals, can be viewed as the projection of the
function W
f
(x, ) in phase space along straight lines parallel to the
axis [inEq. (4.6)] or to the x axis [inEq. (4.7)]. These cases are particular
ones of all possible projections along straight lines of a given function
in phase space. In fact, for any function of (at least) two coordinates,
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
110 C h a p t e r F o u r
R
g
(x
θ
, θ)
g(x, y)
y
x
x
θ
x
θ
y
θ
θ
FIGURE 4.1 Projection scheme for the deﬁnition of the Radon transform.
say, g(x, y), its Radon transform is deﬁned as a generalized marginal
R{g (x, y) , x
, } = R
g
(x
, ) =
+∞
−∞
g (x, y) dy
(4.8)
where, as presented in Fig. 4.1, x
and y
are the coordinates rotated
by an angle . It is easy to see from this ﬁgure that
R
g
(x
, +) = R
g
(−x
, ) (4.9)
Thus, the reduceddomain ∈ [0, ) is usedfor R
g
(x
, ). Note that the
integration in the above deﬁnition is performed along straight lines
characterized, for a given pair (x
, ), by the equation
y =
x
sin
−
x
tan
for = 0,
2
x = x
for = 0
y = x
for =
2
(4.10)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 111
and therefore Eq. (4.8) can be reformulated as
R
g
(x
, ) =
⎧
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎨
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎩
+∞
−∞
g
x,
x
sin
−
x
tan
dx for = 0,
2
+∞
−∞
g (x
, y) dy for = 0
+∞
−∞
g (x, x
) dx for =
2
(4.11)
Thus, when we consider as projected function W
f
(x, ), we can deﬁne
the generalized marginals as the Radon transform of this WDF, namely,
R{W
f
(x, ), x
, } = R
W
f
(x
, ) =
+∞
−∞
W
f
(x, ) d
=
+∞
−∞
W
f
(x
cos − sin, x
sin + cos ) d
=
⎧
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎨
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎩
+∞
−∞
W
f
x,
x
sin
−
x
tan
dx for = 0,
2
+∞
−∞
W
f
(x
, ) d for = 0
+∞
−∞
W
f
(x, x
) dx for =
2
(4.12)
where, in the last expression, we have explicitly considered the equa
tions for the integration lines in the projection. In terms of the original
signal, this transform is called its RadonWigner transform. It is easy
to show that
R
W
f
(x
, ) = RW
f
(x
, ) =
+∞
−∞
+∞
−∞
f
x
cos − sin +
x
2
×f
∗
x
cos − sin −
x
2
×exp[−i 2(x
sin + cos )x
] dx
d (4.13)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
112 C h a p t e r F o u r
By performing a proper change in the integration variables, the fol
lowing more compact expression can be obtained
RW
f
(x
, )
=
⎧
⎪
⎪
⎪
⎪
⎨
⎪
⎪
⎪
⎪
⎩
1
sin
+∞
−∞
f (x) exp
i
x
2
tan
exp
−i 2
x
x
sin
dx
2
for = 0,
2
 f (x
0
= x)
2
for = 0
F
x
/2
=
2
for =
2
(4.14)
From this equation it is clear that
RW
f
(x
, ) ≥ 0 (4.15)
This is a very interesting property, since the WDF cannot be positive
in whole phase space (except for the particular case of a Gaussian
signal). Note also that from Eq. (4.14) a symmetry condition can be
stated, namely,
RW
f
(x
, −) = RW
f
∗ (−x
, ) (4.16)
so that for real signals, that is, f (x) = f
∗
(x)∀x ∈ R, one ﬁnds
RW
f
(x
, −) = RW
f
(−x
, ) (4.17)
and, therefore, for this kind of signal the reduced domain ∈ [0, )
in the Radon transform is clearly redundant. In this case, the range
∈ [0, /2] contains in fact all the necessary values for a full deﬁnition
of the RWT.
Equation (4.14) also allows one to link the RWT with another inte
gral transform deﬁned directly from the original signal, namely, the
fractional Fourier transform (FrFT). This transformation, often con
sidered a generalization of the classic Fourier transform, is given by
F
p
{ f (x), ␣}
= F
p
(␣) =
⎧
⎪
⎪
⎪
⎪
⎪
⎪
⎨
⎪
⎪
⎪
⎪
⎪
⎪
⎩
exp[i (+␣
2
/tan)]
√
i sin
+∞
−∞
f (x)
×exp
i
x
2
tan
exp
−i 2
␣x
sin
dx for = 0,
2
f (␣) for = 0
F(␣) for =
2
(4.18)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 113
where = p/2. From this deﬁnition, it is easy to see that
RW
f
(x
, ) = F
2/
(x
)
2
(4.19)
so that the RWT can be also interpreted as a twodimensional repre
sentation of all the FrFTs of the original function.
Another interesting relationship can be established between the
RWT and the AF associated with the input signal. For our input signal
the AF is deﬁned as
A{ f (x),
, x
} = A
f
(
, x
)
=
+∞
−∞
f
x +
x
2
f
∗
x −
x
2
exp(−i 2
x) dx (4.20)
which can be understood as the twodimensional FT of the WDF, i.e.,
F
2D
{W
f
(x, ),
, x
} =
+∞
−∞
+∞
−∞
W
f
(x, ) exp[−i 2(
x + x
)] dx d
= A
f
(
, −x
) (4.21)
There is a wellknown relationship between the twodimensional FT
of a function and the onedimensional Fourier transformation of its
projections. This link is established through the central slice theorem,
which states that the values of the onedimensional FT of a projection
at an angle give a central proﬁle—or slice—of the twodimensional
FT of the original signal at the same angle. If we apply this theorem
to the WDF, it is straightforward to show that
F{RW
f
(x
, ),
} = A
f
(
cos , −
sin) (4.22)
i.e., the onedimensional FT of the RWT for a ﬁxed projection angle
provides a central proﬁle of the AF A
f
(
, x
) along a straight line
forming an angle − with the
axis. These relationships together
with other links between representations in the phase space are sum
marized in Fig. 4.2.
To conclude this section, we consider the relationship between the
RWTof aninput onedimensional signal f (x) andthe RWTof the same
signal but after passing through a ﬁrstorder optical system. In this
case, the input signal undergoes a canonical transformation deﬁned
through four real parameters (a, b, c, d) or, equivalently, by a 2×2 real
matrix
M =
a b
c d
(4.23)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
114 C h a p t e r F o u r
f(x)
W
W
f
(x, ξ)
A
r
f
(x, x' ) = f (x + x'/2) f
*
(x + x'/2)
F
{
x
'
→
ξ
}
R
RW
f
(x
θ
, θ)
A
f
(ξ', x')
F
–
1
{
ξ
→
x
'}
F
{
x
→
ξ
'}
F–
1
{
ξ
' →
x
}
A
–1
W
–1
R
–1
(1) (2)
(3) (4) F
–
1
{
(
√
ξ
'
2
+
x
'
2 ,
θ
)
p
→
xθ
}
F
{
xθ
→
(
√
ξ
'
2
+
x
'
2 ,
θ
)
p
}

F
p

2
{
x
→
x
θ
}
θ
=
p
π
/
2
FIGURE 4.2 Relationship diagram between the original signal f (x) and
different phasespace representations. F, F
p
, W, A, and Rstand for FT,
FrFT, WDF integral, AF transform and Radon transformation, respectively,
while −1 represents the corresponding inverse operator. (1) WDF and
inverse transform; (2) AF and inverse transform; (3) projection (Radon)
transformation and tomographic reconstruction operator; (4) expression of
the central slice theorem applied to Radon transform and AF; and (, )
p
represents polar coordinates in phase space.
in such a way that the transformed signal g(x) is given by
g(x)
=
⎧
⎪
⎪
⎨
⎪
⎪
⎩
1
√
i b
exp
−i dx
2
b
+∞
−∞
f (x
) exp
−i ax
2
b
exp
i 2
b
x x
dx
b = 0
exp
−i cx
2
a
1
√
a
f
x
a
b = 0
(4.24)
which are the onedimensional counterparts of Eqs. (3.4) and (3.7). We
are restricting our attention to nonabsorbing systems corresponding
to the condition det M = ad −bc = 1.
The application of a canonical transformation on the signal pro
duces a distortiononthe corresponding WDF according to the general
law
W
g
(x, ) = W
f
(ax +b, cx +d) = W
f
(x
,
) (4.25)
where the mapped coordinates are given by
x
=
a b
c d
x
(4.26)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 115
By applying the deﬁnition in Eq. (4.12), it is straightforward to obtain
RW
g
(x
, ) =
⎧
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎨
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎩
+∞
−∞
W
g
x,
x
sin
−
x
tan
dx for = 0,
2
+∞
−∞
W
g
(x
, ) d for = 0
+∞
−∞
W
g
(x, x
) dx for =
2
=
⎧
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎨
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎩
+∞
−∞
W
f
ax +b
x
sin
−
x
tan
, cx
+d
x
sin
−
x
tan
dx for = 0,
2
+∞
−∞
W
f
(ax
+b, cx
+d) d for = 0
+∞
−∞
W
f
(ax +bx
, cx +dx
) dx for =
2
⎫
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎬
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎭
∝ RW
f
(x
,
) (4.27)
where the mapped coordinates for the original RWT are given by
tan
= −
a tan −b
c tan −d
, x
=
x
a sin −b cos
sin
(4.28)
Let us consider in the following examples a spatially coherent light
distribution f (x), with wavelength , that travels along a system that
imposes a transformation in the input characterized by an abcd trans
form. Special attention is usually paid to the cases = 0, /2 since,
according to Eqs. (4.6) and (4.7), the modulus squared of the abcd
transform in Eq. (4.24) and its FT are then obtained, respectively.
1. Coherent propagation through a (cylindrical) thin lens. In this
case the associated M matrix for the transformation of the light
ﬁeld is given by
M
L
=
1 0
1
f
1
(4.29)
with f being the focal length of the lens. Thus, the RWT for the
transformed amplitude light distribution is given in this case by
RW
g
(x
, ) ∝ RW
f
x
,
, tan
= − f
tan
tan − f
,
x
=
x
sin
sin
(4.30)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
116 C h a p t e r F o u r
A careful calculation for the case of = 0 leads to
RW
g
(x
0
, 0) = g(x
0
)
2
∝ RW
f
(x
0
, 0) (4.31)
while for the value = /2 the following result is obtained
RW
g
x
/2
,
2
∝ RW
f
x
/2
sin
,
, tan
= − f
(4.32)
Note that the effect of this propagation through a thin lens of
focal length f is also physically equivalent to the illumination of
the incident light distribution by a spherical wavefront whose
focus is located at a distance = f from the input plane. Thus,
the same results discussedhere canbe appliedstraightforwardly
to that case.
2. Freespace (Fresnel) propagation. If we consider now the Fres
nel approximation for the propagation of a transverse coherent
light distribution f (x) by a distance z, namely,
g (x) =
+∞
−∞
f (x
) exp
¸
i
z
(x
− x)
2
dx
(4.33)
the transformation matrix M is given by
M
F
=
1 −z
0 1
(4.34)
and, therefore, the transformed RWT can be calculated through
the expression
RW
g
(x
, ) ∝ RW
f
(x
,
), tan
= tan −z,
x
=
x
sin +z cos
sin
(4.35)
For the projection with = 0, one obtains
RW
g
(x
0
, 0) = g(x
0
)
2
∝ RW
f
(x
,
), tan
= −z,
x
=
x
0
z
sin
(4.36)
and for the orthogonal projection = /2 the following result
is achieved
RW
g
x
/2
,
2
∝ RW
f
x
/2
,
2
(4.37)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 117
3. Magniﬁer. If a uniform scale factor m is applied to the input
function, the associated M matrix is given by
M
m
=
1
m
0
0 m
(4.38)
In this case, the RWT is transformed according to the law
RW
g
(x
, ) ∝ RW
f
(x
,
), tan
=
1
m
2
tan,
x
=
mx
sin
sin
(4.39)
The vertical and horizontal projections are given here simply by
the following formulas.
RW
g
(x
0
, 0) = g(x
0
)
2
∝ RW
f
x
0
m
, 0
(4.40)
RW
g
x
/2
,
2
∝ RW
f
mx
/2
,
2
4.2.2 Optical Implementation of the RWT:
The RadonWigner Display
Like any other phasespace function, the RWT also enables an optical
implementation that is desirable for applications in the analysis and
processing of optical signals. The correct ﬁeld identiﬁcation requires
a large number of Wigner distribution projections, which raises the
necessity to design ﬂexible optical setups to obtain them. The rela
tionship between the RWT and the FrFT, expressed mathematically
by Eq. (4.19), suggests that the optical computation of the RWT is pos
sible directly from the input function, omitting the passage through
its WDF. In fact, the RWT for a given projection angle is simply the in
tensity registered at the output plane of a given FrFT transformer. For
onedimensional signals, the RWT for all possible projection angles
simultaneously displays a continuous representation of the FrFT of a
signal as a function of the fractional Fourier order p, and it is known
as the RadonWigner display (RWD). This representation, proposed by
Wood and Barry for its application to the detection and classiﬁcation
of linear FM components,
1
has found several applications in optics as
we will see later in this chapter.
Different and simple optical setups have been suggested to im
plement the FrFT, and most have been the basis for designing other
systems toobtainthe RWD. The ﬁrst one describedinthe literature, de
signed to obtain the RWD of onedimensional signals, was proposed
by Mendlovic et al.
4
It is based on Lohmann’s bulk optics systems
for obtaining the FrFT.
5
In this method, the onedimensional input
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
118 C h a p t e r F o u r
function is converted to a twodimensional object by the use of cylin
drical lenses to allowthe constructionof a multichannel processor that
optically implements the calculations of the RWD. The setup consists
of three phase masks separated by ﬁxed distances in free space. The
masks consist of many strips, each one representing a different chan
nel that performs an FrFT with a different order over the input signal.
Each strip is a Fresnel zone plate with a different focal length that is
selected for obtaining the different fractional orderp. Thus, the main
shortcoming of the RWD chart produced by this setup is that it has a
limited number of projection angles (or fractional orders). Besides the
very poor angular resolution, the experimental results obtained in the
original paper are actually very far from the theoretical predictions.
A truly continuous display, i.e., a complete RWD setup, was pro
posed by Granieri et al.
6
This approach is based on the relationship
between the FrFT and Fresnel diffraction,
7, 8
which establishes that
every Fresnel diffraction pattern of an input object is univocally
related to a scaled version of a certain FrFT of the same input. There
fore, if the input function f (x) is registered in a transparency with
amplitude transmittance t(x/s), with s being the construction scale
parameter, then the FrFT of the input can be optically obtained by
freespace propagation of a spherical wavefront impinging on it.
Actually, the Fresnel diffraction ﬁeld U(x, R
p
) obtained at distance
R
p
from the input, which is illuminated with a spherical wavefront
of radius z and wavelength , is related to the FrFT of order p of the
input function F
p
{t (x) , ␣} as follows.
9
U(x, R
p
) = exp
i x
2
¸
z(1 − M
p
) − R
p
z R
p
M
2
p
F
p
t
x
M
p
, x
(4.41)
where M
p
is the magniﬁcation of the optical FrFT. For each fractional
order, the values of M
p
and R
p
are related to the system parameters
s, , and z through
R
p
=
s
2
−1
tan( p/2)
1 +s
2
(z)
−1
tan( p/2)
, (4.42)
M
p
=
1 +tan( p/2) tan( p/4)
1 +s
2
(z)
−1
tan( p/2)
(4.43)
These last equations allow us to recognize that by illumination of an
input transparency with a spherical wavefront converging to an axial
point S, all the FrFTs inthe range [0, 1] canbe obtainedsimultaneously,
apart from a quadraticphase factor and a scale factor. The FrFTs are
axially distributed between the input transparency plane ( p = 0) and
the virtual source (S) plane ( p = 1) inwhichthe optical FTof the input
is obtained. For onedimensional input signals, instead of a spherical
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 119
x
0
x
y
1D input
FrFT of order p
Rp
y
0
Cylindrical
lens
Focal
line
z
S
FIGURE 4.3 Implementation of the FrFT by freespace propagation.
wavefront, we can use a cylindrical one to illuminate the input
(see Fig. 4.3).
Keeping in mind Eq. (4.19), we see the next step is to obtain the
RWDfromthis setup. To do this, we have to ﬁnd an optical element to
form the image of the axially distributed FrFT channels, at the same
output plane simultaneously. Therefore, the focal length of this lens
should be different for each fractional order p. Since in this case the
different axially located FrFTs present no variations along the vertical
coordinate, we can select a different onedimensional horizontal slice
of each one and use it as a single and independent fractionalorder
channel. (see Fig. 4.4).
The setup of Fig. 4.4 takes advantage of the onedimensional na
ture of the input, and it behaves as a multichannel parallel FrFT trans
former, provided that the focal length of the lens L varies with the y
coordinate in the same way as it varies withp. In this way, the problem
can be addressed as follows. For each value of p (vertical coordinate
y) we want to image a different object plane at a distance a
p
from the
lens onto a ﬁxedoutput plane locatedat a
fromthe lens. To obtain this
result, it is straightforward to deduce fromthe Gaussian lens equation
and from the distances in Fig. 4.4 that it is necessary to design a lens
with a focal length that varies with p (vertical coordinate y) according
to
f ( p) =
a
a
p
a
+a
p
=
a
l +(1 +l z
−1
)a
s
2
−1
tan( p/2)
a
−l −(a
+l + z)z
−1
s
2
−1
tan( p/2)
(4.44)
On the other hand, this focal length should provide the exact magniﬁ
cation at each output channel. The magniﬁcation given by the system
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
120 C h a p t e r F o u r
y
Output
plane
Varifocal
lens (L)
a'
x
y(R
0
) x
0
x
y
1D input
y
0
Cylindrical
lens
z
x
y
y(R
0
)
y(R
0
)
a
p
FrFT of order p
R
p
l
FIGURE 4.4 RWD setup (multichannel continuous FrFT transformer).
for each fractional order p is
M
L
( p) =
−a
a
p
=
a
s
2
−1
tan( p/2)
1+s
2
(z)
−1
tan( p/2)
−l
(4.45)
However, for the porder slice of the RWT of the input function to be
achieved, the lens L should counterbalance the magniﬁcation of the
FRT located at R
p
to restore its proper magniﬁcation at the output
plane. Therefore, by using Eq. (4.43), the magniﬁcation provided by
L should be
M
L
( p) =
−1
M
p
= −
1 +s
2
(z)
−1
tan( p/2)
1 +tan( p/2) tan( p/4)
(4.46)
Comparing Eqs. (4.45) and (4.46), we note that the functional depen
dence of both equations on p is different, and, consequently, we are
unable to obtain an exact solution for all fractional orders. However,
an approximate solution can be obtained by choosing the parameters
of the system, namely, s, z, l, , and a
, in such a way that they mini
mize the difference between these functions in the interval p ∈ [0, 1].
One way to ﬁndthe optimumvalues for these parameters is by a least
squares method. This optimization
6
leads to the following constraint
conditions
a
= l
1
2
+
4
, z =
−ls
2
l +s
2
(4.47)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 121
6.00
5.00
4.00
3.00
2.00
1.00
0.00
1.00 0.80
p
0.60 0.40 0.20 0.00
0.00
0.10
0.20
F
o
c
a
l
l
e
n
g
t
h
(
m
)
P
o
w
e
r
(
D
)
0.30
0.40
0.50
FIGURE 4.5 Focal length (solid curve) and optical power (dotted curve) of
the designed varifocal lens L for the values z = 426 mm, l = 646 mm, and
a = 831 mm.
The variation of the focal distance of the lens L with p according to
Eq. (4.44) and its optical power, under the constraints given by Eqs.
(4.47), are representedinFig. 4.5 for the following values: z = 426 mm,
l = 646 mm, and a = 831 mm.
For this particular combination of parameters, the optical power is
nearly linear with p, except for values close to p = 1. This linearity
is also accomplished by some designs of ophthalmic progressive ad
dition lenses in which there is a continuous linear transition between
two optical powers that correspond to the near portion and distance
portion. In the experimental veriﬁcation of the system, a progressive
lens of +2.75 D spherical power and +3 D of addition was used in the
setup of Fig. 4.4 with the abovementioned values for the parameters
z, l, and a. Figure 4.6 illustrates a comparison between the numerical
5.0
4.0
3.0
2.0
1.0
0.0
5.0
4.0
3.0
2.0
1.0
0.0
0.0 0.2 0.4 0.6
Fractional order: p
(a) (b)
x
c
o
o
r
d
i
n
a
t
e
(
m
m
)
0.8 1.0
0.0 0.2 0.4 0.6
Fractional order: p
0.8 1.0
FIGURE 4.6 RWD of a Ronchi grating of 3 lines/mm: (a) exact numerical
simulation; (b) experimental result.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
122 C h a p t e r F o u r
simulations and the experimental results obtained using a Ronchi
grating as input object.
Interestingly, in Fig. 4.6 the values of p that correspond to the self
images, both positive and negative, can be clearly identiﬁed. The op
tical setup designed for the experimental implementation of the RWD
was successfully adapted to several applications, as we show later in
this chapter.
In searching for an RWD with an exact scale factor for all the frac
tional orders, this approach also inspired another proposal
10
in which
a bent structure for the detector was suggested. The result is an exact,
but unfortunately impractical, setup to obtain the RWD. This draw
back was partially overcome in other conﬁgurations derived by the
same authors using the abcd matrix formalism. There, the free propa
gation distances are designed to be ﬁxed or to vary linearly with the
transverse coordinate,
11
so the input plane and/or the output plane
should be tilted instead of bent, resulting in a more realistic conﬁgu
ration, provided that the tilt angles are measured very precisely.
4.3 Analysis of Optical Signals and Systems
by Means of the RWT
4.3.1 Analysis of Diffraction Phenomena
4.3.1.1 Computation of Irradiance Distribution
along Different Paths in Image Space
Determination of the irradiance at a given point in the image space
of an imaging system is a classic problem in optics. The conventional
techniques carry out a ﬁnite partition of the pupil of the system to
sum all these contributions at the observation point.
12–16
This time
consuming procedure needs to be completely repeated for each ob
servation point, or if the aberration state of the systemchanges. In this
section we present a useful technique, based on the use of the RWT
of a mapped version of the pupil of the system, for a much more efﬁ
cient analysis of the irradiance in the image space of imaging systems.
This technique has been successfully applied to the analysis of dif
ferent optical systems with circular
17
as well as square,
18
elliptical,
19
triangular,
19
and even fractal pupils.
20
The method has also been ap
plied to the study of multifacet imaging devices.
21
Let us consider a general imaging system, characterized by an exit
pupil function with generalized amplitude transmittance P( x). The
distance from this pupil to the Gaussian imaging plane is denoted
by f . Note that the function P( x) includes any arbitrary amplitude
variation p( x) and any phase aberration that the imaging system may
suffer from.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 123
y
x
a
f
z
Exit pupil
Focal plane
ar
N
φ
Observation
point
Σ
P
FIGURE 4.7 The imaging system under study.
We now describe the monochromatic scalar light ﬁeld at any point
of the image space of the system in the Fresnel approximation
21
. It is
straightforwardtoshowthat, withinthis approach, theﬁeldirradiance
is given by
I ( x, z) =
1
2
( f + z)
2
×
P
P( x
) exp
¸
−i z x

2
f ( f + z)
exp
¸
−i 2
( f + z)
x · x
d
2
x
2
(4.48)
where is the ﬁeld wavelength, x and z stand for the transverse and
axial coordinates of the observation point, respectively, and
P
rep
resents the pupil plane surface. The origin for the axial distances is
ﬁxed at the axial Gaussian point, as shown in Fig. 4.7.
It is convenient to express all transverse coordinates in normalized
polar form, namely,
x = ar
N
cos , y = ar
N
sin (4.49)
where x and y are Cartesian coordinates and a stands for the maxi
mum radial extent of the pupil. By using these explicit coordinates in
Eq. (4.48), we obtain
¯
I (r
N
, , z)
=
1
2
( f + z)
2
2
0
1
0
¯ p(r
N
,
) exp
¸
i 2W(r
N
,
)
exp
¸
i 2W
20
(z) r
N
2
×exp
¸
−i 2
( f + z)
r
N
r
N
cos(
−)
r
N
dr
N
d
2
(4.50)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
124 C h a p t e r F o u r
where the bar denotes the polar coordinate expression for the corre
sponding function and where we have split out the generalized pupil
P( x) to explicitly show the dependence on the amplitude pupil vari
ations p( x) and the aberration function W(r
N
,
) of the system. The
classic defocus coefﬁcient has also been introduced in this equation,
namely,
W
20
(z) = −
za
2
2 f ( f + z)
(4.51)
In many practical situations the most important contribution to the
aberration function is the primary spherical aberration (SA), whose de
pendence on the pupil coordinates is given by
W
40
(r
N
,
) = W
40
r
N
4
(4.52)
where W
40
is the SA coefﬁcient design constant. In the following rea
soning, we will consider this term explicitly, splitting the generalized
pupil of the system as follows:
¯ p(r
N
,
) exp
¸
i 2W(r
N
,
)
= Q(r
N
,
) exp
¸
i 2W
40
r
N
4
(4.53)
Thus Q(r
N
,
) includes the amplitude variations on the pupil plane
and the aberration effects except for SA. Note that if no aberrations
different from SA are present in the system, Q(r
N
,
) reduces simply
to the pupil mask ¯ p(r
N
,
).
By substituting Eq. (4.53) into Eq. (4.50), we ﬁnally obtain
¯
I (r
N
, , z)
=
1
2
( f + z)
2
2
0
1
0
Q(r
N
,
) exp
i 2W
40
r
N
4
exp
¸
i 2W
20
(z) r
N
2
× exp
¸
−i 2
( f + z)
r
N
r
N
cos(
−)
r
N
dr
N
d
2
(4.54)
Let us nowconsider explicitly the angular integrationinthis equation,
namely,
Q(r
N
, r
N
, , z) =
2
0
Q(r
N
,
) exp
¸
−i 2
( f + z)
r
N
r
N
cos(
−)
d
(4.55)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 125
Thus we arrive at a compact form for the irradiance at a point in the
image space
¯
I (r
N
, , z) =
1
2
( f + z)
2
×
1
0
Q(r
N
, r
N
, , z) exp
i 2W
40
r
N
4
exp
¸
i 2W
20
(z) r
N
2
r
N
dr
N
2
(4.56)
By using the mapping transformation
r
N
2
= s +
1
2
, Q(r
N
, r
N
, , z) = q(s, r
N
, , z) (4.57)
we ﬁnally obtain
¯
I (r
N
, , z) =
1
2
( f + z)
2
×
0.5
−0.5
q(s, r
N
, , z) exp
i 2W
40
s
2
exp
i 2[W
40
+ W
20
(z)] s
ds
2
(4.58)
Note that in this expression all the dependence on the observation
coordinates is concentrated in the mapped pupil q(s, r
N
, , z) and the
defocus coefﬁcient W
20
(z). If we expand the modulus square in this
equation, we ﬁnd
¯
I (r
N
, , z) =
1
2
( f + z)
2
×
0.5
−0.5
0.5
−0.5
q(s, r
N
, , z)q
∗
(s
, r
N
, , z) exp
¸
i 2W
40
(s
2
−s
2
)
×exp
i 2[W
40
+ W
20
(z) ] (s −s
)
ds ds
(4.59)
which by using the change of variables
t =
s +s
2
, u = s −s
(4.60)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
126 C h a p t e r F o u r
can be rewritten as
¯
I (r
N
, , z) =
1
2
( f + z)
2
×
1
−1
0.5
−0.5
q
t +
u
2
, r
N
, , z
q
∗
t −
u
2
, r
N
, , z
×exp
i 2
[W
40
+ W
20
(z) +2W
40
] u
dt du (4.61)
The above integration over the variable u can be clearly identiﬁed as
the WDF of q (s, r
N
, , z) with respect to the ﬁrst variable, as stated in
Eq. (4.1). Thus, it is straightforward to show that
¯
I (r
N
, , z) =
1
2
( f + z)
2
0.5
−0.5
W
q
t, −2
W
40
t −
W
40
+ W
20
(z)
dt
(4.62)
This expression relates the irradiance at any observation point to the
line integral of the function W
q
(x, ) along a straight line in phase
space described by the equation
= −2
W
40
x −
W
40
+ W
20
(z)
(4.63)
as depicted in Fig. 4.8. One can identify this integration as a projection
of the WDF at an angle given by [see Eq. (4.10)]
tan = −
2W
40
(4.64)
FIGURE 4.8 Integration line in phase space.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 127
and at an oriented distance from the origin
x
(z) = −
W
40
+ W
20
(z)
4W
2
40
+
2
(4.65)
in such a way that it is possible to express
¯
I (r
N
, , z) =
1
2
( f + z)
2
RW
q
(x
(z), ) (4.66)
The main conclusion of all this is that it is possible to obtain the ir
radiance at any point in image space through the values of the RWT
of a given function q(s, r
N
, , z) related to the pupil of the system.
Note, however, that this function depends in general on the particular
coordinates r
N
, , and z of the observation point. Thus, a different
function RW
q
(x
, ) has to be considered for different points in image
space. This major drawback can be overcome for special sets of points
or trajectories in image space that share the same associated mapped
pupil q(s, r
N
, , z).
To describe such trajectories in image space, let us express these
lines in parametric form r
N
(z), (z). By substituting Jacobi’s identity
exp(i ␥cos ) =
+∞
¸
n=−∞
i
n
J
n
(␥) exp(−i n) ␥, ∈ R (4.67)
where J
n
(x) represents the Bessel function of the ﬁrst kind and order
n, into Eq. (4.55), it is straightforward to obtain
Q(r
N
, r
N
(z), (z), z) =
+∞
¸
n=−∞
i
n
J
n
−2
( f + z)
r
N
r
N
(z)
Q
n
(r
N
)
×exp[i n(z)] (4.68)
where Q
n
(r
N
) stands for the norder circular harmonic of Q
n
(r
N
,
),
that is,
Q
n
(r
N
) =
2
0
Q(r
N
,
) exp
−i n
d
(4.69)
Note that the dependence on the position parameter z in Eq. (4.68)
is established exclusively in the argument of the Bessel functions—
through r
N
(z)—and the phase exponentials—through (z). Thus, the
only way to strictly cancel this dependence is to consider the trajecto
ries
r
N
(z) = K( f + z), (z) =
o
(4.70)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
128 C h a p t e r F o u r
y
x
a
f
z
Exit pupil
Focal plane
φ
o
Observation
point
Σ
P
α
(z + f ) tanα
FIGURE 4.9 Trajectories in image space.
These curves correspond to straight lines passing through the axial
point at the plane of the exit pupil. Together with the optical axis,
each line deﬁnes a plane that forms an angle
o
with the x axis, as
depicted in Fig. 4.9. Note that the angle ␣ of any of these lines with
the optical axis is given by
tan␣ = Ka (4.71)
For these subsets of observation points, the mapped pupil of the sys
tem can be expressed as
Q(r
N
, r
N
(z), (z), z) = Q
␣,
o
(r
N
)
=
+∞
¸
n=−∞
i
n
J
n
−2a tan␣
r
N
Q
n
(r
N
) exp(i n
o
)
(4.72)
and analogously
r
N
2
= s +
1
2
, q(s, r
N
(z), (z), z) = Q
␣,
o
(r
N
) = q
␣,
o
(s) (4.73)
in such a way that now the corresponding RWT RW
q
␣, o (x
, ) is in
dependent of the propagation parameter z. This is a very interest
ing issue since the calculation of the irradiance at any observation
point lying on the considered line can be achieved from this single
twodimensional display by simply determining the particular co
ordinates (x
(z), ) through Eqs. (4.64) and (4.65). Furthermore, the
proper choice of these straight paths allows one to obtain any desired
partial feature of the whole threedimensional image irradiance distri
bution. Note also that since W
40
is just a parameter inthese coordinates
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 129
and does not affect the function RW
q
␣, o (x
, ), this single display can
be used for the determination of the irradiance for different amounts
of SA. Thus, compared to classic techniques, the reduction in compu
tation time is evident. The axial irradiance distribution is often used
as a ﬁgure of merit for the performance of optical systems with aberra
tions. This distribution can be obtained here as a particular case with
␣ = 0, namely,
I (0, 0, z) =
1
2
( f + z)
2
RW
q
0, 0 (x
(z), ) (4.74)
where
q
0, 0
(s) = Q
0, 0
(r
N
) = Q
0
(r
N
) (4.75)
This result is especially interesting since this mapped pupil, and thus
the associated RWT, is also independent of the wavelength . This fact
represents anadditional advantage whena polychromatic assessment
of the imaging system is needed, as will be shown in forthcoming
sections. Some quantitative estimation of these improvements is pre
sented in Ref. 19.
To prove the performance of this computation method, next we
present the result of the computation of the irradiance distribution
along different lines in image space of two imaging systems, la
beled system I and system II. For the sake of simplicity, we consider
only purely absorbing pupils and no aberrations apart from SA in
both cases. Thus, Q(r
N
,
) reduces to the normalized pupil function
¯ p(r
N
,
). A grayscale representation for the pure absorbing masks
considered for each system is shown in Fig. 4.10.
p
I
( x )
→
x
y
a
x
y p
II
( x )
→
2a/3
a
(b) (a)
FIGURE 4.10 Grayscale picture of the pupil functions for (a) system I and
(b) system II.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
130 C h a p t e r F o u r
0
250
500
750
1000
W
40
= 0
α = 0°
0
250
500
750
1000
0
5
10
15
0
5
10
15
–4.0 –2.0 0.0 4.0
0
2
4
6
8
–4.0 –2.0 0.0 4.0
0
2
4
6
8
W
40
= λ/2
α = 0°
W
40
= 0
α = 0.012°
W
40
= λ/2
α = 0.012°
W
40
= 0
α = 0.024°
W
40
= λ/2
α = 0.024°
I
r
r
a
d
i
a
n
c
e
(
a
.
u
.
)
I
r
r
a
d
i
a
n
c
e
(
a
.
u
.
)
I
r
r
a
d
i
a
n
c
e
(
a
.
u
.
)
Defocus coefficient: W
20
(μm)
2.0
2.0
FIGURE 4.11 Irradiance values provided by system I, along different lines
containing the axial point of the pupil and for two different amounts of SA.
Continuous lines represent the result by the proposed RWT method while
dotted lines stand for the computation by the classic method.
We compute the irradiance values for 256 points along three differ
ent lines passing through the axial point of the pupil, all characterized
by an azimuthal angle
o
= /2. These trajectories are chosen with
tilt angles ␣ = 0.024
◦
, 0.012
◦
and 0
◦
(optical axis). We set a = 10 mm,
z = 15.8 mm, and = 638.2 nm. The function RW
q
␣, o (x
, ) was com
puted for 4096 ×4096 points, and for comparison purposes, the same
irradiance values were computed by using the classic method
12, 13
by partitioning the exit pupil of the imaging system into 1024×1024
radialazimuthal elements. Figure 4.11 shows a joint representation of
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 131
0
250
500
750
1000
W
40
= 0
α = 0°
0
250
500
750
1000
0
5
10
15
0
5
10
15
–4.0 –2.0 0.0 2.0 4.0
0
2
4
6
8
–4.0 –2.0 0.0 2.0 4.0
0
2
4
6
8
W
40
= λ/2
α = 0°
W
40
= 0
α = 0.012°
W
40
= λ/2
α = 0.012°
W
40
= 0
α = 0.024°
W
40
= λ/2
α = 0.024°
I
r
r
a
d
i
a
n
c
e
(
a
.
u
.
)
I
r
r
a
d
i
a
n
c
e
(
a
.
u
.
)
I
r
r
a
d
i
a
n
c
e
(
a
.
u
.
)
Defocus coefficient: W
20
(μm)
FIGURE 4.12 Irradiance values provided by system II, as in Fig. 4.11.
the numerical calculation for system I, when two different values of
the SA are considered. The same results applied now to system II are
presented in Fig. 4.12.
The analysis of these pictures shows that the results obtained with
the RWT method match closely those obtained with the classic tech
nique. In fact, both results differ by less than 0.03 percent. However,
the RWT is much more efﬁcient in this computation process. This is
so because the basic RWT does not require recalculation for any point
in each of the curves. This is also true for any amount of SA. Obvi
ously, the greater the number of observation points, or SAvalues, that
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
132 C h a p t e r F o u r
have tobe considered, the greater the resultant savings incomputation
time.
As a ﬁnal remark on this subject, we want to point out that this
approach can also be applied to other trajectories of interest in im
age space. For instance, short paths parallel to the optical axis in the
neighborhood of the focal plane
17
or straight lines crossing the focal
point can be considered.
22
4.3.1.2 Parallel Optical Display of Diffraction Patterns
In Sec. 4.2.2 we mentioned that the mathematical relationship be
tween Fresnel diffraction and the FrFT is given by Eq. (4.41). This
means that the RWD is itself a continuous display of the evolution
of diffraction patterns of onedimensional objects, and this property
is extremely useful from a pedagogical point of view. In fact, calcu
lations of Fresnel and Fraunhofer diffraction patterns of uniformly
illuminated onedimensional apertures are standard topics in under
graduate optics courses. These theoretical predictions are calculated
analytically for some typical apertures, or, more frequently, they are
computed numerically. The evolution of these diffraction patterns un
der propagation is often represented in a twodimensional display of
gray levels in which one axis represents the transverse coordinate—
pattern proﬁle—and the other axis is related to the axial coordinate—
evolution parameter.
23
This kind of representation illustrates, e.g.,
how the geometrical shadow of the object transforms into the Fraun
hofer diffraction pattern as it propagates, and that the Fraunhofer
diffraction simply is a limiting case of Fresnel diffraction.
24
In addi
tion to the qualitative physical insight that the RWD provides about
diffraction, it can provide a quantitative measurement of a variety of
terms. These include the precise location y
s
andthe relative magniﬁca
tion M
s
of each diffraction pattern. These two terms are quantitatively
deﬁned in terms of the maximum
h
and minimum
0
powers of the
varifocal lens L of the system represented in Fig. 4.5, i.e.,
y
s
=
hs
s +l
2
(
h
−
0
)
, M
s
= 1 +
s
l
2
(
h
−
0
)
(4.76)
where s is the axial coordinate at which the corresponding diffraction
pattern is localized under parallel illumination and h is the extent of
the socalled progression zone of the varifocal lens. Figure 4.13 illus
trates the experimental results registered by a CCD camera using a
double slit as an input object. It can be seen that the RWD is a nice
representationof the evolutionbypropagationof the interference phe
nomena. In fact, the Fraunhofer region of the diffracted ﬁeld clearly
shows the characteristic Young fringes modulated by a sinc function.
To compare the theoretical and experimental results, a cross section of
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 133
FIGURE 4.13 RDW showing the evolution of the ﬁeld diffracted by a double
slit. (a) Experimental result. (b) Cross section of the RWD showing the
intensity proﬁle near the Fraunhofer region. For comparison purposes, the
theoretical sinc envelope of the Young fringes is also shown by the dotted
line.
the experimental RWD for values of y/h close to 1 is also represented
in Fig. 4.13.
Another classic example is diffraction by periodic objects. Here,
selfimaging phenomena, such as the Talbot effect, are interesting and
stimulating and usually attract the students’ attention. As illustrated
earlier in Fig. 4.6, which shows the diffraction patterns of a Ronchi
grating, several selfimaging planes can be identiﬁed. It can be clearly
seenthat, due tothe ﬁnite extent of the gratingat the input, the number
of Talbot images is limited by the socalled walkoff effect. Self imag
ing phenomena are discussed in more detail in Chapter 9 by Markus
Testorf.
In addition to its use as an educational tool for displaying diffrac
tion patterns, the RWD has been used to investigate diffraction by a
variety of different interesting structures including fractal diffraction
screens. Infact, the properties of diffractionpatterns producedbyfrac
tal objects and their potential applications have attracted the attention
of several researchers during recent years because many natural phe
nomena and physical structures, such as phase transition, turbulence,
or optical textures, can be analyzed and described by assuming fractal
symmetry. Most research has been devoted to the study of diffraction
patterns obtainedfromfractal objects in the Fraunhofer region,
25
yet it
is inthe Fresnel regionwhere interesting features appear. For instance,
Fresnel diffraction of a Cantor set
26
shows an irradiance distribution
along the optical axis having a periodicity that depends on the level of
the set. Furthermore, the intensity distributions at transverse planes
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
134 C h a p t e r F o u r
FIGURE 4.14 RWD as a display of all diffraction patterns generated by a
Cantor grating of level 3.
showa partial selfsimilar behavior that is increased when moving to
wardthe Fraunhofer region. For this reason, it is useful torepresent the
evolution of the complex amplitude of onedimensional fractals prop
agating through free space represented on a twodimensional display,
especially if sucha display canbe obtainedexperimentally. Inthis case
one axis represents the transversal coordinate, and the other is a func
tion of the axial coordinate. In fact, according to the analysis carried
out in Ref. 27, the evolution of the diffraction patterns allows one to
determine the main characteristic parameters of the fractal. Therefore,
one of the most important applications of the RWD has been in this
ﬁeld.
28
The RWD obtained for a triadic Cantor grating developed up
to level 3 is shown in Fig. 4.14. Moreover, this result can be favor
ably compared with the results obtained with other displays.
27
The
magniﬁcation provided by the lens L in the experimental setup (see
Fig. 4.4) enables the RWDrepresentation to provide an optimumsam
pling of the diffracted ﬁeld. Near the object, where the diffraction pat
terns change rapidly, the mapping of the propagation distance pro
vides a ﬁne sampling, whereas the sampling is coarse in the far ﬁeld
where the variation of the diffraction patterns with the axial distance
is slow. We note that sampling is the subject of Chapter 10.
4.3.2 Inverting RWT: PhaseSpace Tomographic
Reconstruction of Optical Fields
The WDF is an elegant and graphical way to describe the propagation
of optical ﬁelds through linear systems. Since the WDF of a complex
ﬁeld distribution contains all the necessary information to retrieve the
ﬁeld itself,
29, 30
many of the methods to obtain the WDF (and the AF)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 135
f (x)
(Optical)
1D signal Radon Wigner
spectrum
Inverse RT
(digital)
RW
f
(x, θ) W
f
(x, ξ)
WDF
RWT = ⎜FrFT ⎜
2
FIGURE 4.15 Diagram of the proposed hybrid optodigital method.
couldbe adaptedtosolve the phase retrieval problem. Optical or opto
electronic devices are the most commonlyemployedsystems to obtain
a representation of phasespace functions of onedimensional or two
dimensional complex signals.
31, 32
However, because most detectors
used to this end are only sensitive to the incident intensity, interfero
metric or iterative methods are necessary in general to avoid loss of
information. This is true even for the optically obtained WDF, which
is real but has, in general, negative values and therefore is obtained
froman intensity detector with an uncertainty in its sign. On the other
hand, obtaining the WDF of wave ﬁelds is also possible indirectly
through other representations such as the Radon transform.
33
In this
particular case, a tomographic reconstruction is needed to synthesize
the WDF. With this information it is possible to recover the ampli
tude and the phase of the original ﬁeld distribution solely by means
of intensity measurements. With most experimental setups for phase
retrieval,
29, 30
these measurements have to be taken sequentially in
time while varying the distances between some components in each
measurement. In this way the potential advantage of optics, i.e., paral
lel processing of signal information, is wasted. Consequently, another
interesting application of the setup discussedin Sec. 4.2.2 to obtain the
RWDis the experimental recovery of the WDF by means of an inverse
Radon transformation.
The technique to obtain the WDF from projections is divided into
two basic stages, sketchedinFig. 4.15. Inthe ﬁrst stage, the experimen
tal RadonWigner spectrum of the input function is obtained from a
twodimensional singleshot intensity measurement by the use of the
experimental setup in Fig. 4.4. This optical method beneﬁts from hav
ing no moving parts.
The second part of the proposed method is the digital computation
of the inverse Radon transforms of the experimental RadonWigner
spectrum. The most common algorithms used in tomographic recon
struction are based on the technique known as ﬁltered backprojec
tion. This algorithm is based on the central slice theorem discussed in
Sec. 4.2.1. Thus, from Eqs. (4.21) and (4.22) we have
F{RW
f
(x
, ),
} = F
2D
{W
f
(x, ), (x
cos ,
sin)} (4.77)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
136 C h a p t e r F o u r
3
2
1
0
–1
–2
–3
3
2
1
0
–1
–2
–3
S
p
a
t
i
a
l
c
o
o
r
d
i
n
a
t
e
(
m
m
)
S
p
a
t
i
a
l
c
o
o
r
d
i
n
a
t
e
(
m
m
)
0 15 30 45
Projection angle (°)
(a) (b)
Projection angle (°)
60 75 90 0 15 30 45 60 75 90
FIGURE 4.16 Experimental RadonWigner spectrum of (a) a single slit of
2.2 mm and (b) a binary grating with a linearly increasing spatial frequency.
where the onedimensional FT is performed on the ﬁrst argument
of RW
f
(x
, ). The inversion of this last transformation allows the
recovery of W
f
(x, ) from its projections. Explicitly
34
W
f
(x, ) =
0
C
f
(x cos + sin, ) d (4.78)
with
C
f
(u, ) =
+∞
−∞
F{RW
f
(x
, ),
} 
 exp(i 2
u) d
(4.79)
Equation (4.79) can be clearly identiﬁed as a ﬁltered version of the
original RWT. In this way, from Eq. (4.78), W
f
(x, ) is reconstructed
for each phasespace point as the superposition of all the projections
passing through this point.
The experimental RWD of different onedimensional functions has
been used to reconstruct the WDF from projections. In Fig. 4.16 we
showthe RWDobtained with the optical device described in Sec. 4.2.2
for two different functions, namely, a rectangular aperture (single slit)
andagratingwithalinearlyincreasingspatial frequency(chirpsignal).
To undertake the reconstruction of the WDF through the ﬁltered
backprojection algorithm, it is necessary to consider the complete an
gular region of the RWD, that is, ∈ [0, ). Although we only obtain
optically the RWT for ∈ [0, /2], the symmetry property in Eq. (4.17)
has been used to complete the spectrum. Fromthe experimental RWD
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 137
4 4
2
0
–2
–4
2
0
–2
–4
–2 –1 0 1 2 –2 –1 0 1 2
S
p
a
t
i
a
l
f
r
e
q
u
e
n
c
y
(
a
.
u
.
)
S
p
a
t
i
a
l
f
r
e
q
u
e
n
c
y
(
a
.
u
.
)
Spatial coordinate (mm)
(a)
Spatial coordinate (mm)
(b)
FIGURE 4.17 (a) Theoretical WDF of a single slit. (b) Experimental result for
the tomographic reconstruction of the WDF of the same slit.
in Fig. 4.16, the corresponding WDFs have been obtained using the
ﬁlteredbackprojectionalgorithm. For comparisonpurposes, Figs. 4.17
and 4.18 show both the theoretical and the experimentally recon
structed WDF of the single slit and the chirp grating, respectively.
Note that in Figs. 4.17b and 4.18b some artifacts appear. The lines radi
ating from the center and outward are typical artifacts (ringing effect)
associated with the ﬁltered backprojection method.
35
In spite of this
4 4
2
0
–2
–4
2
0
–2
–4
–2 –1 0 1 2 –2 –1 0 1 2
S
p
a
t
i
a
l
f
r
e
q
u
e
n
c
y
(
a
.
u
.
)
S
p
a
t
i
a
l
f
r
e
q
u
e
n
c
y
(
a
.
u
.
)
Spatial coordinate (mm)
(a)
Spatial coordinate (mm)
(b)
FIGURE 4.18 (a) Theoretical WDF of a binary grating with a linearly
increasing spatial frequency. (b) Experimental tomographic reconstruction of
the WDF of the same grating.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
138 C h a p t e r F o u r
effect, a very good qualitative agreement can be observed between
the results obtained with the theoretical and experimental data. The
asymmetry in Fig. 4.17 is a consequence of the noise in Fig. 4.16a, re
ﬂecting also the asymmetry on the spatial coordinate in this ﬁgure.
In Fig. 4.18 the typical arrowshaped WDF of a chirp function can be
observed in both cases. The slope in the arrowhead that character
izes the chirp rate of the signal is the same for the theoretical and the
experimental results.
Several extensions of the proposed method are straightforward. On
one hand, a similar implementationproposedhere for the WDF canbe
easilyderivedfor the AF, byvirtue of Eq. (4.22). Note also that it is easy
to extend this method to obtain twodimensional samples of the four
dimensional WDF of a complex twodimensional signal by use of a
line scanning system. Moreover, since complex optical wave ﬁelds can
be reconstructed from the WDF provided the inversion formulas, this
approach can be used as a phase retrieval method that is an alterna
tive to the conventional interferometric or iterativealgorithmbased
techniques. In fact, as demonstrated,
36
phase retrieval is possible with
intensity measurements at two close FrFT domains. This approach,
however, requires some a priori knowledge of the signal bandwidth.
In our method, a continuous set of FrFTs is available simultaneously,
and this redundancy should avoid any previous hypothesis about the
input signal.
4.3.3 Merit Functions of Imaging Systems
in Terms of the RWT
4.3.3.1 Axial PointSpread Function (PSF) and Optical
Transfer Function (OTF)
There are several criteria for analyzing the performance of an opti
cal imaging system for aberrations and/or focus errors in which the
onaxis image intensity, or axial pointspread function (PSF), is the
relevant quantity. Among them we mention:
37
Rayleigh’s criterion,
Marechal’s treatment of tolerance, and the Strehl ratio (SR). As Hop
kins suggested,
38
the analysis of Marechal can be reformulated to give
a tolerance criterion based on the behavior of the optical transfer func
tion (OTF) (spatial frequency information) instead of the PSF (space
information). Phasespace functions were also employed to evaluate
some merit functions and quality parameters.
39–41
This point of view
equally emphasizes both the spatial and the spectral information con
tents of the diffractedwave ﬁelds that propagate inthe optical imaging
systems. Particularly, since the information content stored in the FrFT
of an input signal changes from purely spatial to purely spectral as
p varies from p = 0 to p = 1, that is, in the domain of the RWT, it
is expected that the imaging properties of a given optical system, in
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 139
boththe space andspatial frequency domains, couldalso be evaluated
from the RWD.
To derive the formal relationship between the PSF (and the OTF)
and the RWT, let us consider the monochromatic wave ﬁeld, with
wavelength , generated by an optical imaging system characterized
bya onedimensional pupil functiont(x), whena unit amplitude point
source is located at the object plane.In the neighborhood of the image
plane, locatedat z = 0, the ﬁeldamplitude distribution can be written,
according to the Fresnel scalar approximation, as
U (x, z) =
+∞
−∞
t(x
) exp
−i
f
x
2
exp
¸
i
( f + z)
(x
− x)
2
dx
(4.80)
where f is the distance from the pupil to the image plane. The trans
formation of t(x) to obtain the ﬁeld U (x, z) is given by a twostep
sequence of elementary abcd transforms, namely, a spherical wave
front illumination (with focus at = f ) and a freespace propagation
(for a distance f + z). Considering the results presented in Sec. 4.2.1,
the abcd matrix associated with this transform can be found to be
M =
1 0
1
f
1
1 −( f + z)
0 1
=
1 −( f + z)
1
f
−
z
f
(4.81)
and, therefore, the equivalent relationships to that given by Eq. (4.80)
interms of the correspondingRWTs canbe expressedas [see Eq. (4.25)]
RW
U(x, z)
(x
, ) ∝ RW
t
(x
,
)
(4.82)
tan
=
f tan −
2
f ( f + z)
tan −z
, x
=
x
sin +( f + z) cos
sin
In particular, the value = 0 provides the irradiance distribution at
the considered observation point, as stated in Sec. 4.2.1. This function
is the PSF of the imaging system, as a function of the distance z to the
image plane. Thus,
RW
U(x, z)
(x
0
, 0) = U(x
0
, z)
2
= I (x
0
, z) ∝ RW
t
(x
0
,
0
)
(4.83)
tan
0
=
f ( f + z)
z
, x
0
=
x
0
( f + z)
sin
0
For the optical axis (x
0
= 0) the PSF can be expressed as
RW
U(x, z)
(0, 0) = U(0, z)
2
= I (0, z) ∝ RW
t
0, arctan
¸
f ( f + z)
z
(4.84)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
140 C h a p t e r F o u r
A normalized version of this axial irradiance is often used as a ﬁgure
of merit of the performance of the imaging system, namely, the SR
versus defocus, deﬁned as
S(W
20
) =
I (0, z)
I (0, 0)
∝ RW
t
0, arctan
−
h
2
2W
20
(4.85)
where h is the maximum lateral extent of the onedimensional pupil
and W
20
stands for the onedimensional version of the defocus coefﬁ
cient deﬁnedinEq. (4.51). Thus, the function S(W
20
) canbe analyzedin
a polar fashion in the twodimensional domain of the WDF associated
with the pupil function t(x) or, equivalently, in terms of its associated
RWT.
To illustrate this approach, the defocus tolerance of different kinds
of onedimensional pupils was investigated, namely, a clear aperture
(slit) and a pupil with a central obscuration (double slit). The general
form of these pupils can be written as t(x) = rect(x/h) − rect(x/b),
with b = 0 for the uniform aperture. Figure 4.19 shows the RWD
0 x
0 x
x
x
0
0
0.00 0.25 0.50
p
(c) (d)
(a) (b)
p
p p
0.75 1.00 0.00 0.25 0.50 0.75 1.00
0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00
FIGURE 4.19 RWTs: (a) Computer simulation for an aperture with
a = 2.5 mm and b = 0 mm. (b) Experimental result for (a). (c) Computer
simulation for an aperture with a = 2.5 mm and b = 1.3 mm. (d)
Experimental result for (c). The horizontal axis corresponds to the
parameterization of the projection angle = p/2.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 141
0.25 0.25
0 0
1 1
S
(
z
)
S
(
z
)
0.50 0.50 0.75 0.75 1.00 1.00
Fractional order p
(a) (b)
Fractional order p
b = 0 mm
b = 1.3 mm
b = 2 mm
b = 0 mm
b = 1.3 mm
b = 2 mm
FIGURE 4.20 SR versus defocus for circular pupils with pupil function
t(x) = rect(x/h) −rect(x/b): (a) Computer simulation; (b) experimental
results. Again, the projection angle = p/2.
numerically (parts a and c) and experimentally (parts b and d) ob
tained, for two different values of the obscuration b.
According to our previous discussion, the slices of the RWD for
x = 0 give rise to the SR for variable W
20
. These proﬁles are plotted
in Fig. 4.20 for three different pupils. From these results, it can be
observed that, as expected, annular apertures have higher tolerance
to defocus.
The knowledge of the SR is useful to characterize some basic fea
tures of any optical system, such as the depth of focus. However, the
main shortcoming of the SR as a method of image assessment is that
although it is relatively easy to calculate for an optical design pre
scription, it is normally difﬁcult to measure for a real optical system.
Moreover, the quality of the image itself is better described through
the associated OTF. Fortunately, this information can also be obtained
fromthe RWDvia its relationshipwith the AF establishedin Sec. 4.2.1,
since the AF contains all the OTFs H(; W
20
) associated with the opti
cal system with varying focus errors according to the formula
42
H
(; W
20
) = A
t
−( f + z),
2W
20
( f + z)
h
2
(4.86)
In this way, the AF of the pupil function t(x) can be interpreted as a
continuous polar display of the defocused OTFs of the system. Con
versely,
A
t
x
,
= H
−
x
( f + z)
; W
20
= −
h
2
2x
(4.87)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
142 C h a p t e r F o u r
0 ν
0
0 0
ν
ν
ν
0.00 0.25 0.50
p p
p p
(a) (b)
(c) (d)
0.75 1.00 0.00 0.25 0.50 0.75 1.00
0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00
FIGURE 4.21 Computed onedimensional FT of the RWDs shown
in Fig. 4.19.
Thus, by using Eq. (4.22) it is easy to ﬁnd that
F {RW
t
(x
, ),
} = A
t
(
cos , −
sin)
= H
−
cos
( f + z)
; W
20
=
h
2
2
tan
(4.88)
Therefore, the onedimensional FT of the proﬁle of the RWD for a
given value of the fractional order = p/2 corresponds to a defo
cused(scaled) OTF. This representationis quiteconvenient tovisualize
Hopkins’ criterion.
39
Figure 4.21 shows the onedimensional Fourier transforms, taken
with respect to the x axis, of the RWT illustrated in Fig. 4.19. From
the previous analysis, the defocused OTFs are displayed along the
vertical or spatialfrequency axis. These results for the clear aperture
are shown in Fig. 4.22.
The RWD can also be used for calculating the OTF of an optical
system designed to work under polychromatic illumination. In this
case, as we will discuss next, a single RWD can be used to obtain the
set of monochromatic OTFs necessary for its calculation.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 143
1.0 p = 0.5
W
20
/λ = 0.99
p = 0.5
W
20
/λ = 0.99
p = 0.75
W
20
/λ = 0.40
p = 0.75
W
20
/λ = 0.40
p = 1
W
20
/λ = 0
p = 1
W
20
/λ = 0
0.0
O
T
F
O
T
F
0.00
ξ
1.0
0.0
0.00
ξ
1.0
0.0
0.00
ξ
1.0
0.0
0.00
ξ
1.0
0.0
0.00
ξ
1.0
0.0
0.00
ξ
FIGURE 4.22 OTFs obtained from different slices of the intensity
distributions shown in Fig. 4.21 for the case of a uniform aperture, for
different amount of defocus.
4.3.3.2 Polychromatic OTF
As stated above, the RWT associated with the onedimensional pupil
of an imaging system can be used to obtain the OTF of the device,
as a function of the defocus coefﬁcient, through Eq. (4.88). It is worth
notingthat inthis equationthe wavelengthof the incominglight acts
as a parameter inthe determinationof the particular coordinates of the
FTof the RWT, but it does not affect the RWTitself. Thus, changing the
value of simply resets the position inside the same twodimensional
display for the computation of the OTF. The calculation procedure
used in the previous section can be used, therefore, to compute the
transfer function for any wavelength by means of the same RWD. This
approach is based on previous work, where it was shown that the AF
of the generalized pupil function of the system is a display of all the
monochromatic OTFs with longitudinal chromatic aberration.
43
An especially interesting application of this technique is the eval
uation of the spatialfrequency behavior of optical systems working
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
144 C h a p t e r F o u r
under polychromatic evaluation. In fact, the proper generalization of
the OTFbased description to this broadband illumination case allows
one to deﬁne quality criteria for imaging systems working with color
signals.
44, 45
This extension presents, however, some difﬁculties. The
direct comparison of the incoming and the outgoing polychromatic
irradiance distributions does not allow, in general, a similar relation
ship to the monochromatic case to be established. It can be shown,
in fact, that only when the input signal is spectrally uniform can the
frequency contents of both signals be related through a single poly
chromatic OTFfunction, providingthe imagingsystemdoes not suffer
from any chromatic aberrations regarding magniﬁcation.
46, 47
Under
these restrictions, a single polychromatic OTF can be used for relat
ing input and output polychromatic irradiances in spatialfrequency
space. This function is deﬁned as
H(; W
20
) =
H
(; W
20
())S()V() d
S()V() d
(4.89)
where and S() are the spectral range and the spectral power of the
illumination, respectively. The function V() represents the spectral
sensitivity of the irradiance detector used to record the image. Note
also that a new wavelengthdependent defocus coefﬁcient has been
deﬁned, to account for the longitudinal chromatic aberration ␦W
20
()
that the system may suffer from, namely,
W
20
() = W
20
+␦W
20
() (4.90)
where W
20
is the defocus coefﬁcient deﬁned in the previous section.
This OTF cannot account, however, for the chromatic information
of the image, since only a single detector is assumed.
48
Indeed, by fol
lowing the trichromacy of the human eye, three different chromatic
channels are usually employed to properly describe color features in
irradiance distributions, and, consequently, three different polychro
matic OTFs are used, namely,
44, 45
H(; W
20
) =
H
(; W
20
())S()x
d
S()x
d
H(; W
20
) =
H
(; W
20
())S() y
d
S() y
d
(4.91)
H(; W
20
) =
H
(; W
20
())S()z
d
S()z
d
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 145
where x
, y
, and z
are three spectral sensitivity functions associated
with the measured chromaticity. These functions depend, obviously,
on the speciﬁc color detector actually used. In the case of a conven
tional digital color camera, these channels can be associated with the
R, G, and B bands of the three pixel families in the detector array.
On the other hand, when a visual inspection of the ﬁnal image is
considered, these sensitivity functions are the wellknown spectral
tristimulus values of the human eye.
49
Equations (4.91) establish the formulas to describe completely the
response of a system from a spatialfrequency point of view. To
numerically compute the functions described there, the evaluation
of the monochromatic OTFs for a sufﬁcient number of wavelengths
inside the illumination spectrum has to be performed. Since any of
these monochromatic transfer functions can be obtained from a same
single RWD, as stated in the previous section, these computations can
be done in a much more efﬁcient way by use of this twodimensional
display. Furthermore, the same imaging system (i.e., the same pupil
function) but suffering from different longitudinal chromatic aberra
tion can be assessed as well, with no additional computation of the
RWD. This is a critical issue in the saving of computation time which
provides this technique with a great advantage compared to other
classic techniques, as cited above.
To illustrate this technique, we present the result of the computa
tion of the polychromatic OTFs associated with a conventional one
dimensional clearpupil optical system (slit of width h) but suffering
from two different chromatic aberration states (systems I and II from
now on), as shown in Fig. 4.23. We assume that no other aberrations
–5.0
–4.0
–3.0
–2.0
–1.0
0.0
1.0
2.0
L
o
n
g
i
t
u
d
i
n
a
l
c
h
r
o
m
a
t
i
c
a
b
e
r
r
a
t
i
o
n
c
o
e
f
f
i
c
i
e
n
t
:
δ
W
2
0
(
μ
m
)
400 500 600 700
Wavelength: λ (nm)
System I
System II
FIGURE 4.23 Longitudinal chromatic aberration coefﬁcient associated with
the two different correction states of the system under study.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
146 C h a p t e r F o u r
0 100 200 300 400 500
0.0
0.5
1.0
λ = 400 nm
λ = 500 nm
λ = 600 nm
λ = 700 nm
M
o
n
o
c
h
r
o
m
a
t
i
c
O
T
F
:
H
λ
(
ξ
;
δ
W
2
0
(
λ
)
)
Spatial frequency: ξ (mm
–1
)
FIGURE 4.24 Monochromatic OTFs for system I in Fig. 4.23, corresponding
to the imaging plane (W
20
= 0).
are present. This assumption does not imply any restriction of the
method, and the same applies to the onedimensional character of
the imaging system, as can be easily shown. Regarding the geometric
parameters of the system, we ﬁxed h/f = 0.2.
The evaluation of the corresponding monochromatic OTFs for
both aberration states is achieved through the same computation
methodas inthe previous section, namely, throughthe sequential one
dimensional FTof thetwodimensional displayof theRWT RW
t
(x
, ).
Some of these results are shown in Fig. 4.24.
The computation of the polychromatic OTFs is performed next for
both correction states, through the superposition of the monochro
matic ones stated in Eqs. (4.91) for uniform sampling of 36 wave
lengths in the range between 400 and 700 nm. The x
, y
, and z
functions are set to be the spectral tristimulus values of the standard
human observer CIE 1931, while the spectral power for the illumi
nation corresponds to the standard illuminant C
49
. The results for
systemI, corresponding to a defocused plane, and for systemII, at the
image plane, are shown in Fig. 4.25. Note that in both cases the same
RWD is used in the computation, as stated above.
4.3.3.3 Polychromatic Axial PSF
In this section we propose the use of a single twodimensional RWDto
compute the axial irradiance in image space provided by an imaging
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 147
0
0.0
0.5
1.0
P
o
l
y
c
h
r
o
m
a
t
i
c
O
T
F
100 200 300 400 500
Spatial frequency: ξ (mm
–1
)
H
X
(ξ;W
20
= –0.28 μm)
H
Y
(ξ;W
20
= –0.28 μm)
H
Z
(ξ;W
20
= –0.28 μm)
0
0.0
0.5
1.0
P
o
l
y
c
h
r
o
m
a
t
i
c
O
T
F
100 200 300 400 500
Spatial frequency: ξ (mm
–1
)
H
X
(ξ;W
20
= 0)
H
Y
(ξ;W
20
= 0)
H
Z
(ξ;W
20
= 0)
System I System II
FIGURE 4.25 Polychromatic OTFs for (a) system I and (b) system II in
Fig. 4.23, corresponding to a defocused plane (W
20
= −0.28 m) and image
plane, respectively.
system with polychromatic illumination. In fact, the proposed tech
nique is a straightforward extension of what is stated in Sec. 4.3.1.1,
namely, that the axial irradiance distribution I (0, 0, z) provided by a
system with an arbitrary value of SA can be obtained from the single
RWT RW
q
0, 0 (x, ) of the mapped pupil q
0, 0
(s) in Eq. (4.75). When an
object point source is used, this irradiance distributioncorresponds, of
course, to the onaxis values of the threedimensional PSF of the imag
ing system. For notation convenience we denote I
(z) = I (0, 0, z) in
this section.
According to the discussion in Sec. 4.3.3.2, the account for chro
maticityinformationleads toaproper generalizationof the monochro
matic irradiances to the polychromatic case through three functions,
namely,
X(W
20
) =
I
(z)S()x
d
Y(W
20
) =
I
(z)S()x
d (4.92)
Z(W
20
) =
I
(z)S()x
d
where S(), V(), x
, y
, and z
stand for the magnitudes used in
the previous section. The defocus coefﬁcient is deﬁned in Eq. (4.51).
However, it is oftenmore useful todescribe a chromatic signal through
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
148 C h a p t e r F o u r
a combinationof these basic functions. Aconventional choice for these
new parameters is the set
x(W
20
) =
X(W
20
)
X(W
20
) +Y(W
20
) + Z(W
20
)
(4.93)
y(W
20
) =
Y(W
20
)
X(W
20
) +Y(W
20
) + Z(W
20
)
known as chromaticity coordinates, along with the parameter Y(W
20
).
If the sensitivity functions are selected to be the spectral tristimulus
values of the human eye, the Y(W
20
) parameter is known as illumi
nance andit is associatedbasicallywiththe brightness of the chromatic
stimulus. On the other hand, in this case the chromaticity coordinates
provide a joint description for the hue and saturation of the colored
signal.
49
Anyway, as in the previous section, the evaluation of these magni
tudes requires the computationof the monochromatic components for
a sufﬁcient number of spectral components. The use of conventional
techniques, as stated earlier, is not very efﬁcient at this stage, since the
computation performed for a ﬁxed axial point, a given wavelength,
and a given aberration state cannot be applied to any other conﬁgu
ration. The method proposed in Sec. 4.3.1.1 represents a much more
efﬁcient solution since all the monochromatic values of the axial ir
radiance can be obtained, for different aberration correction states,
from a single twodimensional display associated with the pupil of
the system.
To describe this proposal ingreater detail, let us consider the system
presented in Fig. 4.9 with ␣ = 0. According to the formulas in Sec.
4.3.1.1, the axial irradiance distribution in image space, for a given
spectral component, can be expressed as
I
(z) =
1
2
( f + z)
2
RW
q
0, 0 (x
(z), ) (4.94)
where q
0, 0
(s) represents the zeroorder circular harmonic of the pupil
Q(r
N
, ), with s = r
2
N
+
1
2
. The normalized coordinates r
N
and are
implicitly deﬁned in Eq. (4.49). The speciﬁc coordinates (x
(z), ) for
the RWT are given by Eqs. (4.64) and (4.65). Note that for systems
with longitudinal chromatic aberration, the defocus coefﬁcient W
20
is substituted for the wavelengthdependent coefﬁcient in Eq. (4.90).
Note that nowthe whole dependence of the axial irradiance on , W
40
,
and z is established through these coordinates if the function Q(r
N
, )
itself does not depend on wavelength. This is the case when all the
aberrations of the system, apart from SA and longitudinal chromatic
aberration, have a negligible chromatic dependence. This is a very
usual situation in wellcorrected systems, and in this case, every axial
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 149
position, SA and chromatic aberration state, and wavelength can be
studied from the same twodimensional RWD.
Thus, providing that these kinds of systems are analyzed, the poly
chromatic description for the axial image irradiance can be assessed
by the formulas
X(W
20
) =
RW
q
0, 0 (x
(z), )
2
( f + z)
2
S()x
d
Y(W
20
) =
RW
q
0, 0 (x
(z), )
2
( f + z)
2
S() y
d (4.95)
Z(W
20
) =
RW
q
0, 0 (x
(z), )
2
( f + z)
2
S()z
d
where the values of (x
(z), ) for every wavelength, axial position,
and SA amount are given by Eqs. (4.64) and (4.65). Thus, once the
RWDof the function q
0, 0
(s) of the systemis properly computed, these
weighted superpositions can be quickly and easily calculated.
19, 50, 51
As an example for testing this technique, we evaluate the axial re
sponse of a clear circular pupil imaging system, affected by spherical
and longitudinal chromatic aberrations as shown in Fig. 4.26. With
out loss of generality we assume here that the SA coefﬁcient has a ﬂat
behavior for the considered spectral range. Once again, for the sake
of simplicity, we assume that no other aberrations are present.
400 500 600 700
–400
–200
0
200
400
δW
20
A
b
e
r
r
a
t
i
o
n
c
o
e
f
f
i
c
i
e
n
t
(
n
m
)
W
40
Wavelength: λ (nm)
FIGURE 4.26 Aberration coefﬁcients associated with the system under issue.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
150 C h a p t e r F o u r
–1500 –1000 –500 0 500 1000
0.0
1.0
2.0
λ = 400 nm
λ = 590 nm
λ = 700 nm
A
x
i
a
l
i
r
r
a
d
i
a
n
c
e
:
I
λ
(
a
.
u
.
)
Defocus parameter: W
20
(nm)
FIGURE 4.27 Axial irradiance values for the system under study. Solid lines
represent the results by analytical calculation, while superimposed symbols
correspond to the computation through the single RWD technique.
We consider 36 axial positions characterized by defocus coefﬁcient
values in a uniform sequence. We follow the same procedure as in
earlier sections for the digital calculation of the RWD RW
q
0, 0 (x
, ).
It is worth mentioning that for this pupil is possible to achieve an
analytical result for the monochromatic axial behavior of the system
for any value of W
20
, W
40
, and , namely,
12
I
(z) =
¸
a
2
2 f ( f + z)
2
1
W
40
F
¸
W
20
() +2W
40
√
W
40
− F
¸
W
20
()
√
W
40
2
(4.96)
where
F(z) =
z
0
exp
i t
2
2
dt (4.97)
is the complex formof Fresnel integral. This analytical formula is used
here to evaluate the results obtained by the proposed method. Figure
4.27 presents a comparison of these approaches for three different
wavelengths in the visible spectral range. Excellent agreement can be
observed in this ﬁgure.
Finally, we performed the calculation of the axial values for the
chromaticity coordinates and the illuminance, by assuming the same
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 151
0.0 0.8
0.0
0.5
1.0
0.0
0.5
1.0
–1500 –1000 –500 0 500 1000
A
x
i
a
l
i
l
l
u
m
i
n
a
n
c
e
(
a
.
u
.
)
Chromaticity coordinate x(W
20
)
C
h
r
o
m
a
t
i
c
i
t
y
c
o
o
r
d
i
n
a
t
e
y
(
W
2
0
)
Defocus parameter: W
20
(nm)
0.4
FIGURE 4.28 Axial illuminance and chromaticity coordinates for the system
under study. Here solid lines represent the results obtained by the
conventional method, while superimposed symbols correspond to the
computation through the single RWD technique.
settings for the sensitivity functions andthe illuminant as in the previ
ous section. The values obtained with the method presented here are
compared in Fig. 4.28 with the ones obtained by applying the same
classic technique as in Sec. 4.3.3.2. Again, a very good agreement be
tween themcan be seen. Amore detailedcomparison of both methods
is presented in Ref. 19.
4.4 Design of Imaging Systems and Optical
Signal Processing by Means of RWT
4.4.1 Optimization of Optical Systems:
Achromatic Design
We nowpresent a design method for imaging systems working under
polychromatic illumination on a RWT basis. In particular, we ﬁx our
attention on the optimal compensation of the axial chromatic disper
sion of the Fresnel diffraction patterns of a plane object Although this
proposal can be applied to a wide variety of systems, we concentrate
on an optical system specially designed for this purpose. This de
vice allows us to obtain the image of any arbitrary diffraction pattern
with very low residual chromatic aberration.
52, 53
The optical system,
sketchedinFig. 4.29, works under planar broadbandillumination. The
only two optical elements in this device are an achromatic lens, with
focal length f , and an onaxis kinoform zone plate. This element acts,
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
152 C h a p t e r F o u r
Object
Achromatic
lens
R
1
R
o
R
2
Zone
Plate
Achromatic
Fresnel pattern
x
y
x
y
l
f
d'
o
FIGURE 4.29 Achromatic imaging system under study.
from a scalar paraxial diffraction point of view, as a conventional thin
lens with a focal length proportional to the inverse of the wavelength
of the incoming light, i.e.,
Z() = Z
o
o
(4.98)
o
being a reference design wavelength and Z
o
= Z(
o
). We note
that although the effect of residual focuses can be signiﬁcant for those
wavelengths that are different fromthe design wavelength, we do not
consider it here.
Our goal in this section is to achieve the optimal relationship be
tween the geometric distances in the imaging systemto obtain an out
put image corresponding to a given Fresnel pattern with minimum
chromatic aberration. Thus, let us consider a given diffraction pattern
located at a distance R
o
from the object for the reference chromatic
component of wavelength
o
. It is well known that with parallel illu
mination the same diffraction pattern appears for any other spectral
component at a distance from the input mask given by
R() = R
o
o
(4.99)
Inthis way, if thelimits of thespectrumof theincomingradiationare
1
and
2
, the same diffraction pattern is replicated along the optical axis
between the planes characterized by distances R
1
= R(
1
) and R
2
=
R(
2
), providing a dispersion volume for the diffraction pattern under
study. However, if we ﬁxour attentiononthe reference plane locatedat
a distance R
o
fromthe object, for =
o
we obtain a different structure
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 153
for the diffracted pattern, and, therefore, the ﬁnal superposition of all
the spectral components of the incoming light produces a chromatic
blur of the monochromatic result. To analyze this effect, we again use
the RWT approach to describe the spectral components of this Fresnel
pattern. Since the above dispersion volume is transformed, by means
of the imaging system, into a different image volume, it is interesting
to derive the geometric conditions that provide a minimum value for
its axial elongation in the image space. Equivalently, the same RWT
analysis will be performed at the output of the system to analyze the
chromatic blur at the output plane for the reference wavelength
o
.
For the sake of simplicity, we consider a onedimensional ampli
tude transmittance t(x) for the diffracting object. Let us now apply
our approach to calculate the irradiance freespace diffraction pattern
under issue through the RWT of the object mask. If we recall the result
in Eq. (4.36) for z = R
o
, we obtain that each spectral component of this
Fresnel pattern is given by
I
o
(x; ) ∝ RW
t
(x
(), ()) , tan() = −R
o
,
x
() =
x
R
o
sin() (4.100)
In this equation the chromatic blur is considered through the spec
tral variation of the coordinates in the RWT for any given transverse
position x in the diffraction pattern. Thus, for a ﬁxed observation
position there is a region in the Radon space that contains all the
points needed to compute the polychromatic irradiance. If we deﬁne
i
= arctan(
i
R
o
), for i = 1, 2, the width of this region in both Radon
space directions can be estimated as
= 
1
−
2
, x
=
x
R
o
sin
1
1
−
sin
2
2
(4.101)
Note that the smaller this region is, the less is the effect of the chro
matic blur affecting the irradiance at the speciﬁed observation point.
To achieve an achromatization of the selected diffraction pattern, this
region has to be reduced in the output plane of the optical setup.
Let us now ﬁx our attention on the effect of the imaging system on
the polychromatic diffraction pattern under issue. Again, we use the
RWT approach to achieve this description by simply noting that the
systembehaves as an abcd device that links the object plane andthe se
lected output plane. The transformation matrix M
achr
can be obtained
as a sequence of elemental transformations (see Fig. 4.29), namely, free
propagation at a distance l, propagation through the achromatic lens,
free propagation to the focal plane of that element, passage through
the zone plate, and, ﬁnally free propagation at a distance d
o
. The out
put plane is selected as the image plane of the diffraction pattern
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
154 C h a p t e r F o u r
under study for =
o
. Thus, by using the results in Sec. 4.2.1, it is
straightforward to obtain
M() =
⎛
⎜
⎝
1 −
l
f
−
f
Z()
−
f −m
o
f +m
o
l +
f
2
m
o
Z()
1
f
m
o
⎞
⎟
⎠
(4.102)
where the following restriction applies to the ﬁxed desired output
plane
l = R
o
+ f −
f
m
o
−
f
2
Z
o
(4.103)
where m
o
= −d
o
/f is the magniﬁcation obtained at the ﬁxed image
plane (for =
o
). The relationship between the RWTs of the input
object and the output Fresnel pattern for each spectral channel now
canbe establishedbyapplicationof Eqs. (4.27) and(4.28). Inparticular,
by setting = 0 we ﬁnd
I
o
(x; ) ∝ RW
t
x
(),
()
tan
() =
¸
R
o
+
f
2
Z
o
o
−1
, x
=
x
m
o
cos
() (4.104)
Therefore, for the polychromatic description of the output diffraction
pattern we have to sum values of the RWT of the transmittance of the
object in a region in the Radon domain whose size in both dimensions
is given by
= 
max
−
min
, x
=
x
m
o
(cos
max
−cos
min
)
(4.105)
where
max
= max{
() ∈ [
1
,
2
]},
min
= min{
() ∈ [
1
,
2
]}
(4.106)
The speciﬁc values of these limits, which deﬁne the extension of the
integration region in Radon space in the polychromatic case, depend
on the particular values of the geometric design parameters f and
Z
o
of the imaging system. We now try to ﬁnd a case that minimizes
the chromatic blur in the output pattern. It is worth mentioning that
exact achromatization of the pattern is achieved only when
() =
(
o
) ∀ ∈ [
1
,
2
], which cannot be fulﬁlled in practice, as can be
seen from Eq. (4.104). However, a ﬁrstorder approximation to that
ideal correction can be achieved by imposing a stationary behavior
for
() around =
o
. Mathematically, we impose
d
()
d
o
= 0 (4.107)
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 155
(a) (b)
FIGURE 4.30 Grayscale display of the irradiance distribution to be
achromatized: (a) Monochromatic pattern for
o
= 546.1 nm. (b) Broadband
(Hg lamp) irradiance distribution.
or, equivalently,
d tan
()
d
o
= 0 (4.108)
which leads to the optimal constraint
R
o
= −
f
2
Z
o
(4.109)
This condition transforms Eq. (4.103) into
l = 2R
o
+ f −
f
m
o
(4.110)
Thus, the choice of a set of geometric parameters l, f , Z
o
, and d
o
fulﬁlling the two above equations provides a design prescription for
a ﬁrstorder compensation of the chromatic blur in the diffraction
pattern located, for =
o
, at distance R
o
from the object.
54
To illustrate this design procedure and to check the predicted
results, we present an experimental veriﬁcation by using a two
dimensional periodic transmittance as an object, with the same pe
riod p = 0.179 mm in both orthogonal directions. As a Fresnel pat
tern to be achromatized, a selfimaging distribution is selected. In
particular, after parallel illumination with
o
= 546.1 nm, the dis
tance R
o
= 11.73 cm is selected. Figure 4.30a shows a picture of the
irradiance distribution in that situation. In Fig. 4.30b, the irradiance
distribution over the same plane, but when a polychromatic colli
mated beam from a highpressure Hg lamp is used, is presented. The
chromatic blur is clearly seen by comparing these two ﬁgures.
To optimally achromatize this diffraction pattern, we follow the
prescriptions given in the above paragraphs. We use a kinoform lens
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
156 C h a p t e r F o u r
FIGURE 4.31 Grayscale display of the achromatized irradiance distribution.
with Z
o
= −12 cm, and we choose a value d
o
= 10.00 cm. Therefore,
we select a focal distance for the achromatic lens f =
√
−Z
o
R
o
=
11.86 cm, and we place that object at a distance l = 2R
o
+ f + f
2
/d
o
=
49.39 cm from that lens. A grayscale display of the output irradiance
is presented in Fig. 4.31. The comparison between this result and the
monochromatic one inFig. 4.30a shows the highachromatizationlevel
obtained with the optimized system.
4.4.2 Controlling the Axial Response:
Synthesis of Pupil Masks by RWT
Inversion
In Sec. 4.3.1.1 we showed that the axial behavior of the irradiance
distribution provided by a system with an arbitrary value of SA can
be obtained from the single RWT of the mapped pupil q
0, 0
(s) of the
system. In fact, Eq. (4.74) can be considered the keystone of a pupil
design method
55
in which the synthesis procedure starts by perform
ing a tomographic reconstruction of W
q
0, 0 (x, ) from the projected
function I (0, 0, z) representing the irradiance at the axial points—
variableW
20
—for a sufﬁcient set of values of W
40
. Thus, the entire
twodimensional Wigner space can be sampled on a set of lines de
ﬁned by these parameters. The backprojection algorithm converts the
desired axial irradiance for a ﬁxed value of W
40
, represented by a one
dimensional function, to a twodimensional function by smearing it
uniformly along the original projection direction (see Fig. 4.8). Then
the algorithm calculates the summation function that results when all
backprojections are summed over all projection angles , i.e., for all
the different values of W
40
. The ﬁnal reconstructed function W
q
0, 0 (x, )
is obtained by a proper ﬁltering of the summation image.
55
Once the
WDF is synthesized with the values of the input axial irradiances, the
pupil function is obtained by use of Eq. (4.4). Finally, the geometric
mappinginEq. (4.57) is invertedto provide the desiredpupil function.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 157
1.0
0.8
0.6
0.4
0.2
0.0
0.0 0.2 0.4 0.6
(a) (b)
r/a
0.8 1.0
A
m
p
l
i
t
u
d
e
t
r
a
n
s
m
i
t
t
a
n
c
e
1.0
0.8
0.6
0.4
0.2
0.0
0.0 0.2 0.4 0.6
r/a
0.8 1.0
A
m
p
l
i
t
u
d
e
t
r
a
n
s
m
i
t
t
a
n
c
e
FIGURE 4.32 (a) Amplitude transmittance of a desired pupil function.
(b) Phasespace tomographic reconstruction of the same pupil.
To illustrate the method, we numerically simulated the synthesis
of an annular apodizer represented in Fig. 4.32. It has been shown
that its main features are to increase the focal depth and to reduce
the inﬂuence of SA. From this function we numerically determined
ﬁrst the W
q
0, 0 (x, ) function, using the WDF deﬁnition, and thereby
the projected distributions deﬁned by the RWT, obtaining the axial
irradiance distribution for different values of SA. In this case, we used
1024 values for both W
40
/ and W
20
/, ranging from −16 to +16. We
treated these distributions as if they represented the desired axial be
havior for a variable SA, and we reconstructed the WDF by using
a standard ﬁltered backprojection algorithm for the inverse Radon
transform. From the reconstructed WDF we obtained the synthesized
pupil function p( x) by performing the discrete onedimensional in
verse FTof W
q
0, 0 (x, ). The result is shown in Fig. 4.32b. As can be seen,
the amplitude transmittance of the synthesized pupil function closely
resembles the original apodizer in Fig. 4.32a.
4.4.3 Signal Processing through RWT
Throughout this chapter we have discussed the RWT as a mathemat
ical tool that allows us to develop novel and interesting applications
in optics. Among several mathematical operations that can be opti
cally implemented, correlation is one of the most important because it
can be used for different applications, such as pattern recognition and
object localization. Optical correlation can be performed in coherent
systems by use of the fact that the counterpart of this operation in the
Fourier domain is simply the product of both signals. To implement
this operation, several optical architectures were developed, such as
the classic VanderLugt and joint transform correlators.
56, 57
Because
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
158 C h a p t e r F o u r
conventional correlation is a shiftinvariant operation, the correlation
output simplymoves if the object translates at the input plane. Inmany
cases this property is necessary, but there are situations in which the
position of the object provides additional information such as in im
age coding or cryptographic applications, and so shift invariance is a
disadvantage.
The fractional correlation
58, 59
is a generalization of the classic cor
relation that employs the optical FrFT of a given fractional order p
instead of the conventional FT. Conventionally, the fractional corre
lation is obtained as the inverse Fourier transform of the product of
the FrFT of both the reference and the input objects, but for a sin
gle fractional order p at a time. The fractional order involved in the
FrFT controls the amount of shift variance of the correlation. As is
well known, the shiftvariance property modiﬁes the intensity of the
correlation output when the input is shifted. In several pattern recog
nition applications this feature is useful, for example, when an object
should be recognized in a relevant area and rejected otherwise, or
when the recognition should be based on certain pixels in systems
with variable spatial resolution. However, the optimum amount of
variance for a speciﬁc application is frequently difﬁcult to predict,
and therefore more complete information would certainly be attained
from a display showing several fractional correlations at the same
time. Ideally, such a display should include the classic shiftinvariant
correlation as the limiting case. In this section we will showthat such a
multichannel fractional correlator could be easily implemented from
the RWD system presented in Sec. 4.2.2. The resulting optical system
generates a simultaneous display of fractional correlations of a one
dimensional input for a continuous set of fractional orders inthe range
p ∈ [0, 1].
We start by recalling
58
the deﬁnition of the fractional correlation
between two onedimensional functions f (x) and f
(x)
C
p
(x) = F
−1
{F
p
(␣) F
p
∗
(␣), x} (4.111)
It is important to note that with the above deﬁnition the classic cor
relation is obtained if we set p = 1. The product inside the brackets
of Eq. (4.111) can be optically achieved simultaneously for all frac
tional orders, ranging between p = 0 and p = 1, following a twostep
process. In the ﬁrst stage, the RWD of the input is obtained with the
experimental conﬁguration shown in Sec. 4.2.2. A matched ﬁlter can
be obtained at the output plane if, instead of recording the intensity,
we register a hologram of the ﬁeld distribution at this plane with a
reference wavefront at an angle . (see Fig. 4.33).
In the second stage, the obtained multichannel matched ﬁlter is
located at the ﬁlter plane, and the input function to be correlated is
located at the input plane (see Fig. 4.34).
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 159
HeNe
Input plane Filter plane
BS
M
1
M
1
t
z
R
p
l
L
c
a
p
x
h
x
0
H
L
S
θ
a¢
FIGURE 4.33 Multichannel matched ﬁlter registration for a fractional
correlation. The elements are the same as in Sec. 4.2.2, except for BS
(beam splitter) and M1 and M2 (plane mirrors).
Because the transmittance of the holographic ﬁlter has one term
proportional to the complex conjugate of the reference ﬁeld in
Eq. (4.111), for each fractional order channel the ﬁeld immediately
behind the ﬁlter plane has one termproportional to the product of the
complex conjugate of the FrFT of the reference function f
(x) and the
same FrFT of the input function f (x). Thus the multiplicative phase
factor in this equation and the corresponding one of the matched ﬁlter
cancel out. Besides, although the experimental FrFT for a given order
p is approximated owing to the scale error discussed in Sec. 4.2.2, the
experimental fractional correlation can be obtained exactly because
this error affects both F
p
(␣) and F
p
∗
(␣). Finally, the diffracted ﬁeld at
angle is collected by the lens L
c
, which performs a onedimensional
FT. Because each fractional order p ∈ [0, 1] has an independent one
dimensional correlation channel, all the fractional correlations for this
HeNe
Input plane Filter plane
Output plane
BS
t
L
H
L
c
L
c
x
h
x
c
x
0
l
z
a¢
f
f
θ
FIGURE 4.34 Multichannel fractional correlator. The ﬁlter H corresponds to
the one obtained in the setup of Fig. 4.33.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
160 C h a p t e r F o u r
1.0
0.8
0.6
0.4
0.2
0.0
230 235 240 245 250 255 260
Spatial coordinate (pixels)
A
m
p
l
i
t
u
d
e
t
r
a
n
s
m
i
t
t
a
n
c
e
(
a
.
u
.
)
265 270 275 280
FIGURE 4.35 Amplitude transmittance of an input object selected to
perform multichannel fractional correlation.
range of fractional orders are obtained simultaneously at the output
plane. Thus a twodimensional display is obtained in which the frac
tional correlations are ordered in a continuous display along the axis
normal to the plane shown in Fig. 4.34.
The shiftvariant property of the FrFT correlation was conﬁrmed
experimentally in Ref. 60. Here we present a numerical simula
tion using an input object whose amplitude transmittance is shown
in Fig. 4.35. It represents a double nonsymmetric slit with a contin
uous graylevel amplitude transmittance. The continuous transition
between the shiftvariant case p = 0 and the shiftinvariant case p = 1
is conﬁrmed in Fig. 4.36. In this ﬁgure the fractional autocorrelation
of the input is considered, but the reference objects are shifted at the
input plane.
Figure 4.36a shows the fractional correlations when the input is
shifted an amount of onehalf of the object size, and Fig. 4.36b shows
the fractional correlation when the input is shifted an amount equal to
the size of the object. The variant behavior of the fractional correlation
can be clearly seen by the comparison of these ﬁgures. Both displays
coincide near to p = 1 (except for the location of the maxima), but for
lower values of p the fractional correlation is highly dependent on the
magnitude of the shift. As can be seen in the threedimensional plot
in this ﬁgure, for a ﬁxed displacement the correlation peak increases
with p. As expected for p = 1, the correlation peak is the classic one
located at the input position. For values ranging between p = 0.5
and p = 1, the correlation peak did not change appreciably. The
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 161
1.00
0.75
p
p
0.50
0.25
S
p
a
c
e
c
o
o
rd
in
a
te
(x
)
F
r
a
c
t
i
o
n
a
l
o
r
d
e
r
p
F
r
a
c
t
i
o
n
a
l
c
o
r
r
e
l
a
t
i
o
n
i
n
t
e
n
s
i
t
y
0.00
0.0
0
S
p
a
c
e
c
o
o
rd
in
a
te
(x
)
0
0.0
0.2
0.4
0.6
0.8
1.0
F
r
a
c
t
i
o
n
a
l
o
r
d
e
r
p
0.0
0.2
0.4
0.6
0.8
1.0
1.0
0.5
F
r
a
c
t
i
o
n
a
l
c
o
r
r
e
l
a
t
i
o
n
i
n
t
e
n
s
i
t
y
0.0
1.0
0.5
1.00
0.75
0.50
0.25
0.00
0
x
0
x
(a)
(b)
FIGURE 4.36 Multichannel fractional autocorrelation of the function
represented in Fig. 4.35 with (a) a shift in the input plane of onehalf of the
object size and (b) a shift of the whole size of the object.
shiftvariant property becomes evident for values close to p = 0.25.
It can be seen that as the fractional order becomes lower, the peak
degenerates and shifts disproportionately toward the object position.
Thus, the output of the system shows a variable degree of space vari
ance ranging from the pure shift variance case p = 0 to the pure shift
invariance case p = 1, that is, the classic correlation. This kind of
representation provides information about the object, such as classic
correlation, but also quantiﬁes its departure from a given reference
position.
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
162 C h a p t e r F o u r
Acknowledgments
The authors would like to express their gratitude to P. Andr´ es, S.
Granieri, E. Sicre, andE. Silvestre for their contributions inthe research
revisited in this chapter. They also acknowledge ﬁnancial support of
Ministerio de Ciencia y Tecnolog´ıa, Spain (Grants DPI 20068309 and
DPI 200802953).
References
1. J. C. Wood and D. T. Barry, “Radon transformation of timefrequency distri
butions for analysis of multicomponent signals,” Proc. Int. Conf. Acoust. Speech
Signal Process. 4: 257–261 (1992).
2. J. C. Wood and D. T. Barry, “Linear signal synthesis using the RadonWigner
transform,” IEEE Trans. Signal Process. 42: 2105–2111 (1994).
3. L. Cohen, TimeFrequencyAnalysis, PrenticeHall, Upper Saddle River, N.J., 1995.
4. D. Mendlovic, R. G. Dorsch, A. W. Lohmann, Z. Zalevsky, and C. Ferreira, “Op
tical illustration of a varied fractional Fouriertransform order and the Radon
Wigner display,” Appl. Opt. 35: 3925–3929 (1996).
5. A. W. Lohmann, “Image rotation, Wigner rotation, and the fractional Fourier
transform,” J. Opt. Soc. Am. A10: 2181–2186 (1993).
6. S. Granieri, W. D. Furlan, G. Saavedra, and P. Andr´ es, “RadonWigner display:
A compact optical implementation with a single varifocal lens,” Appl. Opt. 36:
8363–8369 (1997).
7. P. PellatFinet, “Fresnel diffraction and the fractionalorder Fourier transform,”
Opt. Lett. 19: 1388–1390 (1994).
8. P. Andr´ es, W. D. Furlan, G. Saavedra, and A. W. Lohmann, “Variable fractional
Fourier processor: A simple implementation,” J. Opt. Soc. Am. A14: 853–858
(1997).
9. E. Tajahuerce, G. Saavedra, W. D. Furlan, E. E. Sicre, and P. Andr´ es, “White
light optical implementation of the fractional fourier transformwith adjustable
order control,” Appl. Opt. 39: 238–245 (2000).
10. Y. Zhang, B. Gu, B. Dong, and G. Yang, “Optical implementations of the Radon
Wigner display for onedimensional signals,” Opt. Lett. 23: 1126–1128 (1998).
11. Y. Zhang, B.Y. Gu, B.Z. Dong, and G.Z. Yang, “Newoptical conﬁgurations for
implementing RadonWigner display: Matrix analysis approach,” Opt. Comm.
160: 292–300 (1999).
12. H. H. Hopkins and M. J. Yzuel, “The computation of diffraction patterns in the
presence of aberrations,” Optica Acta 17: 157–182 (1970).
13. M. J. Yzuel and F. Calvo, “Pointspread function calculation for optical systems
with residual aberrations and nonuniformtransmission pupil,” Optica Acta 30:
233–242 (1983).
14. J. J. Stamnes, B. Spjelkavik, and H. M. Pedersen, “Evaluation of diffraction
integrals using local phase and amplitude approximations,” Optica Acta 30:
207–222 (1983).
15. H. G. Kraus, “Finite element area andline integral transforms for generalization
of aperture function and geometry in Kirchhoff scalar diffraction theory,” Opt.
Eng. 32: 368–383 (1993).
16. L. A. D’Arcio, J. J. M. Braat, and H. J. Frankena, “Numerical evaluation of
diffraction integrals for apertures of complicated shape,” J. Opt. Soc. Am. A11:
2664–2674 (1994).
17. G. Saavedra, W. D. Furlan, E. Silvestre, and E. Sicre, “Analysis of the irradi
ance along different paths in the image space using the Wigner distribution
function,” Opt. Comm. 139: 11–16 (1997).
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
T h e R a d o n  W i g n e r T r a n s f o r m 163
18. W. D. Furlan, G. Saavedra, E. Silvestre, and M. Mart´ınezCorral, “Onaxis ir
radiance for spherically aberrated optical systems with obscured rectangular
apertures: A study using the Wigner distribution function,” J. Mod. Opt. 45:
69–77 (1998).
19. W. D. Furlan, G. Saavedra, E. Silvestre, J. A. Monsoriu, and J. D. Patrignani,
“Assessment of a Wignerdistributionfunctionbased method to compute the
polychromatic axial response given by an aberrated optical system,” Opt. Eng.
42: 753–758 (2003).
20. W. D. Furlan, G. Saavedra, J. A. Monsoriu, andJ. D. Patrignani, “Axial behaviour
of Cantor ring diffractals,” J. Opt. A: Pure Appl. Opt. 5: S361–S364 (2003).
21. W. D. Furlan, M. Mart´ınezCorral, B. Javidi, and G. Saavedra, “Analysis of 3D
integral imaging displays using the Wigner distribution,” J. Disp. Technol. 2:
180–185 (2006).
22. P. Andr´ es, M. Mart´ınezCorral, and J. OjedaCasta˜ neda, “Offaxis focal shift for
rotationally nonsymmetric screens,” Opt. Lett. 18: 1290–1292 (1993).
23. A. Dubra and J. A. Ferrari, “Diffracted ﬁeld by an arbitrary aperture,” Am. J.
Phys. 61: 87–92 (1999).
24. W. D. Furlan, G. Saavedra, and S Granieri, “Simultaneous display of all the
Fresnel diffraction patterns of onedimensional apertures,” Am. J. Phys. 69: 799
(2001).
25. C. Allain and M. Cloitre, “Optical diffraction on fractals,” Phys. Rev. B33: 3566–
3569 (1986).
26. Y. Sakurada, J. Uozumi, and T. Asakura, “Fresnel diffraction by one
dimensional regular fractals,” Pure Appl. Opt. 1: 29–40 (1992).
27. T. Alieva and F. Agull ´ oL´ opez, “Optical wave propagation of fractal ﬁelds,”
Opt. Comm. 125: 267–274 (1996).
28. O. Trabocchi, S. Granieri, and W. D. Furlan, “Optical propagation of fractal
ﬁelds: Experimental analysis in a single display,” J. Mod. Opt. 48: 1247–1253
(2001).
29. M. G. Raymer, M. Beck, and D. F. McAlister, “Complex waveﬁeld reconstruc
tion using phasespace tomography,” Phys. Rev. Lett. 72: 1137–1140 (1994).
30. D. F. McAlister, M. Beck, L. Clarke, A. Mayer, andM. G. Raymer, “Optical phase
retrieval by phasespace tomography andfractionalorder Fourier transforms,”
Opt. Lett. 20: 1181–1183 (1994).
31. Y. Li, G. Eichmann, andM. Conner, “Optical Wigner distributionandambiguity
function for complex signals and images,” Opt. Comm. 67: 177–179 (1988).
32. G. Shabtay, D. Mendlovic, and Z. Zalevsky, “Proposal for optical imple
mentation of the Wigner distribution function,” Appl. Opt. 37: 2142–2144
(1998).
33. R. L. Easton, Jr., A. J. Ticknor, and H. H. Barrett, “Application of the Radon
transform to optical production of the Wigner distribution,” Opt. Eng. 23: 738–
744 (1984).
34. W. D. Furlan, C. Soriano, and G. Saavedra, “Optodigital tomographic recon
struction of the Wigner distribution function of complex ﬁelds,” Appl. Opt. 47:
E63–E67 (2008).
35. A. C. Kak and M. Slaney, Principles of Computerized Tomographic Imaging, IEEE
Press, New York, 1988.
36. U. Gopinathan, G. Situ, T. J. Naughton, and J. T. Sheridan, “Noninterferometric
phase retrieval using a fractional Fourier system,” J. Opt. Soc. Am. A25: 108–115
(2008).
37. M. Born and E. Wolf, Principles of Optics: Electromagnetic Theory of Propaga
tion, Interference and Diffraction of Light, (7th ed., Cambridge University Press,
Cambridge, 1999, Chap. 9.
38. H. H. Hopkins, “The aberration permissible in optical systems,” Proc. Phys. Soc.
B70: 449–470 (1957).
39. H. Bartelt, J. OjedaCasta˜ neda, and E. E. Sicre, “Misfocus tolerance seen by
simple inspection of the ambiguity function,” Appl. Opt. 23: 2693–2696 (1984).
May 22, 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
164 C h a p t e r F o u r
40. W. D. Furlan, G. Saavedra, and J. Lancis, “Phasespace representations as a tool
for the evaluation of the polychromatic OTF,” Opt. Comm. 96: 208–213 (1993).
41. W. D. Furlan, M. Mart´ınezCorral, B. Javidi, and G. Saavedra, “Analysis of
3D integral imaging display using the Wigner distribution,” J. Disp. Technol. 2:
180–185 (2006).
42. K. H. Brenner , A. W. Lohmann, and J. OjedaCasta˜ neda, “The ambiguity func
tion as a polar display of the OTF,” Opt. Comm. 44: 323 (1983).
43. W. D. Furlan, G. Saavedra, and J. Lancis, “Phasespace representations as a tool
for the evaluation of the polychromatic OTF,” Opt. Comm. 96: 208–213 (1993).
44. J. Besc´ os and J. Santamar´ıa, “Formation of color images: Optical transfer func
tions for the tristimulus values,” Photogr. Sci. and Eng. 21: 355–362 (1977).
45. J. Besc´ os, J. H. Altamirano, A. Santisteban, and J. Santamar´ıa, “Digital restora
tion models for color imaging,” Appl. Opt. 27: 419–424 (1988).
46. R. Barnden, “Calculation of axial polychromatic optical transfer function,” Op
tica Acta 21: 981–1003 (1974).
47. R. Barnden, “Extraaxial polychromatic optical transfer function,” Optica Acta
23: 1–24 (1976).
48. M. Takeda, “Chromatic aberration matching of the polychromatic optical trans
fer function,” Appl. Opt. 20: 684–687 (1981).
49. G. Wyszecki and W. S. Stiles, Color Science, Wiley, New York, 1982).
50. W. D. Furlan, G. Saavedra, E. Silvestre, M. J. Yzuel, and P. Andr´ es, “Polychro
matic merit functions in terms of the Wigner distribution function,” Proc. SPIE
2730: 252–255 (1996).
51. W. D. Furlan, G. Saavedra, E. Silvestre, P. Andr´ es, and M. J. Yzuel, “Polychro
matic axial behavior of aberrated optical systems: Wigner distribution function
approach,” Appl. Opt. 36: 9146–9151 (1997).
52. P. Andr´ es, J. Lancis, E. E. Sicre, and E. Bonet, “Achromatic Fresnel diffraction
patterns,” Opt. Comm. 104: 39–45 (1993).
53. P. Andr´ es, J. Lancis, E. Tajahuerce, V. Climent, and G. Saavedra, “Whitelight
optical information processing with achromatic processors,” 1994 OSA Tech.
Digest Series 11: 220–223 (1994).
54. J. Lancis, E. E. Sicre, E. Tajahuerce, and P. Andr´ es, “Whitelight implementation
of the Wignerdistribution function with an achromatic processor,” Appl. Opt.
34: 8209–8212 (1995).
55. W. Furlan, D. Zalvidea, and G. Saavedra, “Synthesis of ﬁlters for speciﬁed axial
irradiance by use of phasespace tomography,” Opt. Comm. 189: 15–19 (2001).
56. A. Vander Lugt, Optical Signal Processing, Wiley, New York, 1992.
57. J. W. Goodman, Introduction to Fourier Optics, McGraw Hill, New York, 1996.
58. D. Mendlovic, H. M. Ozaktas, and A. W. Lohmann, “Fractional correlation,”
Appl. Opt. 34: 303–309 (1995).
59. S. Granieri, R. Arizaga, and E. E. Sicre, “Optical correlation based on the frac
tional Fourier transform,” Appl. Opt. 36, 6636–6645 (1997).
60. S. Granieri, M. Tebaldi, and W. D. Furlan, “Parallel fractional correlation: An
optical implementation,” Appl. Opt. 40: 6439–6444 (2001).
May 22, 2009
17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
108
Chapter Four
closely related to the projections of the WDF in phase space and also intimately connected with AF, as will be shown. The general structure of this chapter is as follows. In Sec. 4.2, a general overview of mathematical properties of the RWT is given, and a summary of different optical setups for achieving it is presented. Next, the use of this representation in the analysis of optical signals and systems is developed in several aspects, namely, the computation of diffraction intensities, the optical display of Fresnel patterns, the amplitude and phase reconstruction of optical ﬁelds, and the calculation of merit function in imaging systems. Finally, in Sec. 4.4, a review of design techniques, based on the utilization of the RWT, for these imaging systems is presented, along with some techniques for optical signal processing.
4.2 Projections of the Wigner Distribution Function in Phase Space: The RadonWigner Transform (RWT)
The RWT was ﬁrst introduced for the analysis and synthesis of frequencymodulated time signals, and it is a relatively new formalism in optics.1,2 However, it has found several applications in this ﬁeld during the last years. Many of them, such as the analysis of diffraction patterns, the computation of merit functions of optical systems, or the tomographic reconstruction of optical ﬁelds, are discussed in this chapter. We start by presenting the deﬁnition and some basic properties of the RWT. The optical implementation of the RWT which is the basis for many of the applications is discussed next. Note, as a general remark, that for the sake of simplicity most of the formal deﬁnitions for the signals used hereafter are restricted to onedimensional signals, i.e., functions of a single variable f (x). This is mainly justiﬁed by the speciﬁc use of these properties that we present in this chapter. The generalization to more than one variable is in most cases straightforward. We will refer to the dual variables x and as spatial and spatialfrequency coordinates, since we will deal mainly with signals varying on space. Of course, if the signal is a function of time instead of space, the terms time and frequency should be applied.
4.2.1 Deﬁnition and Basic Properties
We start this section by recalling the deﬁnition of the WDF associated with a complex function f (x), namely, W{ f (x ), x, } = W f (x, )
+∞
=
−∞
f x+
x 2
f∗ x−
x 2
exp (−i2
x ) dx
(4.1)
May 22, 2009
17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
The RadonWigner Transform
which also can be deﬁned in terms of the Fourier transform (FT) of the original signal
+∞
109
F{ f (x), } = F ( ) =
−∞
f (x) exp (−i2
x) d x
(4.2)
as
+∞
W f (x, ) =
−∞
F
+
2
F∗
−
2
exp (i2
x) d
(4.3)
It is interesting to remember that any WDF can be inverted to recover, up to a phase constant, the original signal or, equivalently, its Fourier transform. The corresponding inversion formulas are3 1 f (x) = ∗ f (x )
+∞
Wf
−∞ +∞
x+x , 2
exp [i2
(x − x )] d
(4.4)
1 F( ) = ∗ F ( )
Wf
−∞
x,
+ 2
exp [−i2 ( − )x] d x (4.5)
Note that these equations state the uniqueness of the relationship between the signal and the corresponding WDF (except for a phase constant). It is straightforward to deduce from these formulas that the integration of the WDF on the spatial or spatialfrequency coordinate leads to the modulus square of the signal or its Fourier transform, respectively, i.e.,
+∞
 f (x) =
2 −∞ +∞
W f (x, ) d
(4.6)
F ( )2 =
−∞
W f (x, ) d x
(4.7)
These integrals, or marginals, can be viewed as the projection of the function W f (x, ) in phase space along straight lines parallel to the axis [in Eq. (4.6)] or to the x axis [in Eq. (4.7)]. These cases are particular ones of all possible projections along straight lines of a given function in phase space. In fact, for any function of (at least) two coordinates,
May 22, 2009
17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
110
Chapter Four
xθ
Rg(xθ, θ)
yθ
y
g(x, y) xθ
θ
x
FIGURE 4.1 Projection scheme for the deﬁnition of the Radon transform.
say, g(x, y), its Radon transform is deﬁned as a generalized marginal
+∞
R {g (x, y) , x , } = Rg (x , ) =
−∞
g (x, y) dy
(4.8)
where, as presented in Fig. 4.1, x and y are the coordinates rotated by an angle . It is easy to see from this ﬁgure that Rg (x , + ) = Rg (−x , ) (4.9)
Thus, the reduced domain ∈ [0, ) is used for Rg (x , ). Note that the integration in the above deﬁnition is performed along straight lines characterized, for a given pair (x , ), by the equation x sin x=x y= y=x − x tan for for for = 0, =0 = 2
2 (4.10)
May 22, 2009
17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04
The RadonWigner Transform
and therefore Eq. (4.8) can be reformulated as
111
⎧ +∞ ⎪ ⎪ x x ⎪ ⎪ g x, − ⎪ ⎪ sin tan ⎪ ⎪−∞ ⎪ ⎪ ⎪ +∞ ⎪ ⎪ ⎨ g (x , y) dy Rg (x , ) = ⎪ ⎪−∞ ⎪ ⎪ ⎪ +∞ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ g (x, x ) d x ⎪ ⎪ ⎩
−∞
dx
for
= 0,
2 (4.11)
for
=0
for
=
2
Thus, when we consider as projected function W f (x, ), we can deﬁne the generalized marginals as the Radon transform of this WDF, namely,
+∞
R{W f (x, ), x , } = RW f (x , ) =
−∞ +∞
W f (x, ) d
=
−∞
W f (x cos − sin , x sin + cos ) d
⎧ +∞ ⎪ ⎪ ⎪ ⎪ W f x, x − x ⎪ ⎪ sin tan ⎪ ⎪ ⎪ −∞ ⎪ ⎪ +∞ ⎪ ⎪ ⎨ W f (x , ) d = ⎪ ⎪ ⎪ −∞ ⎪ ⎪ +∞ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ W f (x, x ) d x ⎪ ⎪ ⎩
−∞
dx
for
= 0,
2
for
=0
for
=
2 (4.12)
where, in the last expression, we have explicitly considered the equations for the integration lines in the projection. In terms of the original signal, this transform is called its RadonWigner transform. It is easy to show that
+∞ +∞
RW f (x , ) = RW f (x , ) =
−∞ −∞
f
x cos − sin + x 2
x 2
×f∗
x cos − sin −
× exp[−i2 (x sin + cos )x ] d x d
(4.13)
17) and. Testorf/1597980/Ch04 112 Chapter Four By performing a proper change in the integration variables. Note also that from Eq.18) . In this case. } ⎧ ⎪ exp[i ( √ 2 /tan )] +∞ + ⎪ f (x) ⎪ ⎪ i sin ⎪ −∞ ⎪ ⎨ x2 × exp i tan exp −i2 = Fp( ) = ⎪ ⎪f( ) ⎪ ⎪ ⎪ ⎪ ⎩ F( ) x sin dx for for for = 0.May 22. ) (4. − ) = RW f ∗ (−x . the range ∈ [0. one ﬁnds RW f (x . for this kind of signal the reduced domain ∈ [0. Equation (4. therefore.14) a symmetry condition can be stated. − ) = RW f (−x . namely. =0 = 2 2 (4. often considered a generalization of the classic Fourier transform. ) (4. that is. (4. the fractional Fourier transform (FrFT).16) so that for real signals. ) ≥ 0 (4. namely. is given by F p { f (x). f (x) = f ∗ (x)∀x ∈ R. ) = ⎧ ⎪ ⎪ ⎪ ⎪ ⎨ 1 sin +∞ −∞ 2 f (x) exp i x2 tan exp −i2 x x sin dx for for = 0. /2] contains in fact all the necessary values for a full deﬁnition of the RWT. RW f (x .14) also allows one to link the RWT with another integral transform deﬁned directly from the original signal. This transformation. =0 = 2 2 ⎪ f (x0 = x)2 ⎪ ⎪ ⎪ ⎩ F x /2 = 2 for (4.15) This is a very interesting property. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. since the WDF cannot be positive in whole phase space (except for the particular case of a Gaussian signal).14) From this equation it is clear that RW f (x . the following more compact expression can be obtained RW f (x . ) in the Radon transform is clearly redundant.
e.. −x ) There is a wellknown relationship between the twodimensional FT of a function and the onedimensional Fourier transformation of its projections. − sin ) (4. From this deﬁnition.e. c.19) so that the RWT can be also interpreted as a twodimensional representation of all the FrFTs of the original function.2. x } = A f ( . ). the input signal undergoes a canonical transformation deﬁned through four real parameters (a . which states that the values of the onedimensional FT of a projection at an angle give a central proﬁle—or slice—of the twodimensional FT of the original signal at the same angle. i. . equivalently. ) exp [−i2 ( x + x )] d x d (4.21) = A f ( . . 4. For our input signal the AF is deﬁned as A{ f (x). This link is established through the central slice theorem. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. These relationships together with other links between representations in the phase space are summarized in Fig. If we apply this theorem to the WDF. it is easy to see that RW f (x .May 22. we consider the relationship between the RWT of an input onedimensional signal f (x) and the RWT of the same signal but after passing through a ﬁrstorder optical system.. To conclude this section. +∞ +∞ F2D {W f (x. by a 2×2 real matrix M= a c b d (4.22) i.20) which can be understood as the twodimensional FT of the WDF. In this case. ) = F2 / 113 (x )2 (4. ). } = Af ( cos . Testorf/1597980/Ch04 The RadonWigner Transform where = p /2. b. it is straightforward to show that F{RW f (x . x ) +∞ = −∞ f x+ x 2 f∗ x− x 2 exp(−i2 x) d x (4.23) . Another interesting relationship can be established between the RWT and the AF associated with the input signal. x ) along a straight line forming an angle − with the axis. the onedimensional FT of the RWT for a ﬁxed projection angle provides a central proﬁle of the AF A f ( . x } = −∞ −∞ W f (x. d) or.
7). ) where the mapped coordinates are given by x = a c b d x (4. (2) AF and inverse transform.24) −i cx 2 a f x a which are the onedimensional counterparts of Eqs.4) and (3.2 Relationship diagram between the original signal f (x) and different phasespace representations. (3) projection (Radon) transformation and tomographic reconstruction operator. and R stand for FT. (4) expression of the central slice theorem applied to Radon transform and AF. W.25) . in such a way that the transformed signal g(x) is given by g(x) = ⎧ ⎪ √1 ⎪ ⎨ ib exp ⎪ ⎪ ⎩ exp −i d x 2 b 1 √ a +∞ −∞ f (x ) exp −i a x 2 b exp i2 b x x dx b=0 b=0 (4.May 22. and ( . ξ) W W–1 R f(x) F 2 {x→xθ} p θ = pπ/2 A A–1 F{ F x θ→ (√ξ } 2 θ x' . A. θ) –1 ξ {(√ '2 + 2 θ) p x' . F p . (3. while −1 represents the corresponding inverse operator. ) = W f (a x + b . WDF integral. '2 + ) p} → Af (ξ'. respectively. x') x θ} R–1 (3) RWf (xθ.26) (4. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. (4) FIGURE 4. x') = f (x + x'/2) f *(x + x'/2) (1) (2) x→ ξ'} F{ x' → 1 ξ} x'} F{ F– {ξ → F –1 {ξ' →x Wf (x. We are restricting our attention to nonabsorbing systems corresponding to the condition det M = a d − bc = 1. The application of a canonical transformation on the signal produces a distortion on the corresponding WDF according to the general law Wg (x. FrFT. cx + d ) = W f (x . (1) WDF and inverse transform. F. Testorf/1597980/Ch04 114 Chapter Four rf (x. AF transform and Radon transformation. ) p represents polar coordinates in phase space.
29) with f being the focal length of the lens. Coherent propagation through a (cylindrical) thin lens. c tan − d x = x sin a sin − b cos (4. it is straightforward to obtain 115 ⎧ +∞ x ⎪ ⎪ W x. ) = ⎪ −∞ ⎪ ⎪ ⎪ +∞ ⎪ ⎪ ⎪ ⎩ Wg (x.24) and its FT are then obtained. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. sin − ⎪ ⎪ −∞ g ⎪ ⎪ ⎪ +∞ ⎨ Wg (x . cx = 0. ) d RWg (x . ) ∝ RW f x . respectively. Thus. according to Eqs. tan =− f tan . the modulus squared of the abcd transform in Eq. 1.May 22. (4.6) and (4. ) (4. the RWT for the transformed amplitude light distribution is given in this case by RWg (x . (4. with wavelength . =0 = 2 2 = ⎧ ⎪ +∞ x x ⎪ ⎪ W a x + b sin − tan ⎪ ⎪ −∞ f ⎪ ⎪ ⎪ ⎪ +d x − x ⎪ dx ⎪ sin tan ⎪ ⎨ +∞ for for for = 0. x = .7). tan − f (4. In this case the associated M matrix for the transformation of the light ﬁeld is given by ML = 1 1 f 0 1 (4. cx + d ) d ⎪ ⎪ ⎪ −∞ ⎪ ⎪ ⎪ ⎪ +∞ ⎪ ⎪ ⎪ ⎪ W f (a x + bx . Testorf/1597980/Ch04 The RadonWigner Transform By applying the deﬁnition in Eq.28) Let us consider in the following examples a spatially coherent light distribution f (x).30) x sin sin .27) where the mapped coordinates for the original RWT are given by tan =− a tan − b . x ) d x −∞ x tan dx for for for . =0 = 2 ⎫ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎬ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎭ ⎪ W f (a x + b . (4.12). that travels along a system that imposes a transformation in the input characterized by an abcd transform. cx + d x ) d x ⎩ −∞ 2 ∝ RW f (x . Special attention is usually paid to the cases = 0. /2 since.
32) Note that the effect of this propagation through a thin lens of focal length f is also physically equivalent to the illumination of the incident light distribution by a spherical wavefront whose focus is located at a distance = f from the input plane. 2 ∝ RW f 2 (4. (4. namely. . the same results discussed here can be applied straightforwardly to that case. one obtains tan = − z.37) .33) the transformation matrix M is given by MF = 1 − z 0 1 (4.36) tan = tan − z. 0) while for the value RWg x /2 . ).May 22. x sin x = sin + z cos For the projection with = 0.31) RWg (x0 . Freespace (Fresnel) propagation. (4. therefore. 2. Thus. If we consider now the Fresnel approximation for the propagation of a transverse coherent light distribution f (x) by a distance z. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.34) and. 0) = g(x0 )2 ∝ RW f (x . the transformed RWT can be calculated through the expression RWg (x . Testorf/1597980/Ch04 116 Chapter Four A careful calculation for the case of = 0 leads to (4. ). 0) = g(x0 )2 ∝ RW f (x0 . +∞ g (x) = −∞ f (x ) exp i z (x − x) 2 d x (4. ) ∝ RW f (x . = /2 the following result x /2 . x0 x = sin z and for the orthogonal projection is achieved RWg x /2 . tan =− f (4.35) RWg (x0 . = /2 the following result is obtained /2 sin 2 ∝ RW f x .
If a uniform scale factor m is applied to the input function.40) RWg x /2 . The ﬁrst one described in the literature.19). 2 2 4. omitting the passage through its WDF. the onedimensional input . For onedimensional signals. and it is known as the RadonWigner display (RWD). the RWT also enables an optical implementation that is desirable for applications in the analysis and processing of optical signals.1 has found several applications in optics as we will see later in this chapter. The correct ﬁeld identiﬁcation requires a large number of Wigner distribution projections. Magniﬁer.4 It is based on Lohmann’s bulk optics systems for obtaining the FrFT. In fact. was proposed by Mendlovic et al. This representation.0 m (4. 0) = g(x0 )2 ∝ RW f . ) ∝ RW f (x . ). expressed mathematically by Eq. Testorf/1597980/Ch04 The RadonWigner Transform 3. proposed by Wood and Barry for its application to the detection and classiﬁcation of linear FM components.5 In this method. The relationship between the RWT and the FrFT. suggests that the optical computation of the RWT is possible directly from the input function. the RWT is transformed according to the law RWg (x .39) The vertical and horizontal projections are given here simply by the following formulas. and most have been the basis for designing other systems to obtain the RWD.2 Optical Implementation of the RWT: The RadonWigner Display Like any other phasespace function. m2 0 (4. which raises the necessity to design ﬂexible optical setups to obtain them. the RWT for a given projection angle is simply the intensity registered at the output plane of a given FrFT transformer. the associated M matrix is given by Mm = 1 m 117 0 m 1 tan . x0 RWg (x0 . designed to obtain the RWD of onedimensional signals. the RWT for all possible projection angles simultaneously displays a continuous representation of the FrFT of a signal as a function of the fractional Fourier order p. (4.May 22. Different and simple optical setups have been suggested to implement the FrFT. ∝ RW f mx /2 . x = mx sin sin tan = (4. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.38) In this case.2.
the values of Mp and R p are related to the system parameters s. with s being the construction scale parameter. The FrFTs are axially distributed between the input transparency plane ( p = 0) and the virtual source (S) plane ( p = 1) in which the optical FT of the input is obtained. Thus. 1] can be obtained simultaneously. the main shortcoming of the RWD chart produced by this setup is that it has a limited number of projection angles (or fractional orders). A truly continuous display. is related to the FrFT of order p of the input function F p {t (x) . apart from a quadraticphase factor and a scale factor.. Each strip is a Fresnel zone plate with a different focal length that is selected for obtaining the different fractional order p. . each one representing a different channel that performs an FrFT with a different order over the input signal. Therefore. the Fresnel diffraction ﬁeld U(x. For onedimensional input signals. and z through Rp = s 2 −1 tan( p /2) . Testorf/1597980/Ch04 118 Chapter Four function is converted to a twodimensional object by the use of cylindrical lenses to allow the construction of a multichannel processor that optically implements the calculations of the RWD. a complete RWD setup. R p ) = exp i x2 z(1 − Mp ) − R p z R p M2 p Fp t x Mp . instead of a spherical . which is illuminated with a spherical wavefront of radius z and wavelength . the experimental results obtained in the original paper are actually very far from the theoretical predictions. The masks consist of many strips.May 22. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Besides the very poor angular resolution. if the input function f (x) is registered in a transparency with amplitude transmittance t(x/s). The setup consists of three phase masks separated by ﬁxed distances in free space. all the FrFTs in the range [0.41) where Mp is the magniﬁcation of the optical FrFT.6 This approach is based on the relationship between the FrFT and Fresnel diffraction.7. was proposed by Granieri et al.x (4.42) Mp = (4. i.9 U(x. } as follows. For each fractional order.e.8 which establishes that every Fresnel diffraction pattern of an input object is univocally related to a scaled version of a certain FrFT of the same input. R p ) obtained at distance R p from the input. 1 + s 2 (z ) −1 tan( p /2) 1 + tan( p /2) tan( p /4) 1 + s 2 (z ) −1 tan( p /2) (4. then the FrFT of the input can be optically obtained by freespace propagation of a spherical wavefront impinging on it. Actually.43) These last equations allow us to recognize that by illumination of an input transparency with a spherical wavefront converging to an axial point S.
this focal length should provide the exact magniﬁcation at each output channel. The setup of Fig.3). 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.44) On the other hand. Testorf/1597980/Ch04 The RadonWigner Transform Cylindrical lens 1D input y0 FrFT of order p y x0 Focal line x 119 Rp z S FIGURE 4. To do this. (see Fig.19). The magniﬁcation given by the system . Since in this case the different axially located FrFTs present no variations along the vertical coordinate. we see the next step is to obtain the RWD from this setup.4).4 that it is necessary to design a lens with a focal length that varies with p (vertical coordinate y) according to f ( p) = a ap a l + (1 + lz−1 )a s 2 −1 tan( p /2) = a + ap a − l − (a + l + z)z−1 s 2 −1 tan( p /2) (4. Keeping in mind Eq. provided that the focal length of the lens L varies with the y coordinate in the same way as it varies with p. To obtain this result. wavefront. the focal length of this lens should be different for each fractional order p. 4.4 takes advantage of the onedimensional nature of the input. For each value of p (vertical coordinate y) we want to image a different object plane at a distance a p from the lens onto a ﬁxed output plane located at a from the lens. Therefore. the problem can be addressed as follows. it is straightforward to deduce from the Gaussian lens equation and from the distances in Fig. In this way. we can use a cylindrical one to illuminate the input (see Fig. and it behaves as a multichannel parallel FrFT transformer. 4. we can select a different onedimensional horizontal slice of each one and use it as a single and independent fractionalorder channel. 4. we have to ﬁnd an optical element to form the image of the axially distributed FrFT channels. (4. 4. at the same output plane simultaneously.May 22.3 Implementation of the FrFT by freespace propagation.
consequently. in such a way that they minimize the difference between these functions in the interval p ∈ [0.43). However. namely.47) . and a . This optimization6 leads to the following constraint conditions a =l 1 + 2 4 . l. z= −ls 2 l + s2 (4.45) and (4.46). we are unable to obtain an exact solution for all fractional orders. an approximate solution can be obtained by choosing the parameters of the system. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. and. the lens L should counterbalance the magniﬁcation of the FRT located at R p to restore its proper magniﬁcation at the output plane. for the porder slice of the RWT of the input function to be achieved.46) Comparing Eqs.45) However. Therefore. (4. Testorf/1597980/Ch04 120 Chapter Four Cylindrical lens 1D input y0 FrFT of order p y y Varifocal lens (L) y x0 y(R0) x Output plane y(R0) x y(R0) x Rp z ap l a' FIGURE 4. we note that the functional dependence of both equations on p is different. .4 RWD setup (multichannel continuous FrFT transformer).May 22. by using Eq. 1]. (4. z. One way to ﬁnd the optimum values for these parameters is by a leastsquares method. the magniﬁcation provided by L should be ML ( p) = −1 1 + s 2 (z ) −1 tan( p /2) =− Mp 1 + tan( p /2) tan( p /4) (4. s. for each fractional order p is ML ( p) = −a = ap a s 2 −1 tan( p /2) 1+s 2 (z ) −1 tan( p /2) −l (4.
4 0.40 0.4 with the abovementioned values for the parameters z.75 D spherical power and +3 D of addition was used in the setup of Fig.0 2.2 0.10 4. Testorf/1597980/Ch04 The RadonWigner Transform 0.0 0.4 0.00 1.0 3. (4.May 22. except for values close to p = 1. 4. l. The variation of the focal distance of the lens L with p according to Eq. (4. are represented in Fig. .8 Fractional order: p 1. the optical power is nearly linear with p.00 5.5 for the following values: z = 426 mm. a progressive lens of +2. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.0 3.8 1.0 0. l = 646 mm. 4.30 0.00 Power (D) 0.00 0.6 RWD of a Ronchi grating of 3 lines/mm: (a) exact numerical simulation. Figure 4. In the experimental veriﬁcation of the system.0 x coordinate (mm) 4.00 0.00 2.0 5. and a = 831 mm.0 0.80 1. under the constraints given by Eqs.00 0.2 0. and a .5 Focal length (solid curve) and optical power (dotted curve) of the designed varifocal lens L for the values z = 426 mm.00 p FIGURE 4.50 Focal length (m) 121 6.20 0. For this particular combination of parameters.00 0. This linearity is also accomplished by some designs of ophthalmic progressive addition lenses in which there is a continuous linear transition between two optical powers that correspond to the near portion and distance portion.0 Fractional order: p (a) (b) FIGURE 4.0 1.60 0.0 0.6 illustrates a comparison between the numerical 5.0 4.00 3.0 0. l = 646 mm.6 0.47). (b) experimental result.0 0.0 2.6 0. and a = 831 mm.40 0.20 0.0 1.44) and its optical power.
3. for a much more efﬁcient analysis of the irradiance in the image space of imaging systems. provided that the tilt angles are measured very precisely. both positive and negative. 4.6 the values of p that correspond to the selfimages. Note that the function P(x) includes any arbitrary amplitude variation p(x) and any phase aberration that the imaging system may suffer from.18 elliptical. This drawback was partially overcome in other conﬁgurations derived by the same authors using the abcd matrix formalism. can be clearly identiﬁed. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. the free propagation distances are designed to be ﬁxed or to vary linearly with the transverse coordinate. characterized by an exit pupil function with generalized amplitude transmittance P(x).3 Analysis of Optical Signals and Systems by Means of the RWT 4.19 and even fractal pupils. This technique has been successfully applied to the analysis of different optical systems with circular17 as well as square. In this section we present a useful technique.May 22.21 Let us consider a general imaging system.1 Computation of Irradiance Distribution along Different Paths in Image Space Determination of the irradiance at a given point in the image space of an imaging system is a classic problem in optics.20 The method has also been applied to the study of multifacet imaging devices. In searching for an RWD with an exact scale factor for all the fractional orders. as we show later in this chapter. The conventional techniques carry out a ﬁnite partition of the pupil of the system to sum all these contributions at the observation point. based on the use of the RWT of a mapped version of the pupil of the system. The optical setup designed for the experimental implementation of the RWD was successfully adapted to several applications. The result is an exact. . or if the aberration state of the system changes. The distance from this pupil to the Gaussian imaging plane is denoted by f . resulting in a more realistic conﬁguration.11 so the input plane and/or the output plane should be tilted instead of bent. in Fig. Testorf/1597980/Ch04 122 Chapter Four simulations and the experimental results obtained using a Ronchi grating as input object. 4. setup to obtain the RWD. Interestingly. There. this approach also inspired another proposal10 in which a bent structure for the detector was suggested.3.1 Analysis of Diffraction Phenomena 4. but unfortunately impractical.19 triangular.1.12–16 This timeconsuming procedure needs to be completely repeated for each observation point.
49) where x and y are Cartesian coordinates and a stands for the maximum radial extent of the pupil. Testorf/1597980/Ch04 The RadonWigner Transform Exit pupil Focal plane y x a arN Observation point φ 123 ΣP f z FIGURE 4. .48) where is the ﬁeld wavelength.48). We now describe the monochromatic scalar light ﬁeld at any point of the image space of the system in the Fresnel approximation21 . 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. The origin for the axial distances is ﬁxed at the axial Gaussian point. ) exp 2 i2 W20 (z) r N 2 −i2 r r N cos( ( f + z) N − ) r N dr N d (4.7 The imaging system under study. 4. By using these explicit coordinates in Eq. It is straightforward to show that. It is convenient to express all transverse coordinates in normalized polar form. z) 1 = 2 ( f + z) 2 × exp 2 1 p (r N . as shown in Fig. namely. y = ar N sin (4.50) . we obtain ¯ I (r N . the ﬁeld irradiance is given by 1 I (x. respectively. x = ar N cos .7. within this approach.May 22. z) = 2 ( f + z) 2 2 × P P(x ) exp −i zx 2 exp f ( f + z) −i2 x·x ( f + z) d2x (4. ¯ 0 0 ) exp i2 W(r N . and P represents the pupil plane surface. (4. x and z stand for the transverse and axial coordinates of the observation point.
) includes the amplitude variations on the pupil plane and the aberration effects except for SA. 0 0 ) exp i2 W40 r N 4 exp 2 i2 W20 (z) r N 2 −i2 r r N cos( ( f + z) N − ) r N dr N d (4.53) Thus Q(r N . ¯ By substituting Eq.53) into Eq. ) = W40 r N 4 (4.54) Let us now consider explicitly the angular integration in this equation. ) = Q(r N . . 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. whose dependence on the pupil coordinates is given by W40 (r N . W20 (z) = − za 2 2 f ( f + z) (4. ). The classic defocus coefﬁcient has also been introduced in this equation. 2 Q(r N . namely. Q(r N .52) where W40 is the SA coefﬁcient design constant. we will consider this term explicitly. splitting the generalized pupil of the system as follows: p (r N .50). namely. ¯ ) exp i2 W(r N . z) = 0 Q(r N . Note that if no aberrations different from SA are present in the system. In the following reasoning. ) reduces simply to the pupil mask p (r N . (4. ) exp −i2 r r N cos( ( f + z) N − ) d (4. Testorf/1597980/Ch04 124 Chapter Four where the bar denotes the polar coordinate expression for the corresponding function and where we have split out the generalized pupil P(x) to explicitly show the dependence on the amplitude pupil variations p(x) and the aberration function W(r N . z) 1 = 2 ( f + z) 2 × exp 2 1 Q(r N . .51) In many practical situations the most important contribution to the aberration function is the primary spherical aberration (SA). ) of the system. ) exp i2 W40 r N 4 (4. r N . (4.May 22. we ﬁnally obtain ¯ I (r N .55) .
z) (4.May 22.5 −0. z) = 0. .5 0. 2 u=s−s (4. r N . . z) = 0.59) which by using the change of variables t= s+s . r N . z)q ∗ (s .5 q (s.57) 2( 1 f + z) 2 2 × −0.5 Q(r N . r N . 2 we ﬁnally obtain ¯ I (r N .5 q (s. . z) = 1 125 2( 1 f + z) 2 2 × 0 Q(r N . .58) Note that in this expression all the dependence on the observation coordinates is concentrated in the mapped pupil q (s. z) exp i2 W40 r N 4 exp i2 W20 (z) r N 2 r N dr N (4. . . we ﬁnd ¯ I (r N .5 2( 1 f + z) 2 i2 W40 (s 2 − s 2 ) × −0. .60) . z) exp i2 [W40 + W20 (z) ] (s − s ) × exp ds ds (4. . If we expand the modulus square in this equation. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. . r N . z) = q (s.56) By using the mapping transformation 1 rN2 = s + . r N . . r N . Testorf/1597980/Ch04 The RadonWigner Transform Thus we arrive at a compact form for the irradiance at a point in the image space ¯ I (r N . r N . z) exp i2 W40 s 2 exp i2 [W40 + W20 (z)] s ds (4. z) and the defocus coefﬁcient W20 (z).
(4. .5 2( 1 f + z) 2 u u .5 t. Thus.64) FIGURE 4.62) This expression relates the irradiance at any observation point to the line integral of the function Wq (x. as stated in Eq. . r N . z 2 2 (4. 4.1).10)] tan = − 2W40 (4. (4.May 22. Testorf/1597980/Ch04 126 Chapter Four can be rewritten as ¯ I (r N .5 q t+ i2 × exp [W40 + W20 (z) + 2 W40 ] u dt du The above integration over the variable u can be clearly identiﬁed as the WDF of q (s.61) × −1 −0. z) = 1 0. −2 W40 t− W40 + W20 (z) dt (4. z) with respect to the ﬁrst variable. rN.5 Wq −0. z q ∗ t − . One can identify this integration as a projection of the WDF at an angle given by [see Eq.8.8 Integration line in phase space. rN. z) = 1 2 ( f + z) 2 0. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. ) along a straight line in phase space described by the equation = −2 W40 x− W40 + W20 (z) (4. . .63) as depicted in Fig. . . it is straightforward to show that ¯ I (r N .
(z). ∈R (4.69) Note that the dependence on the position parameter z in Eq. z). (4. into Eq. z) related to the pupil of the system. Thus. r N (z).68) ). z) = n=−∞ in Jn −2 r r N (z) Qn (r N ) ( f + z) N (4. r N . ) exp −in d (4. By substituting Jacobi’s identity +∞ exp(i cos ) = n=−∞ i n J n ( ) exp (−in ) . . that this function depends in general on the particular coordinates r N .66) The main conclusion of all this is that it is possible to obtain the irradiance at any point in image space through the values of the RWT of a given function q (s. let us express these lines in parametric form r N (z). (z) = o (4.68) is established exclusively in the argument of the Bessel functions— through r N (z)—and the phase exponentials—through (z). Testorf/1597980/Ch04 The RadonWigner Transform and at an oriented distance from the origin x (z) = − W40 + W20 (z) 2 4W40 + 2 127 (4.55). ) f + z) 2 (4. . . z) = 2( 1 RWq (x (z). Note.67) where J n (x) represents the Bessel function of the ﬁrst kind and order n. .65) in such a way that it is possible to express ¯ I (r N . This major drawback can be overcome for special sets of points or trajectories in image space that share the same associated mapped pupil q (s. × exp [in (z)] where Qn (r N ) stands for the norder circular harmonic of Qn (r N . a different function RWq (x . ) has to be considered for different points in image space. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. it is straightforward to obtain +∞ Q(r N . (z). however. (4. Thus. To describe such trajectories in image space. r N . and z of the observation point. the only way to strictly cancel this dependence is to consider the trajectories r N (z) = K ( f + z).May 22.70) . 2 Qn (r N ) = 0 Q(r N . that is.
the proper choice of these straight paths allows one to obtain any desired partial feature of the whole threedimensional image irradiance distribution. r N (z).65). Note also that since W40 is just a parameter in these coordinates . Testorf/1597980/Ch04 128 Chapter Four Exit pupil Focal plane y x α Observation point φo a (z + f ) tanα ΣP f z FIGURE 4. These curves correspond to straight lines passing through the axial point at the plane of the exit pupil.9.May 22. ) is independent of the propagation parameter z.71) For these subsets of observation points. z) = Q . This is a very interesting issue since the calculation of the irradiance at any observation point lying on the considered line can be achieved from this single twodimensional display by simply determining the particular coordinates (x (z). each line deﬁnes a plane that forms an angle o with the x axis. o (s) (4.64) and (4. ) through Eqs. o (r N ) = q .9 Trajectories in image space. z) = Q = n=−∞ . 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. 4. o (x . o (r N ) in Jn −2 a tan r N Qn (r N ) exp (in o) +∞ (4.72) and analogously 1 rN2 = s + . Furthermore.73) in such a way that now the corresponding RWT RWq . r N (z). as depicted in Fig. Together with the optical axis. the mapped pupil of the system can be expressed as Q(r N . (z). 2 q (s. (z). (4. Note that the angle of any of these lines with the optical axis is given by tan = Ka (4.
). This fact represents an additional advantage when a polychromatic assessment of the imaging system is needed. . and thus the associated RWT.0 (s) = Q0. 0. 4. pI( x ) → y pII( x ) → y a x 2a/3 a x (a) (b) FIGURE 4. This distribution can be obtained here as a particular case with = 0. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.0 (x (z). Thus. I (0.May 22. To prove the performance of this computation method. ). ) reduces to the normalized pupil function p (r N . we consider only purely absorbing pupils and no aberrations apart from SA in both cases. labeled system I and system II. this single display can be used for the determination of the irradiance for different amounts of SA. ) f + z) 2 (4. z) = where q 0.75) 2( 129 1 RWq 0. is also independent of the wavelength .10. Thus.74) This result is especially interesting since this mapped pupil. Testorf/1597980/Ch04 The RadonWigner Transform and does not affect the function RWq . A grayscale representation for the pure absorbing masks ¯ considered for each system is shown in Fig. Some quantitative estimation of these improvements is presented in Ref. Q(r N . next we present the result of the computation of the irradiance distribution along different lines in image space of two imaging systems. the reduction in computation time is evident.10 Grayscale picture of the pupil functions for (a) system I and (b) system II. compared to classic techniques.0 (r N ) = Q0 (r N ) (4. as will be shown in forthcoming sections. 19. o (x . For the sake of simplicity. namely. The axial irradiance distribution is often used as a ﬁgure of merit for the performance of optical systems with aberrations.
) 6 4 2 0 –4.0 FIGURE 4.0 –2.2 nm. o (x .0 W40 = 0 α = 0.11 Irradiance values provided by system I.012° W40 = λ/2 α = 0° 5 5 0 8 Irradiance (a.13 by partitioning the exit pupil of the imaging system into 1024×1024 radialazimuthal elements. The function RWq .0 2.024° –2.11 shows a joint representation of . and for comparison purposes.012° 10 10 W40 = 0 α = 0° 1000 750 500 250 0 15 W40 = λ/2 α = 0.0 4. ) was computed for 4096 × 4096 points. Figure 4.May 22. We compute the irradiance values for 256 points along three different lines passing through the axial point of the pupil. Continuous lines represent the result by the proposed RWT method while dotted lines stand for the computation by the classic method. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.u. These trajectories are chosen with tilt angles = 0.0 –4. the same irradiance values were computed by using the classic method12.) 750 500 250 0 15 Irradiance (a.) W40 = 0 α = 0. and = 638.8 mm.0 0.u. Testorf/1597980/Ch04 130 Chapter Four 1000 Irradiance (a. along different lines containing the axial point of the pupil and for two different amounts of SA.024◦ .u. z = 15.012◦ and 0◦ (optical axis). all characterized by an azimuthal angle o = /2.0 Defocus coefficient: W20 (μm) W40 = λ/2 α = 0.024° 0 8 6 4 2 0 4. We set a = 10 mm.0 2. 0.0 0.
In fact. both results differ by less than 0.0 Defocus coefficient: W20 (μm) W40 = λ/2 α = 0.0 4.0 4.u. the greater the number of observation points.0 –4. the numerical calculation for system I.0 0.12 Irradiance values provided by system II.0 –2.012° 131 W40 = λ/2 α = 0° 5 5 0 8 Irradiance (a.024° 0 8 6 4 2 0 2. However. This is so because the basic RWT does not require recalculation for any point in each of the curves.u. that .0 FIGURE 4. the RWT is much more efﬁcient in this computation process.024° –2.u. Obviously.) 6 4 2 0 –4.12.012° 10 10 W40 = 0 α = 0° 1000 750 500 250 0 15 W40 = λ/2 α = 0. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.) W40 = 0 α = 0.0 W40 = 0 α = 0.0 2.11. 4. The analysis of these pictures shows that the results obtained with the RWT method match closely those obtained with the classic technique. when two different values of the SA are considered. Testorf/1597980/Ch04 The RadonWigner Transform 1000 Irradiance (a. or SA values. as in Fig.) 750 500 250 0 15 Irradiance (a. 4.0 0. The same results applied now to system II are presented in Fig.03 percent.May 22. This is also true for any amount of SA.
5. e. The evolution of these diffraction patterns under propagation is often represented in a twodimensional display of gray levels in which one axis represents the transverse coordinate— pattern proﬁle—and the other axis is related to the axial coordinate— evolution parameter. Figure 4.3. To compare the theoretical and experimental results. or.76) where s is the axial coordinate at which the corresponding diffraction pattern is localized under parallel illumination and h is the extent of the socalled progression zone of the varifocal lens. ys = hs s + l 2( h − 0) .23 This kind of representation illustrates.22 4. Ms = 1 + l 2( s h − 0) (4. It can be seen that the RWD is a nice representation of the evolution by propagation of the interference phenomena.41). we want to point out that this approach can also be applied to other trajectories of interest in image space. calculations of Fresnel and Fraunhofer diffraction patterns of uniformly illuminated onedimensional apertures are standard topics in undergraduate optics courses.1. and that the Fraunhofer diffraction simply is a limiting case of Fresnel diffraction. they are computed numerically. how the geometrical shadow of the object transforms into the Fraunhofer diffraction pattern as it propagates.g. i. These include the precise location ys and the relative magniﬁcation Ms of each diffraction pattern. it can provide a quantitative measurement of a variety of terms. more frequently.May 22. a cross section of . This means that the RWD is itself a continuous display of the evolution of diffraction patterns of onedimensional objects. In fact. These two terms are quantitatively deﬁned in terms of the maximum h and minimum 0 powers of the varifocal lens L of the system represented in Fig. 4.2 we mentioned that the mathematical relationship between Fresnel diffraction and the FrFT is given by Eq. (4. 4. the Fraunhofer region of the diffracted ﬁeld clearly shows the characteristic Young fringes modulated by a sinc function. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. In fact. These theoretical predictions are calculated analytically for some typical apertures. the greater the resultant savings in computation time. and this property is extremely useful from a pedagogical point of view. As a ﬁnal remark on this subject. For instance.24 In addition to the qualitative physical insight that the RWD provides about diffraction.2.. short paths parallel to the optical axis in the neighborhood of the focal plane17 or straight lines crossing the focal point can be considered.. Testorf/1597980/Ch04 132 Chapter Four have to be considered.13 illustrates the experimental results registered by a CCD camera using a double slit as an input object.e.2 Parallel Optical Display of Diffraction Patterns In Sec.
Fresnel diffraction of a Cantor set26 shows an irradiance distribution along the optical axis having a periodicity that depends on the level of the set. the properties of diffraction patterns produced by fractal objects and their potential applications have attracted the attention of several researchers during recent years because many natural phenomena and physical structures. Here. 4. such as phase transition. the RWD has been used to investigate diffraction by a variety of different interesting structures including fractal diffraction screens. Self imaging phenomena are discussed in more detail in Chapter 9 by Markus Testorf. turbulence. the theoretical sinc envelope of the Young fringes is also shown by the dotted line. As illustrated earlier in Fig. 4. can be analyzed and described by assuming fractal symmetry. Testorf/1597980/Ch04 The RadonWigner Transform 133 FIGURE 4. such as the Talbot effect. the experimental RWD for values of y/ h close to 1 is also represented in Fig. Most research has been devoted to the study of diffraction patterns obtained from fractal objects in the Fraunhofer region. For comparison purposes. which shows the diffraction patterns of a Ronchi grating. several selfimaging planes can be identiﬁed. (a) Experimental result.25 yet it is in the Fresnel region where interesting features appear.6. Another classic example is diffraction by periodic objects.13. the number of Talbot images is limited by the socalled walkoff effect. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. In addition to its use as an educational tool for displaying diffraction patterns. or optical textures. Furthermore.13 RDW showing the evolution of the ﬁeld diffracted by a double slit. selfimaging phenomena. It can be clearly seen that.May 22. In fact. due to the ﬁnite extent of the grating at the input. the intensity distributions at transverse planes . (b) Cross section of the RWD showing the intensity proﬁle near the Fraunhofer region. For instance. are interesting and stimulating and usually attract the students’ attention.
show a partial selfsimilar behavior that is increased when moving toward the Fraunhofer region. Therefore.28 The RWD obtained for a triadic Cantor grating developed up to level 3 is shown in Fig. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. especially if such a display can be obtained experimentally.14 RWD as a display of all diffraction patterns generated by a Cantor grating of level 3. where the diffraction patterns change rapidly. Since the WDF of a complex ﬁeld distribution contains all the necessary information to retrieve the ﬁeld itself. 4. one of the most important applications of the RWD has been in this ﬁeld.30 many of the methods to obtain the WDF (and the AF) .4) enables the RWD representation to provide an optimum sampling of the diffracted ﬁeld. the evolution of the diffraction patterns allows one to determine the main characteristic parameters of the fractal. the mapping of the propagation distance provides a ﬁne sampling.3.14. In fact. it is useful to represent the evolution of the complex amplitude of onedimensional fractals propagating through free space represented on a twodimensional display. 4.29. Moreover.27 The magniﬁcation provided by the lens L in the experimental setup (see Fig. We note that sampling is the subject of Chapter 10. and the other is a function of the axial coordinate.May 22. Near the object. this result can be favorably compared with the results obtained with other displays. For this reason. Testorf/1597980/Ch04 134 Chapter Four FIGURE 4.2 Inverting RWT: PhaseSpace Tomographic Reconstruction of Optical Fields The WDF is an elegant and graphical way to describe the propagation of optical ﬁelds through linear systems. whereas the sampling is coarse in the far ﬁeld where the variation of the diffraction patterns with the axial distance is slow. 4. 27. according to the analysis carried out in Ref. In this case one axis represents the transversal coordinate.
sin )} (4. The second part of the proposed method is the digital computation of the inverse Radon transforms of the experimental RadonWigner spectrum. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. a tomographic reconstruction is needed to synthesize the WDF. 4. another interesting application of the setup discussed in Sec. obtaining the WDF of wave ﬁelds is also possible indirectly through other representations such as the Radon transform.15 Diagram of the proposed hybrid optodigital method. This algorithm is based on the central slice theorem discussed in Sec. Testorf/1597980/Ch04 The RadonWigner Transform RWT = ⎜FrFT ⎜2 (Optical) Inverse RT (digital) 135 f (x) 1D signal RWf (x.31.33 In this particular case. from Eqs. With most experimental setups for phase retrieval. (4. ). Thus.e. In this way the potential advantage of optics. } = F2D {W f (x. ξ) WDF FIGURE 4. This optical method beneﬁts from having no moving parts. On the other hand.21) and (4. i. in general. With this information it is possible to recover the amplitude and the phase of the original ﬁeld distribution solely by means of intensity measurements. interferometric or iterative methods are necessary in general to avoid loss of information.22) we have F{RW f (x . negative values and therefore is obtained from an intensity detector with an uncertainty in its sign. This is true even for the optically obtained WDF. (x cos . is wasted.32 However. 4.29. the experimental RadonWigner spectrum of the input function is obtained from a twodimensional singleshot intensity measurement by the use of the experimental setup in Fig.77) . Consequently. 4. The technique to obtain the WDF from projections is divided into two basic stages. sketched in Fig. θ) Radon Wigner spectrum Wf (x.30 these measurements have to be taken sequentially in time while varying the distances between some components in each measurement.May 22. which is real but has.4. Optical or optoelectronic devices are the most commonly employed systems to obtain a representation of phasespace functions of onedimensional or twodimensional complex signals. parallel processing of signal information.2 to obtain the RWD is the experimental recovery of the WDF by means of an inverse Radon transformation..1. In the ﬁrst stage.2. The most common algorithms used in tomographic reconstruction are based on the technique known as ﬁltered backprojection. ). could be adapted to solve the phase retrieval problem. 4.2. because most detectors used to this end are only sensitive to the incident intensity.15.
78) with +∞ C f (u.16 Experimental RadonWigner spectrum of (a) a single slit of 2. The inversion of this last transformation allows the recovery of W f (x. Although we only obtain optically the RWT for ∈ [0.2 for two different functions. (4. /2].79) Equation (4. In Fig.17) has been used to complete the spectrum. Explicitly34 W f (x. W f (x. ∈ [0. 4. The experimental RWD of different onedimensional functions has been used to reconstruct the WDF from projections. ). where the onedimensional FT is performed on the ﬁrst argument of RW f (x . To undertake the reconstruction of the WDF through the ﬁltered backprojection algorithm.2 mm and (b) a binary grating with a linearly increasing spatial frequency. In this way.78). ) is reconstructed for each phasespace point as the superposition of all the projections passing through this point. namely. it is necessary to consider the complete angular region of the RWD. ) = −∞ F{RW f (x . from Eq. ). }   exp (i2 u) d (4. ) = 0 C f (x cos + sin . ) d (4. that is.2. From the experimental RWD . ) from its projections. Testorf/1597980/Ch04 136 Chapter Four 3 Spatial coordinate (mm) Spatial coordinate (mm) 30 45 60 75 Projection angle (°) 90 2 1 0 –1 –2 –3 0 15 3 2 1 0 –1 –2 –3 0 15 30 45 60 75 Projection angle (°) 90 (a) (b) FIGURE 4.16 we show the RWD obtained with the optical device described in Sec. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.May 22. a rectangular aperture (single slit) and a grating with a linearly increasing spatial frequency (chirp signal). (4. ). 4.79) can be clearly identiﬁed as a ﬁltered version of the original RWT. the symmetry property in Eq.
Testorf/1597980/Ch04 The RadonWigner Transform 137 4 Spatial frequency (a. in Fig. For comparison purposes. (b) Experimental tomographic reconstruction of the WDF of the same grating. the corresponding WDFs have been obtained using the ﬁltered backprojection algorithm.17 (a) Theoretical WDF of a single slit.May 22.18 show both the theoretical and the experimentally reconstructed WDF of the single slit and the chirp grating.u. Note that in Figs. The lines radiating from the center and outward are typical artifacts (ringing effect) associated with the ﬁltered backprojection method.) 2 0 –2 –4 –2 –1 0 1 2 Spatial coordinate (mm) Spatial frequency (a.u. 4. respectively. . Figs.18 (a) Theoretical WDF of a binary grating with a linearly increasing spatial frequency. 4.u. (b) Experimental result for the tomographic reconstruction of the WDF of the same slit.) 4 2 0 –2 –4 –2 –1 0 1 2 Spatial coordinate (mm) (a) (b) FIGURE 4. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.) 4 2 0 –2 –4 –2 –1 0 1 2 Spatial coordinate (mm) (a) (b) FIGURE 4.16.u.18b some artifacts appear. 4.) 2 0 –2 –4 –2 0 –1 1 2 Spatial coordinate (mm) Spatial frequency (a.17b and 4.35 In spite of this 4 Spatial frequency (a.17 and 4.
Moreover. Testorf/1597980/Ch04 138 Chapter Four effect.May 22. The slope in the arrowhead that characterizes the chirp rate of the signal is the same for the theoretical and the experimental results. Marechal’s treatment of tolerance.3 Merit Functions of Imaging Systems in Terms of the RWT 4. it is expected that the imaging properties of a given optical system.3. In Fig.38 the analysis of Marechal can be reformulated to give a tolerance criterion based on the behavior of the optical transfer function (OTF) (spatial frequency information) instead of the PSF (space information). or axial pointspread function (PSF). since the information content stored in the FrFT of an input signal changes from purely spatial to purely spectral as p varies from p = 0 to p = 1.39–41 This point of view equally emphasizes both the spatial and the spectral information contents of the diffracted wave ﬁelds that propagate in the optical imaging systems. in .1 Axial PointSpread Function (PSF) and Optical Transfer Function (OTF) There are several criteria for analyzing the performance of an optical imaging system for aberrations and/or focus errors in which the onaxis image intensity. since complex optical wave ﬁelds can be reconstructed from the WDF provided the inversion formulas. a similar implementation proposed here for the WDF can be easily derived for the AF.16a. in the domain of the RWT. On one hand. this approach can be used as a phase retrieval method that is an alternative to the conventional interferometric or iterativealgorithmbased techniques. 4. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. (4. and this redundancy should avoid any previous hypothesis about the input signal.3. 4. a continuous set of FrFTs is available simultaneously. Among them we mention:37 Rayleigh’s criterion. 4. Phasespace functions were also employed to evaluate some merit functions and quality parameters. The asymmetry in Fig. Several extensions of the proposed method are straightforward.36 phase retrieval is possible with intensity measurements at two close FrFT domains.3. This approach.17 is a consequence of the noise in Fig. a very good qualitative agreement can be observed between the results obtained with the theoretical and experimental data. 4. reﬂecting also the asymmetry on the spatial coordinate in this ﬁgure. In our method. by virtue of Eq. As Hopkins suggested.18 the typical arrowshaped WDF of a chirp function can be observed in both cases. In fact. that is. Particularly.22). and the Strehl ratio (SR). however. as demonstrated. Note also that it is easy to extend this method to obtain twodimensional samples of the fourdimensional WDF of a complex twodimensional signal by use of a line scanning system. requires some a priori knowledge of the signal bandwidth. is the relevant quantity.
the value = 0 provides the irradiance distribution at the considered observation point. This function is the PSF of the imaging system. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.25)] RWU(x. tan − z (4. a spherical wavefront illumination (with focus at = f ) and a freespace propagation (for a distance f + z). as stated in Sec. Testorf/1597980/Ch04 The RadonWigner Transform both the space and spatial frequency domains. RWU(x. The transformation of t(x) to obtain the ﬁeld U (x. 0) = U(x0 .2. 4.2. therefore. z) = −∞ t(x ) exp −i f x 2 exp i (x − x) 2 d x ( f + z) (4. generated by an optical imaging system characterized by a onedimensional pupil function t(x). the equivalent relationships to that given by Eq. the ﬁeld amplitude distribution can be written.1. (4.In the neighborhood of the image plane. with wavelength . tan 0 0) = f ( f + z) . Thus.84) . let us consider the monochromatic wave ﬁeld.83) 0 For the optical axis (x0 = 0) the PSF can be expressed as RWU(x. according to the Fresnel scalar approximation. as +∞ 139 U (x. 0) = U(0.z) (x . the abcd matrix associated with this transform can be found to be M= 1 1 f 0 1 1 0 − ( f + z) 1 = 1 1 f − ( f + z) −z f (4. To derive the formal relationship between the PSF (and the OTF) and the RWT. ) tan = f tan − 2 f ( f + z) . could also be evaluated from the RWD. (4.80) in terms of the corresponding RWTs can be expressed as [see Eq. z)2 = I (x0 . as a function of the distance z to the image plane. z)2 = I (0.z) (0.May 22. arctan f ( f + z) z (4. namely. z) ∝ RWt 0. when a unit amplitude point source is located at the object plane. 4.1. z) is given by a twostep sequence of elementary abcd transforms.z) (x0 .82) x x = sin sin + ( f + z) cos In particular.81) and. z x0 x0= sin ( f + z) (4. ) ∝ RWt (x . z) ∝ RWt (x 0 . located at z = 0.80) where f is the distance from the pupil to the image plane. Considering the results presented in Sec.
(b) Experimental result for (a). .25 0. Testorf/1597980/Ch04 140 Chapter Four A normalized version of this axial irradiance is often used as a ﬁgure of merit of the performance of the imaging system.00 0. the defocus tolerance of different kinds of onedimensional pupils was investigated.50 p 0. namely. The general form of these pupils can be written as t(x) = rect(x/ h) − rect(x/b).May 22. with b = 0 for the uniform aperture.51).00 0.00 0.25 0.19 RWTs: (a) Computer simulation for an aperture with a = 2.75 1.3 mm. a clear aperture (slit) and a pupil with a central obscuration (double slit). in terms of its associated RWT.5 mm and b = 0 mm. Thus. (d) Experimental result for (c).75 1.85) where h is the maximum lateral extent of the onedimensional pupil and W20 stands for the onedimensional version of the defocus coefﬁcient deﬁned in Eq.5 mm and b = 1. z) ∝ RWt I (0. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. the function S(W20 ) can be analyzed in a polar fashion in the twodimensional domain of the WDF associated with the pupil function t(x) or.00 x 0. (4.50 p 0.19 shows the RWD 0 0 x 0. arctan − h2 2W20 (4.00 (a) (b) 0 0 x 0.00 0. the SR versus defocus. deﬁned as S(W20 ) = I (0. (c) Computer simulation for an aperture with a = 2.00 x 0. The horizontal axis corresponds to the parameterization of the projection angle = p /2. 0) 0. namely. equivalently.00 (c) (d) FIGURE 4.50 p 0.50 p 0.75 1.25 0.75 1. To illustrate this approach. Figure 4.25 0.
numerically (parts a and c) and experimentally (parts b and d) obtained. the main shortcoming of the SR as a method of image assessment is that although it is relatively easy to calculate for an optical design prescription.2.25 0.00 (a) (b) FIGURE 4. the slices of the RWD for x = 0 give rise to the SR for variable W20 . the quality of the image itself is better described through the associated OTF. Testorf/1597980/Ch04 The RadonWigner Transform b = 0 mm b = 1.3 mm b = 2 mm 141 1 S(z) b = 0 mm b = 1.3 mm b = 2 mm 1 S(z) 0 0 0. for two different values of the obscuration b.20 SR versus defocus for circular pupils with pupil function t(x) = rect(x/ h) − rect(x/b): (a) Computer simulation. These proﬁles are plotted in Fig.20 for three different pupils. the AF of the pupil function t(x) can be interpreted as a continuous polar display of the defocused OTFs of the system. However. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.75 Fractional order p 1. Conversely. Fortunately.87) . the projection angle = p /2. W20 = − ( f + z) 2x At x . such as the depth of focus.86) In this way. =H − x h2 .25 0. From these results. since the AF contains all the OTFs H( . (b) experimental results. Moreover. annular apertures have higher tolerance to defocus.00 0.75 Fractional order p 1. 4. W20 ) = At − ( f + z) .May 22. it is normally difﬁcult to measure for a real optical system.50 0. According to our previous discussion. it can be observed that.1. as expected. 4. 2W20 ( f + z) h2 (4. (4. W20 ) associated with the optical system with varying focus errors according to the formula42 H ( . The knowledge of the SR is useful to characterize some basic features of any optical system. Again.50 0. this information can also be obtained from the RWD via its relationship with the AF established in Sec.
In this case. as we will discuss next.00 0. This representation is quite convenient to visualize Hopkins’ criterion. a single RWD can be used to obtain the set of monochromatic OTFs necessary for its calculation.00 0.25 0.19. These results for the clear aperture are shown in Fig.21 Computed onedimensional FT of the RWDs shown in Fig.25 0. taken with respect to the x axis.00 0.00 ν 0. ).25 0.00 (a) (b) 0 0 ν 0.00 (c) (d) FIGURE 4.19. Testorf/1597980/Ch04 142 Chapter Four 0 0 ν 0. 4.50 p 0. The RWD can also be used for calculating the OTF of an optical system designed to work under polychromatic illumination. the defocused OTFs are displayed along the vertical or spatialfrequency axis.88) cos h2 .75 1. W20 = tan ( f + z) 2 Therefore. of the RWT illustrated in Fig. 4.00 0.22. } = At ( =H cos .May 22. the onedimensional FT of the proﬁle of the RWD for a given value of the fractional order = p /2 corresponds to a defocused (scaled) OTF.50 p 0. (4.25 0. .75 1.21 shows the onedimensional Fourier transforms. Thus.75 1. − − sin ) (4.50 p 0.00 ν 0.50 p 0. From the previous analysis. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.75 1.22) it is easy to ﬁnd that F {RWt (x . 4.39 Figure 4. by using Eq.
0 0. to compute the transfer function for any wavelength by means of the same RWD.0 0. It is worth noting that in this equation the wavelength of the incoming light acts as a parameter in the determination of the particular coordinates of the FT of the RWT. The calculation procedure used in the previous section can be used. but it does not affect the RWT itself.00 ξ OTF 0.75 W20/λ = 0.00 ξ FIGURE 4.2 Polychromatic OTF As stated above.43 An especially interesting application of this technique is the evaluation of the spatialfrequency behavior of optical systems working .0 p = 0.21 for the case of a uniform aperture. as a function of the defocus coefﬁcient. where it was shown that the AF of the generalized pupil function of the system is a display of all the monochromatic OTFs with longitudinal chromatic aberration.0 p = 0.0 p = 0.00 1. (4.22 OTFs obtained from different slices of the intensity distributions shown in Fig.00 1. This approach is based on previous work. for different amount of defocus.0 0.0 ξ 0. through Eq.99 0.40 0.May 22.0 p=1 W20/λ = 0 OTF 0. Thus.40 0. therefore.5 W20/λ = 0.0 0.0 p=1 W20/λ = 0 ξ 1.00 ξ 0.88). 4.0 ξ 1. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.3.99 1. 4.3. changing the value of simply resets the position inside the same twodimensional display for the computation of the OTF. Testorf/1597980/Ch04 The RadonWigner Transform 143 1.5 W20/λ = 0. the RWT associated with the onedimensional pupil of an imaging system can be used to obtain the OTF of the device.75 W20/λ = 0.0 p = 0.00 0.
W20 ) = S( )V( ) d (4. This OTF cannot account. It can be shown. This function is deﬁned as H ( .44. and. The function V( ) represents the spectral sensitivity of the irradiance detector used to record the image. by following the trichromacy of the human eye.45 H ( . for the chromatic information of the image. consequently. W20 ( ))S( )V( ) d H( . W20 ) = S( )z d (4. Note also that a new wavelengthdependent defocus coefﬁcient has been deﬁned. since only a single detector is assumed. The direct comparison of the incoming and the outgoing polychromatic irradiance distributions does not allow. a similar relationship to the monochromatic case to be established. Testorf/1597980/Ch04 144 Chapter Four under polychromatic evaluation.48 Indeed. W20 ( ) = W20 + W20 ( ) (4. namely. in fact. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. to account for the longitudinal chromatic aberration W20 ( ) that the system may suffer from. namely. W20 ) = S( )x d H ( .89) where and S( ) are the spectral range and the spectral power of the illumination.46. three different chromatic channels are usually employed to properly describe color features in irradiance distributions.90) where W20 is the defocus coefﬁcient deﬁned in the previous section.47 Under these restrictions. W20 ( ))S( )x d H( . in general. a single polychromatic OTF can be used for relating input and output polychromatic irradiances in spatialfrequency space. W20 ( ))S( ) y d H( . however.91) . however. three different polychromatic OTFs are used.45 This extension presents. the proper generalization of the OTFbased description to this broadband illumination case allows one to deﬁne quality criteria for imaging systems working with color signals. W20 ) = S( ) y d H ( .44. In fact. providing the imaging system does not suffer from any chromatic aberrations regarding magniﬁcation. some difﬁculties.May 22. that only when the input signal is spectrally uniform can the frequency contents of both signals be related through a single polychromatic OTF function. respectively. W20 ( ))S( )z d H( .
0 0. the same imaging system (i.91) establish the formulas to describe completely the response of a system from a spatialfrequency point of view. y .0 Longitudinal chromatic aberration coefficient: δW20 (μm) 1.0 –2. and B bands of the three pixel families in the detector array. To illustrate this technique. when a visual inspection of the ﬁnal image is considered. these channels can be associated with the R. these computations can be done in a much more efﬁcient way by use of this twodimensional display. On the other hand.0 400 500 600 700 Wavelength: λ (nm) System II 145 FIGURE 4. these sensitivity functions are the wellknown spectral tristimulus values of the human eye. Testorf/1597980/Ch04 The RadonWigner Transform where x . 4.0 –1. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. as stated in the previous section. We assume that no other aberrations 2. Since any of these monochromatic transfer functions can be obtained from a same single RWD.23 Longitudinal chromatic aberration coefﬁcient associated with the two different correction states of the system under study. as cited above. This is a critical issue in the saving of computation time which provides this technique with a great advantage compared to other classic techniques. obviously.0 System I –4.49 Equations (4.23. Furthermore.0 –5. These functions depend. the evaluation of the monochromatic OTFs for a sufﬁcient number of wavelengths inside the illumination spectrum has to be performed.. . and z are three spectral sensitivity functions associated with the measured chromaticity.e. To numerically compute the functions described there. as shown in Fig. the same pupil function) but suffering from different longitudinal chromatic aberration can be assessed as well. we present the result of the computation of the polychromatic OTFs associated with a conventional onedimensional clearpupil optical system (slit of width h) but suffering from two different chromatic aberration states (systems I and II from now on). In the case of a conventional digital color camera. on the speciﬁc color detector actually used. G.May 22. with no additional computation of the RWD.0 –3.
91) for uniform sampling of 36 wavelengths in the range between 400 and 700 nm. The results for system I. Testorf/1597980/Ch04 146 Chapter Four 1.25.23. Some of these results are shown in Fig.5 λ = 700 nm Monochromatic OTF: Hλ(ξ.0 λ = 400 nm λ = 500 nm λ = 600 nm 0. The x . Regarding the geometric parameters of the system. namely. are shown in Fig. while the spectral power for the illumination corresponds to the standard illuminant C49 . The computation of the polychromatic OTFs is performed next for both correction states.24. The evaluation of the corresponding monochromatic OTFs for both aberration states is achieved through the same computation method as in the previous section. and z functions are set to be the spectral tristimulus values of the standard human observer CIE 1931. corresponding to a defocused plane.24 Monochromatic OTFs for system I in Fig. corresponding to the imaging plane (W20 = 0). Note that in both cases the same RWD is used in the computation. 4.May 22. ). and for system II.3. through the sequential onedimensional FT of the twodimensional display of the RWT RWt (x . and the same applies to the onedimensional character of the imaging system. This assumption does not imply any restriction of the method. 4. as can be easily shown. we ﬁxed h/ f = 0. as stated above. (4. through the superposition of the monochromatic ones stated in Eqs. are present.δW20(λ)) 0. 4.2. at the image plane.0 0 100 200 300 400 (mm–1) 500 Spatial frequency: ξ FIGURE 4. y . 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.3. 4.3 Polychromatic Axial PSF In this section we propose the use of a single twodimensional RWD to compute the axial irradiance in image space provided by an imaging .
The defocus coefﬁcient is deﬁned in Eq.51).92) I (z)S( )x d where S( ). corresponding to a defocused plane (W20 = −0. X(W20 ) = Y(W20 ) = Z(W20 ) = I (z)S( )x d I (z)S( )x d (4.0 (s) in Eq.1. respectively. z) provided by a system with an arbitrary value of SA can be obtained from the single RWT RWq 0. 0.1.0 Polychromatic OTF HX (ξ.5 System I 0.W20 = 0) HZ (ξ. V( ). However.28 μm) HY (ξ. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. system with polychromatic illumination.3. 4. (4.28 m) and image plane. (4. the account for chromaticity information leads to a proper generalization of the monochromatic irradiances to the polychromatic case through three functions. For notation convenience we denote I (z) = I (0. 4. and z stand for the magnitudes used in the previous section. ) of the mapped pupil q 0.W20 = –0.25 Polychromatic OTFs for (a) system I and (b) system II in Fig.0 (x.W20 = 0) 147 0. it is often more useful to describe a chromatic signal through . namely. Testorf/1597980/Ch04 The RadonWigner Transform 1.0 Polychromatic OTF HX (ξ.W20 = 0) HY (ξ. of course. According to the discussion in Sec.28 μm) 1.2.28 μm) HZ (ξ. namely. to the onaxis values of the threedimensional PSF of the imaging system.W20 = –0.0 0 System II 100 200 300 400 500 Spatial frequency: ξ (mm–1) FIGURE 4. z) in this section.3.May 22.0 0 100 200 300 400 500 Spatial frequency: ξ (mm–1) 0. In fact. that the axial irradiance distribution I (0.75).5 0.23. y . 0. x . 4. When an object point source is used. this irradiance distribution corresponds.3. the proposed technique is a straightforward extension of what is stated in Sec.W20 = –0.
1. 4. This is the case when all the aberrations of the system. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. A conventional choice for these new parameters is the set x (W20 ) = X(W20 ) X(W20 ) + Y(W20 ) + Z(W20 ) Y(W20 ) y (W20 ) = X(W20 ) + Y(W20 ) + Z(W20 ) (4. If the sensitivity functions are selected to be the spectral tristimulus values of the human eye. 4. have a negligible chromatic dependence.0 (x (z).1. for different aberration correction states. the evaluation of these magnitudes requires the computation of the monochromatic components for a sufﬁcient number of spectral components. is not very efﬁcient at this stage.1. for a given spectral component. a given wavelength. According to the formulas in Sec. 4.0 (s) represents the zeroorder circular harmonic of the pupil 2 Q(r N . with s = r N + 1 . ) f + z) 2 (4. This is a very usual situation in wellcorrected systems. Note that for systems with longitudinal chromatic aberration. (4. as in the previous section.64) and (4.3. (4. apart from SA and longitudinal chromatic aberration. from a single twodimensional display associated with the pupil of the system. The normalized coordinates r N and are 2 implicitly deﬁned in Eq.49 Anyway. as stated earlier.49). ) itself does not depend on wavelength. The speciﬁc coordinates (x (z). since the computation performed for a ﬁxed axial point. and z is established through these coordinates if the function Q(r N . in this case the chromaticity coordinates provide a joint description for the hue and saturation of the colored signal. and in this case. On the other hand.65). along with the parameter Y(W20 ). ). let us consider the system presented in Fig. The use of conventional techniques. ) for the RWT are given by Eqs. and a given aberration state cannot be applied to any other conﬁguration. Note that now the whole dependence of the axial irradiance on . can be expressed as I (z) = 2( 1 RWq 0.93) known as chromaticity coordinates.3. every axial . the defocus coefﬁcient W20 is substituted for the wavelengthdependent coefﬁcient in Eq.May 22. W40 .90). (4.9 with = 0.94) where q 0. the axial irradiance distribution in image space. Testorf/1597980/Ch04 148 Chapter Four a combination of these basic functions. To describe this proposal in greater detail.1 represents a much more efﬁcient solution since all the monochromatic values of the axial irradiance can be obtained. the Y(W20 ) parameter is known as illuminance and it is associated basically with the brightness of the chromatic stimulus. The method proposed in Sec.
the polychromatic description for the axial image irradiance can be assessed by the formulas X(W20 ) = Y(W20 ) = Z(W20 ) = RWq 0.65). Without loss of generality we assume here that the SA coefﬁcient has a ﬂat behavior for the considered spectral range.0 (x (z). Once again. we evaluate the axial response of a clear circular pupil imaging system. affected by spherical and longitudinal chromatic aberrations as shown in Fig.19. we assume that no other aberrations are present.May 22. these weighted superpositions can be quickly and easily calculated. 4. once the RWD of the function q 0. .0 (x (z). axial position. Aberration coefficient (nm) 400 200 0 –200 δW20 W40 –400 400 500 600 700 Wavelength: λ (nm) FIGURE 4. ) for every wavelength. ) S( )z d 2 ( f + z) 2 (4.64) and (4.26. (4.26 Aberration coefﬁcients associated with the system under issue. ) S( ) y d 2 ( f + z) 2 RWq 0. and wavelength can be studied from the same twodimensional RWD.95) 149 where the values of (x (z).0 (s) of the system is properly computed.50.0 (x (z). Testorf/1597980/Ch04 The RadonWigner Transform position. and SA amount are given by Eqs. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. for the sake of simplicity. providing that these kinds of systems are analyzed. ) S( )x d 2 ( f + z) 2 RWq 0.51 As an example for testing this technique. Thus. SA and chromatic aberration state. Thus.
) λ = 400 nm λ = 590 nm λ = 700 nm 1. We consider 36 axial positions characterized by defocus coefﬁcient values in a uniform sequence. by assuming the same . Solid lines represent the results by analytical calculation. ).96) where z F (z) = 0 exp i t2 dt 2 (4.0 Axial irradiance: Iλ (a. namely.12 I (z) = a2 2 f ( f + z) 2 1 F W40 W20 ( ) + 2W40 W ( ) √ − F √ 20 W40 W40 2 (4. Figure 4. Finally. W40 .27 Axial irradiance values for the system under study.0 (x .97) is the complex form of Fresnel integral.May 22. We follow the same procedure as in earlier sections for the digital calculation of the RWD RWq 0. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.27 presents a comparison of these approaches for three different wavelengths in the visible spectral range. It is worth mentioning that for this pupil is possible to achieve an analytical result for the monochromatic axial behavior of the system for any value of W20 . while superimposed symbols correspond to the computation through the single RWD technique. Testorf/1597980/Ch04 150 Chapter Four 2. and . Excellent agreement can be observed in this ﬁgure.u. we performed the calculation of the axial values for the chromaticity coordinates and the illuminance. This analytical formula is used here to evaluate the results obtained by the proposed method.0 0.0 –1500 –1000 –500 0 500 1000 Defocus parameter: W20 (nm) FIGURE 4.
settings for the sensitivity functions and the illuminant as in the previous section. sketched in Fig.53 The optical system. 4.28 with the ones obtained by applying the same classic technique as in Sec.0 –1500 –1000 –500 0 500 1000 Defocus parameter: W20 (nm) 0. and an onaxis kinoform zone plate. 4.1 Optimization of Optical Systems: Achromatic Design We now present a design method for imaging systems working under polychromatic illumination on a RWT basis.3.0 0. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Again. Here solid lines represent the results obtained by the conventional method.u. works under planar broadband illumination.28 Axial illuminance and chromaticity coordinates for the system under study.0 151 Axial illuminance (a. . 19.8 0.0 0.May 22. This element acts. 4. with focal length f . The only two optical elements in this device are an achromatic lens.3.2. In particular.5 0.) 0. we ﬁx our attention on the optimal compensation of the axial chromatic dispersion of the Fresnel diffraction patterns of a plane object Although this proposal can be applied to a wide variety of systems. we concentrate on an optical system specially designed for this purpose.4 Design of Imaging Systems and Optical Signal Processing by Means of RWT 4.4. 4. The values obtained with the method presented here are compared in Fig.29. Testorf/1597980/Ch04 The RadonWigner Transform 1.4 Chromaticity coordinate x(W20) FIGURE 4. A more detailed comparison of both methods is presented in Ref.52. This device allows us to obtain the image of any arbitrary diffraction pattern with very low residual chromatic aberration.0 Chromaticity coordinate y(W20) 1. a very good agreement between them can be seen.5 0. while superimposed symbols correspond to the computation through the single RWD technique.
as a conventional thin lens with a focal length proportional to the inverse of the wavelength of the incoming light.98) o being a reference design wavelength and Zo = Z( o ). if we ﬁx our attention on the reference plane located at a distance Ro from the object. the same diffraction pattern is replicated along the optical axis between the planes characterized by distances R1 = R( 1 ) and R2 = R( 2 ).. i. Z( ) = Zo o (4. It is well known that with parallel illumination the same diffraction pattern appears for any other spectral component at a distance from the input mask given by R( ) = Ro o (4. Testorf/1597980/Ch04 152 Chapter Four Object x y Achromatic lens Zone Plate Achromatic Fresnel pattern y x R1 Ro R2 l f d'o FIGURE 4. let us consider a given diffraction pattern located at a distance Ro from the object for the reference chromatic component of wavelength o .e. we do not consider it here.May 22. if the limits of the spectrum of the incoming radiation are 1 and 2 . However. Thus. from a scalar paraxial diffraction point of view. We note that although the effect of residual focuses can be signiﬁcant for those wavelengths that are different from the design wavelength. providing a dispersion volume for the diffraction pattern under study.29 Achromatic imaging system under study. Our goal in this section is to achieve the optimal relationship between the geometric distances in the imaging system to obtain an output image corresponding to a given Fresnel pattern with minimum chromatic aberration.99) In this way. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. for = o we obtain a different structure .
) ∝ RWt (x ( ).May 22. Testorf/1597980/Ch04 The RadonWigner Transform for the diffracted pattern. 2. we again use the RWT approach to describe the spectral components of this Fresnel pattern. passage through the zone plate. namely. propagation through the achromatic lens. into a different image volume. free propagation at a distance l. If we deﬁne i = arctan ( i Ro ). Equivalently. Since the above dispersion volume is transformed. by means of the imaging system. the ﬁnal superposition of all the spectral components of the incoming light produces a chromatic blur of the monochromatic result.100) In this equation the chromatic blur is considered through the spectral variation of the coordinates in the RWT for any given transverse position x in the diffraction pattern. To analyze this effect. The transformation matrix Machr can be obtained as a sequence of elemental transformations (see Fig. x = x Ro sin 1 1 − sin 2 2 (4. we consider a onedimensional amplitude transmittance t(x) for the diffracting object. and. for a ﬁxed observation position there is a region in the Radon space that contains all the points needed to compute the polychromatic irradiance. The output plane is selected as the image plane of the diffraction pattern . the same RWT analysis will be performed at the output of the system to analyze the chromatic blur at the output plane for the reference wavelength o . 4.29). tan ( ) = − Ro . and. Let us now ﬁx our attention on the effect of the imaging system on the polychromatic diffraction pattern under issue. ( )) . If we recall the result in Eq. free propagation to the focal plane of that element. Let us now apply our approach to calculate the irradiance freespace diffraction pattern under issue through the RWT of the object mask. the width of this region in both Radonspace directions can be estimated as = 1 − 2 . ﬁnally free propagation at a distance do . therefore. To achieve an achromatization of the selected diffraction pattern. For the sake of simplicity.36) for z = Ro . x x( )= sin ( ) Ro 153 (4. this region has to be reduced in the output plane of the optical setup. the less is the effect of the chromatic blur affecting the irradiance at the speciﬁed observation point. (4.101) Note that the smaller this region is. we obtain that each spectral component of this Fresnel pattern is given by Io (x. it is interesting to derive the geometric conditions that provide a minimum value for its axial elongation in the image space. Again. we use the RWT approach to achieve this description by simply noting that the system behaves as an abcd device that links the object plane and the selected output plane. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Thus. for i = 1.
as can be seen from Eq. 4. However. ( ) tan ( ) = Ro + f2 Zo o −1 . a ﬁrstorder approximation to that ideal correction can be achieved by imposing a stationary behavior for ( ) around = o . min = min{ ( ) ∈ [ 1 . It is worth mentioning that exact achromatization of the pattern is achieved only when ( ) = ( o ) ∀ ∈ [ 1 . We now try to ﬁnd a case that minimizes the chromatic blur in the output pattern.107) . depend on the particular values of the geometric design parameters f and Zo of the imaging system. for the polychromatic description of the output diffraction pattern we have to sum values of the RWT of the transmittance of the object in a region in the Radon domain whose size in both dimensions is given by = where max max − min . Thus.May 22. it is straightforward to obtain ⎛ ⎜ M( ) = ⎝ 1− l f − f Z( ) 1 f − f − mo f + mo l + mo f 2 mo Z( ) ⎟ (4. In particular. we impose d ( ) d =0 o (4. by setting = 0 we ﬁnd Io (x.28).103) mo Zo where mo = −do / f is the magniﬁcation obtained at the ﬁxed image plane (for = o ). 2 ]. which deﬁne the extension of the integration region in Radon space in the polychromatic case. which cannot be fulﬁlled in practice.104) Therefore. x = x cos ( ) mo (4.106) The speciﬁc values of these limits. 2 ]} (4.1. The relationship between the RWTs of the input object and the output Fresnel pattern for each spectral channel now can be established by application of Eqs. (4.27) and (4.104).105) = max{ ( ) ∈ [ 1 . x = x (cos mo max − cos min ) (4. by using the results in Sec. 2 ]}.102) ⎠ ⎞ where the following restriction applies to the ﬁxed desired output plane f f2 l = Ro + f − − (4. Mathematically. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Testorf/1597980/Ch04 154 Chapter Four under study for = o . ) ∝ RWt x ( ).2. (4.
The chromatic blur is clearly seen by comparing these two ﬁgures. we follow the prescriptions given in the above paragraphs. the distance Ro = 11. at distance Ro from the object. equivalently. (b) Broadband (Hg lamp) irradiance distribution.1 nm. f . (4.108) This condition transforms Eq.1 nm. d tan ( ) d which leads to the optimal constraint Ro = − f2 Zo (4.May 22. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. is presented. Testorf/1597980/Ch04 The RadonWigner Transform 155 (a) (b) FIGURE 4. Figure 4. but when a polychromatic collimated beam from a highpressure Hg lamp is used. In particular.110) Thus.30 Grayscale display of the irradiance distribution to be achromatized: (a) Monochromatic pattern for o = 546. and do fulﬁlling the two above equations provides a design prescription for a ﬁrstorder compensation of the chromatic blur in the diffraction pattern located. To optimally achromatize this diffraction pattern. we present an experimental veriﬁcation by using a twodimensional periodic transmittance as an object. with the same period p = 0. As a Fresnel pattern to be achromatized.30a shows a picture of the irradiance distribution in that situation.73 cm is selected.109) =0 o (4. or. In Fig.30b. a selfimaging distribution is selected. Zo . the irradiance distribution over the same plane.179 mm in both orthogonal directions. after parallel illumination with o = 546. 4.54 To illustrate this design procedure and to check the predicted results. for = o . We use a kinoform lens .103) into l = 2Ro + f − f mo (4. the choice of a set of geometric parameters l.
with Zo = −12 cm. Therefore.8). 4.4. to a twodimensional function by smearing it uniformly along the original projection direction (see Fig.1 we showed that the axial behavior of the irradiance distribution provided by a system with an arbitrary value of SA can be obtained from the single RWT of the mapped pupil q 0. i.86 cm. The comparison between this result and the monochromatic one in Fig. . A grayscale display of the output irradiance is presented in Fig. ) from the projected function I (0. The backprojection algorithm converts the desired axial irradiance for a ﬁxed value of W40 .2 Controlling the Axial Response: Synthesis of Pupil Masks by RWT Inversion In Sec.e.31.55 Once the WDF is synthesized with the values of the input axial irradiances. 4. the entire twodimensional Wigner space can be sampled on a set of lines deﬁned by these parameters. (4. z) representing the irradiance at the axial points— variableW20 —for a sufﬁcient set of values of W40 . 4.1. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.0 (x. 0.57) is inverted to provide the desired pupil function. represented by a onedimensional function.30a shows the high achromatization level obtained with the optimized system.31 Grayscale display of the achromatized irradiance distribution.4). and we choose a value do = 10.3.74) can be considered the keystone of a pupil design method55 in which the synthesis procedure starts by performing a tomographic reconstruction of Wq 0. (4. In fact.00 cm. the pupil function is obtained by use of Eq. 4. and we place that object at a distance l = 2Ro + f + f 2 /do = 49. for all the different values of W40 . Finally. The ﬁnal reconstructed function Wq 0. the geometric mapping in Eq. Testorf/1597980/Ch04 156 Chapter Four FIGURE 4. (4. √ we select a focal distance for the achromatic lens f = −Zo Ro = 11.39 cm from that lens. Thus. Eq..0 (x. 4.May 22.0 (s) of the system. ) is obtained by a proper ﬁltering of the summation image. Then the algorithm calculates the summation function that results when all backprojections are summed over all projection angles .
To implement this operation. The result is shown in Fig.0 157 1. and thereby the projected distributions deﬁned by the RWT. Optical correlation can be performed in coherent systems by use of the fact that the counterpart of this operation in the Fourier domain is simply the product of both signals. From the reconstructed WDF we obtained the synthesized pupil function p(x) by performing the discrete onedimensional inverse FT of Wq 0.0 0.8 1.May 22.0 0.4 r/a 0.8 0. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. 4. As can be seen. In this case.4 r/a 0. Among several mathematical operations that can be optically implemented. (b) Phasespace tomographic reconstruction of the same pupil.4 0.0 (a) (b) FIGURE 4.4 0.0 (x. such as the classic VanderLugt and joint transform correlators.6 0.3 Signal Processing through RWT Throughout this chapter we have discussed the RWT as a mathematical tool that allows us to develop novel and interesting applications in optics. we used 1024 values for both W40 / and W20 / . From this function we numerically determined ﬁrst the Wq 0.2 0.57 Because .2 0. To illustrate the method.32 (a) Amplitude transmittance of a desired pupil function.2 0. the amplitude transmittance of the synthesized pupil function closely resembles the original apodizer in Fig.0 (x. It has been shown that its main features are to increase the focal depth and to reduce the inﬂuence of SA. and we reconstructed the WDF by using a standard ﬁltered backprojection algorithm for the inverse Radon transform. obtaining the axial irradiance distribution for different values of SA. using the WDF deﬁnition.32. We treated these distributions as if they represented the desired axial behavior for a variable SA.32b. we numerically simulated the synthesis of an annular apodizer represented in Fig.32a.4. ranging from −16 to +16. correlation is one of the most important because it can be used for different applications.0 0.0 0. 4.6 0.56. Testorf/1597980/Ch04 The RadonWigner Transform Amplitude transmittance Amplitude transmittance 1. ). 4.0 0. several optical architectures were developed.8 1.0 0.8 0. ) function. 4.2 0.6 0. such as pattern recognition and object localization.6 0.
and so shift invariance is a disadvantage. As is well known. the RWD of the input is obtained with the experimental conﬁguration shown in Sec.2. Conventionally. 4. but there are situations in which the position of the object provides additional information such as in image coding or cryptographic applications. when an object should be recognized in a relevant area and rejected otherwise. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. following a twostep process. In the ﬁrst stage. we register a hologram of the ﬁeld distribution at this plane with a reference wavefront at an angle . and therefore more complete information would certainly be attained from a display showing several fractional correlations at the same time. 4.59 is a generalization of the classic correlation that employs the optical FrFT of a given fractional order p instead of the conventional FT. but for a single fractional order p at a time.2. (4.34). In the second stage. In several pattern recognition applications this feature is useful.2. The fractional order involved in the FrFT controls the amount of shift variance of the correlation. In this section we will show that such a multichannel fractional correlator could be easily implemented from the RWD system presented in Sec. The resulting optical system generates a simultaneous display of fractional correlations of a onedimensional input for a continuous set of fractional orders in the range p ∈ [0.May 22. the correlation output simply moves if the object translates at the input plane. 1]. The product inside the brackets of Eq. instead of recording the intensity. x} (4. or when the recognition should be based on certain pixels in systems with variable spatial resolution. The fractional correlation58. However. We start by recalling58 the deﬁnition of the fractional correlation between two onedimensional functions f (x) and f (x) C p (x) = F −1 {F p ( ) F p ∗ ( ). and the input function to be correlated is located at the input plane (see Fig.111) can be optically achieved simultaneously for all fractional orders. for example.111) It is important to note that with the above deﬁnition the classic correlation is obtained if we set p = 1. the fractional correlation is obtained as the inverse Fourier transform of the product of the FrFT of both the reference and the input objects. the shiftvariance property modiﬁes the intensity of the correlation output when the input is shifted. ranging between p = 0 and p = 1. 4. . (see Fig. the optimum amount of variance for a speciﬁc application is frequently difﬁcult to predict. such a display should include the classic shiftinvariant correlation as the limiting case. 4. Ideally. the obtained multichannel matched ﬁlter is located at the ﬁlter plane. A matched ﬁlter can be obtained at the output plane if.33).2. Testorf/1597980/Ch04 158 Chapter Four conventional correlation is a shiftinvariant operation. In many cases this property is necessary.
May 22. except for BS (beam splitter) and M1 and M2 (plane mirrors). .33.111). Besides. The ﬁlter H corresponds to the one obtained in the setup of Fig. the experimental fractional correlation can be obtained exactly because this error affects both F p ( ) and F p ∗ ( ). 4. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. the diffracted ﬁeld at angle is collected by the lens L c . for each fractional order channel the ﬁeld immediately behind the ﬁlter plane has one term proportional to the product of the complex conjugate of the FrFT of the reference function f (x) and the same FrFT of the input function f (x).33 Multichannel matched ﬁlter registration for a fractional correlation. Thus the multiplicative phase factor in this equation and the corresponding one of the matched ﬁlter cancel out.2. Finally. The elements are the same as in Sec. Because each fractional order p ∈ [0. Testorf/1597980/Ch04 The RadonWigner Transform Input plane x Rp Lc 0 BS HeNe t M1 z l S θ L M1 H Filter plane xh 159 ap a¢ FIGURE 4.2. although the experimental FrFT for a given order p is approximated owing to the scale error discussed in Sec.34 Multichannel fractional correlator. 1] has an independent onedimensional correlation channel.2. which performs a onedimensional FT. Because the transmittance of the holographic ﬁlter has one term proportional to the complex conjugate of the reference ﬁeld in Eq. 4.2. all the fractional correlations for this xc Input plane x z Lc 0 BS HeNe t L H Filter plane xh a¢ θ f Output plane Lc l f FIGURE 4. 4. (4.
0 230 235 240 245 250 255 260 265 270 275 280 Spatial coordinate (pixels) FIGURE 4. range of fractional orders are obtained simultaneously at the output plane. Testorf/1597980/Ch04 160 Chapter Four Amplitude transmittance (a.0 0.34. 4.2 0.5 and p = 1. Both displays coincide near to p = 1 (except for the location of the maxima). Thus a twodimensional display is obtained in which the fractional correlations are ordered in a continuous display along the axis normal to the plane shown in Fig. for a ﬁxed displacement the correlation peak increases with p. It represents a double nonsymmetric slit with a continuous graylevel amplitude transmittance.4 0.6 0. Figure 4.36a shows the fractional correlations when the input is shifted an amount of onehalf of the object size. The continuous transition between the shiftvariant case p = 0 and the shiftinvariant case p = 1 is conﬁrmed in Fig. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.35 Amplitude transmittance of an input object selected to perform multichannel fractional correlation. but for lower values of p the fractional correlation is highly dependent on the magnitude of the shift.u. but the reference objects are shifted at the input plane. the correlation peak did not change appreciably.36b shows the fractional correlation when the input is shifted an amount equal to the size of the object. 4. As expected for p = 1.8 0.) 1. As can be seen in the threedimensional plot in this ﬁgure.May 22. In this ﬁgure the fractional autocorrelation of the input is considered. 60.36.35. For values ranging between p = 0. The . the correlation peak is the classic one located at the input position. and Fig. The variant behavior of the fractional correlation can be clearly seen by the comparison of these ﬁgures. 4. 4. The shiftvariant property of the FrFT correlation was conﬁrmed experimentally in Ref. Here we present a numerical simulation using an input object whose amplitude transmittance is shown in Fig.
the output of the system shows a variable degree of space variance ranging from the pure shift variance case p = 0 to the pure shift invariance case p = 1. Fr 0 x e (x) ac tio 0.6 0. This kind of representation provides information about the object.25 (b) FIGURE 4.0 0.0 0.36 Multichannel fractional autocorrelation of the function represented in Fig.00 161 0. that is.8 0. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E.00 e (x) (a) nsity p 0.50 0.75 1. Thus.8 0.2 0. such as classic correlation.75 nsity Fractional correlation inte p 1.50 Fractional correlation inte 0.0 1. It can be seen that as the fractional order becomes lower.0 1. shiftvariant property becomes evident for values close to p = 0. the classic correlation. Testorf/1597980/Ch04 The RadonWigner Transform 1. 4.0 er p .25.0 Fr ac 0.00 Spac e coo 0 rdina t na lo rd 0.25 0 x 1.4 0.00 tio Spac e coo 0 rdina t na lo er p rd 0.May 22.5 0.4 0. but also quantiﬁes its departure from a given reference position.0 0.2 0. the peak degenerates and shifts disproportionately toward the object position.35 with (a) a shift in the input plane of onehalf of the object size and (b) a shift of the whole size of the object.0 0.5 0.6 0.
Y. Saavedra. 39: 238–245 (2000). L. Am. “Radon transformation of timefrequency distributions for analysis of multicomponent signals. Gu. and C. G.” Appl. R. H. Andr´ s.” J. PrenticeHall. Zalevsky. “The computation of diffraction patterns in the presence of aberrations. L. W. Lett. and the fractional Fourier transform.” J. 139: 11–16 (1997). Pedersen. W.” Appl. 15. “Image rotation. Opt. Lohmann. 13. 23: 1126–1128 (1998). and H. E. Y.J. Braat. They also acknowledge ﬁnancial support of Ministerio de Ciencia y Tecnolog´a. Soc. Testorf/1597980/Ch04 162 Chapter Four Acknowledgments The authors would like to express their gratitude to P. Tajahuerce. “Optical illustration of a varied fractional Fouriertransform order and the RadonWigner display. 19: 1388–1390 (1994). Speech Signal Process. Kraus. Opt. Ferreira. Sicre. and P. B. Andr´ s. Sicre. 32: 368–383 (1993). M. Acoust. W. Barry. J. and A. E. Opt. A. Lett. 5. J. Yang. G. M. Am. G. 10. Soc. H. M. 4. Dorsch. Opt. B. S. G. “Pointspread function calculation for optical systems with residual aberrations and nonuniform transmission pupil. Stamnes. Wigner rotation. “RadonWigner display: e A compact optical implementation with a single varifocal lens. Granieri. Furlan. Hopkins and M. Soc. Silvestre for their contributions in the research revisited in this chapter. 16. D’Arcio. Wood and D. PellatFinet.” Appl. Furlan. “Variable fractional e Fourier processor: A simple implementation. G. “Analysis of the irradiance along different paths in the image space using the Wigner distribution function. Cohen. 42: 2105–2111 (1994). 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. Lohmann.. and E.” Opt. 7.” Opt. D.” Optica Acta 30: 233–242 (1983). Saavedra. e Granieri. D. 12. Dong.” Optica Acta 30: 207–222 (1983). J. W. J. Wood and D. W. H. 9. References 1. Comm. A11: 2664–2674 (1994).” Proc. D. W. 6. Spain (Grants DPI 20068309 and ı DPI 200802953). 2.Z. and H. Int. Andr´ s. Am. Silvestre. 36: 8363–8369 (1997). Yzuel and F. 35: 3925–3929 (1996). B. “Optical implementations of the RadonWigner display for onedimensional signals. Opt. Conf. Zhang. Yzuel. “Evaluation of diffraction integrals using local phase and amplitude approximations. “New optical conﬁgurations for implementing RadonWigner display: Matrix analysis approach. Calvo. Saavedra. A.” IEEE Trans. Comm. Zhang. Gu. 11. E. “Whitee light optical implementation of the fractional fourier transform with adjustable order control. Opt.May 22. T. Andr´ s. Upper Saddle River. Z. N. Furlan. A. A14: 853–858 (1997). J.” Opt. A10: 2181–2186 (1993). B. Barry. 17. and E. D.” Opt. “Finite element area and line integral transforms for generalization of aperture function and geometry in Kirchhoff scalar diffraction theory. and P.” Optica Acta 17: 157–182 (1970). E. “Numerical evaluation of diffraction integrals for apertures of complicated shape. 3. P. P. 4: 257–261 (1992). Sicre. Furlan. D. Y. Signal Process. Mendlovic. 8. Yang. 14. S. “Fresnel diffraction and the fractionalorder Fourier transform. TimeFrequency Analysis. and G. . and G. Saavedra.Z. Eng.” J. Spjelkavik. W. J. Frankena. T. B. J. J. C. J. 160: 292–300 (1999). G. Dong. C.” Opt. E. Lohmann. 1995. “Linear signal synthesis using the RadonWigner transform.
AgulloLopez. and M. Saavedra. Soc. 30. Born and E. G. Comm. 1: 29–40 (1992). Opt. Opt. Interference and Diffraction of Light.” Phys. D. “Analysis of 3D ı integral imaging displays using the Wigner distribution. 45: 69–77 (1998). Sheridan. and M.” Opt. 38. “Noninterferometric phase retrieval using a fractional Fourier system. Soriano. Ferrari. H. G. Mart´nezCorral. Saavedra. Eng. Opt. 26. 29. Phys. Lett. W. B70: 449–470 (1957). Silvestre. G. 23. Conner. “Misfocus tolerance seen by ˜ simple inspection of the ambiguity function. 32. F. J. Monsoriu. M. Furlan. 20. Saavedra. 25. E. G. W. Raymer. “Offaxis focal shift for e ı ˜ rotationally nonsymmetric screens. 48: 1247–1253 (2001). Slaney. Opt. J. Saavedra. Soc. 39. D. 72: 1137–1140 (1994).” J. “Optical phase retrieval by phasespace tomography and fractionalorder Fourier transforms. “Diffracted ﬁeld by an arbitrary aperture. G. D. Barrett. 67: 177–179 (1988). Principles of Optics: Electromagnetic Theory of Propagation. McAlister. and G.” ´ ´ Opt. D. M. M. Asakura. J. Saavedra. Kak and M. and D. A. and W. Situ. Mendlovic. 33. 27. “The aberration permissible in optical systems. C. and J. T. 18: 1290–1292 (1993). “Simultaneous display of all the Fresnel diffraction patterns of onedimensional apertures. and J. Hopkins.. G. A: Pure Appl. J. and G. Sicre. “Complex waveﬁeld reconstruction using phasespace tomography.” Proc. Alieva and F. Eichmann. O. 5: S361–S364 (2003). Silvestre.” Opt. G. D. A. 37. E. Li. A. 61: 87–92 (1999). Opt. New York.” Phys. Disp. 1999. “Optodigital tomographic reconstruction of the Wigner distribution function of complex ﬁelds. Mayer.” Opt. D. Raymer. 1988. Y. A. 125: 267–274 (1996). D. 34. R. Patrignani.” Pure Appl. 24. Andr´ s. Lett. Cloitre. “Assessment of a Wignerdistributionfunctionbased method to compute the polychromatic axial response given by an aberrated optical system. IEEE Press.May 22. B.” Appl. 37: 2142–2144 (1998). Sakurada.” J. Cambridge. Trabocchi. Phys. 21.” Am.” Opt. A25: 108–115 (2008). Furlan. M. G. A. “Optical Wigner distribution and ambiguity function for complex signals and images. 69: 799 (2001). Rev. Bartelt. 19. Javidi.” J. L.” J. “Optical propagation of fractal ﬁelds: Experimental analysis in a single display. Principles of Computerized Tomographic Imaging. “Proposal for optical implementation of the Wigner distribution function. Zalevsky. H. 23: 2693–2696 (1984).” J. “Application of the Radon transform to optical production of the Wigner distribution. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. and J. T. Allain and M. W. Naughton. McAlister. W. Patrignani. Opt. and E. H. E. Mod. F. Cambridge University Press. OjedaCastaneda. C. 36. Technol. 28. “Optical wave propagation of fractal ﬁelds. Furlan. D. Eng. Saavedra. Ticknor. 35. Wolf. L. (7th ed. A. D. Uozumi. Comm. 42: 753–758 (2003).” Am. 2: 180–185 (2006). Opt. Clarke. J. Chap. Furlan. Lett. C. Mart´nezCorral. 23: 738– 744 (1984). Opt. 163 . T. 47: E63–E67 (2008). Mart´nezCorral. Easton. Rev. Furlan. G. H. and H. and J.” Appl. J. W. S. “Onaxis irı radiance for spherically aberrated optical systems with obscured rectangular apertures: A study using the Wigner distribution function. P. Shabtay. “Axial behaviour of Cantor ring diffractals. U. J. and Z. D. and M. D.” Opt. Jr. OjedaCastaneda. Am. Dubra and J. Granieri. Mod. Y. M. W.. “Fresnel diffraction by onedimensional regular fractals. M. J.” Appl. and T. Opt. 22. 9. “Optical diffraction on fractals. Furlan. A. Beck. Gopinathan. Furlan. Phys. 20: 1181–1183 (1994). Testorf/1597980/Ch04 The RadonWigner Transform 18. Monsoriu. and S Granieri. Beck. 31. B33: 3566– 3569 (1986).
” Opt. E. Bescos. Andr´ s.” Optica Acta 21: 981–1003 (1974). A. Furlan. J. W.” Appl. J. Furlan. Furlan. 40: 6439–6444 (2001).” Appl. Javidi. 49. Opt. “Optical correlation based on the fractional Fourier transform. Sicre. SPIE 2730: 252–255 (1996). “Digital restora´ ı tion models for color imaging. “Phasespace representations as a tool for the evaluation of the polychromatic OTF. Saavedra. Comm. W. and P. H. E. “Whitelight implementation e of the Wignerdistribution function with an achromatic processor. 2009 17:52 Phase Space Optics: Fundamentals and Applications/Markus E. W. K. Furlan. G.” Photogr.” Appl. Disp. “Extraaxial polychromatic optical transfer function. 55. D. Arizaga. “Chromatic aberration matching of the polychromatic optical transfer function. 44: 323 (1983). 27: 419–424 (1988). “Formation of color images: Optical transfer func´ ı tions for the tristimulus values. 52. “Phasespace representations as a tool for the evaluation of the polychromatic OTF. OjedaCastaneda. Tajahuerce. Santamar´a. Lohmann. Technol. Color Science. 96: 208–213 (1993). Sci. New York. Andr´ s. Sicre. Climent. “Calculation of axial polychromatic optical transfer function. 36. J. M. “Synthesis of ﬁlters for speciﬁed axial irradiance by use of phasespace tomography. J.” Appl.” Opt.” Appl. 21: 355–362 (1977). and P. D. “Whitelight e optical information processing with achromatic processors. Zalvidea.” Appl. J. P. W. M. Brenner . D. 1982).” Opt. 34: 8209–8212 (1995). Granieri. Granieri. J. 44. and G. Lancis. 48. 20: 684–687 (1981). “Fractional correlation. New York. A. Saavedra. 1992. Introduction to Fourier Optics. Opt. W. and G. R. Lancis. 54. 104: 39–45 (1993). 36: 9146–9151 (1997). S. 34: 303–309 (1995). and A. Barnden. Lancis. 43. Goodman. R. 41. Vander Lugt. H. E. 56. Wiley. “Polychroe matic merit functions in terms of the Wigner distribution function. Andr´ s. P. Wyszecki and W. “Achromatic Fresnel diffraction e patterns. Silvestre. Opt. and W. S. Andr´ s. Comm. Opt. Ozaktas. M. Saavedra. J. 57. A. “The ambiguity func˜ tion as a polar display of the OTF.” J. and J. S. Andr´ s.” Opt. 189: 15–19 (2001). Saavedra. and E. D. Tebaldi.” Opt. M. 96: 208–213 (1993). Lancis. Furlan. 51. W. E. and G. 1996. W. D. Mart´nezCorral. Mendlovic. 58. Comm. D. Takeda.May 22. E. Saavedra.” Appl. Opt. D. Sicre. J. E. Testorf/1597980/Ch04 164 Chapter Four 40. Digest Series 11: 220–223 (1994). 47. H. Lohmann. G.” 1994 OSA Tech. Stiles. Tajahuerce. Bonet. W. E. . Santamar´a. 46. Wiley. G. Lancis. Yzuel. “Parallel fractional correlation: An optical implementation. 6636–6645 (1997). E. 60. Opt. W. Opt. Saavedra. Saavedra. 42. “Analysis of ı 3D integral imaging display using the Wigner distribution. “Polychroe matic axial behavior of aberrated optical systems: Wigner distribution function approach. P. D. Barnden.” Optica Acta 23: 1–24 (1976). Santisteban. Altamirano. and J. M. Furlan. and M. 45. Optical Signal Processing.” Proc. Comm. 2: 180–185 (2006). and J. and E. R. G. V. G. 59. McGraw Hill. E. 53. Furlan. New York. Yzuel. Comm. B. Bescos and J. and J. 50. J. Silvestre. and Eng.
This action might not be possible to undo. Are you sure you want to continue?