Professional Documents
Culture Documents
6093-Vol2 Chap01 v2
6093-Vol2 Chap01 v2
1.1
Schr
odingers wave function
..
........
....
..........................
...
.....
.................
..
.
.......... .... ....
...
....
.
.
.
.
.
... .. ...
.
.
...
.
.
.
.
.
...
.
....... .... ..... .... ..... ....
...
.
.
.
.
.
...
.
.
..
.
.
.
... .. . .. ..
.
.
.
.
.
.
.
.
.
...
.
.
.
.
.. .. .. .. .. ..
........
....
.....................................
... .. .. .. ... ..
... ................
.
.
........................
................
... ... .. ... ... ..
........
.
.
.. .. .. ..
.
.
.
.
.....
.
.
.
.
.
.. ... ... ... .. ...
.......... ..
...........................................................................................................................................................................................................................................................................................................................................................
where the area is (proportional to) that probability. T herefore, the squared
2
wave function (x) is a probability density, and one refers to the wave
function itself as a probability density amplitude. Since we shall surely find
1
dp (p)
=1
(1.1.4)
(1.1.5 )
It helps to build on an analogy that you might want to keep in mind because it will be useful for the visualiz ation of some rather abstract quantum
mechanical statements in terms of geometrical objects. We consider real ncomponent vectors and their numerical description in terms of coeffi cients
(coordinates) that refer to agreed-upon coordinate systems:
r
x.2 .........
...
.
...
...
..
...
.
... ..
... .
... .
... .
.. ..
...
................................................................ x1
r
.......
y2.
...
...
.
...
...
..
...
.
.
...
y1
...
..
........
...
.
........
.
.
.
.
... .
.
.
.
... . ................
................
....
r =
b (x1 , x2 , . . . , xn )
=
b (y1 , y2 , . . . , yn ) .
(1.2.1)
O ne and the same vector has two (or more) numerical descriptions, the
coordinates xj and the coordinates yk . T hese numbers, although not unre-
lated, can be quite diff erent, but they mean th e same v ec tor r . We make
this explicit with the aid of the basis vectors ej (for the x description) and
r =
n
X
xj ej =
j= 1
n
X
yk fk .
(1.2.2)
k= 1
ej ek = jk =
0 if j 6= k ,
fj fk = jk ,
(1.2.3)
xj = ej r
and yk = fk r
(1.2.4)
tell us how we determine the coordinates of r if the basis vectors are given.
is paired with
r =
b (x1 , x2 , . . . , xn ) =
b (y1 , y2 , . . . , yn )
y1
x1
x2 y2
b .
r =
b . =
.. ..
(1.2.5 )
(1.2.6 )
yn
xn
and the two kinds of vectors are related to each other by transposition,
r =r ,
r =r
(1.2.7 )
O ne immediate benefit of distinguishing between row vectors and column vectors is that we can write inner (scalar, dot) products as simple
r =
b (x1 , . . . , xn ) ,
s=
b (u1 , u2 , . . . , un ) ,
(1.2.8 )
u1
u2
X
rs=
xj uj = (x1 , x2 , . . . , xn ) . = r s .
|
{z
} ..
=
b r
un
| {z }
=
bs
(1.2.9 )
|r {z s} = |r s {=z s r } = r| {z s} .
inner product of
two row vectors
products of
th e ty pe
row tim es
colum n
(1.2.10)
T he central identity here is, of course, consistent with the product rule for
transposition, generally: (AB)T = B T AT , here:
( r s )T = s
T T
= sr ,
indeed .
(1.2.11)
U pon combining
r =
xj ej
and xj = ej r = r ej
(1.2.12)
(1.2.13)
into
r =
X
j
r ej ej = r
ej ej
x1 u1 x1 u2 x1 un
x1
x2 u1 x2 u2 x2 un
..
(1.2.14)
r s=
b . (u1 , . . . , un ) = .
.
.
.
.
..
. . ..
..
xn
xn u1 xn u2 xn un
T he particular dyadic that appears in (1.2.13) has the property that when
it multiplies (on the left) the arbitrary vector r , the outcome is this vector
itself: it is the unit dyadic,
X
(1.2.15 )
ej ej .
r 1 = r with 1 =
j
T he notation
reminds us that such a dyadic is like a column vector on
the left, and like a row vector on the right.
T he identification of the unit dyadic is consistent only if it also acts
accordingly on the right. Indeed, it does,
X X
1r =
ej ej r
ej ej r =
j
ej xj = r .
(1.2.16 )
ej ej = 1
(1.2.17 )
expresses the completeness of the set of column vectors ej , and also that of
the set of row vectors ej , because we can expand any arbitrary vector r
ej ek = jk ,
(1.2.18 )
and of completeness in (1.2.17 ) are two sides of the same coin. And, of
course, there is nothing special here about the ej set of vectors, the fk
basis vectors are also orthonormal,
fj fk = jk
(1.2.19 )
and complete,
X
fk fk = 1 .
(1.2.20)
In
r =
X
j
ej xj =
X
k
fk yk
(1.2.21)
fk yk
xj = ej r = ej
k
X
X
(ef )jk yk
ej fk yk =
=
(1.2.22)
with
(ef )jk = ej fk = ej fk ;
(1.2.23)
and likewise
yk =
(f e)kj xj
with
(f e)kj = fk ej .
(1.2.24)
j0
X X
=
(ef )jk (f e)kj 0 xj 0
j0
(1.2.25 )
that
X
(1.2.26 )
must hold. T his is to say that the two transformation matrices are inverses
of each other hardly a surprise.
1-1 U se the definitions of (ef )jk and (f e)kj 0 to verify this relation directly.
P
1-2 What appears in yk = k0 ? kk0 yk0 as the result of converting yk into
xj and then back to yk ? V erify here too that ? kk0 = kk0 .
R ather than converting one description into the other, we can ask how
the two sets of basis vectors are related to each other. Since both sets are
ej fj = O ej
(1.2.27 )
fj ej = O ej ej ,
(1.2.28 )
| {z }
= 1
O =
fj ej .
(1.2.30)
fk kj = fj ,
O ej =
fk ek ej =
|{ z }
k
k
(1.2.31)
= kj
fk O =
fk fj ej =
kj ej = ek ,
|{z }
j
j
(1.2.32)
= kj
so that the same dyadic O also transforms the rows fk into the ek s. T ogether with the transposed statements we thus have
fj = O ej ,
T
fj = ej O
fk O = ek ,
O fk = ek
(1.2.33)
with
O =
X
k
fk ek
and
X
j
ej fj .
(1.2.34)
fj
= O
ej
ek = fk O = ek O
or
= O O fj
O,
(1.2.35 )
O O
= 1 = O
O .
(1.2.36 )
= O
(1.2.37 )
are called orth og onal, in analogy to the corresponding terminology for orthogonal matrices in linear algebra.
1-3
H ow are the transformation matrices (ef )jk and (f e)kj related to the
orthogonal dyadic O ?
Actually, in linear algebra the most basic definition of an orthogonal
transformation is that it leaves all inner products unchanged. T hat is, for
any pair of vectors r , s we should have
" "
r O s O = rs,
(1.2.38 )
" "
Or Os =r s .
(1.2.39 )
rs = r O O s
and
T
and O O
= O
rs = r O
Os
from (1.2.38 )
from (1.2.39 ),
(1.2.40)
10
1-4
P j = ej ej .
P j ? R epeat for
P jk = P j + P k
1.3
with
j 6= k .
11
in some sense and, therefore, x refers to the situation object is at x, exactly . T his, however, is no longer a real physical situation but rather the
overidealiz ed situation of that unphysical limit. As a consequence, ket x
is not actually associated with a physically realiz able state, it is a convenient mathematical fiction. T he unphysical nature is perhaps most obvious
when we recall that the perfect localiz ation of the overidealiz ed limit would
require a control on the quantum object with infinite precision and this
is never available in an actual
real-life experiment.
B y the same token, ket p refers to the overidealiz ed situation of infinitely sharp
again a mathematical fiction, not a physical re momentum,
ality. B oth x and p kets are extremely useful
mathematical objects, but
one must keep in mind that a physical ket alw ays involves a range of x
values and a range of p values. T his range may be small, then we have a
well-controlled position, or a well-controlled momentum, but it is invariably
a finite range.
K ets , x , p , . . . are analogs
vectors. T hey have
of
column-type
their partners in the so-called bras , x, p, . . . , which are analogs of
row-type vectors. When dealing with the real vectors r , r , we related the
T
two kinds to each other by transposition, r = r . N ow, however, the coordinates that is the wave functions (x), (p) are complex-valued.
T herefore, mathematical consistency requires that we supplement transposition with complex conjugation, and thus have h ermitian conjug ation, or,
as the physicists say, we
= .
(1.3.2)
take the adjoint : = ,
"
1 + 2 = 1 + 2 ,
"
3 + 4 = 3 + 4 .
(1.3.3)
In particular the adjoint statements to the decompositions of in (1.3.1)
are
Z
Z
= dx (x) x = dp (p) p .
(1.3.4)
In further analogy with the column-type vectors of Section 1.2, the kets
are also endowed with an inner product, so that the vector space of kets is
an inner-product space or H ilbert space, the name honoring D avid H ilberts
12
contributions. T he notation 1 2 of the inner product as a dot product
is, however, not used at all. In the mathematical literature, inner products
are commonly
as ( , ), so that the inner product of two kets would
" written
appear as 1 , 2 except that mathematicians are not fond of the ket
and bra notation, D iracs stroke of genius.
Instead, one follows the suggestion of (1.2.10) and understands the inner products of two kets, or two bras, as the analogs of row-times-column
products. So the inner product of the ket
Z
1 = dx x 1 (x)
(1.3.5 )
with the ket
2 =
dx x 2 (x)
1 = dx 1 (x) x
with the ket 2 :
= dx 1 (x) dx0 xx0 2 (x0 ) ,
12 =
(1.3.6 )
(1.3.7 )
(1.3.8 )
0
where the integration variable of (1.3.6
) is changed
to x to avoid confusion.
T his is also the inner product of bras 1 and 2. As anticipated
),
in
(1.3.8
one
writes
only one
vertical line in the bra-ket product 1 2 = 1 2 of
bra 1 and ket 2 and speaks of a D irac bracket or simply brac k et.
In accordance with what the reader learned in whichever first course on
quantum mechanics, we expect the inner product (1.3.8 ) to be given by
Z
12 = dx 1 (x) 2 (x) ,
(1.3.9 )
so that we need xx0 such that
Z
dx0 xx0 2 (x0 ) = 2 (x)
for all 2 (x). T hus, we infer
0
xx = (x x0 )
(1.3.10)
(1.3.11)
13
which is to say that x kets and bras are pairwise orthogonal and normaliz ed
to the D irac function. T here is a longer discussion of the function in
Section 4.1 of B asic M atters, and so we are content here with recalling the
basic, defining property, namely,
Z
dx0 (x x0 )f (x0 ) = f (x)
(1.3.12)
for all the functions that are continuous near x.
T he normaliz ation (1.1.2) of the wave function,
Z
dx (x) 2 = 1 ,
(1.3.13)
now appears as
see:
(1.3.14)
dx (x) x dx0 x0 (x0 )
Z
Z
= dx (x)
dx0 xx0 (x0 )
| {z }
= 1,
dx (x)
= (x x0 )
{z
= (x)
(1.3.15 )
In other words, the physical kets , and the physical bras , are of unit
leng th .
We recall also the physical significance of the bracket 12 in (1.3.9 ),
2
after which it is named probability amplitude: Its squared modulus 12
is the probability
prob(2 1) of finding the system in state 1, described
by ket 1 and parameteriz ed by the
wave function 1 (x), if the system is
known to be in state 2, with ket 2 and wave function 2 (x). We should
not fail to note the symmetry possessed by prob(2 1),
2
= prob(1 2) ,
(1.3.16 )
prob(2 1) = 12
which is an immediate consequence of
12 = 21 ,
(1.3.17 )
14
1-5 F or all bras a and all kets b , show that
2
aa bb ,
a b
and state under which condition the equal
0
xx = (x x0 ) ,
(1.3.18 )
ej ek = jk .
(1.3.19 )
ej ej = 1 ,
(1.3.20)
reads
Z
dx x x = 1 .
(1.3.21)
the
operator analog of the unit dyadic 1 but we will not be pedantic about
it and write it just like the number 1. It will always be unambiguously clear
from the context whether we mean the unit operator or the unit number.
L ikewise, the symbol 5 , say, can mean the number 5 or
5 times
the unit
operator, depending on the context. F or instance, in 5 x = x5 we have
15
the number 5 on the left and 5 times the unit operator on the right. T here
will be no confusion arising from this convenience in notation.
B ut we must not forget to verify the completeness relation (1.3.21).
V erification means here just the check that it is consistent
with everything
else we have so far. F or example, is it true that 1 = ? We check:
1 =
Z
dx x x
{z
=1
dx x
Z
dx0 x0 (x0 )
{z
}
}|
dx0 xx0 (x0 )
| {z }
= (x x0 )
dx x (x) = ,
indeed .
(1.3.22)
1 = . Another little calculation:
12 = 112
Z
Z
Z
0
0 0
dx00 x00 2 (x00 )
dx x x
= dx 1 (x) x
Z
Z
Z
0
dx00 x0 x00 2 (x00 )
= dx 1 (x)
dx x x
| {z }
| {z }
(x x0 ) =
dx 1 (x) 2 (x) ,
= (x0 x00 )
{z
{z
= 2 (x )
= 2 (x)
}
}
(1.3.23)
F inally, is 12 = 1? L et us see,
2
1 =
=
dx x x
dx x
dx0 x0 x0
dx0 xx0 x0
| {z }
dx x x = 1 ,
= (x x0 )
indeed.
(1.3.24)
16
adjoint relations: x = x ,
x = x ,
orthonormality: xx0 = (x x0 ) ,
Z
completeness:
dx x x = 1 .
(1.3.25 )
And, by the same token, the corresponding statements hold for the momentum states,
p = p ,
adjoint relations: p = p ,
orthonormality: pp0 = (p p0 ) ,
Z
completeness:
dp p p = 1 ,
(1.3.26 )
because we can repeat the whole line of reasoning with labels x consistently
replaced by labels p.
1.4
xp transformation function
Since
the two integrals in (1.3.1) are diff erent parameteriz ation for the same
ket , there must be well defined relations
the wave functions (x)
between
and (p) and also between the kets x and p . F or the wave functions,
the relations are the F ourier transformations of (1.1.3). We
now
use them
to establish the corresponding statements that relate x and p to each
other.
F irst note what is already implicit in (1.3.22), namely that
Z
x = x dx0 x0 (x0 )
Z
= dx0 xx0 (x0 )
| {z }
= (x)
= (x x0 )
(1.4.1)
or
(x) = x
(1.4.2)
(p) = p .
17
(1.4.3)
T herefore we have
eixp/~
dp
(p)
2~
Z
eixp/~
(1.4.4)
p ,
= dp
2~
and this must be true irrespective of the ket we are considering, so that
x = (x) =
x =
eixp/~
dp
p
2~
(1.4.5 )
(1.4.6 )
p = (p) =
eipx/~
dx
(x)
2~
Z
eipx/~
= dx
x ,
2~
which implies
and
p =
p =
eipx/~
x
dx
2~
eixp/~
dx x
.
2~
(1.4.7 )
(1.4.8 )
(1.4.9 )
xp = x
|
0 eix0 p/~
eixp/~
=
,
dx x
2~
2~
{z R
}
0
dx0 (x x0 )
(1.4.10)
18
that is
eixp/~
,
xp =
2~
(1.4.11)
eixp/~
.
px =
2~
(1.4.12)
(x) = x = x1
Z
= x
dp p p
Z
= dp xp p
=
eixp/~
(p) ,
dp
2~
(1.4.13)
(x x0 ) = xx0 = x1x0
Z
0
dp p p x
= x
Z
= dp xp px0
Z
eixp/~ eipx /~
= dp
2~
2~
Z
dp i(x x0 )p/~
e
=
2~
(1.4.14)
19
as (1.3.9 )
or (1.3.13) tell us that the position wave function (x) = x has the
cm, say),
metrical dimension of the reciprocal square root of a distance
(1/
the metrical
and likewise the momentum wave function (p) = p hasp
dimension of the reciprocal square
root of a momentum (1/ g cm/s, for
is dimensionless, see (1.3.14), the bras
instance).
for
p
x
=
,
a
~/a
(1.4.16 )
p
~/a2 x (1.4.17 )
= dx p = x~/a2
U = dt p = t~/a ~ x = ta .
(1.4.18 )
1-6 V erify
T he hermitian conjugation of (1.3.2) has the same product rule as transposition, see (1.2.11),
(AB) = B A ,
(1.4.19 )
20
x p, so that
U =
dt x = ta ~ p = t~/a
(1.4.20)
(1.4.21)
dt p = t~/a ~
dp p p = 1 .
dt0 (ta t0 a)
= (1/ a) p = t~/ a
(1.4.22)
It is all right indeed, and the right identity in (1.4.21) is demonstrated the
same way. At an intermediate step the identity
(ta t0 a) =
1
(t t0 )
a
(a > 0)
(1.4.23)
21
dx (x) x2 .
xn =
dx (x) xn
dx (x) f (x)
(1.5 .2)
(1.5 .3)
(1.5 .4)
R ecalling that (x) = (x) (x) =
x x , we rewrite these
expressions as
dx x x x ,
Z
2
x2 =
dx x x x ,
x=
Z
..
. Z
dx x f (x) x ,
f (x) =
(1.5 .5 )
thereby isolating the specific state of the system bra on the left, ket on
the right from the quantity that we are taking the average of,
x
x2
f (x)
..
.
Z
dx x x x X ,
dx x x2 x X 2 ,
dx x f (x) x f (X) .
(1.5 .6 )
T he first line introduces the position operator X as the integral of x x x,
22
= dx x x dx0 xx0 x0 x0
| {z }
=
And so forth, we have
= (x x0 )
{z
}
= x x
dx x x2 x = X 2 ,
Xn =
dx x xn x
indeed .
f (X) = dx x f (x) x
(1.5 .7 )
(1.5 .8 )
(1.5 .9 )
1-8
C onsider f (X) = dx x f (x) x and compare f (X)f (X) with
f (X) f (X). Which property must be possessed by f (x) if f (X) is its
own adjoint, f (X) = f (X)? O perators with this property are called
selfadjoint or simply h ermitian (C harles H ermite), whereby we ignore the
subtle diff erence between the two terms in the mathematical literature.
C onclude that all expectation values of a hermitian operator are real. T his
reality property can serve as an alternative definition of what is a hermitian
operator.
1-7
g(P ) = dp p g(p) p
23
(1.5 .11)
(1.5 .12)
for all functions of P that derive from reasonable numerical functions g(p).
O nce again, the reasoning is completely analogous and we need not repeat
it.
1-9 Show the following fundamental relations:
~
xP =
x ,
i x
ix0 P/~
= x + x0 ,
x e
~
p ,
X p =
i p
0
eip X/~p = p + p0 .
(1.5 .13)
(1.5 .14)
average of
the numbers x
..
..
..
..
......
X hXi
| {.z } |{.z }
...
...
...
...
...
...
...
...
...
......
..
...................................................
(1.5 .15 )
expectation value
of position operator X
position operator X
sandwiched
between state bra and state ket
which introduces a new notation, hXi, that emphasiz es the role played by
the position operator X. O ne speaks of the expectation value , a historical
terminology that is, as so
often,
not fully logically but completely standard.
Similarly, we write f
(X) for the
value of the operator
expectation
2
function f (X) and hP i, P , . . . , g(P ) , for the expectation values of
P , P 2 , . . . , g(P ). We have introduced the latter, by analogy, in terms
of integrals involving the momentum wave function (p), but we can, of
course, also refer to the position wave function (x).
24
2first
course on quantum mechanics) for the expectation values hP i and P in
terms of (x).
1-11 D etermine the normaliz ation constant A for the position wave function
(
A sin(2x/L ) for L < x < L ,
(x) =
0
for x > L ,
and then calculate hXi, X 2 , hP i, and P 2 .
1.6
G iven a ket 1 and a bra 2, we can multiply them in either order, thereby
getting
the bracket 21 , a number ,
or the ket-bra 1 2 , an operator .
In the bracket 21 , the ingredients are no longer identifiable because very
many pairs of a bra and a ket
the same number for the bracket. B y
have
contrast, given the ket-bra 1 2, one can identify the ingredient almost
uniquely. T herefore, we cannot expect that there could be a mapping from
the bracket to the ket-bra, but there is a mapping from the ket-bra to the
bracket,
1 2 21 .
(1.6 .1)
read: the trace of 1 2 is 21 .
B efore proceeding, let us look at the analog for column and row vectors:
r s sr = s r ,
(1.6 .2)
25
or
x1
x1 u1 x1 u2 x1 un
x2
x2 u1 x2 u2 x2 un
. (u1 , u2 , . . . , un ) = .
.
.
.
..
. . ..
..
..
xn u1 xn u2 xn un
xn
x1 u1 + x2 u2 + + xn un ,
(1.6 .3)
the diagonal sum of the matrix for r s. C learly, if you only know the value
of this sum, you cannot reconstruct the whole matrix, but given the matrix,
you easily figure out the diagonal sum.
T he linear structure for kets and bras is inherited by the trace. F or
example, consider
1 = 1a + 1b ,
(1.6 .4)
and compare the two ways of evaluating tr 1 2 ,
tr 1 2 = 21 = 21a + 21b
= tr 1a 2 + tr 1b 2 ,
n
o
1 a + 1 b 2
tr 1 2 = tr
(1.6 .5 )
= tr 1a 2 + 1b 2 .
T his generaliz es to
hAi = A
(1.6 .6 )
(1.6 .7 )
A
hAi = A = tr
(1.6 .8 )
|{ z } |{ z }
|{ z } |{ z }
b ra
k et
k et b ra
26
or as
,
hAi = A = tr A
|{ z } |{ z }
|{ z } |{ z }
b ra
k et
(1.6 .9 )
k et b ra
as the mathematical object that refers solely to the
which introduces
state of aff airs of the physical system,
n
o
n
o
.
hAi = tr
A = tr A
..
..
.
.
| { z } ....
... | { z }
.
..
..
..
..
..
.
...
.
...
....
....
.
..........
.
.......
.........
.......
........
........
......... ...............
....
operator
considered
(1.6 .10)
..
..
..
..
..
.
...
.
...
....
....
..
.........
What has been achieved here, is the complete separation of the mathematical entities that refer to the physical property (position, momentum,
functions of them, such as energy), and to the state of aff airs. T here is one
appropriate operator A for the physical property, irrespective of the particular state that is actually the case, and all that characteriz es the actual
situation is contained in the ket-bra
.
(1.6 .11)
T his is also an operator, but not one that describes a physical property,
rather it is the statistical operator that summariz es all statistical aspects
of the system as actually prepared. It is common to refer to the statistical
operator as the state operator or simply the state of the physical system.
In particular, we extract probabilities from as illustrated by the probability of finding the object between x = a and x = b (a < b). We recall
that
Z b
prob(a < x < b) =
dx (x) 2
Za
2
=
dx (x) a,b (x) ,
(1.6 .12)
where
a,b (x) =
if a < x < b ,
elsewhere,
(1.6 .13)
is the mesa func tion for the interval a < x < b. T hus,
Z
prob(a < x < b) =
dx x a,b (x) x
= a,b (X)
= tr a,b (X)
= tr{a,b (X)} ,
27
(1.6 .14)
,
=
(1.6 .15 )
tr{AB} = tr 1 23 4 = 41 23 ,
| { z } |{ z }
k et
b ra
tr{BA} = tr 3 41 2 = 23 41 .
| { z } |{ z }
k et
b ra
(1.6 .16 )
Indeed, they are the same because the numbers 41 and 23 can be
multiplied in either order without changing the value of the product. P lease
note that AB 6= BA, as a rule, but their traces are the same.
28
(1.6 .17 )
for products of three factors and analogous statements about four, five, . . .
factors. All of them are summariz ed in the observation that
the value of a trace does not change when the factors in a
product are permuted cyclically.
T his c yc lic property of th e trace is exploited very often.
, the prodSo far, letter was just an abbreviation for the ket-bra
uct of the ket and the bra describing the actual state of aff airs. L et us now
move on and consider a more general situation:
1 , w
2 , w
3 , w3
...................................................................................
....................................................................................
.....
...
...
...
...
... ...............
.
.
.
....................................
1
..
....
...
..................................
...
..............................................................................................................................................................................
.
..................
...
...
..
.......................................
...
.
...
...
.
.
.
....
.
..................................................................................
2
.
...........
..
........................................................................
source
measurement
stage
outcomes
1 , or in state
We
have
a
source
that
puts
out
the
atoms
either
in
state
2 , or in state 3 , whereby we have no clue what will be the case for the
next atom except that we know that there are definite probabilities w1 , w2 ,
w3 of occurrence for the three states, with w1 + w2 + w3 = 1, of course,
because nothing else can possibly happen.
At the measurement stage we perform a measurement of the physical
property that is associated with operator A. T hus we have an expectation
value that is given by
or
hAi = w1 1A1 + w.2 2A2 +w3 3A3
..
{.z }
.. |
..
..
..
..
..
.....
.....
expectation
value
probability
if 2 is the case
of getting 2
hAi = w1 tr A1 1 + w2 tr A2 2 + w3 tr A3 3
n
o
= tr A 1 w1 1 + 2 w2 2 + 3 w3 3
= tr{A}
(1.6 .18 )
(1.6 .19 )
29
(1.6 .20)
k= 1
T his
summariz es all that we know about the source: there are the states
k and their statistical weights wk .
where the summation can have one or more terms, and it is understood
that all kets and bras are normaliz ed properly,
(1.6 .22)
k k = 1 for all k .
B y measuring suffi ciently many diff erent physical properties, we can establish a body of experimental data that enables us to infer the statistical
operator with the desired precision. T hen we know the statistical properties of the atoms emitted by the source, and this isall we can find out. We
cannot, in particular, establish the ingredients k and their weights wk ,
we can only know . T his is also the only really meaningful thing to know
since it gives us all probabilities for the statistical predictions. N othing else
is needed. N or is anything else available.
When we state that knowledge of does not translate into knowledge
of the ingredients from which it is composed in
X
k wk k ,
(1.6 .23)
=
k
we mean of course that diff erent right-hand sides can give the same . T o
make this point it is quite enough to give one example. T he simplest is
1
1 1 + 2 2
with 12 = 0 ,
(1.6 .24)
=
2
that is just two states mixed with equal weights of 5 0% for each. In view
of their stated orthogonality, the kets
= 1 1 + 2 , = 1 1 2
(1.6 .25 )
2
2
30
establishes
1
+
2
!
1 1 + 2 1 + 2 1 2 1 2
+
=
2
2
2
2
2
1
=
1 1 + 2 2 =
2
1
+
=
2
(1.6 .26 )
(1.6 .27 )
which has diff erent ingredients than the original of (1.6 .24). We speak
of the two blends for one and the same mix ture or mix ed state. All that is
relevant is the mixture , not the particular ways in which one can blend
it.
F or
1
1
=
1 2 + 2 2 =
+
(1.6 .28 )
2
2
one can say that it is as if we had 5 0% of 1 and 5 0% of 2 or one can
say
with equal justification that it is as if we had 5 0% of and 5 0% of
. B ut neither as-if reality is better than the other, both are on exactly
the same footing, and there are many more as-if realities associated with
this .
1
1
1 +i 2 , v =
1 i2 and evaluate
1-13 C onsider u =
2
2
1
u u + v v . What do you conclude?
2
T he basic probability is that of finding a particular state, 0 , say. If
2
is
state k is the case, this probability
0k , as in (1.3.16 ), so more
generally the probability is 00 . It must be nonnegative,
00 0 for any choice of 0 .
(1.6 .29 )
In short: 0, which is a basic property of all statistical operators, their
positiv ity. O ther properties are that is hermitian,
= ,
(1.6 .30)
31
(1.6 .31)
T hen the probabilities of the various outcomes are an an which are
assuredly positive, and their sum must be 1:
X
X
tr an an
a n a n =
1=
n
X
an an = tr{} .
= tr
|
{z
=1
(1.6 .33)
T hat is: tr{} = 1 is just the statement that the probabilities of mutually
exclusive events have unit sum.
T he
situation of only one term is the one we started with,
extreme
. T hen it is possible to almost identify the ingredients. Almost
=
because of the phase arbitrariness,
i i
= e
,
=
e
(1.6 .34)
e , ei
( real)
(1.6 .35 )
is as good as the pair , . It is one
of the advantages of using the statistical operator rather than and that there is no phase arbitrariness
in . T he statistical operator
is unique, its ingredients are not.
is also special because, for such a pure state,
T he situation of =
it is characteristically true that 2 = ; see
2
=
=
|{ z }
1
=
= .
=
(1.6 .36 )
32
P
2
k k wk k then 6= , as is best
X
tr 2 =
tr j wj j k wk k
j,k
X
j,k
2
.
wj wk j k
(1.6 .37 )
N ow, since j j 6= k k if j 6= k (we want really diff erent ingredients),
2
< 1 for j 6= k, so that
we have j k
tr 2 <
X
j,k
X
X
wj wk =
wj
wk = 1 ,
or tr 2 < 1. T hus we have
{z
=1
}|
{z
=1
,
tr 2 = 1 if =
and tr 2 < 1 otherwise .
(1.6 .38 )
2 =
(1.6 .39 )
T herefore, the number tr 2 can serve as a crude measure of the
purity
, and
that
x A is a well defined bra and A p is a well defined ket, and that xAp
is a uniquely specified set of numbers once A is stated. T hese numbers
33
appear in
A = 1A1 =
=
We divide and multiply by
dx x xA
dp p p
dx dp x xAp p .
1
eixp/~ ,
xp =
2~
(1.7 .1)
(1.7 .2)
(1.7 .3)
where
xAp
a(x, p) =
xp
(1.7 .4)
f (X) = dx0 x0 f (x0 ) x0 ,
Z
g(P ) = dp0 p0 g(p0 ) p0
(1.7 .5 )
we have
x x =
and therefore
p p =
A=
dx0 x0 (x0 x) x0 = (X x) ,
dp0 p0 (p0 p) p0 = (P p)
dx dp (X x)a(x, p)(P p) .
(1.7 .6 )
(1.7 .7 )
34
where we must pay due attention to the structure of the previous expression
(1.7.7). There all Xs stand to the left of all P s, and this order must be
preserved when we replace x X, p P in a(x, p).
We have thus achieved even more than what we really needed. Operator
A is now expressed as an ordered function of X and P , for which the
procedure gives a unique answer. Of course, we can interchange the roles
of position and momentum in this argument and can equally well arrive at
a unique P, X-ordered form, where all P operators are to the left of all X
operators in products.
As a simple example, consider A = P X for which
xAp = xP X p
~
~
p
x
=
i x
i p
2
~
=
xp
i
x p
2 !
2
i
i
~
(1.7.9 )
xp xp ,
+
=
i
~
~
where the last step exploits the familiar explicit form of xp in (1.7.2).
Accordingly,
xAp
~
(1.7.10 )
a(x, p) =
= xp +
i
xp
here, and we get the X, P -ordered form
~
A = P X = xp +
= XP i~ .
i x X
(1.7.11)
pP
ordered
The result is, of course, as expected inasmuch as we just get Werner H eisenbergs fundamental com m u ta tion rela tion
[X, P ] = i~ .
(1.7.12)
35
f (X)
,
X
[X, g(P )] = i~
g(P )
,
P
(1.7.13 )
F (X, P )
,
X
[X, F (X, P )] = i~
F (X, P )
,
P
1-16
1-17
f (X, P ) and
f (X, P )?
X P
P X
1 2 2
X ,P .
F ind the X, P -ordered form of the commutator
Is there a diff erence between
i~
, is our next
A simple statistical operator for a pure state, =
example. We have
p = (x)(p)
(1.7.14 )
xp = x
and then
so that
x
xp
=
= 2~ (x) eixp/~(p)
xp
xp
=
(1.7.15 )
(1.7.16 )
H ere
e
iX; P/~
= e
iXP/~
X
1
=
k!
X,P -ordered
k
i
k k
X P
(1.7.17)
(1.7.18)
k= 0
36
= 2~ (P ) eiP ; X/~(X)
(1.7.19 )
= dx dp xAp px
Z
xAp
= dx dp
xp px
(1.7.20 )
xp
| {z } | {z
}
a(x, p) =
= 1/(2~)
so that
tr{A} =
dx dp
a(x, p) .
2~
(1.7.21)
(1.7.22)
(1.7.23 )
Then we have
hAi = tr{A}
Z
Z
= tr
dp p p dx x xA
Z
= dx dp px xAp ,
px = pr(P ; X )x = r(p, x) px
p x
(1.7.24 )
(1.7.25 )
37
and
xAp = xa(X ; P )p = a(x, p) xp .
(1.7.26 )
x p
These just exploit the basic property of ordered operators, namely that X
and P stand next to their respective eigenbras and eigenk ets so that these
operators can be equivalently replaced by their eigenvalues. Then,
Z
(1.7.27)
hAi = dx dp r(p, x)a(x, p) px xp
}
| {z
= 1/(2~)
or
hAi =
dx dp
r(p, x)a(x, p) ,
2~
(1.7.28)
which look s even more lik e a classical phase-space integral of the product of
dx dp
W e y l c o m m u ta to r, B a k e r C a m p b e ll H a u sd o rff re la tio n s
0
The unitary operators eip X/~ and eix P/~ of (1.5 .14 ) do not commute,
except for special values of x0 and p0 , and we can fi nd out what is the
diff erence between applying them in either order by establishing the X, P ordered version of the operator
0
(1.8.1)
38
ix0 P/~ ip0 X/~
p
xAp = x e
e
| { z } | { z }
x + x0 =
= p + p 0
0
0
1
= x + x0 p + p0 =
ei(x + x )(p + p )/~
2~
ixp/~
0
0
0 0
e
eixp /~ eix p/~ eix p /~ ,
=
2~
| { z }|
{z
}
= xp
(1.8.2)
= a(x, p)
0 0
0 0
(1.8.3 )
(1.8.4 )
(1.8.5 )
(1.8.6 )
and then it is true for all polynomials and fi nally for arbitrary functions.
1-19 If a is an eigenk et of A, Aa = a a, then f (A)a = a f (a).
Why? C onclude that U a is an eigenk et of f (U AU ).
39
(1.8.7)
eig(P ) X eig(P ) = X ~g 0 (P ) ,
(1.8.8)
i(p0 X + x0 P )/~
= e
`
ip0 X +
x0
P
p0
/~
for
p0 6= 0
(1.8.9 )
(1.8.10 )
and therefore
e
i(p0 X + x0 P )/~
i x0 2
i x0 2
2~
P
0 2~ p0 P
0
p
= exp ip e
Xe
/~
= e
i x0 2
P
2~ p0
x0 2
P
p0
i
ip0 X/~ 2~
(1.8.11)
i(p0 X + x0 P )/~
= e
i x0 2
P
2~ p0
i
ip0 X/~ 2~
x0 2
P
p0
i
ip0 X/~ 2~
x0 2
P
p0
ip0 X/~
i x0 ip0 X/~ ip0 X/~2
e
= exp
Pe
2~ p0
(1.8.13 )
wherein
0
(1.8.14 )
implying
0
0
ei(p X + x P )/~
= e
i x0 2
P
2~ p0
i
2~
x0
(P
p0
p0 )2 ip0 X/~
e
.
(1.8.15 )
40
The fi rst and second exponentials on the right are functions of P only and
so there is no problem in combining them into one,
e
i x0 2
P
2~ p0
i
2~
x0
(P
p0
p0 )2
= e
i x0
2~ p0
i x0
= e 2~ p0
[P 2 (P 2 p0 )2 ]
(2 P p0 )p0
i
0 0
= eix P/~ e 2 x p /~ .
(1.8.16 )
Accordingly,
0
0 0
0 0
(1.8.17)
where the second equality is that of the right-hand sides in (1.8.1) and
(1.8.2). These are examples of the famous B a k er C a m p bell H a u sdorff rela tion s among exponential functions of operators, named after H enry F .
B ak er, J ohn E . C ampbell, and F elix H ausdorff .
0
0
0
0
1-2 0 What is x ei(p X + x P )/~? What is ei(p X + x P )/~p ? Introduce
the operator
Z
dx0 dp0 i(p0 X + x0 P )/~
e
R=
2~
and fi nd xR and Rp .
1-2 1 S how that
n
o
0
0
00
00
tr ei(p X + x P )/~ ei(p X + x P )/~ = 2~(x0 x00 )(p0 p00 ) .