Beruflich Dokumente
Kultur Dokumente
Broida
QM II
Angular Momentum
As is the case with most operators in quantum mechanics, we start from the classical definition and make the transition to quantum mechanical operators via the
standard substitution x x and p i~. Be aware that I will not distinguish
a classical quantity such as x from the corresponding quantum mechanical operator
x. One frequently sees a new notation such as x
used to denote the operator, but
for the most part I will take it as clear from the context what is meant. I will also
generally use x and r interchangeably; sometimes I feel that one is preferable over
the other for clarity purposes.
Classically, angular momentum is defined by
L = r p.
Since in QM we have
[xi , pj ] = i~ij
it follows that [Li , Lj ] 6= 0. To find out just what this commutation relation is, first
recall that components of the vector cross product can be written (see the handout
Supplementary Notes on Mathematics)
(a b)i = ijk aj bk .
Here I am using a sloppy summation convention where repeated indices are summed
over even if they are both in the lower position, but this is standard when it comes
to angular momentum. The Levi-Civita permutation symbol has the extremely
useful property that
ijk klm = il jm im jl .
Also recall the elementary commutator identities
[ab, c] = a[b, c] + [a, c]b
and
Using these results together with [xi , xj ] = [pi , pj ] = 0, we can evaluate the commutator as follows:
[Li , Lj ] = [(r p)i , (r p)j ] = [ikl xk pl , jrs xr ps ]
= ikl jrs [xk pl , xr ps ] = ikl jrs (xk [pl , xr ps ] + [xk , xr ps ]pl )
= ikl jrs (xk [pl , xr ]ps + xr [xk , ps ]pl ) = ikl jrs (i~lr xk ps + i~ks xr pl )
= i~ikl jls xk ps + i~ikl jrk xr pl = +i~ikl ljs xk ps i~jrk kil xr pl
= i~(ij ks is jk )xk ps i~(ji rl jl ri )xr pl
= i~(ij xk pk xj pi ) i~(ij xl pl xi pj )
= i~(xi pj xj pi ) .
But it is easy to see that
ijk Lk = ijk (r p)k = ijk krs xr ps = (ir js is jr )xr ps
= xi pj xj pi
and hence we have the fundamental angular momentum commutation relation
[Li , Lj ] = i~ijk Lk .
(1.1a)
[Ly , Lz ] = i~Lx
[Lz , Lx ] = i~Ly .
(1.1b)
(1.2b)
Remarkably, this is all we need to compute the most useful properties of angular
momentum.
To begin with, let us define the ladder (or raising and lowering) operators
J+ = Jx + iJy
J = (J+ ) = Jx iJy .
(1.3a)
1
(J+ + J )
2
and
Jy =
1
(J+ J ) .
2i
(1.3b)
(1.4)
In addtion, we have
[Jz , J ] = [Jz , Jx ] i[Jz , Jy ] = i~Jy ~Jx
so that
[Jz , J ] = ~J .
(1.5a)
Furthermore,
2
2
[Jz , J
] = J [Jz , J ] + [Jz , J ]J = 2~J
(1.5b)
(1.6a)
(1.6b)
and
Jz Y = ~Y .
(1.8)
and
J Yl = 0
and
4
l2 .
(1.9b)
n
:= j
2
(1.11)
(1.12)
We say that the states Yjm are angular momentum eigenstates with angular momentum j and z-component of angular momentum m. Note that (1.9b) is now
written
and
J Yjj = 0 .
(1.13)
J+ Yjj = 0
5
or
1
Yjj1 = ~1 J Yjj .
2j
Next we have
(J )2 Yjj = ~2
or
p
p p
2j j(j + 1) (j 1)(j 2) Yjj2 = ~2 (2j)2(2j 1) Yjj2
1
Yjj2 = ~2 p
(J )2 Yjj .
(2j)(2j 1)2
or
1
Yjj3 = ~3 p
(J )3 Yjj .
2j(2j 1)(2j 2)3!
And an exactly analogous argument starting with Yjj and applying J+ repeatedly
shows that we could also write
s
(j m)!
Yjm = ~mj
(J+ )j+m Yjj .
(1.15b)
(2j)!(j + m)!
It is extremely important to realize that everything we have done up to this
point depended only on the commutation relation (1.2a), and hence applies to both
integer and half-integer angular momenta. While we will return in a later section to
discuss spin (including the half-integer case), for the rest of this section we restrict
ourselves to integer values of angular momentum, and hence we will be discussing
orbital angular momentum.
The next thing we need to do is to actually construct the angular momentum
wave functions Ylm (, ). (Since we are now dealing with orbital angular momentum, we replace j by l.) To do this, we first need to write L in spherical coordinates.
One way to do this is to start from Li = (rp)i = ijk xj pk where pk = i~(/xk ),
and then use the chain rule to convert from Cartesian coordinates xi to spherical
coordinates (r, , ). Using
x = r sin cos
y = r sin sin
z = r cos
so that
r = (x2 + y 2 + z 2 )1/2
= cos1 z/r
= tan1 y/x
=
+
+
x
x r x
x
=
xz
y
+ 3
2 cos2
r r r sin x
= sin cos
cos cos
sin
r
r
r sin
with similar expressions for /y and /z. Then using terms such as
Lx = ypz zpy = i~ y
z
z
y
we eventually arrive at
cot cos
Lx = i~ sin
cot sin
Ly = i~ cos
Lz = i~
(1.16a)
(1.16b)
(1.16c)
However, another way is to start from the gradient in spherical coordinates (see
the section on vector calculus in the handout Supplementary Notes on Mathematics)
=
r
1
1
+
+
.
r
r
r sin
are
Then L = r p = i~ r = i~ r (
r ) so that (since
r, and
orthonormal)
1
1
L = i~r
r
r
+
r
+
r
r
r
r sin
1
= i~
sin
If we write the unit vectors in terms of their Cartesian components (again, see the
handout on vector calculus)
= (cos cos , cos sin , sin )
= ( sin , cos , 0)
then
+y
cos
+
z
cot cos
cot sin
L = i~ x
sin
.
(1.17)
i cot
L = ~ei
To find the eigenfunctions Ylm (, ), we start from the definition L+ Yll = 0. This
yields the equation
Yll
Y l
+ i cot l = 0 .
which has the solution F () = eik (up to normalization). But Yll is an eigenfunction of Lz = i~(/) with eigenvalue l~, and hence so is F () (since T () just
cancels out). This means that
i~
eik
= k~eik := l~eik
(1.18)
I will go through all the details involved in doing this integral. You are free to skip
down to the result if you wish (equation (1.21)), but this result is also used in other
physical applications.
First I want to prove the relation
Z
Z
n1
1
n1
n
x cos x +
sinn2 x dx .
(1.19)
sin x dx = sin
n
n
R
by parts (remember the formula u dv = uv
RThis is done as an integration
v du) letting u = sinn1 x and dv = sin x dx so that v = cos x and du =
(n 1) sinn2 x cos x dx. Then (using cos2 x = 1 sin2 x in the third line)
Z
Z
n
sin x dx = sinn1 x sin x dx
n1
= sin
x cos x + (n 1)
= sinn1 x cos x + (n 1)
sinn2 x cos2 x dx
sinn2 x dx (n 1)
sinn x dx .
Now move the last term on the right over to the left, divide by n, and the result is
(1.19).
9
We need to evaluate (1.19) for the case where n = 2l + 1. To get the final result
in the form we want, we will need the basically simple algebraic result
(2l + 1)!! =
(2l + 1)!
2l l!
l = 1, 2, 3, . . .
(1.20)
1 2 3 4 5 6 7 8 9 (n 4) (n 3) (n 2) (n 1) n
2 4 6 8 (n 3) (n 1)
1 2 3 4 5 6 7 8 9 (n 4) (n 3) (n 2) (n 1) n
n1
(2 1)(2 2)(2 3)(2 4) (2 n3
2 )(2 2 )
n!
2
n1
2
( n1
2 )!
=2
(2l l!)2
2l l!
=2
(2l + 1)!!
(2l + 1)!
10
(1.21)
R
where we used 0 sin x dx = 2 and (1.20).
Using this result, (1.18) becomes
and hence
2 (2l l!)2
4 cll
=1
(2l + 1)!
cll = (1)l
(2l + 1)!
4
1/2
1
2l l!
(1.22)
where we included a conventional arbitrary phase factor (1)l . Putting this all
together, we have the top orbital angular momentum state
1/2
1
l
l (2l + 1)!
(sin )l eil .
(1.23)
Yl (, ) = (1)
4
2l l!
To construct the rest of the states Ylm (, ), we repeatedly apply L from equation (1.17) to finally obtain
1/2
1/2
1
(l + m)!
(2l + 1)!
Ylm (, ) = (1)l
4
2l l! (2l)!(l m)!
eim (sin )m
dlm
(sin )2l .
d(cos )lm
(1.24)
Spin
It is an experimental fact that many particles, and the electron in particular, have
an intrinsic angular momentum. This was originally deduced by Goudsmit and Uhlenbeck in their analysis of the famous sodium D line, which arises by the transition
from the 1s2 2s2 2p6 3p excited state to the ground state. What initially appears as a
strong single line is slightly split in the presence of a magnetic field into two closely
spaced lines (Zeeman effect). This (and other lines in the Na spectrum) indicates a
doubling of the number of states available to the valence electron.
To explain this fine structure of atomic spectra, Goudsmit and Uhlenbeck
proposed in 1925 that the electron possesses an intrinsic angular momentum in
addition to the orbital angular momentum due to its motion about the nucleus.
Since magnetic moments are the result of current loops, it was originally thought
that this was due to the electron spinning on its axis, and hence this intrinsic angular
momentum was called spin. However, a number of arguments can be put forth to
disprove that classical model, and the result is that we must assume that spin is a
purely quantum phenomena without a classical analogue.
As I show at the end of this section, the classical model says that the magnetic
moment of a particle of charge q and mass m moving in a circle is given by
q
=
L
2mc
11
(2.1a)
2
S |s ms i = s(s + 1)~ |s ms i
(2.1b)
Sz |s ms i = ms ~|s ms i
p
S |s ms i = ~ s(s + 1) ms (ms 1) |s ms 1i .
(2.1c)
(2.1d)
|i = c+ |
z +i + c |
z i .
(2.2a)
(I am switching to the more abstract notation for the eigenstates because the states
themselves are a rather abstract concept.) We will sometimes drop the subscript s
on ms if there is no danger of confusing this with the eigenvalue of Lz , which we
will also sometimes write as ml . Be sure to realize that particles can have a spin s
that is any integer multiple of 1/2, and there are particles in nature that have spin
0 (e.g., the pion), spin 1/2 (e.g., the electron, neutron, proton), spin 1 (e.g., the
photon, but this is a little bit subtle), spin 3/2 (the s), spin 2 (the hypothesized
graviton) and so forth.
Since the z component of electron spin can take only one of two values ~/2, we
will frequently denote the corresponding orthonormal eigenstates simply by |
z i,
where |
z +i is called the spin up state, and |
z i is called the spin down state.
An arbitrary spin state |i is of the form
12
With this choice of basis, we can now construct the 2 2 matrix representation of
the spin operator S.
Note that the existence of spin has now led us to describe the electron by a
multi-component state vector (in this case, two components), as opposed to the
scalar wave functions we used up to this point. These two-component states are
frequently called spinors.
(z)
The states were specifically constructed to be eigenstates of Sz (recall that
we simultaneously diagonalized J 2 and Jz ), and hence the matrix representation
of Sz is diagonal with diagonal entries that are precisely the eigenvalues ~/2. In
other words, we have
~
h
z |Sz |
z i =
2
so that
"
#
0
~ 1
Sz =
.
(2.4)
2 0 1
(We are being somewhat sloppy with notation and using the same symbol Sz to
denote both the operator and its matrix representation.) That the vectors defined
in (2.3) are indeed eigenvectors of Sz is easy to verify:
"
" #
"
" #
#" #
#" #
0
0
1
0
~ 1
~ 1
~ 1
~ 0
and
.
=
=
2 0 1
2 0
2 0 1
2 1
0
1
To find the matrix representations of Sx and Sy , we use (2.1d) together with
S = Sx iSy so that Sx = (S+ + S )/2 and Sy = (S+ S )/2i. From
p
z ms 1i
S |
z ms i = ~ 3/4 ms (ms 1) |
we have
S+ |
z +i = 0
S |
z +i = ~|
z i
S+ |
z i = ~|
z +i
S |
z i = 0 .
and
13
~
Sy =
2
"
0 i
i
(2.5)
And from (2.4) and (2.5) it is easy to calculate S 2 = Sx2 + Sy2 + Sz2 to see that
"
#
3 2 1 0
2
S = ~
4
0 1
which agrees with (2.1b).
It is conventional to write S in terms of the Pauli spin matrices defined by
S=
where
x =
"
0 1
1 0
y =
"
0
i
i
0
z =
"
(2.6)
Memorize these. The Pauli matrices obey several relations that I leave to you to
verify (recall that the anticommutator is defined by [a, b]+ = ab + ba):
[i , j ] = 2iijk k
i j = iijk k
(2.7a)
for i 6= j
[i , j ]+ = 2Iij
i j = Iij + iijk k
(2.7b)
(2.7c)
(2.7d)
Given three-component vectors a and b, equation (2.7d) also leads to the extremely
useful result
(a )(b ) = (a b)I + i(a b) .
(2.8)
(x)
"
#" #
c
d
(2.9a)
=0
=
.
1
a/ 2 + b/ 2
a 1
b
=
+
=
2 1
2 1
a/ 2 b/ 2
or
= a/ 2 + b/ 2
= a/ 2 b/ 2 .
and
1
b = ( )
2
and
(x)
+ +
15
(x)
(2.10)
"
so that the energies are unchanged. However, because of the spin function, we have
doubled the number of states corresponding to a given energy.
A more mathematically correct way to write these complete states is as the
tensor (or direct) product
|i = |nlml (ms )i := |nlml i |(ms )i .
In this case, the Hamiltonian is properly written as H I where H acts on the
vector space of spatial wave functions, and I is the identity operator on the vector
space of spin states. In other words,
(H I)(|nlml i |(ms )i) := H|nlml i I|(ms )i .
This notation is particularly useful when treating two-particle states, as we will see
when we discuss the addition of angular momentum.
(Remark : You may recall from linear algebra that given two vector spaces V
and V , we may define a bilinear map V V V V that takes ordered pairs
(v, v ) V V and gives a new vector denoted by v P
v . Since this map
Pis bilinear
by definition,
if
we
have
the
linear
combinations
v
=
x
v
and
v
=
yj vj then
i i
P
vv =
xi yj (vi vj ). In particular, if V has basis {ei } and V has basis {ej },
then {ei ej } is a basis for V V which is then of dimension (dim V )(dim V )
and called the direct (or tensor) product of V and V . Then, if we are given
two operators A L(V ) and B L(V ), the direct product of A and B is the
operator A B defined on V V by (A B)(v v ) := A(v) B(v ).)
2.1
I
qv
qvr
area =
r2 =
.
c
2rc
2c
But the angular momentum of the particle is L = mvr so we conclude that the
magnetic moment due to orbital motion is
l =
q
L.
2mc
(2.11)
e
S.
2mc
17
The constant g is found by experiment to be very close to 2. (However, the relativistic Dirac equation predicts that g is exactly 2. Higher order corrections in
quantum electrodynamics predict a slightly different value, and the measurement
of g 2 is one of the most accurate experimental result in all of physics.)
Now we want to show is that the energy of a magnetic moment in a uniform
magnetic field is given by B where for a loop of area A carrying current
I is defined to have magnitude IA and pointing perpendicular to the loop in the
direction of your thumb if the fingers of your right hand are along the direction of
the current. To see this, we simply calculate the work required to rotate a current
loop from its equilibrium position to the desired orientation.
Consider the figure shown below, where the current flows counterclockwise out
of the page at the bottom and into the page at the top.
a/2
FB
FB
a/2
B
Let the loop have length a on the sides and b across the top and bottom, so its area
is ab. The magnetic force on a current-carrying wire is
Z
FB = Idl B
and hence the forces on the opposite a sides of the loop cancel, and the force on
the top and bottom b sides is FB = IbB. The equilibrium position of the loop is
horizontal, so the potential energy of the loop is theR work required to rotate it from
= 0 to some value . This work is given by W = F dr where F is the force that
I must apply against the magnetic field to rotate the loop.
Since the loop is rotating, the force I must apply at the top of the loop is in the
direction of and perpendicular to the loop, and hence has magnitude FB cos .
Then the work I do is (the factor of 2 takes into account both the top and bottom
sides)
Z
Z
Z
W = F dr = 2 FB cos (a/2)d = IabB
cos d = B sin .
0
(2.12)
In this derivation, I never explicitly mentioned the torque on the loop due to B.
However, we see that
kNk = kr FB k = 2(a/2)FB sin(90 + ) = IabB sin(90 + )
= B sin(90 + ) = k Bk
and therefore
Note that W =
N = B.
(2.13)
Mathematical Digression:
Rotations and Linear Transformations
Lets take a look at how the spatial rotation operator is defined. Note that there
are two ways to view symmetry operations such as translations and rotations. The
first is to leave the coordinate system unchanged and instead move the physical
system. This is called an active transformation. Alternatively, we can leave the
physical system alone and change the coordinate system, for example by translation
or rotation. This is called a passive transformation. In the case of an active
transformation, we have the following situation:
x2
x1
Here the vector r is rotated by to give the vector r where, of course, krk = krk = r.
We define a linear transformation T by r = T (r). (This is linear because it is easy
to see that rotating the sum of two vectors is the sum of the rotated vectors.) From
the diagram, the components of r are given by
x
1 = r cos( + ) = r cos cos r sin sin
= (cos )x1 (sin )x2
19
x
2 = r sin( + ) = r sin cos + r cos sin
= (sin )x1 + (cos )x2
or
"
x1
x2
"
cos
sin
sin
cos
#"
x1
x2
(3.1)
But T ei is just another vector in R , and hence it can be expressed in terms of the
basis {ei } as
X
ej aji .
(3.2)
T ei =
j
Be sure to note which indices are summed over in this expression. The matrix (aji )
is called the matrix representation of the linear transformation T with respect
to the basis {ei }. You will sometimes see this matrix written as [T ]e .
It is very important to realize that a linear transformation T takes the ith basis
vector into the ith column of its matrix representation. This is easy to see if we
write out the components of (3.2). Simply note that with respect to the basis {ei },
we have
" #
" #
1
0
e1 =
and
e2 =
0
1
and therefore
T ei = e1 a1i + e2 a2i =
" #
1
0
a1i +
" #
0
1
a2i =
"
a1i
a2i
3
1 0
[T ]v = 0 2 1 .
1 1 1
every w W . Then we say that W is T-invariant (or simply invariant when the
operator is understood).
What can we say about the matrix representation of T under these circumstances? Well, let W have the basis {w1 , . . . , wr }, and extend this to a basis
{w1 , . . . , wr , v1 , . . . , vnr } for V . Since T w W for any w W , we must have
T wi =
r
X
wj aji
j=1
for some set of scalars {aji }. But for any vi there is no such restriction since all we
know is that T vi V . So we have
T vi =
r
X
wj bji +
j=1
nr
X
vk cki
k=1
for scalars {bji } and {cki }. Then since T takes the ith basis vector to the ith column
of the matrix representation [T ], we must have
"
#
A B
[T ] =
0 C
where A is an r r matrix, B is an r (n r) matrix, and C is an (n r) (n r)
matrix. Such a matrix is said to be a block matrix, and [T ] is in block triangular
form.
Now let W be an invariant subspace of V . If it so happens that the subspace
of V spanned by the rest of the vectors {v1 , . . . , vnr } is also invariant, then the
matrix representation of T will be block diagonal (because all of the bji in the above
expansion will be zero). As we shall see, this is in fact what happens when we add
two angular momenta J1 and J2 and look at the representation of the rotation
operator with respect to the total angular momentum states (where J = J1 + J2 ).
By choosing our states to be eigenstates of J 2 and Jz rather than J1z and J2z , the
rotation operator becomes block diagonal rather than a big mess. This is because
rotations dont change j, so for fixed j, the (2j + 1)-dimensional space spanned by
{|j ji, |j j 1i, . . . , |j ji} is an invariant subspace under rotations. This change
of basis is exactly what Clebsch-Gordan coefficients accomplish.
Let us go back to our specific example of rotations. If we define r = T (r), then
on the the one hand we have
X
r =
xj ej
j
21
Since the ej are a basis, they are linearly independent, and we can equate these last
two equations to conclude that
X
aji xi .
(3.3)
xj =
i
"
cos
sin
sin
cos
as in (3.1).
Using (3.3) we can make another extremely important observation.
Since
P the
P
j x
j = i xi xi .
length of a vector is unchanged under rotations, we must have j x
But from (3.3) we see that
X
X
x
j x
j =
aji ajk xi xk
j
X
j
i,k
aji ajk =
aTij ajk = ik .
(3.4a)
It is also quite useful to realize that an orthogonal matrix is one whose rows (or
columms) form an orthonormal set of vectors. If we let Ai denote the ith row of A,
then from (3.4b) we have
X
aij akj = ik .
Ai Ak =
j
22
Similarly, if we let Ai denote the ith column of A, then (3.4a) may be written
X
aji ajk = ik .
Ai Ak =
j
Conversely, it is easy to see that any matrix whose rows (or columns) form an
orthonormal set must be an orthogonal matrix.
the above arguments using kxk2 =
P In the case of complex matrices, repeating
T
i xi xi , it is not hard to show that A A = AA = I where A = A . In this case,
the matrix A is said to be unitary. Thus a complex matrix is unitary if and only if
its rows (or columns) form an orthonormal set under the standard Hermitian inner
product.
To describe a passive transformation, consider a linear transformation P acting
on the basis vectors. In other words, we perform a change of basis defined by
X
i = P (ei ) =
e
ej pji .
j
j a
e
ji =
j,k
(P ej )
aji =
j,k
ek pkj a
ji .
j,k
Equating these two expressions and using the linear independence of the ek we have
X
X
akj pji
pkj a
ji =
j
This extremely important equation relates the matrix representation of an operator A with respect to a basis {ei } to its representation A with respect to a basis
i = P ei . This is called a similarity transformation. In fact, this
{
ei } defined by e
is exactly what you do when you diagonalize a matrix. Starting with a matrix A
relative to a given basis, you first find its eigenvalues, and then use these to find the
corresponding eigenvectors. Letting P be the matrix whose columns are precisely
these eigenvectors, we then have P 1 AP = D where D is a diagonal matrix with
diagonal entries that are just the eigenvalues of A.
4
4.1
Before we begin with the physics, let me first prove an extremely useful mathematical result. For any complex number (or matrix) x, let fn (x) = (1 + x/n)n , and
consider the limit
x n
L = lim fn (x) = lim 1 +
.
n
n
n
Since the logarithm is a continuous function, we can interchange limits and the log,
so we have
x n
x n
ln L = ln lim 1 +
= lim ln 1 +
n
n
n
n
x
ln(1 + x/n)
= lim n ln 1 +
= lim
.
n
n
n
1/n
As n , both the numerator and denominator go to zero, so we apply
lHopitals rule and take the derivative of both numerator and denominator with
respect to n:
x
(x/n2 )/(1 + x/n)
= lim
= x.
2
n
n
1/n
1 + x/n
ln L = lim
(4.1)
X
An
n!
n=0
converges, and we take this as the definition of eA . (For a proof, see one of my
linear algebra books or almost any book on elementary real analysis or advanced
calculus.)
24
Now let us return to physics. First consider the translation of a wave function
(x) by an amount a.
(x)
(x)
x+a
d
.
dx
(4.3)
collection of objects together with a composition law obeying closure, the existence
of an identity, and the existence of an inverse to every object in the collection is
said to form a group, and this group in particular is called the translation group.
The composition of group elements is called group multiplication.
In the case of the translation group, is also important to realize that Ta Tb =
Tb Ta . Any group with the property that multiplication is commutative is called
an abelian group. Another example of an abelian group is the set of all rotations
about a fixed axis in R3 . (But the set of all rotations in three dimensions is most
definitely not abelian, as we are about to see.)
Now that we have shown that the linear momentum operator p generates translations, let us show that the angular momentum operator is the generator of rotations.
In particular, we will show explicitly that the orbital angular momentum operator
L generates spatial rotations in R3 . In the case of the spin operator S which acts
on the very abstract space of spin states, we will define rotations by analogy.
Let me make a slight change of notation and denote the rotated position vector
by a prime instead of a bar. If we rotate a vector x in R3 , then we obtain a new
vector x = R()x where R() is the matrix that represents the rotation. In two
dimensions this is
" # "
#" #
x
cos sin
x
=
.
y
sin
cos
y
If we have a scalar wavefunction (x), then under rotation we obtain a new wavefunction R (x), where (x) = R (R()x) = R (x ). (See the figure below.)
R (x)
x
(x)
i
1 Lz (x) .
~
(4.4)
and
Jz |jmi = m|jmi
and therefore
hj m |U (R)|jmi = 0
unless j = j .
(4.5)
We also make note of the fact that acting with J 2 and Jz in both directions yields
hj m |J 2 |jmi = j (j + 1)hj m |jmi = j(j + 1)hj m |jmi
and
hjm |Jz |jmi = m hjm |jmi = mhjm |jmi
so that (as you should have already known)
hj m |jmi = j j m m .
(4.6)
In other words, the states |jmi form a complete orthonormal set, and the state
U (R)|jmi must be of the form
X
X
(j)
U (R)|jmi =
|jm ihjm |U (R)|jmi =
|jm iDm m ()
(4.7)
m
where
(j)
(4.8)
(Notice the order of susbscripts in the sum in equation (4.7). This is the same as
the usual definition of the matrix
P representation [T ]e = (aij ) of a linear operator
T : V V defined by T ei = j ej aji .)
Since for each j there are 2j + 1 values of m, we have constructed a (2j +
1) (2j + 1) matrix D (j) () for each value of j. This matrix is referred to as the
jth irreducible representation of the rotation group. The word irreducible
means that there is no subset of the space of states {|jji, |j, m 1i, . . . , |j, ji}
that transforms into itself under all rotations U (R()). Put in another way, a
representation is irreducible if the vector space on which it acts has no invariant
subspaces.
Now, it is a general result of the theory of group representations that any representation of a finite group or compact Lie group is equivalent to a unitary representation, and any reducible unitary representation is completely reducible. Therefore,
any representation of a finite group or compact Lie group is either already irreducible or else it is completely reducible (i.e., the space on which the operators act
can be put into block diagonal form where each block corresponds to an invariant
subspace). However, at this point we dont want to get into the general theory
of representations, so let us prove directly that the representations D (j) () of the
rotation group are irreducible.
Recall that the raising and lowering operators J are defined by
p
J |jmi = (Jx iJy )|jmi = j(j + 1) m(m 1) |j, m 1i .
In particular, the operators J dont change the value of j when acting on the states
|jmi. (This is just the content of equation (1.4).)
28
Theorem 4.1. The representations D (j) () of the rotation group are irreducible.
In other words, there is no subset of the space of states |jmi (for fixed j) that
transforms among itself under all rotations.
Proof. Fix j and let V be the space spanned by the 2j + 1 vectors |jmi := |mi.
We claim that V is irreducible with respect to rotations U (R). This means that
given any |ui V , the set of all vectors of the form U (R)|ui (i.e., for all rotations
U (R)) spans V . (Otherwise, if there exists |vi such that {U (R)|vi} didnt span V ,
then V would be reducible since the collection of all such U (R)|vi would define an
invariant subspace.)
To show V is irreducible, let Ve = Span{U (R)|ui} where |ui V is arbitrary but
fixed. For infinitesimal we have U (R()) = eiJ = 1 i J and in particular
U (R(b
x)) = 1 iJx and U (R(b
y)) = 1 iJy . Then
1
1
J |ui = (Jx iJy )|ui =
[1 U (R(b
x))] i
[1 U (R(b
y))]
|ui
i
i
=
1
[1 U (R(b
y))] i + iU (R(b
x)) |ui Ve
Since hm|ui =
6 0 by assumption, it follows that |m + 1i Ve .
We can continue to act on |ui with J+ a total of j m times at which point we
will have shown that |m + j mi = |ji := |jji Ve . Now we can apply J 2j + 1
times to |jji to conclude that the 2j + 1 vectors |jmi all belong to Ve , and thus
Ve = V . (This is because we have really just applied the combination of rotations
(J )2j+1 (J+ )jm to |ui, and each step along the way is just some vector in Ve .)
As the last subject of this section, let us go back and show that e(i/~)S really
represents a rotation in spin space. Let R denote the unitary rotation operator, and
29
AR = RAR = RAR1 .
(4.9)
(This is just the generalization of (2.1c) for the case of spin one-half. It also applies
b m
b ms i = ms |m
b ms i.)
to arbitrary spin if we write (S m)|
b by to obtain another unit vector n
b , and
Let us rotate the unit vector m
consider the operator
b (i/~)S .
e(i/~)S (S m)e
b i we have
Acting on the state e(i/~)S |m
b (i/~)S ]e(i/~)S |m
b i = e(i/~)S (S m)|
b m
b i
[e(i/~)S (S m)e
~
b i .
= e(i/~)S |m
2
b i
|b
n i := e(i/~)S |m
(4.10a)
b := e(i/~)S (S m)e
b (i/~)S
Sn
(4.10b)
~
b )|b
n i
(S n
n i = |b
2
(4.10c)
as it should. This shows that e(i/~)S does indeed act as the rotation operator on
the abstract spin states.
30
b i of the spin
What we have shown then, is that starting from an eigenstate |m
b we rotate by an angle to obtain the rotated state |b
operator S m,
n i that is an
b with the same eigenvalues.
eigenstate of S n
For spin one-half we have S = /2 (with ~ = 1), and using (2.8), you can show
that the rotation operator becomes
ei/2 =
X
(i /2)n
n!
m=0
= I cos
i b sin .
2
2
(4.11)
Example 4.1. Let us derive equations (2.9) using our rotation formalism. We start
b =
b=x
by letting m
z and n
, so that = (/2)
y. Then the rotation operator is
iy sin
4
4
"
"
#
cos /4 sin /4
1 1
=
=
2 1
sin /4
cos /4
1
1
"
b +i = |
|m
z +i
1 1
1
#" #
1
0
1
=
2
" #
1
1
= |
x +i
e
z e
=
2 1
2 1 1
1
0 1
"
# "
#
0 1
1 0 2
=
=
= x
2 2 0
1 0
as it should.
31
4.2
q
L.
2mc
This relation is also true in quantum mechanics for orbital angular momentum, but
for spin angular momentum we must write
s = g
q
S
2mc
(4.12)
gq~
gq
S B(t) =
B(t) .
2mc
4mc
(4.13)
To understand how spin behaves, we look at this motion from the point of view of
the Heisenberg picture, which we now describe. (See Griffiths, Example 4.3 for a
somewhat different approach.)
The formulation of quantum mechanics that we have used so far is called the
Schr
odinger picture (SP). In this formulation, the operators are independent of
time, but the states (wave functions) evolve in time. The stationary state solutions
to the Schrodinger equation
H|(t)i = i~
|(t)i
t
(4.14)
(4.15a)
(4.15b)
i
O iHt/~
[H, OH ] + eiHt/~
e
.
~
t
If O has no explicit time dependence (for example, the operator pt has explicit time
dependence), then this simplifies to the Heisenberg equation of motion
i
dOH
= [H, OH ] .
dt
~
(4.16)
gq
dS(t)
=
S(t) B(t) = s (t) B(t) .
dt
2mc
33
(4.17)
This is just the operator version of the classical equation that says the time rate of
change of angular momentum is equal to the applied torque. From this equation we
see that the spin of a positively charged particle (q > 0) will precess in the negative
sense about B, while the spin of a negatively charged particle (q < 0) will precess
in the positive sense.
B
S
s
dSy (t)
gqB0
=
Sx (t)
dt
2mc
dSz (t)
= 0 . (4.18)
dt
gqB0
2mc
we can combine the Sx and Sy equations to obtain
0 =
(4.19)
d2 Sx (t)
= 02 Sx (t)
dt2
and exactly the same equation for Sy (t). These equations have the solutions
Sx (t) = a cos 0 t + b sin 0 t
Sy (t) = c cos 0 t + d sin 0 t
Sz (t) = Sz (0) .
Clearly, Sx (0) = a and Sy (0) = c. Also, from the equations of motion (4.18) we
have 0 Sy (0) = (dSx /dt)(0) = 0 b and 0 Sx (0) = (dSy /dt)(0) = 0 d so that our
solutions are
Sx (t) = Sx (0) cos 0 t + Sy (0) sin 0 t
Sy (t) = Sy (0) cos 0 t Sx (0) sin 0 t
Sz (t) = Sz (0) .
34
(4.20)
These can be written in the form Sx (t) = A cos(0 t+x ) and Sy (t) = A sin(0 t+y )
where A2 = Sx (0)2 +Sy (0)2 . Since these are the parametric equations of a circle, we
see that S precesses about the B field, with a constant projection along the z-axis.
For example, suppose the spin starts out at t = 0 in the xy-plane as an eigenstate
of Sx with eigenvalue +~/2. This means that
hSx (0)i = h
x + |Sx (0)|
x +i =
~
2
~
cos 0 t
2
~
hSy (t)i = sin 0 t
2
hSz (t)i = 0 .
(4.21)
This clearly shows that the spin stays in the xy-plane, and precesses about the zaxis. The direction of precession depends on the sign of 0 , which in turn depends
on the sign of q as defined by (4.19).
How does this look from the point of view of the SP? From (4.14), the spin state
evolves according to
|(t)i = eiHt/~ |(0)i = ei(gqBt/2mc)Sz /~ |(0)i = ei0 tSz /~ |(0)i .
Comparing this with (4.10a), we see that in the SP the spin precesses about
z
with angular velocity 0 .
Now let us consider the motion of an electron in an inhomogeneous magnetic
field. Since for the electron we have q = e < 0, we see from (4.12) that s is antiparallel to the spin S. The potential energy is Hs = s B, and as a consequence,
the force on the electron is
F = Hs = (s B) .
Suppose we have a Stern-Gerlach apparatus set up with the inhomogeneous B
field in the z direction (i.e., pointing up):
~
~
(cos 0 T )
x (sin 0 T )
y.
2
2
(4.22)
Interestingly, for the inhomogeneous field this expectation value turns out to be
zero.
To see this, let the particle enter the apparatus with its spin along the x
direction,
and let its initial spatial wave function be the normalized wave packet 0 (r, t). In
terms of the eigenstates of Sz , the total wave function at t = 0 is then (from (2.9a))
"
#
0 (r, 0)
1
(r, 0) =
.
(4.23a)
2 0 (r, 0)
When the particle emerges at t = T , it is a superposition of localized spin up wave
packets + (r, T ) and spin down wave packets (r, T ):
"
#
+ (r, T )
1
.
(4.23b)
(r, T ) =
2 (r, T )
Since the spin up and spin down states are well separated, there is no overlap
between these wave packets, and hence we have
+ (r, T ) (r, T ) 0 .
(4.23c)
36
and therefore
Z
Z
| (r, T )| d3 r =
2
|+ (r, T )| d3 r =
|0 (r, 0)| d3 r
2
| (r, T )| d3 r .
(4.24)
~
|0 (r, 0)|2 x
(4.25)
as claimed.
Why does the uniform field give a non-zero value for hSiT whereas the inhomogeneous field gives zero? The answer lies with (4.19). Observe that the precessional
frequency 0 depends on the field B. In the Stern-Gerlach apparatus, the B field is
weaker up high and stronger down low. This means that particles lower in the beam
precess at a faster rate than particles higher in the beam. Since (4.25) is essentially
37
an average of (4.22), we see that the different phases due to the different values of
0 will cause the integral to vanish.
Of course, this assumes that there is sufficient variation 0 in precessional
frequencies that the trigonometric functions average to zero. This will be true as
long as
(0 )T 2 .
In fact, this places a constraint on the minimum amount of inhomogeneity that the
S-G apparatus must have in order to split the beam. From (4.19) we have
0 =
e
e B0
B0
z
me c
me c z
where z is the distance between the poles of the magnet. Therefore, in order for
the apparatus to work we must have
e B0
2
me c z T z .
The basic idea is simply to express the eigenstates of the total angular momentum
operator J of two particles in terms of the eigenstates of the individual angular
momentum operators J1 and J2 , where J = J1 + J2 . (Alternatively, this could be
the addition of spin and orbital angular momenta of a single particle, etc.) Each
particles angular momentum operator satisfies (where the subscript a labels the
particle and where we take ~ = 1)
[Jai , Jaj ] = i
3
X
ijk Jak
k=1
=i
3
X
k=1
3
X
ijk J1k + i
3
X
ijk J2k = i
k=1
3
X
k=1
ijk Jk
k=1
and hence J is just another angular momentum operator with eigenstates |jmi.
Next we want to describe simultaneous eigenstates of the total angular momentum. From [J1 , J2 ] = 0, it is clear that [J1 2 , J2 2 ] = 0, and since [Ja 2 , Jaz ] = 0, we
see that we can choose our states to be simultaneous eigenstates of J1 2 , J2 2 , J1z
and J2z . We denote these states by |j1 j2 m1 m2 i.
38
J 2 = J1 + J2
2
= J1 2 + J2 2 + 2J1 J2
= J1 2 + J2 2 + 2(J1x J2x + J1y J2y + J1z J2z ) .
It is now easy to see that [J 2 , Ja 2 ] = 0 because [J1 , J2 ] = 0 and [Ja 2 , Jai ] = 0 for
i = x, y, z. We denote these simultaneous eigenstates by |j1 j2 jmi.
However, let me emphasize that even though [J 2 , Jz ] = 0, it is not true that
2
[J , Jaz ] = 0. This means that we can not specify J 2 in the states |j1 j2 m1 m2 i, and
we can not specify either J1z or J2z in the states |j1 j2 jmi.
We know that the angular momentum operators and states satisfy (with ~ = 1)
J 2 |jmi = j(j + 1)|jmi
(5.1a)
Jz |jmi = m|jmi
p
J |jmi = j(j + 1) m(m 1)|jm 1i .
(5.1b)
(5.1c)
Let us denote the two individual angular momentum states by |j1 m1 i and |j2 m2 i.
Then the two-particle basis states are denoted by the various forms (see Part I,
Section 1 in the handout Supplementary Notes on Mathematics)
|j1 j2 m1 m2 i = |j1 m1 i|j2 m2 i = |j1 m1 i |j2 m2 i .
When we write hj1 j2 m1 m2 |j1 j2 m1 m2 i we really mean
(hj1 m1 | hj2 m2 |)(|j1 m1 i |j2 m2 i) = hj1 m1 |j1 m1 ihj2 m2 |j2 m2 i .
Since these two-particle states form a complete set, we can write the total combined angular momentum state of both particles as
X
(5.2)
|j1 j2 m1 m2 ihj1 j2 m1 m2 |j1 j2 jmi .
|j1 j2 jmi =
j1 j2
m1 m2
However, we now show that many of the matrix elements hj1 j2 m1 m2 |j1 j2 jmi vanish,
and hence the sum will be greatly simplified.
For our two-particle operators we have
Jz = J1z + J2z
and
39
J = J1 + J2 .
While we wont really make any use of it, let me point out the correct way to write
operators when dealing with tensor (or direct) product states. In this case we should
properly write operators in the form
J = J1 12 + 11 J2 .
Then acting on the two-particle states we have, for example,
Jz (|j1 m1 i |j2 m2 i) = J1z 1 + 1 J2z (|j1 m1 i |j2 m2 i)
As I said, while we wont generally write out operators and states in this form, you
should keep in mind what is really going on.
Now, for our two-particle states we have
Ja 2 |j1 j2 m1 m2 i = ja (ja + 1)|j1 j2 m1 m2 i
Jaz |j1 j2 m1 m2 i = ma |j1 j2 m1 m2 i
where a = 1, 2. Taking the matrix element of J1 2 acting to both the left and right
we see that
hj1 j2 m1 m2 |J1 2 |j1 j2 jmi = j1 (j1 + 1)hj1 j2 m1 m2 |j1 j2 jmi
or
[j1 (j1 + 1) j1 (j1 + 1)]hj1 j2 m1 m2 |j1 j2 jmi = 0 .
if j1 6= j1 or j2 6= j2 .
Next, from Jz = J1z + J2z we can let J1z + J2z act to the left and Jz act to the
right so that
hj1 j2 m1 m2 |Jz |j1 j2 jmi = (m1 + m2 )hj1 j2 m1 m2 |j1 j2 jmi
= hj1 j2 m1 m2 |j1 j2 jmim .
This shows that
hj1 j2 m1 m2 |j1 j2 jmi = 0
40
unless m = m1 + m2
Finally, we regard the values of j1 and j2 as fixed and understood, and we simply
write the total angular momentum state as
X
|jmi =
|m1 m2 ihm1 m2 |jmi .
(5.3)
m1 m2
m1 +m2 =m
hm1 m2 |m1 m2 i = m1 m1 m2 m2 .
and
Therefore, taking the inner product of equation (5.3) with itself we see that
X
|hm1 m2 |jmi|2 = 1 .
(5.4)
m1 m2
m1 +m2 =m
Given j1 and j2 , this holds for any resulting values of j and m. (This is really
just another way of saying that the columns of the unitary transition matrix are
normalized.)
Now, we know that j1 m1 j1 and j2 m2 j2 . Therefore, from
m = m1 + m2 , the maximum value of m must be j1 + j2 . And since j m j,
it follows that the maximum value of j is jmax = j1 + j2 . Corresponding to this
maximum value j1 + j2 of j, we have a multiplet of 2(j1 + j2 ) + 1 values of m, i.e.,
(j1 + j2 ) m j1 + j2 . On the other hand, since there are 2j1 + 1 possible
values of m1 , and 2j2 + 1 possible values of m2 , the total number of (not necessarily
distinct) possible m1 + m2 = m values is (2j1 + 1)(2j2 + 1).
The next highest possible value of m is j1 + j2 1 so that there is a j state with
j = j1 + j2 1 and a corresponding multiplet of 2(j1 + j2 1) + 1 = 2(j1 + j2 ) 1
possible m values. We continue lowering the j values by one, and for each such
value j = j1 + j2 k there are 2(j1 + j2 k) + 1 possible m values. However, the
total number of m values in all multiplets must equal (2j1 + 1)(2j2 + 1).
41
Example 5.1. Consider the case j1 = 1, j2 = 2. Then the total number of possible
m values is
(2j1 + 1)(2j2 + 1) = 3 5 = 15 .
We can arrange these values of m = m1 + m2 in a table with m1 across the top and
m2 down the side:
m1
1
0
1
2 3 2 1
1 2 1
0
m2
0
1
0 1
1
0
1
2
1
2
3
2
The distribution of m values is as follows:
value of m = m1 + m2 : 3
number of occurances: 1
2 1
2 3
0 1 2 3
3
3
2
1
k=0
(5.5)
k=n
k=
mn+1
(m + n)
2
42
we have
b
X
k=0
b
X
k + (b + 1)
k=0
(5.6)
Equations (5.1c), (5.3), (5.4) and (5.6) are all we need to calculate the ClebschGordan coefficients. The procedure is best illustrated by a specific example.
Example 5.2. Let us consider the case j1 = j2 = 1/2 because you (should) already
know the answer. In this case we have m1 = m2 = 1/2, so according to equation
(5.6) we know that 0 j 1.
First we have jmax = j1 + j2 = 1, and the maximum value of m is m1 + m2 = 1
also. From equation (5.3) we have the single term
E 1 1 ED 1 1 E
1 1 =
1 1
2 2 2 2
D
1 1 E2
2 2 1 1 = 1 .
This specifies the topmost state up to a phase factor. Choosing this phase to be 1
(the Condon-Shortley convention), we are left with
D1 1 E
1 1 = +1 .
2 2
43
(5.7a)
where the state on the left-hand side of this equation is the |jmi state, and the
state on the right-hand side is the |m1 m2 i state.
To construct the next lowest state, we act on the left-hand side of this last result
with J , and on the right-hand side with J = J1 + J2 . Using equation (5.1c),
we obtain
r
E
1E
3 1 1 1 E 1
21 0 =
+
+
4 4
2 2
2
2
or
E
1E
1 1 1 E 1
+
.
(5.7b)
1 0 =
2 2
2
2
2
1 0 +
1 0 =
1 0
2
2 2
2
2 2
2 2
we see that
or
1 E D 1 1 E
1
1 0 =
1 0 = .
2
2
2 2
2
We now act on equation (5.7b) with J = J1 + J2 again to obtain
"r
#
E
1
1 E 1
1E
3 1 1
21 1 =
+
+
4 4
2
2
2
2
2
and hence
D1
E 1
1E
1 1 =
2
2
(5.7c)
E
1
1
1 1 = 1 .
2
2
This completes the j = 1 multiplet which, for obvious reasons, is called the triplet
state.
Now for the j = 0 state. We know from equation (5.3) that
E 1
1 ED 1
1 E 1 1 ED 1 1 E
0 0 +
0 0 =
0 0
2
2 2
2
2 2
2 2
1 1E
1
1E
a+
b.
:=
2
2
2 2
D
To solve for a and b, we first see that from equation (5.4) we have a2 + b2 = 1.
Next, we note that the |jmi = |1 0i state is orthogonal to the |jmi = |0 0i state,
and therefore
1
0 = h1 0|0 0i = (a + b)
2
44
so that a = b. Therefore 2a2 = 1 so that a = b = 1/ 2 and we have
E
1 E 1 1 E
1 1
0 0 =
2
2 2
2 2
(5.7d)
1 1 = 21 12
E
n
E
Eo
triplet symmetric states:
1 0 = 12 12 12 + 21 21
E
E
1 1 = 1 1
2
2
E
1 E 1 1 E
1 1
.
singlet antisymmetric state: 0 0 =
2
2 2
2 2
While we will soon return to discuss the symmetry of wave functions in detail,
it should be fairly clear that exchanging electrons 1 and 2 in any of the triplet
states leaves the wave function unchanged, and hence the states are referred to as
symmetric. Similarly, exchanging electrons 1 and 2 in the singlet state changes
the overall sign of the state, and hence this is called an antisymmetric state.
45