Beruflich Dokumente
Kultur Dokumente
Source:
http://ocw.mit.edu/courses/chemistry/5-61-physical-chemistry-fall-2007/lecture-notes/
F dv
()
v t
v t =
v0
dv = t0 m
dt F = ma = m dt
dr
()
r t
r t =
r0
dr =
t0
vdt dt = v
• Blackbody radiation
• Photoelectric effect
Quantum Mechanics!
But Thomson experimentally observes electrons as particles with charge & mass.
y = displacement
induced by
deflector voltage
e
Integrating v y = t
me
(
[Note v y t = 0 = 0] )
e t f
2
e
e
Modern day value is = 1.758x1011 C/kg
me
mist of micron-size
oil droplets in air
x-rays n-
Fc Fg Ff
19
(very close to today’s value e = 1.602x10 C)
27
Hydrogen mass was known: mH = 1.66x10 kg electron is subatomic!!
(c) Where are the electrons? What’s the structure of the atom?
Angstrom (10 -10 m) atomic size scale already inferred from gas kinetics
First “jellium” model didn’t last long
n+
(jelly)
e
(1) He2+ nucleus very small, << 10 -10 m (Rutherford estimated 10 -14 m)
(2) Au atoms are mostly empty!
1 Ze2 1 Ze2
E = K.E. + P.E. = me v +
2
= <0
2 4 0 r
2 4 0 r
v (m/s)
(Hz = cycles/s) =
2 r (m/cycle)
circumference of orbit
For H:
n1 = 3 n1 = 2 n1 = 1
n2 = 3 4 5 6… n2 = 2 3 4…
For the H atom, Rydberg showed that the spectrum was consistent with the
simple formula:
1 1
(cm -1 ) = R 2 2
n1 n2
with n1 = 1, 2, 3, ... and n2 = n1 + 1, n1 + 2, n1 + 3, ...
R = 1.097x105 cm -1 (Rydberg constant)
n1 = 1 Lyman series
n1 = 2 Balmer series
n1 = 3 Paschen series
(2) The electrons can oscillate (& radiate) equally well at any
frequency
8 kT 2
( )
d = ,T d =
c 3
d 2
Rayleigh- ()
2 divergence at high freq.
Jeans Law lots of emission in UV, x-ray regions!
“UV catastrophe” – not observed in exp’t
( )
Exp’t T3
T2 T1 < T2 < T3
T1
E
(2) The energy an integral multiple of
(the # of oscillators n)
E n
or E = nh
constant
Planck used statistical mechanics (5.62) to derive the expression for black body
radiation
8 h 3
( )
d = ,T d =
1
c3 eh kT 1
d
3
( )
1
eh kT 1
Planck’s constant
UV light e-
Exp’t:
K.E. = _mv2
metal
I
I 0
threshold
E = h h Planck’s constant
E K.E. energy of
of e- photon = h
0
= work function
of metal
K.E.
K.E. = h( – 0)
0 = / h
Summary:
(1) Atoms can exist in stable “states” without radiating. The states have
discrete energies En, n = 1, 2, 3,..., where n = 1 is the lowest energy
state (the most negative, relative to the dissociated atom at zero
energy), n = 2 is the next lowest energy state, etc. The number “n” is
an integer, a quantum number, that labels the state.
h
(3) Angular momentum is quantized: = n where =
2
Angular momentum
L=rp = L
For circular motion:
L
L is constant if r and p are constant
r
l = mrv is a constant of the motion
p = mv
Other useful properties
1
Prepared By Kevin Boudreaux
12 of 227
MIT Open Courseware Compillation
Source:
http://ocw.mit.edu/courses/chemistry/5-61-physical-chemistry-fall-2007/lecture-notes/
v
( )
= 2 r rot = r rot
velocity
circumference frequency angular
(m/s)
(m/cycle) (cycles/s) frequency
(rad/s)
= mvr = mr 2 rot
= I rot
1 2 p2 1 m2 r 2 v 2 2
K.E. = mv = K.E. = =
2 2m 2 mr 2 2I
Ze2 n2 2
r=
4 0 mv 2
r=
Z
(
4 0
me2
) The radius is quantized!!
2
( 4 ) 0
me2
a0 the Bohr radius
2
Prepared By Kevin Boudreaux
13 of 227
MIT Open Courseware Compillation
Source:
http://ocw.mit.edu/courses/chemistry/5-61-physical-chemistry-fall-2007/lecture-notes/
1 Ze2 1 Z 2 me4
E= En = 2 Energies are quantized!!!
2 4 0 r n 8 02 h2
me4 1 1
(
cm -1
) =
hc
En2
En1
= 2 3 2 2
hc 8 0 h c n1 n2
me4
Rydberg formula ! with R = 2 3 = 109,737 cm -1
8 0 h c
Measured value is 109,678 cm-1 (Slight difference due to model that gives
nucleus no motion at all, i.e. infinite mass.)
3
Prepared By Kevin Boudreaux
14 of 227
MIT Open Courseware Compillation
Source:
http://ocw.mit.edu/courses/chemistry/5-61-physical-chemistry-fall-2007/lecture-notes/
Light as a wave
0
0
x (const. t) or
t (const. x)
2
( )
x,t = 0 sin x ct
( )
2
Define k= (“wavevector” magnitude)
sin(kx)
Constructive
+ = 2sin(kx) interference
sin(kx)
(in phase)
sin(kx)
Destructive
+ =0 interference
sin(kx + )
(out of phase by /2)
D
l
s
Interference
pattern
light, screen observed
D
= D=
s s
Light as a particle
Compton exp‛t
e-
x-rays ( 1Å)
e-
Experimentally:
e-
Prepared By Kevin Boudreaux
16 of 227
MIT Open Courseware Compillation
Source:
http://ocw.mit.edu/courses/chemistry/5-61-physical-chemistry-fall-2007/lecture-notes/
hc hc
E = < =E
h h
p= = for the light
c
(B) Matter
Matter as waves (deBroglie, 1929, Nobel Prize for his Ph.D. thesis!)
Same relationship between momentum and wavelength for light and for matter
h h
p= = de Broglie wavelength
p
Diffraction
e- crystal pattern
Consequences (I)
+ +
Heisenberg Uncertainty Principle xpx
2
x
peak-null
D
distance
s
l
light,
D
= D=
s s
This means the electrons must go through the slit with some range of
velocity components vx
v
x xv
v
vx D
= =
v s x
xvx = v or xpx = p
h
But from deBroglie =
p
xpx = h
xpx
2
1
px vx x106 m s
2x 2mx 2
Basically, if we know the e- is in the atom, then we can’t know its velocity at all!
Bohr had assumed the electron was a particle with a known position and velocity.
To complete the picture of atomic structure, the wavelike properties of the
electron had to be included.
Time-independent
2 2
2m x 2
() ()
+ V x x = E x () Schrödinger
equation
We cannot prove the Schrödinger equation. But we can motivate why it might be
reasonable.
A v
( ) ( )
1 x,t = Asin kx t is a right-traveling wave.
2
k= = 2 = v
( )
2 x,t = Asin kx + t .( )
Both are solutions to the wave equation
( )=
2 x,t 1 x,t
2
( )
x 2 v 2 t 2
( )
Further, the sum x,t = 1 x,t + 2 x,t ( ) ( ) of left and right traveling waves is
also a solution.
( ) ( ) ( )
x,t = A sin kx t + sin kx + t = 2 Asin kx cos t ( ) ( )
This is a stationary wave or standing wave. Its peaks and nulls remain stationary.
( )
x,0 = 2 Asin kx ( ) t=0
1 2 2
x, = 2 A
8
sin kx t = /4 ( )
2
1 2 t = /2
x, = 0
4
t = 3/4
3 2 2
x, = 2 A
8
sin kx ( )
2 nodes
t = /
1 2
( ) ( )
x, = 2 A 1 sin kx
2
As in a vibrating violin string, the node positions are independent of time. Only
the amplitude of the fixed waveform oscillates with time.
( ) () ( )
x,t = x cos t
()
In the particular case above, x = 2 Asin kx . ( )
( )=
2 x,t 1 x,t
2
( )
2
x 2 v 2
t 2
= 2 x,t = k 2 x,t
v
( ) ( ) ( k=v )
( )
Plugging in x,t = x cos t () ( )
2 x ( ) = k 2
2
x 2
2
()
x =
x
()
2 x( ) = p 2
de Broglie relation =
h
p
x 2
x ()
2 x ( )= p
2
x 2
2
( x)
( )
But p = 2m K.E. = 2m E V x
2
( ) (assuming t-independent potential)
2 x
2
()
2m x 2
+ V x x = E x () () ()
time-independent Schrödinger equation in one dimension
• a physical picture involving wave and particle duality of light and matter !
2 x
2
()
2m x 2
+ V x x = E x () () ()
We can find two independent solutions 1 x and 2 x () ()
The general solution is a linear combination
()
x = A1 x + B2 x() ()
A and B are then determined by boundary conditions on x and x . () ()
Additionally, for physically reasonable solutions we require that x and ()
x () be continuous function.
2 x
2
()
2m x 2
= E x ()
2mE 2 k 2
Define k = 2
2
or E =
2m
p2
()
V x = 0, E =
2m
p 2 = 2 k 2 p = k
h 2
de Broglie p= k=
2 x ( ) = k
The wave eq. becomes
x 2
2
( x)
with solutions ()
x = Acos kx + B sin kx( ) ( )
Free particle no boundary conditions
2 k 2
any A and B values are possible, any E = possible
2m
So any wavelike solution (traveling wave or standing wave) with any wavelength,
wavevector, momentum, and energy is possible.
()
V x = ( x < 0, x > a )
V ( x) = 0 (0 x a ) V( x )
0
0 x a
2 x
2
()
2m x 2
= E x ()
2 x ( ) = k
x 2
2
( x) with same definition
2mE 2 k 2
k2 = or E =
2 2m
Continuity of x () ()
0 = a = 0 ()
(i) () ()
0 = Acos 0 + B sin 0 = 0 () A=0
(ii) ()
a = B sin ka = 0 ( )
Can’t take B = 0 (no particle anywhere!)
Must have ( )
sin ka = 0 ka = n n = 1, 2, 3,...
n
k is not continuous but takes on discrete values k =
a
Thus integer evolves naturally !!
n x
( )
0 x a = B sin
a
n = 1,2,3,...
2 k 2 n2 h2
E= En =
2m 8ma 2
E5 = 25E1
2
( 2 h
)
2
En+1 En = n + 1 n
8ma 2
E4 = 16E1
E3 = 9E1
(
En+1 En = 2n + 1 E1 )
E2 = 4E1
E1 = h2 8ma 2
(b) ()
The wavefunction x is sinusoidal, with the number of nodes
increased by one for each successive state
4 = a 2 (
4 = B sin 4 x a ) (3 nodes)
3 = 2a 3 (
3 = B sin 3 x a ) ( 2 nodes)
n = 2a n # nodes = n - 1
a=
( ) (1 node )
2
2 = B sin 2 x a
1 = 2a 1 = B sin ( x a ) (0 nodes)
0 x a
We’ve solved some simple quantum mechanics problems! The P-I-B model is a
good approximation for some important cases, e.g. pi-bonding electrons on
aromatics.
Summary so far:
( ) ( )
V x < 0, x > a = x < 0, x > a = 0
n x
V (0 x a ) = 0 ( 0 x a ) = B sin
V( x )
n
a
0 n2 h2 n 2a
0 a En = k= = n = 1,2,3,...
x 8ma 2 a n
() () ()
2
x = * x x is a probability distribution or probability density
for the particle
()
2
x dx is the probability of finding the particle in the interval
between x and x + dx
This is a profound change in the way we view nature!! We can only know the
probability of the result of a measurement – we can’t always know it with
certainty! Makes us re-think what is “deterministic” in nature.
()
x2 2
x1
x dx = probability of finding particle in interval
() ( )
2 a 2
x dx =
0
x dx = 1 Normalization condition
a n x 2
0
B 2 sin 2
a
dx = 1 B=
a
n x
()
n x =
2
a
sin
a
n = 1,2,3,... Normalized wavefunction
()
2
( ) ( ) Interpretation of x
2
4 = 2 a sin 4 x a 2
based on measurement
( ) ( )
2
1 = 2 a sin 2 x a
0.35
0.3
0.25
e.g.
0.2
0.15
0.1
0.05
0
2 4 6 8 10 12 14
x = xPx
()
2
Px x dx
()
2
x =
x x dx
Similarly
()
2
x2 =
x 2 x dx
() ()
x =
* x x x dx
= * ( x ) x ( x )dx
x2 2
2 a n x
x =
a 0
x sin 2
a
dx
a
Integrate by parts x =
2
(x x )
2
i
Square of deviation of ith measurement from average
value <x >
(x x )
2
i
x2 the Variance in x
(x x )
2 2
Note i
= x2 x = x2
12
x = x2 x
2
x = x
() () () ()
a
x2 =
0
* x x 2 x dx
* x x x dx
2
2 a n x 2 a 2 n x
=
x 2 sin 2
dx
0 x sin dx
a 0 a a
a
2 2
a a2 a
=
2
3 2 n
( ) 4
x 2
( )
n
2
2
a
= 2
2
( ) 3
x 2
4 n
12
( )
2
a n
x = x = 2
2 n 3
( )
Now suppose we want to test the Heisenberg Uncertainty Principle for the
particle in a box.
12
p and p 2 to get p = p = p 2 p
2
We need
() ()
But do we write p =
* x p x dx ?
ˆ x =g x
Af () ()
operator acts on function to get a new function
d x2 2x
e.g. =
dx 3 3
Special Case
If ()
ˆ x =af x
Af ()
number (constant)
2 d 2
2
+ V x x () () = E x ()
2m dx
2 d 2
or Ĥ = E with ()
Ĥ x =
2m dx 2
+V x () (in 1D)
p2
()
If V x = 0 , then E = K.E. =
2m
( p̂ )
2
d2
( p̂ )
2
Ĥ = = 2
2m dx 2
( p̂ ) ( pˆ ) ( p̂ )
2
means i.e. the operator acts sequentially on the function
d d d2
( )( )
p̂ p̂ = i i = 2 2
dx dx dx
d
p̂ = i Momentum operator (in 1D)
dx
p
for Particle in a Box
2 2
d
d
() () () ()
2
2p = p 2 p = * x i x dx * x i x dx
dx
dx
Note order is now very important! Operator acts only on the function to its
right.
d
() ()
p = * x i
dx
x dx
a
2 1 2 n x
d
2
12
n x
=
0
sin
a a
i
dx
a
sin a dx
=0
a 2 1 2 n x
d d 2
12
n x
=
sin i dx i dx a sin a dx
2
p
0
a a
2 2
22 a n x n n x 22 n a
2 n x
=
a
0 sin a a sin a dx = a a
0 a dx
sin
n2 2 2
=
a2
n2 h2 n2 h2
Note p 2
= = 2m = 2mE as expected
4a 2 8ma 2
E = K.E. since V(x) = 0
n2 2 2
( )
2
=2
p
= p
a2
12 12
( )
( )
2 2
a n n n
xp = 2 = 2
2 n 3
( )
a 2 3
always > 1
xp as expected from Heisenberg Uncertainty Principle
2
a = *Â d
2
#1] (a) The physically relevant quantity is
( ) ( ) ( )
2
* r,t r,t = r,t probability density at time t
and position r
(b) ( )
r,t must be normalized
* d = 1
(c) ( )
r,t must be well behaved
12
2 n x
() ()
Ĥ x n x = En n x () n ()
x =
a
sin
a
e.g. ()
n x above with momentum operator
d 2 n x
12
() d
dx
()
p̂n n x = i n x = i
sin
dx
a
a
2 1 2 n x
pn
sin
a
a
2 2 2 2
= + + (3D)
2m x 2 y 2 z 2
d d
lx = ypz zp y = i y z
dz dy
d d
l y = zpx xpz = i z x
dx dz
d d
lz = xp y ypx = i x y
dy dx
() () ()
 f x + g x = Af ˆ x + Ag
ˆ x () and
() ()
 cf x = c f x
( )
1 Â 2 d = 2 Â 1 d
( ) ( ) d
 d = Â
( )
a d = a
d
a = a*
then m
n d = 0 if m n
Proof:
( )
m  n d = n  m d
an m n d = am n m d
(a n
am )
n d = 0
m
(a n
am )
m
n d = 0
= 0 if = 0 if n m
n=m
4
1 4 As much + as - area
3
Eigenfunctions 1 3
of
2
1 2
1
0 x a 0 x a
m
n d = mn
Krönecker delta
1 if m = n (normalization)
mn =
0 if m n (orthogonality)
= c11 + c22
2 2
Then a measurement of A returns either a1 or a2 , with probability c1 or c2
respectively, and making the measurement changes the state to either 1 or 2 .
a1 1 (probability c12 )
a2 2 (probability c22 )
measure
a = n  n d = an n n d = an
(c ) ( )
a = Aˆ d = 1 1
+ c22 Aˆ c11 + c22 d = c12 a1 + c22 a2
()
ˆ x =g x
Af () ( p̂x̂ ) f ( x ) = i dxd
( x ) f ( x )
= ix
d
dx
() ()
d
f x if x = ix i f x
dx
()
d
 = ix i = p̂x̂
dx
( )
ˆ x = x i d f x = ix d f x
Bf () () () ()
dx
dx
d
B̂ = ix = x̂p̂
dx
x̂p̂ p̂x̂
Define commutator
Â, B̂ = 0
• This means that we can know the exact values of both observables
A and B simultaneously (no uncertainty limitation).
Â, B̂ 0
e.g. x̂, p̂ = i 0 xp
2
examples
m1 m2 B (magnetic
field)
A diatomic molecule μ
(spin
magnetic
moment)
E (electric
field)
Classical H.O.
m
k
X0 X
Hooke’s Law: (
f = k X X 0 kx )
(restoring force)
d2x d2x k
f = ma = m 2 = kx + x=0
dt dt 2 m
Solve diff. eq.: General solutions are sin and cos functions
() ( )
x t = Asin t + B cos t ( ) =
k
m
or can also write as
() (
x t = C sin t + )
where A and B or C and are determined by the initial conditions.
e.g. ()
x 0 = x0 ()
v 0 =0
spring is stretched to position x 0 and released at time t = 0.
Then
() ()
x 0 = A sin 0 + B cos 0 = x0 () B = x0
()
v 0 =
dx
dt
()
= cos 0 sin 0 = 0 () A=0
x=0
So ()
x t = x0 cos t ( )
k
Mass and spring oscillate with frequency: =
m
and maximum displacement x0 from equilibrium when cos(t)= ±1
Energy of H.O.
Kinetic energy K
2
1 dx
1 1 1
( ) ( )
2
K = mv 2 = m = m
x0 sin t = kx02 sin 2 t
2 2 dt 2 2
Potential energy U
()
f x =
dU
dx
()
U = f x dx = ( kx )dx =
1
2
kx 2
=
1 2
2
( )
kx0 cos 2 t
Total energy = K + U = E
E=
1 2
2
( )
kx0 sin 2 t + cos 2 t ( ) E=
1 2
kx
2 0
x (t )
x 0(t )
0 t
-x0(t)
U K
1 2
kx E
2 0
0 t
X
U
X
X0 A + B separated atoms
1 d 2U 1 d 3U
( ) ( ) dU
(X X ) (X X ) (X X )
2 3
U X = U X0 + 0
+ 0
+ 0
+
dX X = X0
2 dX 2 X = X0
3! dX 3 X = X0
Redefine x = X X 0 and ( )
U X = X0 = U x = 0 = 0 ( )
1 d 2U 1 d 3U
()
U x =
dU
dx
x+
2 dx 2
x +
2
3! dx 3
x3 +
x=0 x=0 x=0
real potential
H.O. approximation
dU
At eq. =0
dx x=0
1 d 2U
U x () 2 dx 2
x2
1 2
2
kx
x=0
X1 XCOM X2 X
M = m1 + m2 total mass
xrel
m1 m2
μ= reduced mass
m1 + m2
m1 X 1 + m2 X 2
X COM = COM position
m1 + m2
xrel = X 2 X 1 x relative position
2 2 2 2
1 dX 1 dX 1 dX 1 dx
K = m1 1 + m2 2 = M COM + μ
2 dt 2 dt 2 dt 2 dt
1 2
U= kx
2
2 2
1 dX 1 dx 1
E = K + U = M COM + μ + kx 2
2 dt 2 dt 2
2
1 dX
Etrans = M COM
2 dt
2
1 dx 1
Evib = μ + kx 2
2 dt 2
2 d 2 1 2
()
Ĥ x = 2
+ kx x = E x () ()
2m dx 2
K U
Rewrite as:
d 2 x ( ) + 2m E 1 kx
dx 2
2
2
()
x = 0
2
()
f x = e x
2
2
TRY: (gaussian function)
d2 f x ( ) = e x 2 2
+ 2 x 2 e x () ()
2
2
2
= f x + 2 x 2 f x
dx
d2 f x ( ) + f
or rewriting,
dx 2 ( x) 2
()
x 2 f x = 0w
2mE mk
= and 2 =
2 2
k
E=
2 m
k 1 k
Recall = or =
m 2 m
1 1
E= = h
2 2
()
x = Nf x = Ne x ()
2
2
where N is the normalization constant
14
2
()
x2
x dx = 1 N 2
e =1 N =
14
() e x
2
0 x =
2
0 x ()
1 1 1
E0 =
= h E0 =
2 2 2
x
()
Note 0 x is symmetric. It is an even function: 0 x = 0 x () ( )
There are no nodes, & the most likely value for the oscillator displacement is 0.
So far we have just one eigenvalue and eigenstate. What about the others?
14
() e x
1
2
0 x =
2
E0 = h
2
14
1
()
1 x =
2
( 2 x ) e
12 x 2 2
E1 =
3
2
h
14
1
2 ()x =
8
( 4 x 2
)
2 e x
2
2
E2 =
5
2
h
14
1
3 ()x =
48
(8 32
x 3 12 1 2 x e x ) 2
2
E3 =
7
2
h
12
km
with = 2
14
()
n x =
1
1/ 2
(
H n 1 2 x e x ) 2
2
n = 0,1,2,...
( 2 n!)
n
Normalization Gaussian
Hermite polynomial (pronounced “air-MEET”)
( )
H0 y = 1 even ( n = 0)
H ( y) = 2 y
1
odd ( n = 1)
H ( y) = 4 y 2
2
2
even ( n = 2 )
H ( y ) = 8y 12 y
3
3
odd ( n = 3)
H ( y ) = 16 y 48y
4
4 2
+ 12 even ( n = 4 )
() ()
2
n x n x
7
3 x () E3 =
2
5
2 x () E2 =
2
3
1 x () E1 =
2
0 x () E0 =
2
1
Energies are En = n + h
2
Energy levels are evenly spaced
1 1
( 2
)
En+1 En = n + 1 + h n + h = h
2
regardless of n
1
There is a “zero-point” energy E0 = h
2
( )=
d odd (
d even )=
dx
(even ) dx
( odd )
( odd ) dx = 0 (even ) dx = 2 (even ) dx
0
d
() () ()
x n
=
n x x n x dx = 0 p n
= n dx
n x dx = 0
ih
odd odd
IR spectroscopy
H.O. selection rules
n’ = 1
Vibrational transition h
n=0
2
dμ
dx n n '
Intensity I nn x dx
1 2
Classical H.O.: Total energy ET = kx
2 0
oscillates between K and U.
ET
The classical oscillator with energy ET can never exceed this displacement, since
if it did it would have more potential energy than the total energy.
1 2
kx
()
2
()
2 2
3 x 12 x
()
()
2
2 x
()
2
1 x At high n, probability
density begins to look
classical, peaking at turning
() points.
2
0 x
()
2
Prob. of (x > x 0, x < -x 0):
3 x 1
()
2
2
e x dx
2
02 x dx = 2
1 2 1 2
()
2
2 x 2
()
2
= e y dy = erfc 1
1 2 1
()
2
Prob. of (x > x 0, x < -x 0) = erfc(1)
0 x
= 0.16
Significant probability!
x
Prepared By Kevin Boudreaux
58 of 227
MIT Open Courseware Compillation
Source:
http://ocw.mit.edu/courses/chemistry/5-61-physical-chemistry-fall-2007/lecture-notes/
The oscillator is “tunneling” into the classically forbidden region. This is a purely
QM phenomenon!
()
x ~ sin kx ( ) ()
x ~ e x
Finite
V0
barrier
()
x ~ sin kx + ( )
x
Even though the energy is less than the barrier height, the wavefunction is
nonzero within the barrier! So a particle on the left may escape or “tunnel” into
the right hand side.
2 d 2
Inside barrier:
2m 2
+ V0
x = E x () ()
dx
d 2 x ( ) = 2m (V E )
or
dx 2
0
2 ()
x 2 x ()
1
(
2m V0
E ) 2
( )
12
Note V0
E and m1 2
If barrier is not too much higher then the energy and if the mass is light, then
tunneling is significant.
Important for protons (e.g. H-bond fluctuations, tautomerization)
2 2 1
= c0 0 + c1 1 with c0 + c1 = 1 (normalization), e.g. c0 = c1 =
2
Full time-dependent eigenstates can be written as
( )
0 x,t = 0 x e ()
i 0 t
( ) ()
1 x,t = 1 x e
i1t
where
1 1 3 3
0 = E0 = vib 0 = vib 1 = E1 = vib 1 = vib
2 2 2 2
( )
x,t =
1
e
i 0 t
()
0 x +
1
e
i1t
() () ( )
1 x = c0 t 0 x + c1 t 1 x () ( )
2 2
where ()
c0 t =
1
e
i 0 t
c1 t = () 1
e
i1t
2 2
( ) ( )
x,t x,t =
1
2
0
x e ()
i 0 t
+
1
x e
i1t
()
0 x e i 0 t +
1 x e i1t
() ()
=
0
0 +
1
1 +
1
0 e ( 1 0 ) +
0
1e ( 1 0 ) =
02 +
12 + 2
0
1 cos vib t
1
2
i t i t 1
2 ( )
t = 0, 4, 2, 3 4,
( )
2
x,t
() ( )
2
1 x
()
2
0 x
()
x
2 0 1 cos ( t )
t = 0, 4, 2, 3 4,
( ) ( )
x =
x,t x̂ x,t dx
=
1
2
() () ()
0 x ei 0 t +
1 x ei1t
x
0 x e i 0 t +
1 x e i1t
dx ()
=
0 x
0 dx +
1 x
1 dx +
1 x
0 e ( 1 0 ) dx +
0 x
1e ( 1 0 ) dx
1 i t i t
2
( )
<x >0 = 0 <x >1 = 0 = cos vib t
0 x
1 dx
12
2
1
() ( ) ()
4
1 2
x 0 x =
xe
x 2 = 2
2
1 x
( ) ( ) () ( ) ( )
1 2
1 2
1 2
0 x 1 dx = 2
02 dx = 2 x t = 2 cos vib t
Normalization Gaussian
Hermite polynomial
( )
H0 y = 1 even ( n = 0)
H ( y) = 2 y
1
odd ( n = 1)
H ( y) = 4 y
2
2
2
even ( n = 2 )
H ( y ) = 8y
12 y
3
3
odd ( n = 3)
H ( y ) = 16 y
48y
4
4 2
+ 12 even ( n = 4 )
( ) ( ) ( ) ( )
n y2 n y2 d
y2
1 2 ye e +
1 e e = 2 yH n y
H n+1 y
dy dy n dy n+1
( ) ( )
H n+1 y
2 yH n y + 2nH n
1 y = 0 ( )
Substituting ( ) ( )
2 yH n y = H n+1 y + 2nH n
1 y above gives ( )
dH n y ( ) = 2nH
dy n
1 ( y)
Use these relations to solve for momentum <p>(t)
( ) ( )
p =
x,t p̂ x,t dx
=
1
2
() () ()
0 x ei 0 t +
1 x ei1t
pˆ
0 x e i 0 t +
1 x e i1t
dx ()
=
0 pˆ
0 dx +
1 pˆ
1 dx +
1 pˆ
0 e ( 1 0 ) dx +
0 pˆ
1e ( 1 0 ) dx
1 i t i t
2
<p >0 = 0 <p >1 = 0
1 1
d
() (
x ) e ()
4 2
x 2 2
0 x = =
1 x
dx 2
1 1
i t i t
1 p̂
0 e ( )
dx = i e ( 1 0 )
1
1 dx = i e vib
i 1 0 t
2 2
2
2
d
dx
1 x =
d
dx
()
2
( ) d
N 1 H1 1 2 x e
x 2 = 1 2 N 1 H1 y e
y 2
dy
2
( )
d d
with y 1 2 x dy = 1 2 dx dx =
1 2 dy = 1 2
dx dy
d
d
() ( ) ( )
2 2
1 x = 1 2 N1 H1 y e
y 2
yH1 y e
y 2
dx dy
d
( )
H y = 2nH 0 y = 2H 0 y
dy 1
( ) ( )
( )
yH1 y =
1
2
( )
1
( )
2nH 0 y + H 2 y = H 0 y + H 2 y
2
( ) ( )
1
d
() ( )
1
( ) 1
() ()
2 2
1 x = 1 2 N1 H 0 y e
y 2
H 2 y e
y 2 = 1 2 N1 0 x
2 x
dx 2 N0 2N 2
0 p̂
1e ( 1 0 ) dx = e ( 1 0 ) i ( ) d
i t i t
0
dx
dx 1
1
( )
= e ( 1 0 ) i 1 2 N1
1
i t
0
0 dx 0 2
dx
N0 2N 2
1
N1 i t
= e ( 1 0 ) i 1 2 ( )
2
i t
= i e vib
N0 2
Finally
1
1
()
2 i t
p t = i
e vib
e vib
1
2 2
i t
( ) =
2 sin t
2
vib( )
We can rearrange the Schrödinger equation for the HO into an interesting form ...
1 d
2
2
1
+ ( m x ) = p 2 + ( m x )2 = E
2m i dx 2m
with
1
H= p 2 + ( m x )2
2m
which has the same form as
u 2 + v 2 = ( iu + v )(
iu + v ) .
1
a± ( ip + m x )
2m
( a a ) f ( x ) = ( ip + m x )(
ip + m x ) f ( x )
1
+
2m
1
p 2 + ( m x )
im (xp
px) f (x)
2
=
2m
= { 1
2m
( i
)
p 2 + ( m x )
[ x, p ] f ( x )
2
2 }
a
a+ =
1
2m
( 1
p 2 + ( m x ) + =
2 1
2
H+ )
1
2
1
H = a
a+
2
1
H = a+ a
+
2
or
1
a± a ± = E
2
1 1
1
= a+ a
a+ +
= a+ a+ a
+ 1 +
= a+ a+ a
+
+
2
2
2
= a+ ( H + ) = ( E + )( a+ )
H ( a+ ) = (E + ) ( a+ )
1
= a
a
a+
1
= a
( H
) = a
( E
)
2
H ( a
) = ( E
) ( a
)
So, these are operators connecting states and if we can find one state then we can
use them to generate other wavefunctions and energies. In the parlance of the
trade the a± are known as LADDER operators or
a
0 = 0
1 d
+ m x 0 = 0
2m dx
d m m 2
00 =
xdx ln 0 =
2h
x + A0
m x2 1
( x) = A e
0 0
2
and E0 =
2
Now that we are firmly planted on the bottom rung of the ladder, we can
utilize a+ repeatedly to obtain other wavefunctions, n , and energies, En. That is,
m
n
2 x2 1
n ( x ) = An ( a+ ) e , with En = n +
2
Thus, for 1 we obtain
1 1
m 4 2m 2
m2 x2
1 ( x ) = A1
xe
What are the proportionality factors cn and dn ? For any functions f (x) and g(x)
f (a± )gdx =
(a f ) g dx
Proof:
1 d
f (a± g) dx =
(2m )1 2
f + m x
g dx
dx
recall that
1 1 d 1 d
12 [
a± = ip + mwx ] = i i dx + mwx = (2m )1 2 dx + mwx
(2m ) (2m )1 2
Integrate by parts
1 d
± dx + mwx
(a f ) g dx
f (a± g) dx = dx = f gdx =
(2mw)
So we can write
(a± n ) *(a± n )dx = (a a±
n ) * n dx
We now use
1 1
a± a ± n = En n and En = n + w
2 2
1 1
a+ a
± n = n + n
2 2
and therefore
a+ a
n = n n
And
1 1
a
a+
n = n + n
2
2
a
a+ n = ( n + 1) n
We can now calculate cn :
(a+ n ) * (a+ n )dx = cn
n+1 n +1dx =
( a
a+ n ) * n dx = (n + 1)
n n dx
2 * *
cn = n +1
dn = n
a+ n = ( n + 1) n+1
12
a
n = n1 2 n
1
1 1
n +1 = a+ n n
1 = a
n
( n + 1)12
n1 2
We can now use these equations to generate other wavefunctions. Thus, if we start
with 0 we obtain:
1
n=0 1 = a+ 0 = a+ 0
( 0 + 1)1 2
1 1
n =1 2 = a+ 1 = ( a+ )2 0
2 2
1 1
n=2 3 = a+ 2 = a+3 0
2 +1 3 2
1 1
n=3 4 = a+ 3 = a+4 0
3+1 4 3 2
So that n is
1
( a+ ) 0
n
n =
n!
Using the a± operators we can show this condition also holds for the HO
wavefunctions. The proof is as follows.
( a+ a
) dx = n m* n dx
*
m n
( n
m ) m* n dx = 0
n dx = 0
*
m
Potential Energy of the Harmonic Oscillator: We can now use the a± operators to
perform some illustrative calculations. Consider the potential energy associated with
the HO.
1 2 1
V= kx = m 2 x 2
2 2
and therefore
1 1
V = m 2 x 2 = m 2 n* x̂ 2 n dx
2 2
12 12
m
x̂ =
2m
( a+ + a
) p̂ = i
2
( a+
a
)
and
x2 =
2m
( a+ + a
) =
2
( a a + a+ a
+ a
a+ + a
a
)
2m + +
Dirac Notation: Before we evaluate this expression let’s introduce some new
notation that will make life simpler for us on future occasions. Instead of writing
the integral between ± we use brackets to denote this integral. The first
half is called a “bra” and the second a “ket”. That is, “bra”-c-“ket” notation is
m* n dx = mn
where the present of m* is understood. Using this notation, integrals for
x , p , and x 2 assume the form
m* x̂ n dx = m x̂ n and
m* p̂ n dx = m p̂ n
1 1 1
V = m 2 x 2 = m 2 n* x̂ 2 n dx = m 2 n x̂ 2 n
2 2 2
1
V = m 2 n a+ a+ n + n a+ a
n + n a
a+ n + n a
a
n
2m 2
yielding
h 1
V = n + n + 1
= n+
4 2 2
It is important to not get totally embroiled in the equations and neglect the
chemistry and physics. Accordingly, we should ask the question as to the physical
significance of this formula ?
Ehrenfest’s Theroem
12
x (t) = ( 2 ) cos ( vibt ) = cos ( vibt )
1 2
2 μ
and
1
12 12
μ
p (t) = i
2
2
(e i vib t
e
i vib t
) =
2
sin ( vibt )
d x (t)
For our original expression was …
dt
d x (t) d *
d *
d
= x̂ dx = x̂ dx + * x̂ dx
dt dt dt
dt
Recall the time dependent Schrödinger equation is
1
i = H or = H
t t i
Inserting these results into the expression above yields
d x (t) 1
( ) 1
x̂ ( Ĥ ) dx
*
= Ĥ x̂ dx + *
dt i i
=
1
i
( ) i
(
* x̂Ĥ Ĥx̂ dx = * Ĥx̂ x̂Ĥ dx
)
Evaluating the commutator (assuming that H is the HO Hamiltonian) we find
…
2 d 2 2 d 2
( )
Ĥx̂ x̂Ĥ f (x) =
2 μ dx 2
+ U(x)
ˆ
x f (x) x̂
2 μ dx 2
+ U(x) f (x)
2 df 2 i i
= 2 = p̂x = p̂x
2 μ dx μ μ
And therefore
d x (t) i *
dt
( 1
)
= Ĥx̂ x̂Ĥ dx = * p̂x dx
m
Thus, we can now obtain p̂x (t) without the lengthy calculation contained in
the HO lecture notes.
d x (t) d
12
μ
12
12
μ
p̂x (t) = sin ( t )
2
d F (t) i *
dt
(
= ĤF̂ F̂Ĥ dx
)
For further information see McQuarrie Problems 4-43 and 4-44, p 187-188.
∇ 2 WKH /DSODFLDQ
ª =2 2 º
« − 2m ∇ +V ( x̂, ŷ, ẑ ) »ψ ( x, y, z ) = Eψ ( x, y, z )
¬ ¼
=2 2
Ĥ = − ∇ +V ( x̂, yˆ, zˆ ) +DPLOWRQLDQ RSHUDWRU LQ '
2m
ª =2 ∂ 2 º ª =2 ∂ 2 º ª =2 ∂ 2 º
H ( x, y, z ) = « −
ˆ +V ( ˆ
x ) + −
» « 2m ∂y 2 +V ( ˆ
y ) + −
» « 2m ∂z 2 +V ( ˆ
z ) »
¬ 2m ∂x
2 x y z
WKHQ ¼ ¬ ¼ ¬ ¼
= Ĥ x + Ĥ y + Ĥ z
6FKU|GLQJHU·V (T EHFRPHV
ª Ĥ x + Ĥ y + Ĥ z ºψ ( x, y, z ) = Eψ ( x, y, z )
¬ ¼
)LUVW WHUP
Ĥ xψ x ( x )ψ y ( y )ψ z ( z ) = ψ y ( y )ψ z ( z ) Ĥ xψ x ( x ) = ψ y ( y )ψ z ( z ) E xψ x ( x )
= E xψ x ( x )ψ y ( y )ψ z ( z )
Ĥ ψ = E ψ
ª Hˆ x + Hˆ y + Hˆ z º ª¬ψ x ( x )ψ y ( y )ψ z ( z )º¼ = ( E x + E y + E z ) ª¬ψ x ( x )ψ y ( y )ψ z ( z )º¼
¬ ¼
E = Ex + E y + Ez
7KXV LI WKH +DPLOWRQLDQ KDV WKLV VSHFLDO IRUP WKH HLJHQIXQFWLRQV RI WKH '
+DPLOWRQLDQ DUH MXVW SURGXFWV RI WKH HLJHQIXQFWLRQV RI WKH ' +DPLOWRQLDQ DQG
WKH VLWXDWLRQ LV HTXLYDOHQW WR GRLQJ WKUHH VHSDUDWH ' SUREOHPV
Ĥ = Ĥ x + Ĥ y + Ĥ z
7KH ' SURGXFW VWDWHV DUH QDWXUDOO\ QRUPDOL]HG LI WKH ' ZDYHIXQFWLRQV DUH
QRUPDOL]HG
³³³ψ ( x )ψ ( y )ψ ( z )ψ ( x )ψ ( y )ψ ( z ) dx dy dz =
* * *
x y z x y z
³ψ ( x )ψ ( x ) dx ³ψ ( y )ψ ( y ) dy ³ψ ( z )ψ ( z ) dz = 1
* * *
x x y y z z
[ [
1RWLFH WKDW IRU ψ ZH DUH IUHH WR FKRRVH DQ\ HLJHQIXQFWLRQ ψ nx RI Ĥ x
³³³ψ 1* ( x )ψ 1* ( y )ψ 1* ( z )ψ 3 ( x )ψ 1 ( y )ψ 1 ( z ) dx dy dz
x y z x y z
³ψ 1* ( x )ψ 3 ( x ) dx ³ψ 1* ( y )ψ 1 ( y ) dy ³ψ 1* ( z )ψ 1 ( z ) dz = 0
x x y y z z
[ [
%HFDXVH WKH SURGXFW VWDWHV DUH RUWKRJRQDO DQG QRUPDOL]HG WKH\·UH RUWKRQRUPDO
DQG ZH VXPPDUL]H WKLV E\ ZULWLQJ
³³³ ( ) (
ψ nxnynz * x, y, z ψ mxmymz x, y, z dx dy dz = δ nx ,mxδ ny ,myδ nz ,mz )
([DPSOH ' +DUPRQLF 2VFLOODWRU
©π ¹
x y z x z nz z
ª§ 1· º ª§ 1· º ª§ 1· º
Enxn y nz = «¨ n x + ¸ =ωx » + «¨ n y + ¸ =ω y » + «¨ n x + ¸ =ωz »
¬© 2¹ ¼ ¬© 2¹ ¼ ¬© 2¹ ¼
( mk x )
12
kx
αx = ωx = etc.
= m
1RWLFH DJDLQ WKDW IRU WKH ' SUREOHP ZH KDYH TXDQWXP QXPEHUV DQG WKH
HQHUJ\ DQG ZDYHIXQFWLRQ GHSHQG RQ DOO WKUHH VLPXOWDQHRXVO\
'HJHQHUDFLHV
,Q ' WKHUH DUH D QXPEHU RI LQWHUHVWLQJ WKLQJV WKDW FDQ KDSSHQ WKDW ZH GLGQ·W
VHH LQ ' 2QH H[DPSOH RI WKLV LV WKDW LQ ' LW LV SRVVLEOH IRU WZR GLIIHUHQW
HLJHQIXQFWLRQV RI WKH +DPLOWRQLDQ WR KDYH WKH VDPH HQHUJ\ :KHQ WKLV KDSSHQV
WKHVH WZR VWDWHV DUH FDOOHG GHJHQHUDWH ,I WKHUH DUH WKUHH IRXU« VWDWHV ZLWK
WKH VDPH HQHUJ\ WKH\ DUH VDLG WR EH WKUHHIROG IRXUIROG « GHJHQHUDWH 7KLV
QHYHU KDSSHQHG IRU WKH 3DUWLFOH LQ D %R[ RU WKH +DUPRQLF 2VFLOODWRU ,Q IDFW
RQH FDQ VKRZ WKDW IRU ERXQG VWDWHV LQ ' RQH QHYHU KDV GHJHQHUDF\ HYHU\ VWDWH
KDV LWV RZQ HQHUJ\ +RZHYHU LQ ' ZH FDQ ILQG GHJHQHUDF\ YHU\ HDVLO\ )RU
H[DPSOH LI RXU VSULQJ FRQVWDQWV DUH DOO WKH VDPH
kx = k y = kz ≡ k ωx = ω y = ωz ≡ ω
§ 3·
Enxn y nz = =ω ¨ n x + n y + n z + ¸
© 2¹
7KH JURXQG VWDWH KDV DQ HQHUJ\ E000 = 23 =ω 7KHUH LV RQO\ RQH ZD\ , FDQ JHW WKLV
HQHUJ\ n x = 0, n y = 0, nz = 0 VR LW LV QRW GHJHQHUDWH +RZHYHU WKHUH DUH WKUHH
ZD\V , FDQ JHW WKH ILUVW H[FLWHG VWDWH HQHUJ\ 5
2
=ω n x = 1, ny = 0, nz = 0
n x = 0, ny = 1, nz = 0 RU n x = 0, ny = 0, nz = 1 6R ZH ILQG WKDW LI ZH FKRRVH DOO WKH
VSULQJ FRQVWDQWV WR EH HTXDO
3=ω
E000 = nondegenerate level
2
5=ω
E100 = E010 = E001 = 3-fold degenerate level
2
HWF
:H ZLOO W\SLFDOO\ GUDZ WKLV ZLWK D SLFWXUH OLNH
(
«
( 1RQGHJHQHUDWH
ZKHUH D DQG E DUH FRQVWDQWV 7KHQ LW WXUQV RXW WKDW ψ LV DOVR DQ HLJHQVWDWH RI
WKH +DPLOWRQLDQ 7R VHH WKLV
Ĥψ = Ĥ ( a ψ 010 + b ψ 001 ) = a Hˆ ψ 010 + b Hˆψ 001
=a 5=ω
2 ψ 010 + b 5=ω
2 ψ 001 = 5=2ω ( a ψ 010 + b ψ 001 )
= 5=2ω ψ
7KLV LOOXVWUDWHV WKH LPSRUWDQW SRLQW WKDW DQ\ VXP RI GHJHQHUDWH HLJHQVWDWHV LV
DOVR DQ HLJHQVWDWH RI WKH +DPLOWRQLDQ ZLWK WKH VDPH HLJHQYDOXH
6HWWLQJ WKH VSULQJ FRQVWDQWV HTXDO DPRXQWV WR DVVXPLQJ WKH ZHOO LV V\PPHWULF
ZLWK UHVSHFW WR [ \ DQG ] ,I WKH V\PPHWU\ LV ´EURNHQµ LH k x ≠ k y ≠ k z WKHQ
WKH DQG VWDWHV ZLOO EHFRPH QRQGHJHQHUDWH ,Q WKLV FDVH ZH ZLOO
XVXDOO\ VD\ WKDW WKH GHJHQHUDF\ KDV EHHQ ´OLIWHGµ RU WKDW WKH GHJHQHUDWH VWDWHV
KDYH EHHQ VSOLW 7KH ODWWHU ODQJXDJH FRPHV IURP WKH SLFWRULDO YLHZ LI WKH VSULQJ
FRQVWDQWV DUH RQO\ VOLJKWO\ GLIIHUHQW WKHQ WKH HQHUJ\ OHYHOV PLJKW ORRN OLNH
( (
( (
( (
(
+HUH \RX FDQ VHH WKDW WKH Q OHYHOV DUH DOPRVW GHJHQHUDWH WKH\·YH EHHQ
´VSOLWµ DQG WKH Q OHYHOV DUH DOPRVW GHJHQHUDWH EXW QRW TXLWH EHFDXVH WKH
IRUFH FRQVWDQWV DUH VOLJKWO\ GLIIHUHQW 1RWLFH WKDW LW LV SRVVLEOH WR EUHDN VRPH
GHJHQHUDFLHV EXW NHHS RWKHUV )RU H[DPSOH LI ZH FKRRVH
kx = k y ≡ k ≠ kz ωx = ω y ≡ ω ≠ ωz
7KHQ WKH HQHUJLHV EHFRPH
§ 1·
Enxn y nz = =ω ( n x + n y + 1) + =ωz ¨ nz + ¸
© 2¹
DQG ZH ILQG
1
E000 = =ω + =ωz nondegenerate level
2
1
E100 = E010 = 2=ω + =ωz 2-fold degenerate level
2
3
E001 = =ω + =ωz nondegenerate level
2
F V ( x, y, z ) = Vx ( x ) +V y ( y ) +Vz ( z )
Vx ( x ) = 0 0≤ x≤a
Vy ( y ) = 0 0≤ y≤b
Vz ( z ) = 0 0≤ z≤c
D \ Vx ( x ) ,V y ( y ) ,Vz ( z ) = ∞ otherwise
E
[
,QVLGH WKH ER[ 2XWVLGH WKH ER[
=2 § ∂ 2 ∂2 ∂2 ·
− + + ψ ( x, y, z ) = Eψ ( x, y, z ) ψ (x, y, z )= 0
2m ¨© ∂x 2 ∂y 2 ∂z 2 ¸¹
ψ ( x , y , z ) = ψ n ( x )ψ n
x y
( y )ψ n ( z )
z
= 2 § n x2 n y nz2 ·
2
En x n y n z = E n x + E n y + E nz = ¨ + + ¸
8m ¨© a 2 b 2 c 2 ¸¹
'HJHQHUDFLHV
'HJHQHUDFLHV RFFXU LQ D VLPLODU IDVKLRQ IRU WKH 3L% HJ LI D E F LQ RXU '
ER[
h2
2 ( x
En x n y n z = n 2 + n 2y + nz2 )
8ma
3h 2
E111 = LV QRQGHJHQHUDWH
8ma 2
1
§ 8 ·2 §π x · §π y · §π z ·
ψ 111 ( x, y, z ) = ψ 1 ( x )ψ 1 ( y )ψ 1 ( z ) = ¨ 3 ¸ sin ¨ ¸ sin ¨ ¸ sin ¨ ¸
©a ¹ © a ¹ © a ¹ © a ¹
%XW«
h2
2 (
E211 = E121 = E112 = 22 +12 + 12 ) IROG GHJHQHUDF\
8ma
( 1RQGHJHQHUDWH
( ( (
IROG GHJHQHUDWH
( ( (
1RQGHJHQHUDWH
(
h2 § 1 1 4 · h2 § 1 4 1·
E112 = ¨ 2 + 2 + 2 ¸ ≠ E121 = ¨ 2+ 2+ 2¸
8m © a b c ¹ 8m © a b c ¹
6XPPDU\
' ER[
1
§ 8 ·2 § n xπ x · § n yπ y · § n zπ z ·
ψ nxn ynz ( x, y , z ) = ¨ ¸ sin ¨ ¸ sin ¨ ¸ sin ¨ ¸
© abc ¹ © a ¹ © b ¹ © c ¹
h 2 § n x2 n y nz2 ·
2
5LJLG 5RWDWLRQV
&RQVLGHU WKH URWDWLRQ RI WZR SDUWLFOHV DW D IL[HG GLVWDQFH R IURP RQH DQRWKHU
ω
r1 + r2 ≡ R
P
U m1r1 = m2 r2 FHQWHU RI PDVV &20
&20 U P
7KHVH WZR SDUWLFOHV FRXOG EH DQ HOHFWURQ DQG D SURWRQ LQ ZKLFK FDVH ZH·G EH
ORRNLQJ DW D K\GURJHQ DWRP RU WZR QXFOHL LQ ZKLFK FDVH ZH·G EH ORRNLQJ DW D
GLDWRPLF PROHFXOH &ODVVLFDOO\ HDFK RI WKHVH URWDWLQJ ERGLHV KDV DQ DQJXODU
PRPHQWXP Li = I iω ZKHUH ω LV WKH DQJXODU YHORFLW\ DQG ,L LV WKH PRPHQW RI
LQHUWLD Ii = mri 2 IRU WKH SDUWLFOH 1RWH WKDW LQ WKH &20 WKH WZR ERGLHV PXVW
KDYH WKH VDPH DQJXODU IUHTXHQF\ 7KH FODVVLFDO +DPLOWRQLDQ IRU WKH SDUWLFOHV LV
L12 L2 2 1
= m1r12ω 2 + m2 r22ω 2 = ( m1r12 + m2 r22 ) ω 2
1 1
H= +
2I1 2I 2 2 2 2
,QVWHDG RI WKLQNLQJ RI WKLV DV WZR URWDWLQJ SDUWLFOHV LW ZRXOG EH UHDOO\ QLFH LI ZH
FRXOG WKLQN RI LW DV RQH HIIHFWLYH SDUWLFOH URWDWLQJ DURXQG WKH RULJLQ :H FDQ GR
WKLV LI ZH GHILQH WKH HIIHFWLYH PRPHQW RI LQHUWLD DV
m1 m2
I = m1r12 + m2 r22 = μ r02 μ=
m1 + m2
ZKHUH LQ WKH VHFRQG HTXDOLW\ ZH KDYH QRWHG WKDW WKLV WZR SDUWLFOH V\VWHP
EHKDYHV DV D VLQJOH SDUWLFOH ZLWK D UHGXFHG PDVV μ URWDWLQJ DW D GLVWDQFH R
IURP WKH RULJLQ 7KXV ZH KDYH
]
1 2 L2
H = Iω =
2 2I
ZKHUH LQ WKH VHFRQG HTXDOLW\
ZH KDYH GHILQHG WKH DQJXODU μ
U
PRPHQWXP IRU WKLV HIIHFWLYH θ
SDUWLFOH L = Iω 7KH SUREOHP
LV QRZ FRPSOHWHO\ UHGXFHG WR D
\
ERG\ SUREOHP ZLWK PDVV μ φ
SRVLWLRQV UHODWLYH WR RQH DQRWKHU ² VD\ WKH DWRPV LQ D FU\VWDO ² DQG ZH URWDWH WKH
ZKROH DVVHPEO\ ZLWK DQ DQJXODU YHORFLW\ ω DERXW D JLYHQ D[LV r WKHQ E\ D VLPLODU
PHWKRG ZH FDQ UHGXFH WKH FROOHFWLYH URWDWLRQ RI DOO RI WKH REMHFWV WR WKH
URWDWLRQ RI D VLQJOH ´HIIHFWLYHµ REMHFW ZLWK D PRPHQW RI LQHUWLD ,U ,Q WKLV
PDQQHU ZH FDQ WDON DERXW URWDWLRQV RI D PROHFXOH RU D ERRN RU D SHQFLO ZLWKRXW
KDYLQJ WR WKLQN DERXW WKH PRYHPHQW RI HYHU\ VLQJOH HOHFWURQ DQG TXDUN
LQGLYLGXDOO\
+HQFHRQH JHWV GLIIHUHQW DQVZHUV GHSHQGLQJ RQ ZKDW RUGHU WKH URWDWLRQV DUH
SHUIRUPHG LQ *LYHQ RXU H[SHULHQFH ZLWK TXDQWXP PHFKDQLFV ZH PLJKW GHILQH DQ
RSHUDWRU R̂x R̂y WKDW URWDWHV DURXQG [ \ 7KHQ ZH ZRXOG ZULWH WKH DERYH
H[SHULPHQW VXFFLQFWO\ DV Rˆ Rˆ ≠ Rˆ Rˆ 7KLV UDWKHU SURIRXQG UHVXOW KDV QRWKLQJ
x y y x
i j k
Lˆ = rˆ × p̂ = ( x̂, yˆ, zˆ ) × ( p̂ x , p̂ y , pˆ z ) = x̂ ŷ ẑ
p̂ x p̂ y p̂z
= ( yp ˆ ˆ y ) i + ( zp
ˆ ˆ z − zp ˆ ˆ z ) j + ( xp
ˆ ˆ x − xp ˆˆx )k
ˆ ˆ y − yp
∴ Lˆx = yp
ˆ ˆ z − zp
ˆˆ y Lˆ y = zp
ˆˆ x − xp
ˆˆz Lˆ y = xp
ˆ ˆ y − yp
ˆˆx
1RZ LW WXUQV RXW WKDW WKHVH DQJXODU PRPHQWXP RSHUDWRUV GR QRW FRPPXWH )RU
H[DPSOH
+ L̂x L̂y = + ( ŷp̂ z − ẑp̂ y )(ẑp̂ x − x̂p̂ z ) = + ŷp̂ z ẑp̂ x − ẑp̂ y ẑp̂ x − yˆ pˆ z xˆpˆ z + ẑp̂ y x̂p̂ z
− L̂y L̂x = −(ẑp̂ x − x̂p̂ z )( ŷp̂ z − ẑp̂ y ) = −ẑp̂ x ŷp̂ z + ẑp̂ x ẑp̂ y + xˆpˆ z yˆ pˆ z − x̂p̂ z ẑp̂ y
[ ]
Lˆ x , Lˆ y = ŷp̂ x [ p̂ z , ẑ ] + x̂p̂ y [ẑ, p̂ z ]
⎯⎯⎯
x→y
→ ª¬ L̂z , Lˆ x º¼ = i=Lˆ y
y→z
z→x
%HFDXVH WKH DQJXODU PRPHQWXP RSHUDWRUV GR QRW FRPPXWH WKH GR QRW VKDUH D
FRPPRQ VHW RI HLJHQIXQFWLRQV 7KXV ZH DUULYH DW WKH LPSRUWDQW FRQFOXVLRQ WKDW
D V\VWHP FDQQRW VLPXOWDQHRXVO\ KDYH ZHOOGHILQHG DQJXODU PRPHQWXP DURXQG
WKH [ \ DQG ] D[HV 7KH EHVW ZH FDQ KRSH IRU LV ZHOOGHILQHG DQJXODU
PRPHQWXP DURXQG D[LV
:H DUH QRZ OHIW ZLWK WKH IDLUO\ GLIILFXOW SUREOHP RI VROYLQJ IRU WKH HLJHQYDOXHV
(O DQG HLJHQVWDWHV Ylm RI WKH ULJLG URWRU 1RWLFH WKDW WKH HLJHQYDOXHV RI WKH
ª 1 ∂ § ∂ · 1 ∂2 º
L̂2 = L̂2x + L̂2Y + L̂2z L̂2 = − = 2 « ¨ sin θ ¸+
¬ sin θ ∂θ © ∂θ ¹ sin 2 θ ∂φ 2 »¼
$V D UHVXOW WKH 6FKU|GLQJHU HTXDWLRQ IRU WKH ULJLG URWRU EHFRPHV
−= 2 ª 1 ∂ § ∂ · 1 ∂2 º m
2» l (
« sin θ ∂θ ¨ sin θ ¸ + Y θ , φ ) = ElYl m (θ , φ )
2I ¬ © ∂θ ¹ sin θ ∂φ ¼
2
$V LW WXUQV RXW WKHUH DUH VRPH YHU\ QLFH RSHUDWRU DOJHEUD WULFNV WKDW DOORZ XV
WR JHW WKH HLJHQYDOXHV (O TXLFNO\ 7KHVH WULFNV DUH DQDORJRXV WR WKH UDLVLQJ DQG
ORZHULQJ RSHUDWRU PDQLSXODWLRQV ZH XVHG WR VROYH WKH +DUPRQLF RVFLOODWRU DQG ZH
ZLOO XVH WKHP WR GHULYH WKH HLJHQYDOXHV LQ WKH QH[W OHFWXUH 7KH UHVXOW LV WKDW
WKH HQHUJLHV RI WKH ULJLG URWRU REH\
=2
El = l ( l +1) l = 0, 1, 2,...
2I
2QH FRQIXVLQJ SRLQW DERXW DQJXODU PRPHQWXP LV WKDW IRU GLIIHUHQW NLQGV RI
DQJXODU PRPHQWXP ZH XVH GLIIHUHQW OHWWHUV 6R ZKLOH O LV WKH OHWWHU FKRVHQ IRU
DQ HOHFWURQ PRYLQJ DURXQG WKH QXFOHXV - LV WKH FKRVHQ OHWWHU IRU WKH URWDWLRQ RI
D GLDWRPLF PROHFXOH
=2
EJ = J ( J +1) J = 0, 1, 2,...
2I
7KXV WKH VSDFLQJ EHWZHHQ WKH HQHUJ\ OHYHOV LQFUHDVHV ZLWK LQFUHDVLQJ J XQOLNH
IRU WKH KDUPRQLF RVFLOODWRU ZKHUH WKH\ ZHUH HTXDOO\ VSDFHG
=2 =2
E J +1 − E J = ª( J +1)( J + 2 ) (
− J J +1)º¼ = ( J + 1)
2I ¬ I
)XUWKHU IRU HDFK J ZH KDYH PXOWLSOH SRVVLEOH YDOXHV RI m DOVR NQRZQ DV K LQ
VRPH FRQWH[WV:
m = 0, ± 1, ± 2,..., ±J
7KXV HDFK HQHUJ\ OHYHO LV JIROG GHJHQHUDWH 3K\VLFDOO\ m UHIOHFWV WKH
FRPSRQHQW RI WKH DQJXODU PRPHQWXP DORQJ WKH ]GLUHFWLRQ )RU IL[HG J WKH
GLIIHUHQW YDOXHV RI m UHIOHFW WKH GLIIHUHQW GLUHFWLRQV WKH DQJXODU PRPHQWXP
YHFWRU FRXOG EH SRLQWLQJ ² IRU ODUJH
SRVLWLYH m WKH DQJXODU PRPHQWXP LV ]
PRVWO\ DORQJ ] LI m LV ]HUR WKH m = +J
DQJXODU PRPHQWXP LV RUWKRJRQDO WR ]
3K\VLFDOO\ ZH NQRZ WKDW WKH HQHUJ\
RI WKH URWDWLRQ GRHVQ·W GHSHQG RQ
\
WKH GLUHFWLRQ 7KLV LV UHIOHFWHG LQ m=0
WKH IDFW WKDW WKH HQHUJ\ GHSHQGV [
RQO\ RQ J ZKLFK PHDVXUHV WKH OHQJWK m = −J
RI WKH YHFWRU QRW LWV GLUHFWLRQ
$QJXODU 0RPHQWXP
6LQFH L̂2 DQG L̂z FRPPXWH WKH\ VKDUH FRPPRQ HLJHQIXQFWLRQV 7KHVH IXQFWLRQV DUH
H[WUHPHO\ LPSRUWDQW IRU WKH GHVFULSWLRQ RI DQJXODU PRPHQWXP SUREOHPV ² WKH\
GHWHUPLQH WKH DOORZHG YDOXHV RI DQJXODU PRPHQWXP DQG IRU V\VWHPV OLNH WKH 5LJLG
5RWRU WKH HQHUJLHV DYDLODEOH WR WKH V\VWHP 7KH ILUVW WKLQJV ZH ZRXOG OLNH WR NQRZ
DUH WKH HLJHQYDOXHV DVVRFLDWHG ZLWK WKHVH HLJHQIXQFWLRQV :H ZLOO GHQRWH WKH
HLJHQYDOXHV RI L̂2 DQG L̂z E\ α DQG β UHVSHFWLYHO\ VR WKDW
L̂2Y β (θ , φ ) = α Y β (θ , φ )
α αLˆ Y β (θ , φ ) = β Y β (θ , φ )
z α α
)RU EUHYLW\ LQ ZKDW IROORZV ZH ZLOO RPLW WKH GHSHQGHQFH RI WKH HLJHQVWDWHV RQ θ
DQG φ VR WKDW WKH DERYH HTXDWLRQV EHFRPH
Lˆ 2Yαβ = α Yαβ Lˆ zYαβ = β Yαβ
,W LV FRQYHQLHQW WR GHILQH WKH UDLVLQJ DQG ORZHULQJ RSHUDWRUV QRWH WKH VLPLODULW\ WR
WKH +DUPRQLF RVFLOODWRU
L̂± ≡ L̂ x ± iL̂ y
:KLFK VDWLVI\ WKH FRPPXWDWLRQ UHODWLRQV
ª L̂+ , L̂− º = 2=L̂z ª L̂z , L̂± º = ± =L̂± ª L̂± , L̂2 º = 0
¬ ¼ ¬ ¼ ¬ ¼
7KHVH UHODWLRQV DUH UHODWLYHO\ HDV\ WR SURYH XVLQJ WKH FRPPXWDWLRQ UHODWLRQV ZH·YH
DOUHDG\ GHULYHG
ªˆ ˆ º ˆ ªˆ ˆ º ˆ ªˆ ˆ º ˆ ª ˆ2 ˆ º
¬ L x , L y ¼ = i=Lz ¬ L y , Lz ¼ = i=L x ¬ Lz , L x ¼ = i=Ly ¬ L , Lz ¼ = 0
)RU H[DPSOH
ª L̂z , L̂± º = ª L̂z , L̂ x º ± i ª L̂z , L̂ y º
¬ ¼ ¬ ¼ ¬ ¼
(
= i=L y ± i ( −i=L x ) = ± = L x ± iLy )
= ± =L̂±
7KH UDLVLQJ DQG ORZHULQJ RSHUDWRUV KDYH D SHFXOLDU HIIHFW RQ WKH HLJHQYDOXH RI L̂z
Lˆ z (Lˆ ±Yαβ ) = ( ª¬ Lˆ z , Lˆ ± º¼ + Lˆ ± Lˆ z )Yαβ = ( ± =Lˆ ± + Lˆ ± β )Yαβ = ( β ± = ) (Lˆ ±Yαβ )
7KXV L̂+ L̂− UDLVHV ORZHUV WKH HLJHQYDOXH RI L̂z E\ = KHQFH WKH QDPHV 6LQFH
WKH UDLVLQJ DQG ORZHULQJ RSHUDWRUV FRPPXWH ZLWK L̂2 WKH\ GR QRW FKDQJH WKH YDOXH
RI α DQG VR ZH FDQ ZULWH
Lˆ ±Yαβ ∝ Yαβ ±=
DQG VR WKH HLJHQYDOXHV RI L̂z DUH HYHQO\ VSDFHG
:KDW DUH WKH OLPLWV RQ WKLV ODGGHU RI HLJHQYDOXHV" 5HFDOO WKDW IRU WKH KDUPRQLF
RVFLOODWRU ZH IRXQG WKDW WKHUH ZDV D PLQLPXP HLJHQYDOXH DQG WKH HLJHQVWDWHV FRXOG
7KLV UHVXOW VLPSO\ UHIOHFWV WKH IDFW WKDW LI \RXWDNH DQ\ REVHUYDEOH RSHUDWRU DQG
VTXDUH LW \RXPXVW JHW EDFN D SRVLWLYH QXPEHU 7R JHW D QHJDWLYH YDOXH IRU WKH
DYHUDJH YDOXH RI L̂2x RU L̂2y ZRXOG LPSO\ DQ LPDJLQDU\ HLJHQYDOXH RI L̂ x RU L̂ y ZKLFK LV
LPSRVVLEOH VLQFH WKHVH RSHUDWRUV DUH +HUPLWLDQ %HVLGHV ZKDW ZRXOG DQ LPDJLQDU\
DQJXODU PRPHQWXP PHDQ" :H QRZ DSSO\ WKH DERYH HTXDWLRQ IRU WKH VSHFLILF
ZDYHIXQFWLRQ Yαβ
³ ( ) ³ (
0 ≤ Yαβ * L̂2x + L̂2y Yαβ = Yαβ * L̂2 − L̂2z Yαβ )
= ³ Yαβ * (α − β 2 ) Yαβ
=α −β2
+HQFH β 2 ≤ α DQG WKHUHIRUH − α ≤ β ≤ α :KLFK PHDQV WKDW WKHUH DUH ERWK
PD[LPXP DQG PLQLPXP YDOXHV WKDW β FDQ WDNH RQ IRU D JLYHQ α ,I ZH GHQRWH WKHVH
YDOXHV E\ βPD[ DQG βPLQ UHVSHFWLYHO\ WKHQ LW LV FOHDU WKDW
Lˆ +Yαβmax = 0 Lˆ −Yαβmin = 0
:H FDQ WKHQ XVH WKLV NQRZOHGJH DQG VRPH DOJHEUD WULFNV WULFN WR GHWHUPLQH WKH
UHODWLRQVKLS EHWZHHQ α DQG βPD[ RU βPLQ )LUVW QRWH WKDW
Lˆ − Lˆ +Yαβ max = 0 Lˆ + Lˆ −Yαβ min = 0
:H FDQ H[SDQG WKLV H[SOLFLWO\ LQ WHUPV RI L̂ x DQG L̂ x
( )
Lˆ2x + Lˆ2y − i( Lˆ y Lˆ x − Lˆ x Lˆ y ) Yαβ max = 0 (
Lˆ2x + Lˆ2y + i( Lˆ y Lˆ x − Lˆ x Lˆ y ) Yαβ min = 0 )
+RZHYHU WKLV LV QRW WKH PRVW FRQYHQLHQW IRUP IRU WKH RSHUDWRUV EHFDXVH ZH GRQ·W
NQRZ ZKDW L̂ x RU L̂ y JLYHV ZKHQ DFWLQJ RQ Yαβ +RZHYHU ZH FDQ UHZULWH WKH VDPH
H[SUHVVLRQ LQ WHUPV RI L̂2 DQG L̂z
(
Lˆ2x + Lˆ2y ± i( Lˆ y Lˆ x − Lˆ x Lˆ y ) )
L̂2 − L̂2z −i=Lˆ z
6R WKHQ ZH KDYH
( )
L̂2 − L̂2z − =L̂z Yαβ max = 0 ( L̂ − L̂ + =L̂ ) Yαβ = 0
2 2
z z
min
(α − β 2
max − =β max = 0 ) ( α − β + =β ) = 0
2
min min
ZKHUH LQ WKH ODVW OLQH ZH KDYH VLPSO\ GHILQHG D QHZ YDULDEOH l WKDW LV GLPHQVLRQOHVV
QRWLFH WKDW = KDV WKH XQLWV RI DQJXODU PRPHQWXP 6R FRPELQLQJ WKHVH PLQLPXP
DQG PD[LPXP YDOXHV ZH KDYH WKDW − =l ≤ β ≤ =l )XUWKHU VLQFH ZH FDQ JHW IURP WKH
ORZHVW WR WKH KLJKHVW HLJHQYDOXH LQ LQFUHPHQWV RI = E\ VXFFHVVLYH DSSOLFDWLRQV RI
WKH UDLVLQJ RSHUDWRU LW LV FOHDU WKDW WKH GLIIHUHQFH EHWZHHQ WKH KLJKHVW DQG
ORZHVW YDOXHV > =j − ( −=j ) = 2=l @ PXVW EH DQ LQWHJHU PXOWLSOH RI = 7KXV l LWVHOI PXVW
HLWKHU EH DQ LQWHJHU RU D KDOILQWHJHU
ZKHUH ZH KDYH UHSODFHG α ZLWK l DQG β ZLWK m VR WKDW Yαβ EHFRPHV Ylm $OVR LQ WKH
ILUVW HTXDWLRQ ZH KDYH QRWHG WKDW 0 ≤ L̂2 = = 2l ( l + 1) LPSOLHV l ≥ 0 7KHVH DUH WKH
IXQGDPHQWDO HLJHQYDOXH HTXDWLRQV IRU DOO IRUPV RI DQJXODU PRPHQWXP
1RWLFH WKDW WKHUH LV D GLIIHUHQFH KHUH IURP ZKDW ZH VDZ IRU WKH ULJLG URWRU
7KHUH ZH KDG
=2
EJ = J ( J +1) J = 0, 1, 2,...
2I
ZKHUH DV D UHPLQGHU WKH TXDQWXP QXPEHU J IRU WKH ULJLG URWRU LV HTXLYDOHQW WR WKH
TXDQWXP QXPEHU l GHILQH DERYH +HUH WKH GHSHQGHQFH RI WKH HQHUJ\ RQ J ²
E ∝ J ( J + 1) LV WKH VDPH DV ZH IRXQG LQ RXU GHULYDWLRQ IRU L̂2 7KH IDFWRU RI 1 / 2I
VLPSO\ DULVHV IURP WKH IDFW WKDW WKH ULJLG URWRU +DPLOWRQLDQ LV L̂2 / 2I UDWKHU WKDQ
L̂2 7KH UHDO GLIIHUHQFH LV WKDW KDOI LQWHJHU YDOXHV RI J GR QRW DSSHDU IRU WKH
ULJLG URWRU $W ILUVW \RXPLJKW WKLQN WKLV PHDQV ZH PDGH D PLVWDNH LQ RXU
GHULYDWLRQ DERYH DQG WKDW l VKRXOG RQO\ EH DQ LQWHJHU DQG QRW D KDOI LQWHJHU
+RZHYHU WKHUH LV QR HUURU 7KH GLIIHUHQFH DULVHV EHFDXVH RXU GHULYDWLRQ DERYH LV
YDOLG IRU DQ\ NLQG RI DQJXODU PRPHQWXP 7KXV ZKLOH FHUWDLQ YDOXHV RI l PD\ QRW
DSSHDU IRU FHUWDLQ W\SHV RI DQJXODU PRPHQWXP HJ WKH\ GRQ·W RFFXU IRU WKH ULJLG
URWRU ZH ZLOO VHH ODWHU RQ WKDW WKH\ FDQ DSSHDU IRU RWKHU W\SHV RI DQJXODU
PRPHQWXP 0RVW QRWDEO\ HOHFWURQV KDYH DQ LQWULQVLF VSLQ DQJXODU PRPHQWXP ZLWK
l = 12 7KXV ZKLOH LQGLYLGXDO V\VWHPV PD\ KDYH DGGLWLRQDO UHVWULFWLRQV RQ WKH
DOORZHG YDOXHV RI l DQJXODU PRPHQWXP VWDWHV DOZD\V REH\ WKH DERYH HLJHQYDOXH
UHODWLRQV
$1*8/$5 020(1780
1RZ WKDW ZH KDYH REWDLQHG WKH JHQHUDO HLJHQYDOXH UHODWLRQV IRU DQJXODU
PRPHQWXP GLUHFWO\ IURP WKH RSHUDWRUV ZH ZDQW WR OHDUQ DERXW WKH DVVRFLDWHG
ZDYH IXQFWLRQV 5HWXUQLQJ WR VSKHULFDO SRODU FRRUGLQDWHV ZH UHFDOO WKDW WKH
DQJXODU PRPHQWXP RSHUDWRUV DUH JLYHQ E\
§ ∂ ∂ ·
L̂x = −i= ¨ − sin φ − cot θ cos φ ¸
© ∂θ ∂φ ¹
§ ∂ ∂ ·
L̂y = −i= ¨ cos φ − cot θ sin φ
© ∂θ ∂φ ¸¹
∂
L̂z = −i=
∂φ
2 ª 1 ∂ § ∂ · 1 ∂2 º
Lˆ2 = Lˆ2x + Lˆ2Y + Lˆ2z ˆ
L = −= «
2
¨ sin θ ¸+
¬ sin θ ∂θ © ∂θ ¹ sin 2 θ ∂φ 2 »¼
,Q WHUPV RI WKHVH RXU RULJLQDO 6FKU|GLQJHU (TXDWLRQ IRU ULJLG URWDWLRQV ZDV
ˆ L̂2 m
HYl = Yl = ElYl m
m
2I
−= ª 1 ∂ §
2
∂ · 1 ∂2 º m
2» l (
« ¨ sin θ ¸+ 2 Y θ , φ ) = ElYl m (θ , φ )
2I ¬ sin θ ∂θ © ∂θ ¹ sin θ ∂φ ¼
2
ZKHUH l ZDV WKH TXDQWXP QXPEHU IRU L̂ DQG m ZDV WKH TXDQWXP QXPEHU IRU L̂z
2
7DNLQJ ZKDW ZH OHDUQHG LQ WKH ODVW VHFWLRQ DERXW WKH HLJHQYDOXHV RI L̂ DQG L̂z
ZH FDQ VD\ WKDW DW PRVW ZH FDQ KDYH
l = 0, 12 ,1, 23 , 2,... m = −l, −l + 1,..., l
:H ZLOO VHH WKDW WKHUH LV DQ DGGLWLRQDO UHVWULFWLRQ RQ WKH SRVVLEOH YDOXHV RI l LQ
WKH SUHVHQW FDVH EXW WKHVH DUH WKH SRVVLEOH YDOXHV IRU WKH TXDQWXP QXPEHUV ,Q
WHUPV RI WKH TXDQWXP QXPEHUV ZH KDYH WKH HLJHQYDOXH UHODWLRQV
Lˆ2Yl m = L2Yl m = = 2l ( l +1) Yl m
L̂zYl m = =mYl m
1RZ WKH IXQFWLRQV Ylm WKDW VDWLVI\ WKHVH UHODWLRQV IRU ULJLG URWDWLRQV DUH
FDOOHG 6SKHULFDO +DUPRQLFV ,W LV SRVVLEOH WR GHULYH WKH VSKHULFDO KDUPRQLFV E\
VROYLQJ WKH ' GLIIHUHQWLDO HTXDWLRQ DERYH 0F4XDUULH JRHV WKURXJK D IDLUO\
FRPSOHWH GHULYDWLRQ DQG ZH RXWOLQH WKDW VROXWLRQ LQ WKH DSSHQGL[ WR WKHVH QRWHV
EHORZ 7KH UHVXOW LV WKDW
Pl ( x ) LV DQ DVVRFLDWHG /HJHQGUH
m
ZKHUH Alm LV D QRUPDOL]DWLRQ FRQVWDQW DQG
3RO\QRPLDO 7KH ILUVW IHZ $VVRFLDWHG /HJHQGUH 3RO\QRPLDOV DUH
P00 ( cos θ ) = 1 P10 ( cos θ ) = cos θ
P11 ( cos θ ) = sin θ P20 ( cos θ ) = 1
2 ( 3cos θ −1)
2
∂φ ∂φ
2 m
7KH\ DUH DOVR HLJHQIXQFWLRQV RI L̂ DV FDQ EH SURYHQ IRU DQ\ JLYHQ Yl
DIWHU VRPH DOJHEUD E\ FRPSXWLQJ Lˆ Y DQG YHULI\LQJ WKDW WKH UHVXOW LV
2 m
l
MXVW = l ( l +1) Yl
2 m
• :H FDQ QRZ VHH ZK\ KDOILQWHJHU YDOXHV RI l DUH QRW DOORZHG KHUH 5HFDOO
WKDW φ LV WKH DQJOH LQ WKH [\ SODQH DQG LW YDULHV IURP WR π :KDW
VKRXOG KDSSHQ WR Yl
m
(θ , φ ) ZKHQ φ → φ + 2π " 2I FRXUVH WKH YDOXH RI WKH
ZDYHIXQFWLRQ VKRXOG QRW FKDQJH EHFDXVH E\ LQFUHPHQWLQJ φ E\ π ZH·YH
MXVW PRYHG WKH SDUWLFOH DURXQG LQ D IXOO FLUFOH 7KXV IRU WKH ZDYH
IXQFWLRQ WR EH VLQJOHYDOXHG ZH PXVW KDYH
Yl m (θ , φ ) = Yl m (θ , φ + 2π )
Alm Pl m ( cos θ ) eimφ = Alm Pl m ( cos θ ) eim(φ +2π )
eimφ = eim(φ +2π )
eimφ = eimφ eim 2π
1 = eim 2π
m = an integer
WKXV m PXVW EH DQ LQWHJHU QR PDWWHU ZKDW YDOXH RI l ZH FKRRVH +RZHYHU
VLQFH WKH PLQLPDO YDOXH IRU m LV l l PXVW DOVR EH DQ LQWHJHU 7KLV
FRQWLQXLW\ DUJXPHQW LV WKH UHDVRQ ZK\ KDOI LQWHJHU YDOXHV RI l DUH QRW
DOORZHG IRU ULJLG URWDWLRQV
• 1RWH WKDW WKHUH DUH D IHZ LQWHUHVWLQJ DOJHEUDLF IHDWXUHV RI WKH VSKHULFDO
KDUPRQLFV WKH φ SDUW RI WKH ZDYHIXQFWLRQ GRHV QRW GHSHQG RQ l 7KH
lWK RUGHU /HJHQGUH SRO\QRPLDO DOZD\V LQYROYHV VXPV RI SURGXFWV RI VLQHV
DQG FRVLQHV VXFK WKDW WKH VXP RI WKH VLQH DQG FRVLQH SRZHUV LV OHVV WKDQ
RU HTXDO WR l IRU m ≠ 0 WKH VSKHULFDO +DUPRQLFV DUH FRPSOH[ DQG
Ylm * = Yl−m 7KXV RQH FDQ REWDLQ WZR UHDO IXQFWLRQV IURP HDFK ±m SDLU YLD
Rlm = 1
2 (Y
l
m
+ Yl−m ) I lm = 1
i 2 (Y
l
m
− Yl−m )
7KHVH IHDWXUHV DUH KHOSIXO LQ WU\LQJ WR LGHQWLI\ ZKHQ JLYHQ DQ DUELWUDU\
IXQFWLRQ RI WKH DQJOHV ZKLFK VSKHULFDO KDUPRQLFV PLJKW FRQWULEXWH WR WKDW
IXQFWLRQ
• 7\SLFDOO\ WKH VSKHULFDO +DUPRQLFV DUH DVVRFLDWHG ZLWK OHWWHUV DV \RX KDYH
VHHQ LQ \RXU SUHYLRXV FKHPLVWU\ FRXUVHV 7KXV l LV ¶V· l LV ¶S· l LV ¶G·
«
• ,Q WKH DEVHQFH RI D SRWHQWLDO DV LV WKH FDVH IRU ULJLG URWDWLRQV WKH
VSKHULFDO +DUPRQLFV DUH 2l+1-IROG GHJHQHUDWH
l m Yl m ' s 2l + 1
Y00
Y1−1 , Y10 , Y11
Y2−2 , Y2−1 , Y20 , Y21 , Y22
Y −3 , Y −2 , Y −1 , Y 0 , Y 1 , Y 2 , Y 3
3 3 3 3 3 3 3
ª ∂ § ∂ · 2IE 2 º ∂2
«sin θ ∂θ ¨ sin θ ∂θ ¸ + = 2 sin θ » Y (θ , φ ) = − ∂φ 2 Y (θ , φ )
¬ © ¹ ¼
RQO\ θ RQO\ φ
:H·YH VHSDUDWHG WKH YDULDEOHV MXVW DV LQ WKH ' KDUPRQLF RVFLOODWRU
∴ 7U\ Y (θ , φ ) = Θ (θ ) Φ (φ ) DV D VROXWLRQ
2IE
'HILQH β≡ ( note β ∝ E )
=2
ª ∂ § ∂ · 2 º ∂2
«sin θ ∂θ ¨ sin θ ∂θ ¸ + β sin θ » Θ (θ ) Φ (φ ) = − ∂φ 2 Θ (θ ) Φ (φ )
¬ © ¹ ¼
sin θ ∂ § ∂ · 1 ∂2
θ ¸ ( )
Θ θ + β θ = − Φ (φ )
2
¨ sin sin
Θ (θ ) ∂θ © ∂θ ¹ Φ (φ ) ∂φ 2
RQO\ θ RQO\ φ
6LQFH θ DQG φ DUH LQGHSHQGHQW YDULDEOHV HDFK VLGH RI WKH HTXDWLRQ PXVW EH
HTXDO WR D FRQVWDQW ≡ P
1 ∂2
Φ (φ ) = − m 2 ,
Φ (φ ) ∂φ 2
sin θ ∂ § ∂ ·
¨ sin θ ¸ Θ (θ ) + β sin θ = m
2 2
DQG ,,
Θ (θ ) ∂θ © ∂θ ¹
∂ 2 Φ (φ )
= − m 2 Φ (φ )
∂φ 2
6ROXWLRQV DUH Φ (φ ) = Am e
imφ
and A− m e − imφ
Φ (φ + 2π ) = Φ (φ )
2π
1RUPDOL]DWLRQ ³ Φ ∗ (φ )Φ (φ ) d φ = 1
0
1 imφ
Φ (φ ) = e m = 0, ± 1, ± 2, ± 3,...
2π
1RZ OHW·V ORRN DW Θ (θ ) 1HHG WR VROYH ,,
sin θ ∂ § ∂ ·
¨ sin θ ¸ Θ (θ ) + β sin θ = m
2 2
Θ (θ ) ∂θ © ∂θ ¹
dx
&KDQJH YDULDEOHV x = cos θ Θ (θ ) = P ( x ) = dθ
− sin θ
6LQFH 0 ≤ θ ≤ π −1 ≤ x ≤ +1
sin θ
d
dθ
ª dΘ º
«sin θ dθ » +
¬ ¼
(β sin2 θ − m2 ) Θ (θ ) = 0
ZKLFK ZH FDQ UHZULWH
d 2Θ
sin 2 θ
dθ 2 + sin θ cos θ
dΘ
dθ (
+ β sin 2 θ − m 2 Θ (θ ) = 0 )
/HW x = cosθ DQG Θ(θ ) = P (x )
( )
dΘ dP dx dP 12 dP
= = − sin θ = − 1− x 2
dθ dx dθ dx dx
d 2Θ d ª dΘ º ª dx º d ª
( )
1 2 dP º
= = « − 1− x 2
dθ 2 dθ « dθ » « dθ » dx
¬ ¼ ¬ ¼ ¬ dx »¼
ª º
« 1 2 d 2P »
= − sin θ «
x dP
1 2 dx − 1− x
2
( » )
(
« 1− x 2
¬
) dx 2 »
¼
= −x
dP
dx (
+ 1− x 2
d 2P
dx 2
)
6XEVWLWXWLQJ WKHVH UHVXOWV LQWR /HJHQGUH·V HTXDWLRQ JLYHV
( ) d 2P
( ) ( ( ) )
2
2 dP
1− x 2 2 − 2x 1− x dx + β 1− x − m P ( x ) = 0
2 2
dx
'LYLGH E\ (1− x 2 ) WR REWDLQ WKH /HJHQGUH HTXDWLRQ LQ D FRQYHQLHQW IRUP
dP ª m2 º
(
1− x 2 ) d 2P
dx 2
− 2x + «
dx «¬
β − » P(x) = 0
1− x 2 »¼
7KH VROXWLRQV WR WKLV HTXDWLRQ DUH NQRZQ EXW YHU\ PHVV\ 7KH\ DUH FDOOHG WKH
DVVRFLDWHG /HJHQGUH SRO\QRPLDOV Pl
m
(x) 1RWH WKDW WKH\ RQO\ GHSHQG RQ _m_
EHFDXVH WKH HTXDWLRQ GHSHQGV RQ m2
Pl
m
(x )= P (cosθ )
l
m
( ) P (cos θ )= (3cos θ − 1)
1
P00 cos θ = 1 2
0 2
2
P (cos θ )= cos θ
1
0
P (cos θ )= 3cos θ sin θ
2
1
P (cos θ )= sin θ
1
1
P (cos θ )= 3sin θ
2
2 2
HWF
6R Θ (θ )= A P
m
(cosθ )
ª§ 2l + 1· l − m !º 2
Alm = «¨ »
( )
lm l
«© 2 ¸¹ l + m !»
¬ ¼ ( )
ZKHUH Alm LV WKH QRUPDOL]DWLRQ FRQVWDQW
2
( )
π
Alm2 ³ ª Pl cos θ º sin θ dθ = 1
m
0 ¬ ¼
ψ lm (r0 ,θ , φ )= Yl m (θ , φ )= Θl (θ )Φ m (φ )
m
Yl m
( )
ª§ 2l + 1· l − m !º
θ ,φ = «¨
(
» Pl m cos θ eimφ
) 2
( )
¸
«© 4π ¹ l + m !»
¬ ¼ ( )
7KHVH IXQFWLRQV DUH WKH VSKHULFDO KDUPRQLFV
( ) () ()
Yl m θ , φ = Θ l θ Φ m φ
m
( )
ª§ 2l + 1· l − m !º 2
Yl θ , φ = «¨
m
( )
» Pl m cos θ eimφ ( )
«© 4π ¹̧ l + m !»
¬ ¼ ( )
l = 0, 1, 2,... m = 0, ± 1, ± 2, ± 3,... ± l
( 4π ) © 16π ¹
12
1 1
§ 3 ·2 § 15 · 2
Y10 = ¨ ¸ cos θ Y2±1 = ¨ ¸ sin θ cos θ e ±iφ
© 4π ¹ © 8π ¹
1 1
§ 3 · 2 § 15 · 2
Y11 = ¨ ¸ sin θ eiφ Y2±2 = ¨ ¸ sin θ e
2 ±2iφ
© 8π ¹ © 32π ¹
1
§ 3 ·2
Y1−1 = ¨ ¸ sin θ e −iφ
© 8π ¹
³³ Y (θ ,φ )Y (θ ,φ )sin θ dθ dφ = δ δ mm′
m ′∗
Yl m ·V DUH RUWKRQRUPDO l′ l
m
ll ′
1 if l = l′ 1 if m = m′ normalization
.U|QHFNHU GHOWD δ ll′ = ® δ mm′ = ®
¯0 if l ≠ l′ ¯0 if m ≠ m′ orthogonality
SPHERICAL HARMONICS
( )
Yl m , = l m
m
() ()
1
( )
2 l + 1
l m !
Yl , =
m
( )
Pl m cos eim
2
( )
4 l + m !
( )
l = 0, 1, 2,... m = 0, ± 1, ± 2, ± 3,... ± l
( 4 )
12
16
1 1
3 2 15 2
Y10 = cos Y2±1 = sin cos e± i
4
8
1 1
3 2 15 2
Y11 = sin ei Y2±2 = sin 2 e±2i
8
32
1
3 2
Y11 = sin e i
8
1 if l = l 1 if m = m normalization
Krönecker delta ll = mm =
0 if l l 0 if m m orthogonality
Energies: ˆ m = E Y m)
(eigenvalues of HYl lm l
Recall
2IE
= 2 = l l +1 J J +1
( ) (
J = 0, 1, 2,... )
2
E
EJ =
2I
J J +1 ( )
62
J=3 E3 =
I
(
Y30 , Y3±1 , Y3±2 , Y3±3 7x degenerate )
32
J=2 E2 =
I
(
Y20 , Y2±1 , Y2±2 5x degenerate )
2
J=1 E1 =
I
(
Y10 , Y10 2x degenerate )
J=0 E0 = 0 (
Y00 nondegenerate )
Degeneracy of each state (
g J = 2J + 1 )
from m = 0, ± 1, ± 2,..., ± J
2 2
E J +1 E J = ( )(
J +1 J + 2 J J +1 =
2I
) (
I J +1 ) ( )
Transitions between rotational states can be observed through
spectroscopy, i.e. through absorption or emission of a photon
+ +
h Absorption
EJ EJ+1
- -
+ +
or Emission h
EJ EJ-1
- -
2
dμ
Strength of transition I JJ
dx J (ξ ⋅ μ) J
dx
2
E photon = h
J J +1
photon
J J +1
= Erot = E J +1 EJ =
I
( J + 1) photon
J J +1
=
4
h
2
I
( J + 1)
Define
h
B 2
rotational constant (Hz)
8 I
and
h
B 2
rotational constant (cm-1)
8 cI
J J +1 (
(Hz) = 2B J + 1 ) J J +1
(cm -1 ) = 2B J + 1 ( )
J=3 E3 = 12Bhc
E23 = 6Bhc
J=2 E2 = 6Bhc
E12 = 4Bhc
J=1 E1 = 2Bhc
E01 = 2Bhc
This gives rise to a rigid rotor absorption spectrum with evenly spaced lines.
J=0
2B
( )
J +1 J +2 J J +1 = 2B J + 1 + 1 2B J + 1 = 2B ( )
Use this to get microscopic structure of diatomic molecules directly from
the absorption spectrum!
h m1 m2
2B = I = μ r02 μ=
4 2 cI m1 + m2
1 1
h 2 h 2
r0 = 2 -1
(B in cm ) or r0 = 2 (B in Hz)
8 cB μ 8 B μ
HYDROGEN ATOM
Schrodinger equation in 3D spherical polar coordinates:
2 1 2
2
r +
1
2 μ
r 2 r r r 2 sin
sin +
1
( )
r, , + U r, , r, , = E r, ,
r 2 sin 2 2
( ) ( ) ( )
Ze2
with Coulomb potential U (r) =
4 0 r
Rewrite as
2 2
r
r r
()
+ 2 μr 2
U r E
r, , + L̂2
r, , = 0 ( ) ( )
r is separable is separable
( )
r, , = R r Yl m , () ( )
with L̂2Yl m ( , ) = l ( l + 1)Y ( , )
2
l
m
l = 0,1,2,...
2 d 2 dR r l l + 1
2
() ( )
2 μr 2 dr
r +
dr
2 μr 2
+ U r E R r = 0 () ()
()
Solutions R r are the H atom radial wavefunctions
32
Z
()
R r = 2
a0
e
Zr a0
exponential decay away from nucleus
with
Energy eigenvalues:
Z 2 e2 Z 2 μe4
E= = n = 1,2,3,...
8 0 a0 n2 8 02 h2 n2
Radial eigenfunctions:
12
( )
n l 1 ! 2Z
l +3 2
Rnl ()
r =
na
r le
Z r na0 2l +1 2 Zr
Ln+ l
( ) na0
3
2n n + l !
0
( )
where L2ln++1l 2Zr na0 are the associated Laguerre functions, the first few of which are:
n=1 l=0 L11 = 1
n=2 l=0 L12 = 2! 2 Zr
a0
l =1 L33 = 3!
2 2
n=3 l=0 L13 = 3! 3 2Zr + 2Z r 2
a0 9a0
l =1 L34 = 4! 4 2Zr
3a0
Normalization:
( ) ( )
2
Spherical harmonics 0
d d sin Yl m* , Yl m , = 1
0
() ()
Radial wavefunctions 0
dr r 2 Rnl* r Rnl r = 1
( ) () ( )
nlm r, , = Rnl r Yl m ,
( ) ( )
2
d sin d r 2 dr nlm r, , n ' l ' m ' r, , = nn ' ll ' mm '
*
0 0 0
() ( )
since components Rnl r Yl m , are normalized and orthogonal.
3/ 2
1 Z
n=1 l=0 m=0 100 = e = 1s
a0
3/ 2
Z
n=2 l=0 m=0 200 =
1
(2 ) e / 2
= 2s
32 a0
3/ 2
Z
1
l =1 m=0 210 = e / 2 cos = 2 p
32 a0
z
3/ 2
Z 1
l =1 m = ±1 21±1 = e / 2 sin e± i
64 a0
n En (a.u.)
1 -1/2
2 -1/8
3 -1/18
4 -1/32
5 -1/50
2 -1/8
1 -1/2 -109,680
(A)
(1012 Hz)
( )
2
Electron probability density given by r, ,
Probability that a 1s electron lies between r and r + dr of the nucleus:
( ) ( ) ( )
2 1 2r / a0 2
0
d d sin 1s* r, , 1s r, , r 2 dr = 4 a03
0
e r dr
P ORBITALS: wavefunctions
( ) (r a ) e
1/ 2 r / 2a0
m = 0 case: 210 = 2 p = 32 a03 0
cos
z
radial nodes n - l -1 = 0 ()
(note difference from 2s: Rnl r depends on l as well as n)
angular nodes l =1
total nodes n -1 = 1
( ) (r a ) e
1/ 2 r / 2a0
2 p = 32 a03 0
sin cos
m = ±1 case: Linear combinations give
x
= ( 32 a ) ( r a ) e
1/ 2 r / 2a0
2p 3
0 0
sin sin
y
r Rnl ( r )
2 2 a0
()
0.6
0.2
0.5
0.1
2s
0.4
0
1s 0 5 10 15 20 25 30
0.3
0.2 0.2
0.1 0.1
2p
0 0
0 5 10 15 0 5 10 15 20 25 30
0.2
0.1
3s
0
0 5 10 15 20 25 30
0.1
3p
0
0 5 10 15 20 25 30
0.1
3d
0
0 5 10 15 20 25 30
s = r/a0
μ=− e
L
2me
eBz
Ĥ = Ĥ 0 + L̂z
2me
Z 2 e2 eB
E= 2
+ z m
8 0 a0 n 2me
Applied
magnetic field
No magnetic
m = +1
field
2p m=0
m = -1
Energy
1s m=0
2p
1s Emission spectra
One line Three lines
Complex functions 211 and 21+1 are eigenfunctions of L̂z with eigenvalues ±m .
(/(&7521 63,1
8KOHQEHFN DQG *RXGVPLW VKRZHG WKDW WKHVH ZHUH WZR DQJXODU
PRPHQWXP VWDWHV ² WKH HOHFWURQ KDV LQWULQVLF DQJXODU PRPHQWXP ² 63,1
DQJXODU PRPHQWXP
3DXOL ([FOXVLRQ 3ULQFLSOH QR PRUH WKDQ HOHFWURQV SHU RUELWDO RU
QR WZR HOHFWURQV ZLWK DOO WKH VDPH TXDQWXP QXPEHUV $GGLWLRQDO TXDQWXP
QXPEHU QRZ FDOOHG PV ZDV SRVWXODWHG
7KHRUHWLFDO -XVWLILFDWLRQ
'LUDF GHYHORSHG UHODWLYLVWLF TXDQWXP WKHRU\ GHULYHG HOHFWURQ
VSLQ DQJXODU PRPHQWXP
1
Ŝ 2α = s ( s + 1) = 2α Ŝ 2 β = s ( s + 1) = 2 β s= always
2
1 1
Ŝ zα = ms =α msα = Ŝ z β = ms =β msβ = −
2 2
³ α α dσ = ³ β β d σ = 1 σ ≡ spin variable
* *
³ α β dσ = ³ β α d σ = 0
* *
Ψ ( r, θ , φ , σ ) = ψ ( r, θ , φ ) α (σ ) or ψ ( r, θ , φ ) β (σ )
HJ IRU + DWRP WKH JURXQG VWDWH WRWDO ZDYHIXQFWLRQV LQ DWRPLF XQLWV DUH
12 12
§ Z3 · § Z3 ·
Ψ = ¨ ¸ e − Zrα Ψ = ¨ ¸ e − Zr β
©π ¹ ©π ¹
1 1
100 100−
2 2
ZKLFK DUH RUWKRJRQDO DQG QRUPDOL]HG 1RWH WKH TXDQWXP QXPEHUV DUH QRZ
QOPP6
The basic idea is that we can write any electron spin state as a linear
combination of the two states α and β:
ψ ≡ cαα + cβ β
Note that, for now, we are ignoring the spatial part of the electron
wavefunction (e.g. the angular and radial parts of ψ). You might ask how we
can be sure that every state can be written in this fashion. To assure
yourself that this is true, note that for this state, the probability of finding
2 2
the electron with spin “up” (“down”) is cα ( cβ ). If there was a state that
could not be written in this fashion, there would have to be some other spin
state, γ, so that
ψ ≡ cαα + cβ β + cγ γ
2
In this case, however, there would be a probability cγ of observing the
electron with spin γ – which we know experimentally is impossible, as the
electron only has two observable spin states.
http://ocw.mit.edu/courses/chemistry/5-61-physical-chemistry-fall-2007/lecture-notes/
⎛ cα ⎞
ψ ≡ cαα + cβ β
ψ≡⎜ ⎟ →
⎝ cβ ⎠
Note that this is not a vector in physical (x,y,z) space but just a convenient
way to arrange the coefficients that define ψ. In particular, this is a nice
way to put a wavefunction into a computer, as computers are very adept at
dealing with vectors.
Now, our goal is to translate everything that we might want to do with the
wavefunction ψ into something we can do to the vector ψ . By going through
this stepbystep, we arrive at a few rules
Integrals are replaced with dot products. We note that the overlap
between any two wavefunctions can be written as a modified dot product
between the vectors. For example, if φ ≡ dαα + d β β then:
1 0 0 1
∫ φ ψ dσ = dα cα ∫ α α dσ + dα cβ ∫ α β dσ + dβ cα ∫ β α dσ + dβ cβ ∫ β β dσ
* * * * * * * * *
= dα *cα + d β *cβ
⎛ cα ⎞
(
= dα * )
dβ
* ⎜ ⎟
⎝ cβ ⎠
≡ φ† iψ
Where, on the last line, we defined the adjoint of a vector:
†
⎛ x1 ⎞
⎜ x ⎟ ≡ x1 x2 .
⎝ 2⎠
* *
( )
Thus, complex conjugation of the wavefunction is replaced by taking the
adjoint of a vector. Note that we must take the transpose of the vector
and not just its complex conjugate. Otherwise when we took the overlap we
would have (column) x (column) and the number of rows and columns would
not match.
⎛ cα ⎞
∫ φ ψ dσ = 0
*
→ (
φ iψ = dα * )
d β * ⎜ ⎟ = dα *cα + d β *cβ = 0
⎝ cβ ⎠
Now, the one challenge is that we have to determine which matrix belongs to
a given operator – that is, we need to fill in the question marks above. As an
exercise, we will show how this is done by working out the matrix
representations of Ŝ 2 , Ŝz , Ŝx and Ŝy .
We can use this information to obtain the unknown constants in the matrix
S2 :
3 � 3 � ⎛c d ⎞⎛1⎞ 3 2 ⎛1⎞
Ŝ 2α = � 2α → S 2α = � 2α ⇒ ⎜ ⎟⎜ ⎟ = 4 � ⎜ ⎟
4 4 ⎝e f ⎠⎝0⎠ ⎝0⎠
⎛ c ⎞ ⎛ 3 �2 ⎞
⇒⎜ ⎟=⎜4 ⎟
⎝e⎠ ⎝ 0 ⎠
so
c = 43 � 2 e=0
Operating on β
3 2 � 3 � ⎛c d ⎞⎛0⎞ 3 2 ⎛0⎞
Ŝ 2 β = �β → S 2α = � 2α ⇒ ⎜ ⎟⎜ ⎟ = 4 � ⎜ ⎟
4 4 ⎝e f ⎠⎝1⎠ ⎝1⎠
⎛d⎞ ⎛ 0 ⎞
⇒⎜ ⎟=⎜3 2⎟
⎝ f ⎠ ⎝4� ⎠
Thus:
d=0 f = 34 � 2
So for S 2 we have
3 ⎛1 0⎞
S2 = �2 ⎜
4 ⎝ 0 1 ⎟⎠
We can derive Sz in a similar manner. We know its eigenstates as well:
Ŝzα = �2 α Ŝz β = − �2 β
So that the matrix must satisfy:
� ⎛c d ⎞⎛1⎞ � ⎛1⎞
Ŝzα = α → ⎜e f ⎟⎜0⎟ = 2 ⎜0⎟
2 ⎝ ⎠⎝ ⎠ ⎝ ⎠
⎛c⎞ ⎛ ⎞
�
⇒⎜ ⎟=⎜2⎟
⎝e⎠ ⎝0⎠
which gives us
�
c= e=0
2
Analogous operations on β give
� ⎛c d ⎞⎛0⎞ ⎛0⎞
Ŝz β = − β → ⎜ ⎟⎜ ⎟ = − 2 ⎜ ⎟
�
2 ⎝e f ⎠⎝1⎠ ⎝1⎠
⎛d⎞ ⎛ 0 ⎞
⇒⎜ ⎟=⎜ �⎟
⎝ f ⎠ ⎝− 2 ⎠
�
d=0 f =−
2
Compiling these results, we arrive at the matrix representation of Sz :
� ⎛1 0 ⎞
Sz = ⎜
2 ⎝ 0 −1 ⎟⎠
Now, we need to obtain S x and S y , which turns out to be a bit more tricky.
We remember from our operator derivation of angular momentum that we
can rewrite the S x and S y in terms of raising and lowering operators:
1 1
Sx = ( S + + S- ) Sy = (S+ − S- )
2 2i
where we know that
Ŝ+ β = c+α Ŝ+ α = 0 and Ŝ−α = c− β Ŝ− β = 0
where c+ and c are constants to be determined. Therefore for the raising
operator we have
⎛ c d ⎞ ⎛ 0 ⎞ ⎛ c+ ⎞ ⎛c d ⎞⎛1⎞ ⎛0⎞
Ŝ+ β = c+α → ⎜ ⎟⎜ ⎟ = ⎜ ⎟ Ŝ+ α = 0 → ⎜e f ⎟⎜0⎟ = ⎜0⎟
⎝ e f ⎠ ⎝ 1 ⎠ ⎝ 0 ⎠ ⎝ ⎠⎝ ⎠ ⎝ ⎠
⇒ d = c+ f = 0 ⇒c=0 e=0
And for the lowering operator
⎛ c d ⎞ ⎛ 1 ⎞ ⎛ 0 ⎞ ⎛c d ⎞⎛0⎞ ⎛0⎞
Ŝ−α = c− β → ⎜ ⎟⎜ ⎟ = ⎜ ⎟ Ŝ− β = 0 → ⎜e f ⎟⎜1⎟ = ⎜0⎟
⎝ e f ⎠ ⎝ 0 ⎠ ⎝ c− ⎠ ⎝ ⎠⎝ ⎠ ⎝ ⎠
⇒ c = 0 e = c − ⇒d=0 f =0
Thus
⎛0 1⎞ ⎛0 0⎞
S + = c+ ⎜ ⎟ S - = c− ⎜ ⎟
⎝0 0⎠ ⎝1 0⎠
3 � 2 c+
c
⇒ �2 = + + +
4 4 4 4
2
⇒ c+ = � 2
⇒ c+ = �
where on the last line, we have made an arbitrary choice of the sign of c+ .
Thus we arrive at the final expressions for S x and S y
� ⎛0 1⎞ � ⎛ 0 −i ⎞
⎜ Sx =
⎟ Sy = ⎜
2 ⎝1 0⎠ 2 ⎝ i 0 ⎟⎠
In summary, then, the matrix representations of our spin operators are:
� ⎛0 1⎞ � ⎛ 0 −i ⎞ � ⎛1 0 ⎞ 3 2 ⎛1 0⎞
Sx = Sy = Sz = S2 = �
2 ⎜⎝ 1 0 ⎟⎠ 2 ⎜⎝ i 0 ⎟⎠ 2 ⎝⎜ 0 −1 ⎟⎠ 4 ⎜⎝ 0 1 ⎟⎠
Given our representations of S x , S y and Sz now we can work out any property
we like by doing some linear algebra. For example, we can work out the
commutator of S x and S y :
� ⎛ 0 1 ⎞ � ⎛ 0 −i ⎞ � ⎛ 0 −i ⎞ � ⎛ 0 1 ⎞
⎡⎣S x ,S y ⎤⎦ = S x S y − S yS x = ⎜ −
2 ⎝ 1 0 ⎟⎠ 2 ⎝⎜ i 0 ⎟⎠ 2 ⎜⎝ i 0 ⎟⎠ 2 ⎝⎜ 1 0 ⎟⎠
� 2 ⎡⎛ 0 1 ⎞ ⎛ 0 −i ⎞ ⎛ 0 −i ⎞ ⎛ 0 1 ⎞ ⎤
= ⎢⎜ −
4 ⎣⎝ 1 0 ⎠ ⎝ i 0 ⎟⎠ ⎜⎝ i 0 ⎟⎠ ⎝⎜ 1 0 ⎟⎥
⎟⎜
⎠⎦
�2 ⎡⎛ i 0 ⎞ ⎛ −i 0 ⎞ ⎤ � 2 ⎛ i 0 ⎞
= ⎢⎜ ⎟−⎜ ⎟⎥ = ⎜ ⎟
4 ⎣⎝ 0 −i ⎠ ⎝ 0 i ⎠ ⎦ 2 ⎝ 0 −i ⎠
⇒ ⎡⎣S x ,S y ⎤⎦ = i�S z
It is comforting to see that the matrices that represent the angular
momentum operators obey the commutation relations for angular momentum!
It is similarly possible to work out the eigenvalues of Ŝ 2 , Ŝz and Ŝx by
computing the eigenvalues of S x , S y and Sz ; average values can be obtained
by taking (row)*x(matrix)x(column) products. We are thus in a position to
compute anything we want for these operators. As it turns out, any operator
in this 2x2 space can be written as a linear combination of S x , S y , Sz and S2 .
So, in some sense, we now have all the information we could possibly want
about spin½ systems.
possible place to practice our skills at using matrix mechanics. We will see
later on that more complicated systems can also be described by matrix
vector operations. However, in these cases the matrices will have many
more elements – sometimes even infinite numbers of elements! Therefore, it
is best to get some practice on these simple 2x2 matrices before taking on
the more complicated cases.
HELIUM ATOM
Now that we have treated the Hydrogen like atoms in some detail, we now
proceed to discuss the nextsimplest system: the Helium atom. In this
situation, we have tow electrons – with coordinates
z r1 and r2 – orbiting a nucleus with charge Z = 2
located at the point R. Now, for the hydrogen atom
we were able to ignore the motion of the nucleus
r2 by transforming to the center of mass. We then
R obtained a Schrödinger equation for a single
y
effective particle – with a reduced mass that was
very close to the electron mass – orbiting the
x origin. It turns out to be fairly difficult to
r1
transform to the center of mass when dealing with
three particles, as is the case for Helium. However, because the nucleus is
much more massive than either of the two
electrons (MNuc ≈ 7000 mel) it is a very good z
approximation to assume that the nucleus sits at
the center of mass of the atom. In this
approximate set of COM coordinates, then, R=0 r2
and the electron coordinates r1 and r2 measure
y
the between each electron and the nucleus.
Further, we feel justified in separating the
motion of the nucleus (which will roughly x
correspond to rigidly translating the COM of the
r1
atom) from the relative d the electrons orbiting
the nucleus within the COM frame. Thus, in what follows, we focus only on the
motion of the electrons and ignore the motion of the nucleus.
We will treat the quantum mechanics of multiple particles (1,2,3…) in much the
same way as we described multiple dimensions. We will invent operators r̂1 , r̂2 ,
r̂3 , … and associated momentum operators p̂1 , p̂ 2 , p̂3 …. The operators for a
given particle (i) will be assumed to commute with all operators associated with
any other particle (j):
[r̂1, p̂ 2 ] = [ p̂2 , r̂3 ] = [r̂2 , r̂3 ] = [ p̂1, p̂3 ] = ... ≡ 0
Meanwhile, operators belonging to the same particle will obey the normal
commutation relations. Position and momentum along a given axis do not
commute:
⎡ x̂1, p̂x ⎤ = i� ⎡ ŷ , p̂ ⎤ = i� ⎡ ẑ , p̂ ⎤ = i�
⎣ 1⎦ ⎣ 1 y1 ⎦ ⎣ 1 z1 ⎦
while all components belonging to different axes commute:
xˆ1 yˆ1 = yˆ1 xˆ1 p̂z1 yˆ1 = yˆ1 pˆ z1 pˆ z1 p̂x1 = pˆ x1 pˆ z1 etc.
As you can already see, one of the biggest challenges to treating multiple electrons
is the explosion in the number of variables required!
In terms of these operators, we can quickly write down the Hamiltonian for the
Helium atom:
Attraction
Repulsion
This Hamiltonian looks very intimidating, mainly because of all the constants (e, me,
ε0, etc.) that appear in the equation. These constants result from our decision to
use SI units (meter, gram, second) to express our lengths, masses and energies.
This approach is quite awkward when the typical mass we’re dealing with is 1028
grams, the typical distance is 1010 meters and the typical energy unit is 1018
Joules. It is therefore much simpler to work everything out in what are called
atomic units. In this system of units we choose our unit of mass to be equal to the
electron mass, me, our unit of charge to be equal to the electron charge, e, and our
unit of angular momentum to be � . Further, we choose to work in electrostatic
units, so that the permittivity of free space (4π ε0) is also 1. The result of these
four choices is twofold. First of all, the Hamiltonian of the Helium atom (and
indeed of any atom or molecule) simplifies greatly because all the constants are
unity and can be omitted in writing the equations:
p̂12 p̂22 Z Z
Ĥ ≡ + − − + 1
2 2 r̂1 r̂2 r̂ − r̂
1 2
∇12 ∇22
=− − −Z −Z + 1
2 2 r̂1 r̂2 r̂ − r̂
1 2
This will greatly simplify our algebra in the future and we will typically use atomic
units from this point forward. The second benefit of this choice of units is that
now our units of energy, mass, distance, angular momentum, etc. are of the
Charge e =1 1.06x10−19 C
energies that are of order unity, distances that are of order unity, momenta
that are of order unity … without having to keep track of so much scientific
notation.
Once Schrodinger had solved the Hydrogen atom, it was generally believed that
the solution of the Helium atom would follow not long afterward. However,
Scientists have tried for decades to solve this three body problem without
succeeding. Very, very accurate approximations were developed, but no exact
solutions were found. As it turns out, even with the simplifications described
above it is impossible to determine the eigenstates of the Helium atom. This
situation is common in chemistry, as most of the problems we are interested in
cannot be solved exactly. We therefore must resort to making approximations, as
we will do throughout this course.
∇12
Z Z ∇22
∇12 Z ∇22 Z ˆ ˆ
Ĥ ind = + − − = − + − ≡ H1 + H 2
2 2 r̂1 r̂2 2 r̂1 2 r̂2
Electron One Electron Two
Thus, if we neglect the interaction between the electrons, the Hamiltonian reduces
to the sum of two hydrogenic atom Hamiltonians (with Z=2). Based on our
experience with separable Hamiltonians in the past (Hamiltonians Add �
Wavefunctions Multiply � Energies Add), we can immediately write down the
correct form for the eigenfunctions and eigenvalues of this Hamiltonian:
Ψ n l m s ;n l m s (r1,σ 1; r2 ,σ 2 ) = ψ n l m s (r1,σ 1 )ψ n l m s (r2 ,σ 2 )
11 1 1 2 2 2 2 11 1 1 22 2 2
2 2
Z Z
En1n2 = En1 + En2 = − 2
−
2n1 2n2 2
where, in the second line, we have made use of atomic units ( � =me=e=4π ε0=1) to
simplify our energy expression. Thus, by neglecting the electron repulsion (an
approximation) we move from a problem that is impossible to solve to one that we
can already solve easily.
However, the outstanding question is: how good is this approximation? The easiest
way to test this is to look at the ground state. According to the Pauli exclusion
principle, we can’t put two electrons in the same state. However, as we learned in
freshman chemistry, we can put one electron in 1s with spin up and the other in 1s
with spin down:
Ψ1sα ;1sβ (r1,σ 1; r2 ,σ 2 ) = ψ100α (r1,σ 1 )ψ100 β (r2 ,σ 2 )
where the leading factor of 1/√2 ensures that the new wavefunction is normalized.
We check that this is antisymmetric:
Ψ1sα ;1sβ ( r1 , σ 1 ; r2 , σ 2 ) ≡
1
2
( )
ψ 1sα ( r1 , σ 1 )ψ 1sβ ( r2 , σ 2 ) − ψ 1sα ( r2 , σ 2 )ψ 1sβ ( r1 , σ 1 )
⎯⎯⎯
1↔2
→
1
2
( )
ψ 1sα ( r2 , σ 2 )ψ 1sβ ( r1 , σ 1 ) − ψ 1sα ( r1 , σ 1 )ψ 1sβ ( r2 , σ 2 )
=−
1
2
( )
ψ 1sα ( r1 , σ 1 )ψ 1sβ ( r2 , σ 2 ) − ψ 1sα ( r2 , σ 2 )ψ 1sβ ( r1 , σ 1 )
= −Ψ1sα ;1sβ ( r1 , σ 1 ; r2 , σ 2 )
Does this new wavefunction give a better energy? As it turns out, this change by
itself does nothing for the energy prediction. The new wavefunction is a linear
combination of two degenerate eigenstates of the independent electron
Hamiltonian. As we learned before, any sum of degenerate eigenstates is also an
eigenstate with the same eigenvalue. So, while quantum mechanics says we have to
make the wavefunction antisymmetric, antisymmetry by itself does not affect our
energy for Helium.
The simplest way for us to improve our estimate of the helium ground state energy
is to consider not the eigenvalues of our approximate Hamiltonian with our
approximate eigenfunctions, but instead look at the average energy of our
approximate function with the exact Hamiltonian. That is to say, a better
approximation to the energy can be got from
Ĥ = ∫ Ψ*1sα ;1sβ Ĥ Ψ 1sα ;1sβ dτ1dτ 2
where dτ1 = dr1 dσ1 and similarly for dτ2. We refer to this picture as an
independent electron approximation. Within the wavefunction the electrons
behave as if they do not interact, because we have retained the separable form.
However, in computing the energy, we fold these interactions back in in an
approximate way by computing the average energy including the interaction.
2
Ψ 1sα ;1sβ
Ĥ = −4 + ∫ dτ1dτ 2
r1 − r2
The first term is simply the noninteracting electron energy from above. The
second term is the average value of the electronelectron repulsion. Because the
integrand is strictly positive (as one would expect for electron repulsion) this new
term will only raise the average energy, which is a good thing, since our non
interacting picture gave an energy that was 30 eV too low! We can further expand
the repulsion term based on the antisymmetric form of the wavefunction. First,
we note that we can factorize the antisymmetric wavefunction into a space part
times a spin part:
1
(ψ 1s ( r1 ) α (σ1 )ψ 1s ( r2 ) β (σ 2 ) − ψ 1s ( r2 ) α (σ 2 )ψ 1s ( r1 ) β (σ1 ) )
⇒ ψ 1s ( r1 )ψ 1s ( r2
)
1
2
(α (σ1 ) β (σ 2 ) − β (σ1 ) α (σ 2 ) )
With these definitions, it is easy to verify that the space and spin wavefunctions
are individually normalized. Note, in the absence of a magnetic field, you will
always be able to write the eigenfunctions of the Hamiltonian in this form because
H is separable into a space and spin part
Ĥ = Ĥspace + Ĥ spin
With the spin part being (trivially) zero. As a result, eigenfunctions of H will always
be products of a space part and a spin part as above. With this space/spin
separation in hand, we can simplify the repulsion term:
2
r1 − r2
r1 − r2 r1 − r2
Ψ space ( r1 , r2 )
2
(σ 1 , σ 2 )
2
= ∫
Ψ spin dσ 1dσ 2 ∫ r1 − r2
dr1dr2
1
Ψ space ( r1 , r2 )
2
= ∫ r1 − r2
dr1dr2
Ψ space 5Z 5
∫ r1 − r2
dr1dr2 =
8
=
a.u. = +34 eV
4
Adding this average repulsion term to our noninteracting energy gives a ground
state estimate of 108.8 eV +34 eV= 74.8 eV, which is only 4 eV off of the correct
energy. The energy is still not very close, but at least we are making progress.
As we have seen already, the independent electron picture is not all that accurate
for describing atoms. However, chemists are very pragmatic and realize that the
ease of solving noninteracting problems is extremely valuable and as we will soon
see, it gives us a picture (molecular orbital theory) that allows us to describe a
wide range of chemistry. Therefore, chemists are extremely reluctant to abandon
an independent particle picture. Instead, a great deal of work has gone into making
more accurate models based on independent particles – either by making more
sophisticated corrections like the one above or by coming up with a different non
interacting Hamiltonian that gives us a better independent particle model. We will
spend the next several lectures discussing the qualitative features of electronic
structure and bonding that come out of this picture.
In the last lecture, we learned that the independent particle model gives a reasonable
description of the ground state energy of the Helium atom. Before moving on to talk
about manyelectron atoms, it is important to point out that we can describe many
more properties of the system using the same type of approximation. By using the
same independent particle prescription we can come up with wavefunctions for
excited states and determine their energies, their dependence on electron spin, etc.
by examining the wavefunctions themselves. That is to say, there is much we can
determine from simply looking at Ψ without doing any significant computation.
We will use the excited state 1s2s configuration of Helium as an example. For the
ground state we had:
Ψ space (r1,r2 ) × Ψspin (σ1,σ 2 )
1s
2s
? ?
1s
2s
? ?
1s
2s
? ?
1s
Where the question marks indicate that we need to determine the space and spin
wavefunctions that correspond to these stick diagrams. It is fairly easy to come up
with a reasonable guess for each configuration. For example, in the first case we
might write down a wavefunction like:
ψ1sα (r1σ1 )ψ 2sα (r2σ 2 ) =ψ1s (r1 )ψ 2s (r2 ) α (σ1 ) α (σ 2 )
However, we immediately note that this wavefunction is not antisymmetric. We can
perform the same trick as before to make an antisymmetric wavefunction out this:
⇒ 1 ⎛ψ
(r1 )ψ 2s (r2 )α (σ1 )α (σ 2 ) −ψ1s (r2 )ψ 2s (r1 )α (σ 2 )α (σ1 ) ⎟⎞⎠
2 ⎜⎝ 1s
Ψ space Ψ spin
Applying the same principle to the 1s↑2s↓ configuration gives us a bit of trouble:
⇒ 1 ⎛ψ
2 ⎜⎝ 1s (r1 )ψ 2s (r2 )α (σ1 ) β (σ 2 ) −ψ1s (r2 )ψ 2s (r1 )α (σ 2 ) β (σ1 ) ⎟⎠⎞ ≠ ΨspaceΨspin
Hence, the pure ↑↓ configuration can’t be separated in terms of a space part and a
spin part. We find a similar result for 1s↓2s↑:
⇒ 1 ⎛ψ
2 ⎜⎝ 1s (r1 )ψ 2s (r2 ) β (σ1 )α (σ 2 ) −ψ1s (r2 )ψ 2s (r1 ) β (σ 2 )α (σ1 ) ⎟⎞⎠ ≠ ΨspaceΨspin
Since we know the wavefunction should separate, we have a problem. The solution
comes in realizing that for an open shell configuration like this one, the 1s↑2s↓ and
1s↓2s↑ states are degenerate eigenstates and so we can make any linear combinations
of them we like and we’ll still obtain an eigenstate. If we make the “+” and ““
combinations of 1s↑2s↓ and 1s↓2s↑ we obtain:
Ψ space Ψ spin
which separates nicely. Performing similar manipulations for the ↓↓ configuration and
taking care to make sure that all our spatial and spin wavefunctions are individually
normalized allows us to complete the table we set out for the 1s2s excited states:
2s
1s
1
(
ψ ( r )ψ ( r ) + ψ1s ( r2 )ψ 2s ( r1 ) ×
2 1s 1 2s 2
) 1
2
(
α (σ1 ) β (σ 2 ) − β (σ1 ) α (σ 2 ) )
and
2s 1
(
ψ ( r )ψ ( r ) −ψ1s ( r2 )ψ 2s ( r1 ) ×
2 1s 1 2s 2
) 1
2
(
α (σ1 ) β (σ 2 ) + β (σ1 ) α (σ 2 ) )
1s
2s
⇒ 1
2 (ψ1s (r1 )ψ 2s (r2 ) −ψ1s (r2 )ψ 2s (r1 )) × β (σ1 ) β (σ 2 )
1s
We notice several things about these wavefunctions:
• While the overall wavefunction is always antisymmetric by construction, the
spatial part can be either antisymmetric (cases 1, 3 and 4) or symmetric (case
2). This effect is compensated for in the spin part, which can also be
antisymmetric (case 2) or symmetric (cases 1,3 and 4). The resulting
wavefunction always has a symmetric part times an antisymmetric part,
resulting in an antisymmetric wavefunction.
• The spin part of Case 2 is exactly the same as the spin part of the ground state
of the helium atom. Thus, just as we thought of the electrons in the ground
state as being “paired”, we say the electrons in Case 2 are paired.
• The spatial parts of three of the states above (cases 1,3 and 4) are the same.
Case 2 has a different spatial part. Because the Hamiltonian only depends on
spatial variables and not spin, we immediately conclude that 1,3 and 4 will be
degenerate states – even when we take into account the electronelectron
interaction. State 2, however, will generally have a different energy once we
account for interactions. In common spectroscopic parlance the three
degenerate states are called a triplet and the unique state is called a singlet.
Further, because these states arise from degenerate spin states, they are
called singlet and triplet spin states.
So we expect the triplet to be lower. How much lower? To answer this question, we
have to compute the average energies of the singlet and triplet wavefunctions. Recall
that the spin part never matters for the energy:
∫ Ψ Ĥ Ψdrdσ = ∫ Ψ ∫ ∫
Ψ spin* Ĥ Ψ space Ψ spin drdσ = Ψ spin* Ψ spin dσ Ψ space* Ĥ Ψ space dr
* *
space
= ∫Ψ
1
space Ĥ Ψ space dr
*
The influence of the spin wavefunction is only indirect: if the spin part is
antisymmetric (e.g. singlet) then the spatial part must be symmetric and vice versa.
To simplify our algebra, it is convenient to create the obvious shorthand notation:
( )
ΨT = 1 ψ 1s ( r1 )ψ 2s ( r2 ) −ψ 1s ( r2 )ψ 2s ( r1 ) ≡ 1 (1s2s − 2s1s )
2 2
ΨS = 1 (ψ 1s ( r1 )ψ 2s ( r2 ) +ψ 1s ( r2 )ψ 2s ( r1 ) ) ≡ 1
(1s2s + 2s1s )
2 2
where we just need to remember that the first function in a product will be the one
that has electron “1” while the second will have electron “2”. Proceeding then:
1 ⎛ ⎞
ES /T = ∫ Ψ S /T * Ĥ Ψ S /T dr1dr2 = ( ) ˆ + Hˆ + 1 ⎟ (1s 2s ± 2s1s ) dr1dr2
2 ∫ 1s 2s ± 2 s1s ⎜H
⎜
1 2
r1 − r2 ⎟
⎝ ⎠
1 ⎛ ⎞
( ) ⎜ −2 − 1 + 1 ⎟ (1s2s ± 2s1s ) dr1dr2
2∫
= 1s 2 s ± 2 s1s
⎜ 2 r1 − r2 ⎟
⎝ ⎠
5 1 1
= − + ∫ (1s 2s ± 2s1s ) (1s2s ± 2s1s ) dr1dr2
2 2 r1 − r2
On the second line, we have used the fact that both ΨS and ΨT are eigenstates of the
independent particle Hamiltonian (by construction) and on the third line, we have
taken the independent particle energy outside the integral because ΨS and ΨT are
normalized. Thus, we see that the average energy takes on the familiar form of
(noninteracting energy)+(interactions). The interaction term can be simplified
further:
1
2∫
(1s2s ± 2s1s ) r −1 r (1s2s ± 2s1s ) dr1dr2
1 2
1 1 1 1 1
2∫
⇒ 1s2s 1s 2s −1s2s 2s1s − 2s1s 1s2s + 2s1s 2s1s dr1dr2
r1 − r2 r1 − r2 r1 − r2 r1 − r2
We note that the first and last terms are the same if we just interchange the dummy
integration variables:
1 1
∫ 1s (r1 ) 2s (r2 ) r1 − r2 1s (r1 ) 2s ( r2 ) dr1dr2 ⎯⎯⎯→∫ 1s ( r2 ) 2s (r1 ) r2 − r1 1s (r2 ) 2s (r1 ) dr2 dr1
1↔2
1
= ∫ 2s ( r1 )1s ( r2 ) 2s ( r1 )1s ( r2 ) dr1dr2 ≡ J12
r1 − r2
Meanwhile the second and third terms are also the same:
1 1
∫ 1s (r1 ) 2s (r2 ) r1 − r2 2s (r1 ) 1s (r2 ) dr1dr2 ⎯⎯⎯→∫ 1s ( r2 ) 2s (r1 ) r2 − r1 2s (r2 ) 1s ( r1 ) dr2dr1
1↔2
1
= ∫ 2s ( r1 ) 1s ( r2 ) 1s ( r2 ) 2s ( r1 ) dr1dr2 ≡ K12
r1 − r2
These integrals are called Coulomb(J) and exchange(K) integrals, respectively. Both
are positive numbers (because they arise from electron repulsion) and it can be
rigorously proved that J>K always (i.e. no matter what functional form the 1s and 2s
wavefunctions have). Thus, in terms of J and K the energies of the singlet and triplet
states become:
5
ES /T = − + J12 ± K12
2
Thus we see that, as expected, the singlet state is higher in energy than the
triplet. In fact, we can even give a numerical estimate for the splitting by evaluating
K12. Plugging in the forms of the 1s and 2s orbitals of helium and doing the integrals,
we obtain K12=32/729=1.2 eV and a splitting of 2K12=2.4 eV. The latter is quite a bit
larger than the experimental singlettriplet splitting in helium, which comes out to
only .8 eV. Once again, we see the independent particle model gives us a qualitatively
correct picture (i.e. the sign of the splitting is correct and of the right order of
magnitude) but we fail to obtain quantitative results. We therefore arrive at the
following qualitative picture of the 1s2s excited state of Helium:
It is interesting to note that the exchange interaction results from the fact that the
electrons are indistinguishable. Notice that, if we had not antisymmetrized our
wavefunctions, the spatial part would have just been a direct product 1s2s instead of
the symmetric/antisymmetric 1s2s±2s1s combinations we obtained for the singlet and
triplet. In the former case, the electrons are being treated as distinguishable (e.g.
electron “1” is always 1s while electron “2” is always 2s) and the exchange term
disappears from the interaction:
1 1 1
2 ∫ (1s 2s ± 2s1s )
r1 − r2
(1s 2s ± 2s1s ) dr1dr2 ⎯⎯⎯⎯⎯
distinguishable
⎯→ ∫ 1s2s
r1 − r2
1s 2sdr1dr2 = J12
Clearly exchange – which arises from the cross terms on the left – is absent on the
right. Thus, the K integrals only arise when we have terms in the wavefunction where
two electrons have exchanged places. Hence the name “exchange.” It is important to
note that, next to the Pauli exclusion principle, this is the biggest impact that
antisymmetry has on chemistry.
There was a lot of interest in class about how we reconcile the fact that, in other
chemistry courses you’ve been taught that there are only two spin states for a pair of
electrons: ↑↓ and ↑↑. The former represented the singlet state and the latter the
triplet state. You referred to the singlet electrons as being “paired” and the triplets
as being “unpaired.” However, how do these strange spin states we’ve derived connect
with the “paired” and “unpaired” ideas? To answer this question, we first of all we
should note that neither of the antiparallel states we’ve derived is strictly ↑↓.
Instead, they look like ↑↓±↓↑, with ↑↓-↓↑ being the singlet and ↑↓+↓↑ being part of the
triplet. The idea that the singlet state is ↑↓ is a white lie that we tell in order to
simplify our arguments: as long as the subtle difference between ↑↓ and ↑↓-↓↑ isn’t
important, we can get away with explaining much (though not all) chemistry by treating
the singlet state as ↑↓.
In the more precise picture we’ve derived here, it the spin part of the wavefunction
determines whether the electrons are paired or not. An electron pair has the
characteristic spin part αβ−βα. That is to say, paired electrons form a singlet. Spin
parts that look like αα, αβ+βα, or ββ are unpaired triplet configurations. As we have
seen above, pairing two electrons raises the energy through the exchange integral.
In some situations, this is called the “pairing energy.” The counterintuitive thing that
we have to relearn is that αβ+βα does not describe an electron pair. In every way it
behaves like αα or ββ : the energies are the same and (as you will show on the
2
homework) the eigenvalues of Ŝtotal are the same. This idea really does not fit into the
simple qualitative picture of triplet states being ↑↑, but it is nonetheless true.
The fact that there are three elements of the triplet state is not a coincidence. As
you will show, the eigenvalues of Sˆtotal
2
for the triplet states are all 2� 2 , which is
consistent with a total spin of S=1, because the eigenvalues of Sˆ 2 would then be total
Thus far, we have learned that the independent particle model (IPM) gives a
qualitatively correct picture of the eigenstates of the helium atom. What about
atoms with more than two electrons, such as lithium or carbon? As it turns out, the
IPM is capable of giving a realistic picture of atomic structure in essentially an
analogous fashion to the helium case. To begin with, we set up our coordinates so that
the nucleus is at the origin and the N electrons are at positions r1, r2, r3, …rN. In terms
of these variables, we can quickly write down the many-electron Hamiltonian (in atomic
units):
1 N 2 Z N N N
1
Hˆ = − ∑∇ j −∑ + ∑∑
2 j=1 j=1 r̂ j i=1 j >i r̂i − r̂ j
1 N N Z N ⎛ Z ⎟
⎞ N
Ĥ =− 2
∑ ∇ − ∑ ⎜
= ∑ − ∇ − 1 2 ≡ ∑ ĥ
j = 1 ⎝⎜
NI 2 j = 1 j j = 1 r̂ 2 j r̂ ⎟ i
j j ⎠ j =1
where each of the independent Hamiltonians ĥi describes a single electron in the field
of a nucleus of charge +Z. Based on our experience with separable Hamiltonians, we
can immediately write down the eigenstates of this Hamiltonian as products with
energies given as sums of the independent electron energies:
Ψ = ψ k (1)ψ k ( 2 )ψ k ( 3) ...ψ k (N ) E = Ek + Ek + Ek +... + Ek
1 2 3 N 1 2 3 N
Where (1) is a shorthand for (r1,σ1) and ki ≡ {ni ,li , mi , si } specifies all the quantum
numbers for a given hydrogen atom eigenstate. Of course, there is a problem with
these eigenstates: they are not antisymmetric. For the Helium atom, we fixed this by
making an explicitly antisymmetric combination of two degenerate product states:
ψ1sα (1) ψ1sα ( 2 )
1
(ψ (1)ψ1sβ ( 2) −ψ1sα ( 2)ψ1sβ (1)) =
2 1sα
1
2ψ
1sβ (1) ψ1sβ ( 2)
where on the right we have noted that this antisymmetric product can also be written
as a determinant of a 2x2 matrix. As it turns out, it is straightforward to extend this
idea to generate an N particle antisymetric state by computing an NxN determinant
called a Slater Determinant:
ψ k (1) ψ k (1) ψ k (1) � ψ k (1)
1 2 3 N
ψ k ( 2) ψ k ( 2) ψ k ( 2) � ψ k ( 2)
1 2 3 N
Ψ (1,2,..., N ) =
1
N ! k1 ( )
ψ 3 ψ k ( 3) ψ k ( 3) � ψ k ( 3)
2 3 N
� � � � �
ψk (N ) ψk (N ) ψk (N ) � ψk (N )
1 2 3 N
As you can imagine, the algebra required to compute integrals involving Slater
determinants is extremely difficult. It is therefore most important that you realize
several things about these states so that you can avoid unnecessary algebra:
• A Slater determinant corresponds to a single stick diagram. This is easy to
see by example:
2px 1sα (1) 1s β (1) 2sα (1) 2 pxα (1)
1sα ( 2 ) 1s β ( 2 ) 2sα ( 2 ) 2 pxα ( 2 )
2s ⇒ Ψ (1, 2, 3, 4 ) =
1sα ( 3 ) 1s β ( 3 ) 2sα ( 3 ) 2 pxα ( 3 )
1s 1sα ( 4 ) 1s β ( 4 ) 2sα ( 4 ) 2 pxα ( 4 )
It should be clear that we can extend this idea to associate a determinant with
an arbitrary stick diagram. Further, recall that for the excited states of
helium we had a problem writing certain stick diagrams as a (space)x(spin)
product and had to make linear combinations of certain states to force things
to separate. Because of the direct correspondence of stick diagrams and
Slater determinants, the same pitfall arises here: Slater determinants
sometimes may not be representable as a space)x(spin) product, in which
case a linear combination of Slater determinants must be used instead.
This generally only happens for systems with unpaired electrons, like the 1s↑2s↓
configuration of helium or the …2px↑2py↓ configuration of carbon.
• A Slater determinant is anitsymmetric upon exchange of any two electrons.
We recall that if we take a matrix and interchange two its rows, the
determinant changes sign. Thus, interchanging 1 and 2 above, for example:
ψ k ( 2 ) ψ k ( 2 ) ψ k ( 2 ) � ψ k ( 2 )
1 2 3 N
Ψ ( 2,1,..., N ) =
1
N ! k1 ( )
ψ 3 ψ k ( 3) ψ k ( 3) � ψ k ( 3)
2 3 N
� � � � �
ψ k ( N ) ψ k ( N ) ψ k ( N ) � ψ k ( N )
1 2 3 N
ψ k ( 2 ) ψ k ( 2 ) ψ k ( 2 ) � ψ k ( 2 )
Interchange
1 2 3 N
ψ 3 ψ k ( 3) ψ k ( 3) � ψ k ( 3) = −Ψ (1,2,..., N )
1
N ! k1 ( )
Rows 1&2 → −
⎯⎯⎯⎯⎯⎯
2 3 N
� � � � �
ψ k ( N ) ψ k ( N ) ψ k ( N ) � ψ k ( N )
1 2 3 N
A similar argument applies to any pair of indices, and so the Slater determinant
antisymmetric under any i↔j interchange.
• The determinant is zero if the same orbital appears twice. We recall that
if we take a matrix and interchange two of its columns, the determinant also
changes sign. Assuming k1=k3 above:
ψ k (1) ψ k (1) ψ k (1) � ψ k (1)
1 2 1 N
ψ k ( 2 ) ψ k ( 2 ) ψ k ( 2 ) � ψ k ( 2 )
1 2 1 N
Ψ (1,2,..., N ) =
1
N ! k1 ( )
ψ 3 ψ k ( 3) ψ k (3) � ψ k (3)
2 1 N
� � � � �
ψ k ( N ) ψ k ( N ) ψ k ( N ) � ψ k ( N )
1 2 1 N
Interchange
ψ k ( 2) ψ k ( 2) ψ k ( 2) � ψ k ( 2)
1 2 1 N
ψ 3 ψ k ( 3) ψ k ( 3) � ψ k ( 3) = −Ψ (1,2,..., N )
1
N ! k 1 ( )
Columns
⎯⎯⎯⎯⎯⎯⎯ 1&3→ −
2 1 N
� � � � �
ψk (N ) ψk (N ) ψk (N ) � ψk (N )
1 2 1 N
The only way a number can be equal to its opposite is if it is zero. Thus, in
enforcing antisymmetry the determinant also enforces the Pauli exclusion
principle. Thus, Slater determinants immediately lead us to the aufbau
We have only completed half of the independent particle picture at this point. We
have the noninteracting energy; what remains is the computation of the average
energy. To compute this, we would need to do a 2N-dimensional integral involving a
Slater Determinant on the left, the Hamiltonian in the middle (including all interaction
terms) and a Slater Determinant on the right. The book-keeping is quite tedious, but
can be worked out in the most general case. The result is that the energy breaks
down into terms that we already recognize:
Noninteracting Average
Energy Repulsion
N N
Ĥ = ∑ Ei + ∑ J�ij − K� ij
i=1 i< j
This energy expression has a nice intuitive feel to it: we get contributions from the
interaction of each electron orbital with the nucleus (first term) as well as from the
mutual repulsion of each pair of orbitals (second term). The repulsion has the
characteristic “Coulomb minus Exchange” form because of the antisymmetry of the
Slater determinant, which leads to a minus sign every time we exchange 1 and 2. For
example, when we expand the determinant in terms of product functions, we will have
a term like:
X ≡ ψ k1 (1)ψ k2 ( 2 )ψ k3 ( ? ) �ψ k N ( ? )
we will have a corresponding term with 1↔2 with a minus sign:
Y ≡ ψ k1 ( 2 )ψ k2 (1)ψ k3 ( ? ) �ψ k N ( ? )
When we compute the average repulsion we will have:
1 1
≈ ( X − Y ... ) ( X − Y ... ) dr1 dr2 ...
∫∫
r12 r12
The +XX and +YY terms will give us Coulomb integrals and the cross terms -XY and
-YX will give us exchange integrals. We note that, as defined, the Coulomb and
Exchange terms are both positive, so that the exchange integral always reduces the
repulsion energy for a Slater determinant. These arguments do not constitute a proof
of the energy expression above – they are merely intended to give you a general
feeling that this expression is plausible. Deriving the average energy actually takes
quite a bit of time and careful bookkeeping of different permutations (e.g. 1↔2↔4
versus 4↔2↔3 …). If you want to delve deeper into these kinds of things, we
recommend you take 5.73, which covers much of the material here in greater depth.
In any case, most of the time we need not worry about the details of how the average
energy expression is derived for a general determinant. Usually what we want to do is
use the formula to compute something interesting rather than re-derive it. Toward
this end we note that the Coulomb and exchange integrals above involve integration
over spin variables, which we can be done trivially. Basically, the integration over σ1
(σ2) will give unity if the left and right functions for electron 1 (electron 2) have the
same spin, and zero otherwise. For the Coulomb integral, the left and right functions
are the same, so the spin integration always gives 1 and we can get rid of the spin
integration:
1
J = ∫∫ ψ * (1)ψ * ( 2 ) ψ (1)ψ ( 2 ) dr dr
ij i j r −r i j 1 2
1 2
where here we use the shorthand i ≡ {ni , mi , li } to represent all the spatial quantum
numbers for the ith orbital, and similarly for the jth one. For the exchange integrals,
the left and right functions are different, and so the exchange integral is only
nonzero if the ith orbital and the jth orbital have the same spin part:
⎧ * 1
⎪ ∫∫ ψ i (1)ψ j ( 2 ) ψ ( 2 )ψ (1) dr dr if s = s
*
− i j 1 2 i j
K =⎨ r
1 2
r
ij
⎪
⎩ 0 Otherwise
In practice (for example on a Problem Set) it is useful to begin with the general
expression for the independent particle energy and then integrate out the spin part
to get a final expression that only involves the spatial integrals J and K rather than
the space-spin integrals J� and K� . For example, say we are interested in the
1s↑1s↓2s↑ configuration of lithium. The energy is given by:
N N
Hˆ = ∑ E + ∑ J� − K�
i ij ij
i =1 i< j J1s1s 0 J1s2s K1s2s J1s2s 0
= E + E + E + J� − K� + J� − K� + J� − K�
1s 1s 2s 1sα ;1s β 1sα ;1s β 1sα ; 2 sα 1sα ; 2 sα 1s β ; 2 sα 1s β ; 2 sα
= 2E + E + J + 2J −K
1s 2s 1s1s 1s 2 s 1s 2 s
Thus, in going from space-spin integrations to space integrations, we acquire a few
factors of 2, but the formula still looks similar to the expressions we’ve seen before.
ΔE = E1s2 2 p − E1s2 2s = 2E + E +J + 2J −K −
1s 2 px 1s1s 1s2 px 1s2 px
Veff(r)
R Small R Large
How are we going to solve for these eigenstates? It should be clear that
looking for exact solutions is going to lead to problems in general. Even for
H2+ the solutions are extremely complicated and for anything more complex
than H2+exact solutions are impossible. So we have to resort to
approximations again. The first thing we note is that if we look closely at
our intuitive picture of the H2+ eigenstates above, we recognize that these
molecular eigenstates look very much like the sum of the 1s atomic orbitals
for the two hydrogen atoms. That is, we note that to a good approximation
we should be able to write:
ψ el ( r ) ≈ c11s A ( r ) + c2 1sB ( r )
where c1 and c2 are constants. In the common jargon, the function on the
left is called a molecular orbital (MO), whereas the functions on the right
are called atomic orbitals (AOs). If we write our MOs as sums of AOs, we
are using what is called the linear combination of atomic orbitals (LCAO)
approach. The challenge, in general, is to determine the “best” choices for c1
and c2 – for H2+ it looks like the best choice for the ground state will be
c1=c2. But how can we be sure this is really the best we can do? And what
about if we want something other than the ground state? Or if we want to
describe a more complicated molecule like HeH+2?
Note that in practice you will not know these eigenstates. The important
point is that no matter what function you choose you can always expand it in
terms of the infinite set of orthonormal eigenstates of Ĥ .
∫ψ *ψ dτ = ∑ an am ∫ φn *φm dτ = ∑ an amδ mn = ∑ an
2
* *
=1
n ,m n ,m n
= ∑ an*am Emδ mn = ∑ an En
2
n ,m n
Now, subtracting the ground state energy from the average
E0 = 1• E0 = ∑ an E0
2
⇒ Eavg − E0 = ∑ an En − ∑ an E0 = ∑ an ( En − E0 ) ≥ 0 since En ≥ E0
2 2 2
n n n
Where, in the last line we have noted that the sum of terms that are non
negative must also be nonnegative. It is important to note that the
equals sign is only obtained if an=0 for all states that have En>E0. In this
situation, ψ actually is the ground state of the system (or at least one
ground state, if the ground state is degenerate).
The variational method does two things for us. First, it gives us a means of
comparing two different wavefunctions and telling which one is closer to the
ground state – the one that has a lower average energy is the better
approximation. Second, if we include parameters in our wavefunction
variation gives us a means of optimizing the parameters in the following way.
Assume that ψ depends on some parameters c – such as is the case for our
LCAO wavefunction above. We’ll put the parameters in brackets ψ[c] – in
order to differentiate them from things like positions that are inside of
parenthesis ψ(r).Then the average energy will depend on these parameters:
ψ [ c ] Ĥψ [c ] dτ
(c) = ∫
*
Eavg
∫ψ [c] *ψ [c] dτ
Note that, using the variational principle, the best parameters will be the
ones that minimize the energy. Thus, the best parameters will satisfy
∂Eave ( c ) ∂ ∫ ψ [ c ] Hψ [c ] dτ
*
ˆ
= =0
∂ci ∂ci ∫ ψ [c ]*ψ [c ] dτ
Thus, we can solve for the optimal parameters without knowing anything
about the exact eigenstates!
Let us apply this in the particular case of our LCAOMO treatment of H2+.
Our trial wavefunction is:
ψ el [ c ] = c11s A + c2 1sB
where c=(c1 c2). We want to determine the best values of c1 and c2 and
the variational theorem tells us we need to minimize the average energy
to find the right values. First, we compute the average energy. The
numerator gives:
∫ψ el Ĥ elψ el dτ = ∫ ( c11sA + c21sB ) Ĥ ( c11sA + c21sB ) dτ
* *
Now, we want to minimize this average energy with respect to c1 and c2.
Taking the derivative with respect to c1 and setting it equal to zero [Note:
when dealing with complex numbers and taking derivatives one must treat
variables and their complex conjugates as independent variables. Thus d/dc1
has no effect on c1*]:
∂E avg c1* H11 + c2* H 21
=0= *
∂c1 c1 S11c1 + c1* S12 c2 + c2* S21c1 + c2* S22 c2
c1* H11c1 + c1* H12 c2 + c2* H 21c1 + c2* H 22 c 2
− (c*
S + c2* S21 )
(c S c + c S c + c2 S21c1 + c2 S22 c2 )
2 1 11
* * * *
1 11 1 1 12 2
=0= *
∂c2 c1 S11c1 + c1* S12 c2 + c2* S21c1 + c2*
S22 c2
c1* H11c1 + c1* H12 c2 + c2* H 21c1 + c2* H 22 c2
− (c*
S + c2* S22 )
(c S c + c S c + c2 S21c1 + c2 S22 c2 )
2 1 12
* * * *
1 11 1 1 12 2
∂Eavg
= 0 ⇔ H ic = Eavg Sic Eq. 2
∂c *
Taking the derivatives with respect to c1* and c2* is mathematically
equivalent to taking the derivatives with respect to c1 and c2 [because we
can’t change c1 without changing its complex conjugate, and vice versa].
Thus, the two matrix equations (Eqs. 1&2) above are precisely equivalent, and
the second version is a little more familiar. We can make it even more
familiar if we think about the limit where 1sA and 1sB are orthogonal (e.g.
when the atoms are very, very far apart). Then we would have for the
Overlap matrix:
⎛ 1s 1s dτ 1s A 1sB dτ ⎞ ⎛ 1 0 ⎞
S≡ ⎜ ∫
A A
⎟=∫ =1
⎜⎜ ⎟⎟ ⎝⎜ 0 1 ⎠⎟
⎝ ∫
1sB 1s A dτ 1sB 1sB dτ
⎠ ∫
Thus, in an orthonormal basis the overlap matrix is just the identity matrix
and we can write Eq. 2 as:
∂Eavg
= 0 ⇔ H ic = Eavg c
∂c *
Now this equation is in a form where we certainly recognize it: this is an
eigenvalue equation. Because of its close relationship with the standard
eigenvalue equation, Eq. 2 is usually called a Generalized Eigenvalue
Equation.
In any case, we see the quite powerful result that the Variational theorem
allows us to turn operator algebra into matrix algebra. Looking for the
lowest energy LCAO state is equivalent to looking for the lowest eigenvalue
of the Hamiltonian matrix H. Further, looking for the best c1 and c2 is
equivalent to finding the lowest eigenvector of H.
Let’s go ahead and apply what we’ve learned to obtain the MO coefficients c1
and c2 for H2+. At this point we make use of several simplifications. The
offdiagonal matrix elements of H are identical because the Hamiltonian is
Hermitian and the orbitals are real:
(∫ 1s ) = ∫ 1s Hˆ 1s dτ ≡ V
*
∫ 1s A Hˆ el 1sB dτ = Hˆ el 1s A dτ
*
B B el A 12
Meanwhile, the diagonal elements are also equal, but for a different reason.
The diagonal elements are the average energies of the states 1sA and 1sB. If
these energies were different, it would imply that having the electron on one
side of H2+ was more energetically favorable than having it on the other,
which would be very puzzling. So we conclude
∫ 1s A Ĥ el 1s A dτ ∫
= 1sB Ĥ el 1sB dτ ≡ ε
Finally, we remember that 1sA and 1sB are normalized, so that
∫ 1s 1s
A A dτ ∫
= 1sB 1sB dτ = 1
and because the 1s orbitals are real, the offdiagonal elements of S are also
the same:
∫ ∫
S12 = 1s A 1sB dτ = 1sB 1s A dτ = S21 .
Incorporating all these simplifications gives us the generalized eigenvalue
equation:
⎛ ε V12 ⎞ ⎛ c1 ⎞ ⎛ 1 S12 ⎞ ⎛ c1 ⎞
⎜V ⎟ ⎜ ⎟ = Eavg ⎜ ⎟⎜ ⎟
⎝ 12 ε ⎠ ⎝ c2 ⎠ ⎝ S21 1 ⎠ ⎝ c2 ⎠
All your favorite mathematical programs (Matlab, Mathematica, Maple,
MathCad…) are capable of solving for the generalized eigenvalues and
eigenvectors, and for more complicated cases we suggest you use them.
However, this case is simple enough that we can solve it by guessand test.
Based on our physical intuition above, we guess that the correct eigenvector
will have c1=c2. Plugging this in, we find:
⎛ ε V12 ⎞ ⎛ c1 ⎞ ⎛ 1 S12 ⎞ ⎛ c1 ⎞
⎜ ⎟ ⎜ ⎟ = Eavg ⎜ ⎟⎜ ⎟
⎝ V12 ε ⎠ ⎝ c1 ⎠ ⎝ S21 1 ⎠ ⎝ c1 ⎠
⎛ (ε + V12 ) c1 ⎞ ⎛ (1 + S12 ) c1 ⎞
⇒⎜ ⎟ = Eavg ⎜ ⎟
⎝ (ε + V12 ) c1 ⎠ ⎝ (1 + S12 ) c1 ⎠
ε + V12
⇒ Eavg = ≡ Eσ
1 + S12
Thus, our guess is correct and one of the eigenvectors of this matrix has
c1=c2. This eigenvector is the σbonding state of H2+, and we can write down
the associated orbital as:
ψ elσ = c11s A + c2 1sB = c11s A + c11sB ∝ 1s A + 1sB
where in the last expression we have noted that c1 is just a normalization
constant. In freshman chemistry, we taught you that the σbonding orbital
existed, and this is where it comes from.
We can also get the σ∗antibonding orbital from the variational procedure.
Since the matrix is a 2x2 it has two unique eigenvalues: the lowest one
(which we just found above) is bonding and the other is antibonding. We can
again guess the form of the antibonding eigenvector, since we know it has
the characteristic shape +/, so that we guess the solution is c1=c2:
⎛ ε V12 ⎞ ⎛ c1 ⎞ ⎛ 1 S12 ⎞ ⎛ c1 ⎞
⎜ ⎟⎜ ⎟ = Eavg ⎜ ⎟⎜ ⎟
⎝ V12 ε ⎠ ⎝ − c1 ⎠ ⎝ S21 1 ⎠ ⎝ − c1 ⎠
⎛ (ε − V12 ) c1 ⎞ ⎛ (1 − S12 ) c1 ⎞
⇒⎜ ⎟ = Eavg ⎜ ⎟
⎝ − (ε − V12 ) c1 ⎠ ⎝ − (1 − S12 ) c1 ⎠
ε − V12
⇒ Eavg = = Eσ *
1 − S
12
so, indeed the other eigenvector
has c1=c2. The corresponding
antibonding orbital is given by: ψσ(r)
ψ elσ * = c11s A + c2 1sB = c11s A − c11sB ∝ 1s A − 1sB
H2+:
ε − V12
Eσ∗=
1 − S12
E1sA=ε E1sB=ε
ε + V12
Eσ=
1 + S12
On the left and right, we draw the energies of the atomic orbitals (1sA and
1sB) that make up our molecular orbitals (σ and σ*) in the center. We note
that when the atoms come together the energy of the bonding and
antibonding orbitals are shifted by different amounts:
ε − V12 ε − V12 ε (1 − S12 ) ε S12 − V12
Eσ * − E1s = −ε = − =
1 − S12 1 − S12 1 − S12 1 − S12
ε + V12 ε (1 + S12 ) ε + V12 ε S12 − V12
E1s − Eσ * = ε − = − =
1 + S12 1 + S12 1 + S12 1 + S12
Now, S12 is the overlap between two 1s orbitals. Since these orbitals are
never negative, S12 must be a positive number. Thus, the first denominator
is greater than the second, from which we conclude
ε S − V12 ε S12 − V12
Eσ * − E1s = 12 > = E1s − Eσ *
1 − S12 1 + S12
Thus, the antibonding orbital is destabilized more than the bonding orbital is
stabilized. This conclusion need not hold for all diatomic molecules, but it is
a good rule of thumb. This effect is called overlap repulsion. Note that in
the special case where the overlap between the orbitals is negligible, S12
goes to zero and the two orbitals are shifted by equal amounts. However,
when is S12 nonzero there are two effects that shift the energies: the
physical interaction between the atoms and the fact that the 1sA and 1sB
orbitals are not orthogonal. When we diagonalize H, we account for both
these effects, and the orthogonality constraint pushes the orbitals upwards
in energy.
For the simple case of the oneelectron bond in H2+ we have seen that using
the LCAO principle together with the variational principle led to a recipe for
computing some approximate orbitals for a system that would be very
difficult to solve analytically. To generalize this to the more interesting
case of many electrons, we take our direction from our experience with the
independent particle model (IPM) applied to atoms and we build up
antisymmetrized wavefunctions out of the molecular orbitals. This is the
basic idea behind molecular orbital theory – there are many variations on the
central theme, but the same steps are always applied. Rather than go step
bystep and deal with H2 and then Li2 and then LiH … we will instead begin by
stating the general rules for applying MO theory to any system and then
proceed to show some illustrations of how this works out in practice.
⎛ c1 ⎞
⎜ ⎟
� c
ψ =⎜ 2 ⎟
⎜ ... ⎟
⎜ ⎟
⎝ cN ⎠
We note that for the sake of accuracy it is never a bad idea to include
more AO functions than you might think necessary – more AO functions
will always lead to more accurate results. The price is that the more
accurate computations also tend to be more complicated and time
consuming. To illustrate, note that we could have chosen to write the H2+
MOs as linear combinations of four functions – the 1s and 2s states on
each atom:
ψ = c11s A + c2 1sB + c3 2s A + c4 2sB
Now, when we use the variational principle to get the coefficients of the
lowest MO , c0, we are guaranteed that there is no set of coefficients
that will give us a lower energy. This is the foundation of the variational
method. Note that one possible set of coefficients is c3=c4=0, in which
case our 4function expansion reduces to the 2function expansion above.
Thus, the variationally optimal 4function MO will always have an energy
less than or equal to the optimal 2function MO. As a result, the
expansion with four functions allows the approximate MO to get closer to
the ground state energy. This makes sense, as the four AO expansion
has more flexibility than the constrained two AO expansion used
previously. The reason we didn’t use the four function expansion from
the beginning is that all the algebra is twice as difficult when we use four
functions as two: the vectors are twice as long, the matrices are twice as
big…. At least for a first try, it is generally good to start with the
smallest conceivable set of AOs for performing a calculation. If higher
accuracy is required, a longer expansion can be tried.
2) Compute the relevant matrix representations
For H2+ we had to compute two matrices – the Hamiltonian and the
overlap, which were both 2by2 by virtue of the two AO basis functions:
⎛ 1s A Ĥ el 1sB dτ ⎞
H≡⎜
∫
⎛ H11 H12 ⎞ ⎜ 1s A Ĥ el 1s A dτ ∫ ⎟
⎟≡⎜ ⎟
⎝ ∫
⎝ H 21 H 22 ⎠ ⎜ 1sB Ĥ el 1s A dτ ∫
1sB Ĥ el 1sB dτ ⎟
⎠
⎛ 1s A 1sB dτ ⎞
S≡⎜
∫
⎛ S11 S12 ⎞ ⎜ 1s A 1s A dτ ∫ ⎟
⎟≡⎜ ⎟
S
⎝ 21 S22 ⎠
⎝ ∫
⎜ 1sB 1s A dτ 1sB 1sB dτ ⎟∫ ⎠
ψ α ( r ) = ∑ ciα φiAO ( r
)
i =1
ψ1
with a determinant:
N
E = ∑ Ei
i=1
K� ij ≡ ∫∫ψ i* (1)ψ *j ( 2) ( ) j( ) 1 2 1 2
1 ψ 2 ψ 1 dr dr dσ dσ
r1 − r2 i
In fact, as we will see later on, there are even more elaborate ways to
obtain the energy from an MO calculation. When we work things out
by hand, the noninteracting picture is easiest and we will usually work
in that approximation when dealing with MOs.
Diatomic molecules
As a first application of MO theory, it is useful to consider firstrow
diatomic molecules (B2, C2, N2,O2, CO,CN, NO, etc.), which actually map
rather nicely on to an MO picture. We’ll go stepby step for the generic
“AB” diatomic to show how this fits into the MO theory framework.
However, for all the diatomics above, the 1s orbitals on both atoms will
be doubly occupied. Since we will primarily be interested in comparing the
MO descriptions of different diatomics the eternally occupied 1s orbital
will have no qualitative effect on our comparisons. It is therefore
customary to remove the 1s orbitals from the expansion:
ψ = c1 2s A + c2 2sB + c3 2 pzA + c4 2 pzB + c5 2 p yA + c6 2 p yB + c7 2 p xA + c8 2 pxB
The latter approximation is referred to as the valence electron or
frozen core approximation. The advantage is that it reduces the length
of our vectors from 10 to 8.
A B
symmetric (even) with respect to reflection
X Reflection Y Reflection
Hamiltonian + +
s + +
pz + +
py +
px +
This knocks out a bunch of integrals for us. For example, looking at
reflection about x:
s Hpˆ
∫
= ( + )( + )( −) = − ⇒ 0
A/ B xA / B
∫p zA / B
ˆ
Hp xA / B = ( + )( + )( − ) = − ⇒ 0
∫p yA / B
ˆ
Hp xA / B = ( + )( + )( − ) = − ⇒ 0
∫p xA / B
ˆ
Hp xA / B = ( − )( + )( − ) = + ⇒ ≠0
We have analogous expressions for reflection about y zero several more
⎛ s A Hs ⎞
⎜ ∫ ˆ
A ∫s ˆ
A HsB ∫s ˆ
A Hp zA ∫s ˆ
A HpzB 0 0 0 0
⎟
⎜ ⎟
∫ ˆ
⎜ sB Hs A ∫s ˆ
B HsB ∫s ˆ
B HpzA ∫s ˆ
B Hp zB 0 0 0 0 ⎟
⎜ ⎟
⎜
∫
⎜ pzA Hsˆ
A ∫p ˆ
zA HsB ∫p ˆ
zA Hp zA ∫p ˆ
zA Hp zB 0 0 0 0 ⎟
⎟
H≡⎜
∫
⎜ pzB Hsˆ
A ∫p ˆ
zB HsB ∫p ˆ
zB Hp zA ∫pzB Hp zB
ˆ 0 0 0 0 ⎟
⎟
⎜ ⎟
⎜
0 0 0 0 ∫p ˆ
yA Hp yA ∫p ˆ
yA Hp yB 0 0
⎟
⎜ ⎟
⎜ 0 0 0 0 ∫p ˆ
yB Hp yA ∫p ˆ
yB Hp yB 0 0 ⎟
⎜ ⎟
⎜
⎜
0 0 0 0 0 0 ∫p ˆ
xA Hp xA ∫ ˆ
pxA Hp xB ⎟
⎟
⎜
⎝
0 0 0 0 0 0 ∫p ˆ
xB Hp xA ∫ ˆ
pxB Hp xB ⎟
⎠
which we write:
⎛ H11 H12 H13 H14 0 0 0 0 ⎞
⎜H H 22 H 23 H 24 0 0 0 0 ⎟⎟
⎜ 21
⎜ H 31 H 32 H 33 H 34 0 0 0 0 ⎟
⎜ ⎟
H H 42 H 43 H 44 0 0 0 0 ⎟
H ≡ ⎜ 41
⎜ 0 0 0 0 H 55 H 56 0 0 ⎟
⎜ ⎟
⎜ 0 0 0 0 H 65 H 66 0 0 ⎟
⎜ 0 0 0 0 0 0 H 77 H 78 ⎟
⎜⎜ ⎟
⎝ 0 0 0 0 0 0 H 87 H 88 ⎟⎠
It is easy to show that the overlap matrix has the same overall shape
And similarly for the overlap matrix. The nice thing about block diagonal
matrices is you can reduce a large eigenvalue to several smaller ones. In
this case, our matrices break down into a 4by4 block (sA, sB, pzA, pzB) a
2by2 block (pyA,pyB) and another 2by2 block (pxA,pxB). All the rest of
the matrix is zero. As a result, we can decompose the above 8by8 into
three separate eigenvalue problems:
A) The first eigenvalue problem to be solved is a 4by4:
α α
⎛ H11 H12 H14 ⎞ ⎛ c1 ⎞
H13 ⎛ S11 S12 S13 S14 ⎞ ⎛ c1 ⎞
⎜H ⎟ ⎜ α⎟ ⎜ ⎜ ⎟
⎜ 21 H 22 H 23 H 24 ⎟ ⎜ c2 ⎟ α ⎜ S21 S 22 S23 S24 ⎟⎟ ⎜ c2α ⎟
⎜ ⎟ = E
⎜ H 31 H 32 H 33 H 34 ⎟ c3α ⎜ S31 S32 S33 S34 ⎟ ⎜ c3α ⎟
⎜ ⎟⎜ ⎟ ⎜ ⎟⎜ ⎟
⎝ H 41 H 42 H 43 H 44 ⎠ ⎜ c α ⎟ ⎝ S41 S42 S43 S44 ⎠ ⎜ c α ⎟
⎝ 4 ⎠ ⎝ 4 ⎠
which will give us four molecular orbitals that can be written as linear
combinations of the first four AOs (sA, sB, pzA, pzB)
ψ α = c1α 2s A + c2α 2sB + c3α 2 pzA + c4α 2 pzB
Because these orbitals are symmetric with respect to reflection about
both x and y, they will look something like the H2+ bonding and
antibonding orbitals, and so they are referred to as σorbitals. For
example, we can make the +/ combinations of the 2s orbitals to
obtain one bonding orbital and one antibonding:
σ1orbital
+ A B
A B
σ1∗orbital
A B
A B
σ2orbital
A B
where we label the upper orbital σ* because of the nodes between the
nuclei, whereas the σ orbital has no nodes between the nuclei. Note
10
that these +/ combinations are just to illustrate what the orbitals
will look like; in order to get the actual molecular orbitals we would
need to diagonalize the 4by4 and get the eigenvectors. However, if
we do that for a molecule like N2 we actually get orbitals that look
strikingly similar to the ones above:
σ2*
σ2
σ1*
σ1
11
πyorbital
A B
+
πy*orbital
A B
A B
πy
πy*
12
as for the πy orbitals above, expect that the πx orbitals come out of
the page.
4) Occupy the orbitals based on a stick diagram. The most important
thing here is to know the energetic ordering of the orbitals. This would
come out of actually evaluating the nonzero matrix elements in matrices
above and then solving the generalized eigenvalue problem, which is
tedious to do by hand. As a general rule however, there are only two
commonly found MO diagrams for diatomics:
σ2∗ σ2∗
πx*,πy* πx*,πy*
σ2
πx,πy
πx,πy
σ2
σ1 σ1
Hence, the only question is whether the second σbonding orbital is above
or below the πbonding orbitals. In practice, the σorbital (which has
significant pz character) is stabilized as you move from left to right along
the periodic table, with the σorbital being less stable for atoms to the
left of and including nitrogen and more stable for atoms to the right of
N. Once we have the orbital energy diagram in hand, we can assign the
electrons based on stick diagrams. For example, for CO we have 10
valence electrons and we predict a stick diagram for the ground state
like the one at left below .Meanwhile for NO, which has 11 valence
electrons, we have the stick diagram shown on the right.
13
σ2∗ σ2∗
πx*,πy* πx*,πy*
σ2
πx,πy
πx,πy
σ2
σ1∗ σ1∗
σ1 σ1
CO MO diagram NO MO diagram
We note one important feature we get directly out of the stick
diagrams: the highest occupied molecular orbital (HOMO) and the
lowest unoccupied molecular orbital (LUMO). For example, in CO the
HOMO is a πbonding orbital, whereas the LUMO is a π* antibonding
orbital. Meanwhile, for NO the HOMO and LUMO are both π*
orbitals. These orbitals determine reactivity in a crude fashion, as
when electrons are taken out of the molecule, they are removed from
the HOMO, and when electrons are added, they are added to the
LUMO.
5) Compute the energy. Here we can say very little about diatomics,
because we don’t even know the orbital energies exactly, making it
difficult to predict the energies of the whole molecule. If we knew
the orbital energies, the total energy for CO, for example, would be:
ECO=2Eσ1+2Eσ1∗+2Eπx+2Eπy+2Eσ2
As we don’t know these orbital energies, we cannot evaluate the
accuracy of this independent electron model for diatomics. However,
the bond order is a useful descriptor that correlates very well with
the MO energy. For a diatomic, the bond order is simply:
((# Bonding Electrons)(# Antibonding Electrons))/2
The factor of two reflects the requirement of two electrons for
forming a bond. A higher bond order implies a stronger bond and a
lower bond order a weaker bond. Thus, MO theory predicts CO will
have a stronger bond (bond order 3) than NO (bond order 2.5), which
14
Now, you may have been taught in previous courses that because there are
other resonance structures you can draw for this molecule, such as:
Conjugated molecules of tend to be planar, so that we can place all the atoms
in the xy plane. Thus, the molecule will have reflection symmetry about the
zaxis:
Now, for diatomics, we had reflection symmetry about x and y and this gave
rise to πx and πy orbitals that were odd with respect to reflection and σ
orbitals that were even. In the same way, for planar conjugated systems the
orbitals will separate into σ orbitals that are even with respect to reflection
and πz orbitals that are odd with respect to reflection about z. These πz
orbitals will be linear combinations of the pz orbitals on each carbon atom:
N
Etot = ∑ Ei
i =1
To illustrate how we apply Hückel in practice, let’s work out the energy of
benzene as an example.
1
6 2
5 3
4
1) Each of the MOs is a linear combination of 6 pz orbitals
⎛ c1µ ⎞
⎜ µ⎟
⎜ c2 ⎟
6 ⎜ cµ ⎟
ψ µ = ∑ ciµ pzi → c µ = ⎜ 3µ ⎟
i=1 ⎜ c4 ⎟
⎜ cµ ⎟
⎜⎜ 5µ ⎟⎟
⎝ c6 ⎠
2) It is relatively easy to work out the Hamiltonian. It is a 6by6 matrix.
The first rule implies that every diagonal element is α:
⎛α ⎞
⎜ ⎟
⎜ α ⎟
⎜ α ⎟
H=⎜ ⎟
⎜ α ⎟
⎜ α ⎟
⎜⎜ ⎟
⎝ α ⎟⎠
The only other nonzero terms will be between neighbors: 12, 23, 34, 45,
56 and 61. All these elements are equal to β:
⎛α β β⎞
⎜ ⎟
⎜β α β ⎟
⎜ β α β ⎟
H=⎜ ⎟
⎜ β α β ⎟
⎜ β α β⎟
⎜
⎜ β ⎟
⎝ β α ⎟⎠
All the rest of the elements involve nonnearest neighbors and so are zero:
⎛α β 0 0 0 β⎞
⎜
⎜β α β 0 0 0 ⎟⎟
⎜0 β α β 0 0⎟
H=⎜ ⎟
⎜0 0 β α β 0⎟
⎜0 0 0 β α β⎟
⎜⎜ ⎟
⎝β 0 0 0 β α ⎟⎠
3) Finding the eigenvalues of H is easy with a computer. We find 4 distinct
energies:
E6=α−2β
E4=E5=α−β
E2=E3=α+β
E1=α+2β
The lowest and highest energies are nondegenerate. The second/third and
fourth/fifth energies are degenerate with one another. With a little more
work we can get the eigenvectors. They are:
⎛ +1⎞ ⎛ +1 ⎞ ⎛ +1⎞ ⎛ +1⎞ ⎛ +1 ⎞ ⎛ +1⎞
⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟ ⎜ ⎟
⎜ −1⎟ ⎜ −2 ⎟ ⎜0⎟ ⎜0⎟ ⎜ +2 ⎟ ⎜ +1⎟
1 ⎜ +1⎟ 5 1 ⎜ +1 ⎟ 4 1 ⎜ −1⎟ 3 1 ⎜ −1⎟ 2 1 ⎜ +1 ⎟ 1 1 ⎜ +1⎟
c6 = ⎜ ⎟ c = ⎜ ⎟ c = ⎜ ⎟ c = ⎜ ⎟ c = ⎜ ⎟ c = ⎜ ⎟
6 ⎜ −1⎟ 12 ⎜ +1 ⎟ 4 ⎜ +1⎟ 4 ⎜ −1⎟ 12 ⎜ −1 ⎟ 6 ⎜ +1⎟
⎜ +1⎟ ⎜ −2 ⎟ ⎜0⎟ ⎜0⎟ ⎜ −2 ⎟ ⎜ +1⎟
⎜⎜ ⎟⎟ ⎜⎜ ⎟⎟ ⎜⎜ ⎟⎟ ⎜⎜ ⎟⎟ ⎜⎜ ⎟⎟ ⎜⎜ ⎟⎟
⎝ −1⎠ ⎝ +1 ⎠ ⎝ −1⎠ ⎝ +1⎠ ⎝ −1 ⎠ ⎝ +1⎠
The pictures at the bottom illustrate the MOs by denting positive (negative)
lobes by circles whose size corresponds to the weight of that particular pz
orbital in the MO. The resulting phase pattern is very reminiscent of a
particle on a ring, where we saw that the ground state had no nodes, the
first and second excited states were degenerate (sine and cosine) and had
one node, the third and fourth were degenerate with two nodes. The one
difference is that, in benzene the fifth excited state is the only one with
E6=α−2β
E4=E5=α−β
E2=E3=α+β
E1=α+2β
Now, we get to the interesting part. What does this tell us about the
bonding in benzene? Well, first we note that benzene is somewhat more
stable than a typical system with three double bonds would be. If we do
Hückel theory for ethylene, we find that a single ethylene double bond has
an energy
EC =C = 2α + 2β
Thus, if benzene simply had three double bonds, we would expect it to have a
total energy of
E = 3EC =C = 6α + 6 β
which is off by 2β. We recall that β is negative, so that the πelectrons in
benzene are more stable than a collection of three double bonds. We call
this aromatic stabilization, and Hückel theory predicts a similar stabilization
of other cyclic conjugated systems with 4N+2 electrons. This energetic
stabilization explains in part why benzene is so unreactive as compared to
other unsaturated hydrocarbons.
We can go one step further in our analysis and look at the bond order. In
Hückel theory the bond order can be defined as:
occ
Oij ≡ ∑ ciµ c µj
µ =1
This definition incorporates the idea that, if molecular orbital µ has a bond
between the ith and jth carbons, then the coefficients of the MO on those
carbons should both have the same sign (e.g. we have pzi + pzj). If the orbital
⎛ +1 ⎞ ⎛ + 1 ⎞ ⎛ +1 ⎞ ⎛ +2 ⎞ ⎛ +1 ⎞ ⎛ 0 ⎞
= 2⎜ ⎟×⎜ ⎟ + 2⎜ ⎟×⎜ ⎟ + 2⎜ ⎟×⎜ ⎟
⎝ 6⎠ ⎝ 6⎠ ⎝ 12 ⎠ ⎝ 12 ⎠ ⎝ 4⎠ ⎝ 4⎠
1 2 2
=2 +2 =
6 12 3
Thus, the C1 and C2 formally appear to share 2/3 of a πbond [Recall that we
are omitting the σorbitals, so the total bond order would be 1 2/3 including
the σ bonds]. We can repeat the same procedure for each CC bond in
benzene and we will find the same result: there are 6 equivalent πbonds,
each of order 2/3. This gives us great confidence in drawing the Lewis
structure we all learned in freshman chemistry:
You might have expected this to give a bond order of 1/2 for each CC π
bond rather than 2/3. The extra 1/6 of a bond per carbon comes directly
from the aromatic stabilization: because the molecule is more stable than
three isolated πbonds by 2β, this effectively adds another πbond to the
system, which gets distributed equally among all six carbons, resulting in an
increased bond order. This effect can be confirmed experimentally, as
benzene has slightly shorter CC bonds than nonaromatic conjugated
systems, indicating a higher bond order between the carbons.
orbitals that arise in directional bonding. These results not only mean MO
theory is a useful tool – in practice these discoveries have led to MO theory
becoming part of the way chemists think about molecules.
The basic idea of electronic structure theory is that, within the Born
Oppenheimer approximation, we can fix the M nuclei in our molecule at some
positions RI. Then, we are left with the Hamiltonian for the electrons
moving in the effective field created by the nuclei:
N N M N
ZI 1
Ĥ ≡ − 1
2 ∑ ∇ -∑∑
i=1
2
i
i=1 I =1
+∑
r̂i − R I i< j r̂i − r̂j Eq. 1
Where the first term is the kinetic energy of all N electrons, the second
term is the attraction between the electrons and nuclei and the third is the
pairwise repulsion between all the electrons. The central aim of electronic
structure theory is to find all the eigenfunctions of this Hamiltonian. As
we have seen, the eigenvalues we
Reaction get will depend on our choice of
Unstable
Barrier the positions of the nuclei –
intermediate
Eel(R1,R2,R3,…RM). As was the
case with diatomics, these
Eel(R1,R2)
energies will tell us how stable
the molecule is with a given
R2 Equilibrium configuration of the nuclei {RI} –
Conformation if Eel is very low, the molecule will
R1 be very stable, while if Eel is high,
the molecule will be unstable in
that configuration. The energy Eel(R1,R2,R3,…RM) is called the potential
energy surface, and it contains a wealth of information, as illustrated in the
picture at above. We can determine the equilibrium configuration of the
molecule by looking for the minimum energy point on the potential energy
surface. We can find metastable intermediate states by looking for local
minima – i.e. minima that are not the lowest possible energy states, but which
are separated from all other minima by energy barriers. In both of these
cases, we are interested in points where ∇Eel = 0 . Further, the potential
surface can tell us about the activation energies between different minima
and the pathways that are required to get from the “reactant” state to the
“product” state.
We note that while the Hamiltonian above will have many, many eigenstates,
in most cases we will only be interested in the lowest eigenstate – the
electronic ground state. The basic reason for this is that in stable
molecules, the lowest excited states are usually several eV above the ground
state and therefore not very important in chemical reactions where the
available energy is usually only tenths of an eV. In cases where multiple
electronic states are important, the
Hamiltonian above will give us
separate potential surfaces E1el, E2el, σ* potential
E3el … and separate wavefunctions surface
1 2 3
Ψ el, Ψ el, Ψ el. The different
potential surfaces will tell us about
the favored conformations of the
molecules in the different electronic σ potential
states. We have already seen this Erxn surface
for H2+. When we solved for the
electronic states, we got two
eigenstates: σ and σ*. If we put the electron in the σ orbital, the molecule
was bound and had a potential surface like the lower surface at right.
However, if we put the electron in the σ∗ orbital the molecule was not bound
and we got the upper surface.
So, at the very least our task is clear cut: solve for the eigenstates of Eq. 1.
Unfortunately, this task is also impossible in practice, even on a computer.
Thus, over the years chemists have developed a vast array of sophisticated
techniques that allow us to at least approximate these solutions to within a
tolerable degree of accuracy. Learning all the details of these
approximations would require a course unto itself: the derivations of the
individual approximations are exceedingly complex, and the sheer number of
different approximations that can be made is quite impressive. These
detailed derivations teach us a lot about what molecules and properties we
should expect our approximations to work for and how we should think about
improving our calculations in cases where the theory fails. However, the
thing that has really brought computational chemistry into the mainstream is
the fact that one does not have to understand every nuance of a method
in order to know how to use it successfully. It suffices to have a simple,
qualitative understanding of how each method works and when it can be
applied. Then, coupling that knowledge with a little technical proficiency at
using commercial chemistry software packages allows us to run fairly
sophisticated calculations on our desktop computer. The next two lectures
are intended to prepare us to run these types of calculations.
First, we note that nearly all the popular approximations are still based on
MO theory – MO theory on steroids in some cases, but MO theory
nonetheless. Thus, there are still 5 steps in the calculation
1) Choose an Atomic Orbital Basis
2) Build the Relevant Matrices
3) Solve the Eigenvalue Problem
4) Occupy the orbitals based on a stick diagram
5) Compute the energy
In a typical calculation, the computer automatically handles steps 24
automatically – we don’t have to tell it anything at all. It is sometimes
helpful to know that the computer is doing these things (e.g. The calculation
crashed my computer. What was it doing when it crashed? Oh, it was trying
to solve the eigenvalue problem.) but we essentially never have to do them
ourselves. This leaves two steps (1 and 5) that require some input from us
for the calculation to run properly
The first point here is that for real electronic structure calculations, you
typically use a basis set that is much larger than your intuition might tell
you. For example, for H2+ we guessed that we would get a decent result if
we wrote:
ψ = c11sA + c2 1sB
A basis of this type would be called a minimal basis set, because it is the
smallest one that is even close to right. In a more accurate calculation, you
might use a basis that looks more like:
ψ = c11s A + c2 1sB + c3 2s A + c4 2sB + c5 2 pxA + c6 2 pxB + c7 2 p yA
+ c8 2 p yA + c9 2 pzA + c10 2 pzB + c11 3s A + c12 3sB
The reason we use such extended basis sets arises from a point that was
discussed earlier concerning MO theory. Because our results are variational,
a bigger basis always gets us a lower energy, and therefore a longer AO
expansion always gets us closer to the ground state energy. In the worst
case, the unimportant basis functions will just have coefficients that are
very near zero. While such extended basis sets would be a significant
problem if we were working things out by hand, computers have no problem
dealing with expansions involving even 10,000 basis functions.
The second important point about the atomic orbitals we use is that they are
not hydrogenic orbitals. The reason for this is that the twoelectron
integrals involving hydrogenic orbitals cannot all be worked out analytically,
making it very difficult to complete Step 2. Instead of hydrogenic orbitals –
which decay like e−r – we will use Gaussian orbitals that decay like e−α r .
2
while using Gaussians may mean we have to use a few extra AOs, if we use
enough of them we should be able to get the same answer.
So we plan to use relatively large Gaussian basis sets for our calculations.
How exactly do we choose those basis sets? Thankfully, a significant amount
of trialanderror research has distilled the choices down to a few key basis
set features. Depending on the problem at hand and the accuracy desired
we only need to consider three aspects of the AO basis.
Unfortunately, the commonly used names for basis sets follow somewhat
uneven conventions. The basic problem is that many different folks develop
basis sets and each group has their own naming conventions. At the end of
the table above, we’ve listed a few names of commonly used SZ,DZ and TZ
basis sets. There aren’t any commonly used QZ basis sets, because once
Diffuse Functions
Occasionally, and particularly when dealing with anions, the SZ/DZ/TZ/…
hierarchy converges very slowly. For anions, this is because the extra
electron is only very weakly bound, and therefore spends a lot of time far
from the nucleus. It is therefore best to include a few basis functions that
decay very slowly to describe this extra electron. Functions of this type
are called “diffuse” functions. They are still Gaussians ( e−α r ), but the value
2
of α is very, very small causing the atomic orbital to decay slowly. Similar to
the situation for polarization functions, diffuse functions can be added in a
mixandmatch way to standard basis sets. Here, the notation “+” or “aug“
is added to a basis set to show that diffuse functions have been added.
Thus, we have basis sets like 321++G, 631+G(d,p), augTZP.
Things also become more complicated when dealing with second row
transition metals. Here, relativistic effects become important, because the
Schrödinger equation predicts that the 1s electrons in these atoms are
actually moving at a fair fraction of the speed of light. Under these
circumstances, the Schrödinger equation is not strictly correct and we need
As one progresses further up the periodic table, fewer and fewer basis sets
are available, simply because less is known about their chemistry.
This is just a very brief overview of what basis sets are available and how
good each one is expected to be. The general idea of using basis sets is to
use larger and larger basis sets until the quantity we are computing stops
changing. This is based on the idea that we are really using the AO
expansion to approximate the exact solution of the Schrödinger equation. If
we had an infinite basis, then we would get the exact answer, but with a
large and finite basis we should be able to get close enough. Note, however,
that the calculation will typically take more time in a larger basis than in a
smaller one. Thus, we really want to make the basis just big enough to get
the answer right, but no larger.
N N
ENI = ∑ Ei = ∑ ∫ ψ i (1) Ĥψ i (1) dτ
i=1 i=1
Where, on the right hand side we have noted that we can write the NI
energy as a sum of integrals involving the orbitals. We already know from
looking at atoms that this isn’t going to be good enough to get us really
accurate answers; the electronelectron interaction is just too important.
In real calculations, one must choose a method for computing the energy
from among several choices.
K� ij ≡ ∫∫ψ i* (1)ψ *j ( 2) ( ) j( ) 1 2 1 2
1 ψ 2 ψ 1 dr dr dσ dσ
r1 − r2 i
Since the energy contains the average repulsion, we expect our results will
be more accurate. However, there is an ambiguity in this expression. The
IPM energy above is correct for a determinant constructed out of any set of
orbitals {ψ i } and the energy will be different depending on the orbitals we
choose. For example, we could have chosen a different set of orbitals, {ψ 'i } ,
and gotten a different energy:
N
E ' NI = ∑ E 'i + ∑ J� 'ij − K� 'ij
i =1
How do we choose the best set of orbitals then? HartreeFock uses the
variational principle to determine the optimal orbitals. That is, in HF we find
the set of orbitals that minimize the independent particle energy. These
orbitals will be different from the noninteracting orbitals because they will
take into account the average electronelectron repulsion terms in the
Hamiltonian. Thus, effects like shielding that we have discussed
qualitatively will be incorporated into the shapes of the orbitals. This will
tend to lead to slightly more spread out orbitals and will also occasionally
10
change the ordering of different orbitals (e.g. σ might shift below π once
interactions are included).
Now, the molecular orbitals (and hence the energy) are determined by their
coefficients. Finding the best orbitals is thus equivalent to finding the best
coefficients. Mathematically, then, we want to find the orbitals that make
the derivative of the IPM energy zero:
∂EIPM ∂ N
∂ciα
= ∑ Ei + ∑ J�ij − K� ij = 0
∂ciα i=1
In order to satisfy this condition, one typically resorts to an iterative
procedure, where steps 25 of our MO procedure are performed repeatedly:
1) Choose an AO Basis
11
There exists a functional Ev[ρ] such that, given the ground state
density, ρ0, Ev[ρ0]=E0 where E0 is the exact ground state energy.
Further, for any density, ρ’, that is not the ground state density,
Ev[ρ’]>E0.
The unfortunate point is that we don’t know the form of the functional Ev.
We can prove it exists, but we can’t construct it. However, from a
pragmatic point of view, we do have very good approximations to Ev, and the
basic idea is to choose an approximate (but perhaps very, very good) form
for Ev and then minimize the energy as a function of the density. That is, we
dEv
look for the point where = 0 . Based on Eq. 2 above, we see that ρ just
dρ
12
depends on the MOs and hence on the MO coefficients, so once again we are
dEv
looking for the set of MO coefficients such that = 0 . Given the
dc
similarity between DFT and HF, it is not surprising that DFT is also solved
by self consistent field iterations. In fact, in a standard electronic
structure code, DFT and HF are performed in exactly the same manner (see
flow chart above). The only change is the way one computes the energy and
dE
.
dc
Now, as alluded to above, there exist good approximations (note the plural)
to Ev. Just as was the case with approximate AO basis sets, these
approximate energy expressions have strange abbreviations. We won’t go
into the fine differences between different DFT energy expressions here.
I’ll simply note that roughly, the quality of the different functionals is
expected to follow:
LSDA < PBE ≈ BLYP < PBE0 ≈ B3LYP
Thus, LSDA is typically the worst DFT approximation and B3LYP is typically
among the best. I should mention that this is just a rule of thumb; unlike
the case for basis sets where we were approaching a welldefined limit, here
we are trying various uncontrolled approximations to an unknown functional.
Experience shows us that B3LYP is usually the best, but this need not always
be the case.
Finally, we note that the speed of a DFT calculation is about the same as
Hartree Fock – both involve selfconsistent field iterations to determine the
best set of orbitals, and so both take about the same amount of computer
time. However, for the same amount of effort, you can get quite accurate
results. As a rule of thumb, with B3LYP you can get energies correct to
within 3 kcal/mol and distances correct to within .01 Å.
PostHartree Fock Calculations
Here, the idea is to employ wavefunctions that are more flexible than a
Slater determinant. This can be done by adding up various combinations of
Slater determinants, by adding terms that explicitly correlate pairs of
electrons (e.g. functions that depend on r1 and r2 simulataneously) and a
variety of other creative techniques. These approaches are all aimed at
13
As a general rule, postHF calculations are much, much more expensive than
HF or DFT and also require bigger basis sets: whereas HF might converge
with a DZ or TZ basis, a postHF calculation might require QZ or even 5Z.
Hence, they should only be attempted for relatively small molecules where
high accuracy is required
Combining what we have learned, then, the approximations we can make fit
nicely into a twodimensional parameter space:
Exact
Method
Answer
CCSD(T)
CCSD
MP2
Feasible
DFT Calculations
CASSCF
HF Basis
14
On the one axis, we have the basis sets we can choose from. On the other,
we have the different methods for approximating the energy. The get close
to the exact answer, we need to employ a large basis set and an accurate
energy method. Unfortunately, both increasing the size of the basis and
improving the method tend to slow our calculations down. Given that we don’t
want to wait years and years to find out the result of a calculation, modern
computers therefore limit how accurate our answers can be (as illustrated
with the red line above). As we become experts at what is and is not
feasible with current computing power, we become better able to get good
accuracy for a variety of problems.
Now, in chemistry we typically deal with the part of the spectrum from
ultraviolet (λ≈100 nm) to radio waves (λ≈10 m)1. Meanwhile, a typical molecule
1
There are a few examples of spectroscopic measurements in the XRay region. In these
cases, the wavelength can be very small and the dipole approximation is not valid.
is about 1 nm in size. Let us assume that the molecule is sitting at the origin.
Then, in the 1 nm3 volume occupied by the molecule we have:
k·r ≈ |k| |r| ≈ 2p/(100 nm) 1 nm = .06
Where we have assumed UV radiation (longer wavelengths would lead to even
smaller values for k·r). Thus, k·r is a small number and we can expand the
electric and magnetic fields in a power series in k·r:
E(r,t)≈E0 [cos(k·0-ωt)+O(k·r)]≈E0 cos(ωt)
So now suppose that we have a molecule we are interested in, and it has a
Hamiltonian, Ĥ 0 , when the field is off. Then, when the field is on, the
Hamiltonian will be
Ĥ ( t ) = Ĥ 0 + Ĥ E ( t ) + Ĥ B ( t )
Actually, in most cases, the simultaneous effects of electric and magnetic
fields are not important and we will consider one or the other:
Ĥ ( t ) ≡ Ĥ 0 + Ĥ1 ( t ) Ĥ1 ( t ) ≡ Ĥ E ( t ) or Ĥ B ( t ) .
Thus, in the presence of light, a molecule feels a timedependent
Hamiltonian. This situation is quite different with what we have discussed
so far. Previously, our Hamiltonian has been time independent and our job
has simply reduced to finding the eigenstates of Ĥ . Now, we have a
Hamiltonian that varies with time, meaning that the energy eigenvalues and
eigenstates of Ĥ also change with time. What can we say that is meaningful
about a system that is constantly changing?
Thus, things like the probability density do not change because multiplying
by the complex conjugate wipes out the phase factor:
φn ( t ) = {e−iE t / �φn ( 0 )} * e−iE t / �φn ( 0 ) =eiE t / �φn ( 0 ) * e−iE t / �φn ( 0 ) = φn ( 0 )
2 2
n n n n
To work out these rates, we first work out the time dependence of some
arbitrary state, ψ(t). We can expand ψ(t) as a linear combination of the
eigenstates:
ψ ( t ) = ∑ cn ( t ) φn ( t )
n
where cn(t) are the coefficients to be determined. Next, we plug this into
the TDSE:
i�ψ� ( t ) = Ĥψ ( t )
∂
⇒ i� ∑ cn ( t ) φn ( t ) = Ĥ ∑n cn ( t ) φn ( t )
∂t n
( )
⇒ i� ∑ c�n ( t ) φn ( t ) + cn ( t ) φ�n ( t ) = ∑ cn ( t ) Ĥ 0 + Ĥ1 ( t ) φn ( t )
n n
⇒ i� ∑ c�n ( t ) φn ( t ) −
n
iEn
�
( )
cn ( t ) φn ( t ) = ∑ cn ( t ) En + Ĥ1 ( t ) φn ( t )
n
n n n
(
⇒ i� ∑ c�n ( t ) φn ( t ) + ∑ En cn ( t ) φn ( t ) = ∑ cn ( t ) En + Ĥ1 ( t ) φn ( t ) )
⇒ i� ∑ c�n ( t ) φn ( t ) = ∑ cn ( t ) Ĥ1
( t ) φn ( t )
n n
Next, we multiply both sides by the final state we are interested in (φf*) and
then integrate over all space. On the left hand side, we get:
i� ∫ φ f * ( t ) ∑ c�n ( t )φn ( t ) dτ = i�∑ c�n ( t ) ∫ φ f * ( t ) φn ( t ) dτ = i�c� f ( t )
n n
δnf
Meanwhile, on the right we get:
∫ φ f ( t ) ∑ cn ( t )Hˆ 1 ( t ) φn ( t ) dτ = ∑ cn ( t )∫ φ f ( t ) Ĥ1 ( t ) φn ( t ) dτ
* *
n n
Up to this point, we haven’t used the form of H1 at all. We note that we can
rewrite the lightmatter interaction as:
Ĥ1 ( t ) = V̂ cos (ω t )
qgˆ
where, for electric fields V̂ ≡ −erˆ i E 0 and for magnetic fields Vˆ ≡ − Ii B .
2m
In either case, we can rewrite the cosine in terms of complex exponentials:
Ĥ1 ( t ) = V̂ 1
2 (e iωt
+ e− iωt )
Plugging this into Eq.1 above gives:
( )
i�c� f ( t ) = ∑ ∫ φ f * ( t ) 12 V̂ eiωt + e−iωt φn ( t ) dτ cn ( t )
n
= ∑ ∫ φf * (0) e
iE f t /� 1
2 ( )
V̂ eiωt + e −iωt e−iEnt /�φn ( 0 ) dτ cn ( t )
n
= ∑ ∫ φ f * ( 0 ) 12 V̂ φn ( 0 ) dτ
n
(e −i ( En −E f −�ω ) t / �
+e
−i ( En −E f +�ω ) t / �
)c n (t )
=∑
n
1
2 V fn (e −i ( En −E f −�ω ) t / �
+e
−i ( En −E f +�ω ) t / �
)c n (t )
These two assumptions lead to an equation for the coefficients of the form:
i�c� f
( t ) = ∑ 12 V fn e
n
( − i( En −E f −�ω ) t / �
+e
−i ( En −E f +�ω ) t / �
)c n (t )
⇒ i�c� f ( t ) = 12 V fi e ( −i ( Ei −E f −�ω ) t / �
+e
−i ( Ei −E f +�ω ) t / �
) c (t )
i
= 12 V fi e ( −i ( Ei −E f −�ω ) t / �
+e
−i ( Ei −E f +�ω ) t / �
)
Now we can integrate this new equation to obtain c f ( t ) :
( ) dt
T
i�c f ( T ) = V fi ∫ e
−i ( Ei −E f −�ω ) t / � −i ( Ei −E f +�ω ) t / �
1
2 +e
0
(e ) dt
T
V fi
⇒ c f (T ) =
−i ( Ei −E f −�ω ) t / � −i ( Ei −E f +�ω ) t / �
2i� ∫
+e Eq. 2
0
2 2
∫ (e ) dt
T
V fi
Pf ( T ) = c f ( T ) =
2 −i ( Ei −E f −�ω ) t / � −i ( Ei −E f +�ω ) t / �
+e
4� 2 0
∫ (e ) dt
T
V fi −i ( Ei −E f −�ω ) t / � −i ( Ei −E f +�ω ) t / �
Pf = lim +e
4� 2 T →∞
0
and in fact, we are usually not interested in probabilities, but rates, which
are probabilities per unit time:
2 2
∫ (e ) dt
T
V fi 1 −i ( Ei −E f −�ω ) t / � −i ( Ei −E f +�ω ) t / �
W fi = lim
2 T →∞
+e
4� T 0
This integral looks very difficult. However, it is easy to work out with
pictures because it is almost always zero. Note that both the real and
imaginary parts of the integrand oscillate. Thus, we will be computing the
integral of something that looks like:
Thus, as long as the integrand oscillates, the positive regions will cancel out
the negative ones and the integral will be zero. There only two situations
where the integrand is not oscillatory: Ei − E f − �ω = 0 (in which case the
first term is unity) and Ei − E f + �ω = 0 (in which case the second term is
unity). We can therefore write
2
V fi
W fi ∝ ⎡δ ( Ei − E f − �ω ) + δ ( Ei − E f + �ω ) ⎤
4� ⎣ 2 ⎦
where δ(x) is a function that is defined to be nonzero only when x=0. This
result is called Fermi’s golden rule. It gives us a way of predicting the rate
of any i→f transition in any molecule induced by an electromagnetic field of
The way these terms are interpreted are as follows: in the first case, the
light increases the energy in the system by exactly one photon worth of
energy. Here, we think of a photon being absorbed by the molecule and
exciting the system. In the second case, the light reduces the energy of
the system by exactly one photon worth of energy. Here, we think of the
molecule emitting a photon and relaxing to a lower energy state. The fact
that photon emission by a molecule can be induced by light is called
stimulated emission, and is the principle on which lasers are built: basically,
when you shine light on an excited molecule, you get more photons out than
you put in.
VIBRATIONAL SPECTROSCOPY
As we’ve emphasized many times in this course, within the Born
Oppenheimer approximation,
the nuclei move on a potential Harmonic
energy surface (PES) Approximation
determined by the electrons. R
For example, the potential R0 A + B separated atoms
felt by the nuclei in a V(R)
diatomic molecule is shown in
cartoon form at right. At low
energies, the molecule will sit
near the bottom of this
potential energy surface. In
this case, no matter what the equilibrium bond length
detailed structure of the potential is, locally the nuclei will “feel” a nearly
harmonic potential. Generally, the motion of the nuclei along the PES is
called vibrational motion, and clearly at low energies a good model for the
nuclear motion is a Harmonic oscillator.
Applying Fermi’s Golden Rule, we find that when we irradiate the molecule,
the probability of a transition between the ith and fth Harmonic oscialltor
states is:
2
V fi
W fi ∝ ⎡δ ( Ei − E f − �ω ) + δ ( Ei − E f + �ω )⎤
4� 2 ⎣ ⎦
2 µω 2 µω
�e 2
( (i + 1) δ + i δ f ,i −1 )
2 2
⇒ V fi = ER f ,i +1 i + 1 φi +1 iφi−1
2 µω
where above it should be clarified that in this expression “i" never refers to
√1 – it always refers to the initial quantum number of the system. Thus, we
immediately see that a transition will be forbidden unless the initial and
final states differ by one quantum of excitation. Further, we see that
the transitions become more probable for more highly excited states. That
is, Vfi gets bigger as i gets bigger.
Combining the explicit expression for the transition matrix element with
Fermi’s Golden Rule again gives:
e2
W fi ∝
8�µω
( ) (
ER ( i + 1) δ f ,i +1 + i δ f ,i −1 ⎡⎣δ Ei − E f − �ω + δ Ei − E f + �ω ⎤⎦
2
) ( )
⇒ W fi ∝ E R
2
{(i + 1) δ f ,i +1
⎡⎣δ ( Ei − Ei+1 − �ω ) + δ ( Ei − Ei+1 + �ω )⎤⎦
0
⇒ W fi ∝ ER
2
{(i + 1) δ f ,i+1 ⎡⎣δ ( −�ωo − �ω ) + δ ( −�ωo + �ω )⎤⎦
+ i δ f ,i −1 ⎡⎣δ ( �ωo − �ω ) + δ ( �ωo + �ω ) ⎤⎦ }
0
⇒ W fi ∝ ER
2
{(i + 1) δ f ,i +1 + i δ f ,i −1 }δ ( �ωo − �ω )
Thus, we see that a harmonic oscillator will only absorb or emit photons
of frequency �ωo , where ωo is the frequency of the oscillator. Thus, if
we look at the absorption spectrum, for example, we will see absorption at
only one frequency:
ωo Light Frequency (ω)
Absorption
Intensity
a nice correlation between the “strength” of the bond and the frequency at
which it will absorb.
Heterogeneity
The primary reason the real spectrum above looks different than the model
is because the real spectrum was taken in solution. The model is correct for
a single diatomic, or for many, many copies of identical diatomic molecules.
However, in solution, ever molecule is just slightly different, because every
molecule has a slightly different arrangement of solvent molecules around it.
These solvent molecules subtly change the PES, slightly shifting the
vibrational frequency of each molecule and also modifying the transition
dipole a bit. Thus, while a single hydrogen fluoride molecule might have a
spectrum like the model above, a solution with many HF molecules would look
something like:
Absorption
Intensity
Going over to the situation where there are 1023 HF molecules and
recognizing that our spectra will tend to add the intensity of lines that are
closer together than our spectrometer can differentiate, we anticipate that
for a diatomic molecule in solution, the vibrational spectrum should look
something like:
ωo Light Frequency (ω)
Absorption
Intensity
Anharmonicity
Another reason real spectra differ from our model is that assuming the PES
is harmonic is only a model. If we want high accuracy, we need to account
for anharmonic terms in the potential:
1 1 1
V ( R ) = mωo 2 ( R − R0 ) + α ( R − R0 ) + β ( R − R0 ) + ...
2 3 4
2 6 24
One can investigate the quantitative effects of the anharmonic terms on the
spectrum by performing variational calculations. However, at a basic level
there are two ways that anharmonic terms impact vibrational spectra:
1) The energy differences between adjacent states are no longer
constant. Clearly, the eigenvalues of an anharmonic Hamiltonian
will not be equally spaced – this was a special feature of the
Harmonic system. Thus, for a real system we should expect the
0→1 transition to have a slightly different frequency than 1→2,
which in turn will be different that 2→3 …. Generally, the higher
transitions have lower (i.e. redshifted) energies because of the
shape of the molecular PES – rather than tend toward infinity at
large distances as the harmonic potential does, a molecular PES
tends toward a constant dissociation limit. Thus, the higher
eigenstates are lower in energy than they would be for the
corresponding harmonic potential. Taking this information, we
would then expect a single anharmonic oscillator to have a
spectrum something like:
ωo Light Frequency (ω)
2→ 3
Absorption
1→ 2
Intensity
0→ 1
where we note that while the rate of, say, 1→2 is about twice that
of 0→1 (because the transition dipole is twice as big) the intensity
of 0→1 is greater because the intensity is (Probability of i being
occupied)x(Rate of i→f) and at room temperature the system
spends most of its time in the ground vibrational state. The 1→2,
2→3 … lines in the spectrum are called hot bands.
2) Anharmonicity relaxes the Δn=±1 selection rule. Note that the
rules we arrived at were based on the fact that â +φi ∝ φi+1 . This is
only true for the Harmonic oscillator states. For anharmonic
eigenstates â +φi ∝ φi+1 + ε1φi+2 + ..... . Thus transitions with Δn=±2,±3…
will no longer be forbidden for anharmonic oscillators. Rather, in
the presence of a bit of anharmonicity, they will be weakly allowed.
Combining this observation with point 1) above results in a more
complete picture for what the IR spectrum of an anharmonic
oscillator should look like:
ωo 2ωo Light Frequency (ω)
2→ 4
2→ 3
Absorption 1→ 3
1→ 2
Intensity
0→ 2
0→ 1
Polyatomic Molecules
The final difference between the model above and a general IR spectrum is
that in chemistry, we are not always dealing with diatomic molecules. For a
polyatomic molecule, we can still think of the potential as a Harmonic
potential, but it has to be manydimensional – it has to depend on several
variables R1, R2, R3, …. The most general Harmonic potential we can come up
with is then of the form:
V ( R1 , R2 , R3 ,...) = 12 k11 R12 + 12 k12 R1 R2 + 12 k13 R1 R3 + .... + 12 k21 R2 R1 + 12 k22 R2 2 + 12 k23 R2 R3 + ....
+ 12 k31 R3 R1 + 12 k32 R3 R2 + 12 k33 R32 + ...
where it is important to notice the cross terms involving, say R1 and R2,
which couple the different vibrations. At first sight, it seems like we
⎛ k11' 0 0 ... ⎞
⎜ ⎟
0 k22' 0 ... ⎟
� ixˆ = yˆ T iK ′iy
x̂T iK K′ = ⎜
⎜ 0 0 k33' ... ⎟
⎜⎜ ⎟
⎝ ... ... ... ... ⎟⎠
Further, we will assume there is a matrix U that transforms from x to y
y = Uix ⇒ y T = xT i U T
where in the second equality, we recall the general rule that the transpose
of a product is the product of the transposes, but in the opposite order.
Combining these two equations:
� ixˆ = yˆ T iK ′iy = xˆ T iUT iK ′iUixˆ
x̂T iK
⇒K � = UT iK ′iU
The last equation is a common problem encountered in linear algebra: the
� ) and place it in diagonal form (the right
quest to take a given matrix ( K
hand side). For a symmetric matrix like K � the solution to this problem is
well known: the diagonal entries of K ′ are the eigenvalues of K � and the
columns of the transformation matrix U are the eigenvectors of K � . The
transformed variables y are called the normal modes. These modes are
linear combinations of the local degrees of freedom R1, R2, R3, … that we
started out with. Thus, while the initial motions might correspond clearly to
local stretching of one bond or bending of an angle, the normal modes will
generally be complicated mixtures of different molecular motions. We can
visualize this in the simple case of two degrees of freedom. The local modes
R1, R2 can be thought of as the two orthogonal axes in a plane. Meanwhile,
the normal modes y1, y2 are also orthogonal axes, but rotated from the
original set:
R2 R2
y2 Normal modes
Local modes
y1
R1 R1
The local modes have interactions between each other: local stretches are
coupled to local bends, etc. As a result, the Hamiltonian is not separable in
terms of the local modes R1, R2. However, by design the Hamiltonian is
separable when written in terms of the normal modes:
p̂i 2 1
Ĥ = ∑ + kii′ ŷi 2
i 2 2
’
The eigenvalues, kii , tell us the “stiffness” of the PES in the particular
direction yi. Note that the Hamiltonian above is exactly the same
Hamiltonian as the one we started with. The coupling terms have simply
been rotated away by changing the coordinates. As discussed before, we can
immediately interpret the spectra of this Hamiltonian in terms of a sum of
many independent oscillators. Thus, for a polyatomic molecule within the
harmonic approximation we expect to see lines at each of the normal mode
frequencies:
ω1 ω2 ω3 ω4 ω5 Light Frequency (ω)
Absorption
Intensity
Where we have noted that the different oscillators will typically also have
different transition dipoles. (For obvious reasons, in vibrational
spectroscopy the square of the transition dipole is often called the
oscillator strength) We can, of course, combine this polyatomic picture with
the anharmonicity effects above to get a more general picture that looks
like:
2ω1 2ω2 ω5-ω1 ω2+ω3
ω1 ω2 ω3 ω4 ω5 Light Frequency (ω)
Absorption
Intensity
where we predict the existence of various hotbands and overtones for each
of the normal mode oscillators in the molecule. Note that while the
overtones always involve multiple quanta, the quanta need not come from the
same normal mode – hence we expect not only overtones at 2ω1 , but also a
combination bands at ω2 + ω3 and ω5 − ω1 . The picture above is qualitatively
correct for the IR spectrum of a single molecule. In solution, heterogeneity
leads to a smearing out and broadening of the peaks, leading to the complex
IR fingerprints we are used to.
magnetic field of magnitude (1-σ)B0 and the associated Hamiltonian and spin state energies
will become:
E± = ± 12 γ �B0 (1 − σ ) ≡ ± 12 �ω0 (1 − σ )
This is illustrated in the figure above. Note the sign of the Hamiltonian is chosen so that the
α state (spin parallel to B0) is lower in energy than the β state ( spin antiparallel to B0).
Now, in the simplest NMR experiment, we probe this system with an oscillating magnetic field
perpendicular to the static field. By convention, we take this field to be along the x axis:
Ĥ1 ( t ) = − m
ˆ iB1 ( t ) = −γ Iˆ iB1 ( t ) = −γ Iˆx Bx cos (ωt )
We use Fermi’s Golden Rule to describe the spectrum of the spin in the oscillating field. The
selection rule is:
2 2 2
V fi = ∫ φ f * miB1φi dτ = γ Bx i ∫ φ f * Iˆxφi dτ
Now, we recall that Iˆx can be written in terms of the raising and lowering operators for
angular momentum:
(
Iˆx ∝ Iˆ+ + Iˆ− )
So that:
( )
2 2
V fi ∝ γ Bx i ∫ φ f * Iˆ+ + Iˆ− φi dτ
We immediately see that the integral is nonzero only if the initial and final spin states
differ by ±1 quantum of angular momentum (i.e. ΔM = ±1 ) , because the operator must either
raise or lower the eigenvalue of Iˆz . Thus, there are two possible transitions: α→β and β→α.
Futher, the energy conservation rule tells us that these transitions will only occur when the
photon energy matches the energy gap between the two states. As a result, we can
immediately draw the spectrum of a single shielded spin:
Intensity
This is perhaps not all that shocking: there is only one transition here, and so we might have
guessed that the spectrum would involve the frequency of that transition. However, we note
two generalizations of this result. First, we note that if we had chosen to apply the
oscillating field parallel to the static field, we would not have generated any transitions; we
only changed the spin state because we could decompose the xoscillating field into raising
and lowering operators. If the field was zoscillating, then we would have had
2 2
V fi ∝ γ Bx i ∫ φ f * Iˆzφi dτ
which is only nonzero for the trivial α→α and β→β transitions. Second, we note that if the
spin was bigger than ½ (e.g. a spin3/2 nucleus) then our selection rule above would be
precisely the same. Thus, we would have allowed transitions − 23 ↔ − 12 , − 12 ↔ + 12 and
+ 12 ↔ + 32 and all of these transitions would occur at the same frequency. Thus, spin3/2
transitions like − 23 ↔ + 12 or − 32 ↔ + 32 are strictly forbidden.
Now, as noted above, depending on their environment, different protons will be shielded
differently, resulting in a spectrum that will look qualitatively like:
Intensity
We note that the transition moment above is independent of the chemical environment: it
does not depend on shielding or any other property of the molecule. Thus, the area under an
NMR peak is strictly proportional to the number of spins that have transitions at that
frequency. This stands in contrast to IR spectroscopy, where the intensity of each oscillator
depended on the character of the oscillator, the initial state ….
ωο
E↑↓ = E1↑ + E2↓ = − [(1 − σ 1 ) − (1 − σ 2 )]
ωο
E↑↑ = E1↑ + E2↑ = − [(1 − σ 1 ) + (1 − σ 2 )]
where we have assumed for simplicity that σ1< σ2 so that E↓↑ > E↑↓ . Now, the selection rule is
the of the same form as for a single spin, but Ix decomposes into a sum of Ix for spin 1 and an
Ix for spin 2:
( )
2 2 2
V fi = γ Bx i ∫ φ f * Iˆxφi dτ = γ Bx i ∫ φ f * Iˆ1x + Iˆ2 x φi dτ
( )
2
∝ ∫ φ f * Iˆ1+ + Iˆ1− + Iˆ2+ + Iˆ2− φi dτ
The remaining integral is only nonzero if ΔM 1 = ±1 or ΔM 2 = ±1 , because the operators must
raise or lower the spin state of either spin 1 or spin 2 (but not both). If we wanted to change
both spins, we would need an operator like Î1+ Î 2− , which would allow us to raise 1 while also
lowering 2. Since we do not have any of these cross terms, we conclude only one or the other
spin can flip in an allowed transition – any twospin transitions are forbidden.
Combining these results for two uncoupled spins, we obtain the picture at left. We note that
the ↑↑↔↓↓ and ↓↑↔↑↓ transitions are forbidden, since they require flipping both spins
simultaneously. For the allowed transitions, we can easily work out
E↓↓
the energies:
ΔM2=±1 ω ω
ΔM1=±1 E↓↓ − E↑↓ = + ο [(1 − σ 1 ) + (1 − σ 2 )] + ο [(1 − σ 1 ) − (1 − σ 2 )] = ωο (1 − σ 1 )
2 2
E↓↑ ωο ωο
E↑↓
E↓↑ − E↑↑ = + [(1 − σ 1 ) − (1 − σ 2 )] + [(1 − σ 1 ) + (1 − σ 2 )] = ωο (1 − σ 1 )
2 2
ΔM1=±1 ωο ωο
ΔM2=±1 E↓↓ − E↓↑ = + [(1 − σ 1 ) + (1 − σ 2 )] − [(1 − σ 1 ) − (1 − σ 2 )] = ωο (1 − σ 2 )
2 2
ωο ωο
E↑↑ E↑↓ − E↑↑ = − [(1 − σ 1 ) − (1 − σ 2 )] +
[(1 − σ 1 ) + (1 − σ 2 )] = ωο (1 − σ 2 )
2 2
So we have only two transition energies, corresponding to each of the isolated transitions,
just as predicted above:
Intensity
Where we note that there are actually two degenerate transitions contributing to each line.
⎛ J⎞
= ⎜ E1↓ +E2↓ + ⎟ φ↓↓
⎝ 4 ⎠
E↑↑ E↑↑
Uncoupled J Coupling
Where we have noted that states where the spins are parallel shift upward in energy and
those where the spins are antiparallel shift down, and we have exaggerated the magnitude of
the shift for visual effect. Note that the selection rules do not change, because the states
have not changed – only the energies are different with the coupling on. The energies of the
allowed transitions are:
ω ⎡ J⎤ ω ⎡ J⎤ J
E↓↓ − E↑↓ = + ο ⎢(1 − σ 1 ) + (1 − σ 2 ) + ⎥ − ο ⎢ −(1 − σ 1 ) + (1 − σ 2 ) − ⎥ = ωο (1 − σ 1 ) +
2 ⎣ 4⎦ 2 ⎣ 4⎦ 2
Intensity
J J
where here we have noted the physical fact that J is typically much smaller than the
difference in chemical shielding σ between distinct protons. Thus, we see that the splitting
of NMR peaks is mediated by the coupling between the nuclear spins. This coupling is
typically mediated via the electrons – nucleus 1 pushes on the electrons, which are delocalized
and in turn push on nucleus 2. While one can routinely compute these couplings via DFT or
HF, it is much more common to use empirical rules to determine which protons will be coupled
and how large we expect the coupling to be. We should note that the magnitude of the J
splitting is independent of the magnetic field strength. Meanwhile, the Larmor frequency
increases with increasing B0. Thus, in a strong enough magnet, the peaks with shielding near σ1
will be very far from those with shielding σ2.
cα ( t ) = e 2
cα ( 0 ) cβ ( t ) = e 2
cβ ( 0 )
where we will assume for simplicity that the initial values, cα ( 0 ) , cβ ( 0 ) are real. Thus, the
magnitude of each coefficient is constant with time; we only acquire a phase factor for each
coefficient. However, these coefficients completely describe the time evolution of an
arbitrary spin state in the static magnetic field.
Now that we have solved for the coefficients of the time dependent wavefunction, let’s look
at some interesting properties of the system. First, let’s compute the zcomponent of the
spin:
⎛� ⎞
⎜ 2 0 ⎟ ⎛ cα ( t ) ⎞ �
Î z ( t ) = ( cα ( t ) * cβ ( t ) *) ⎜
�
⎟⎟ = ⎡⎢ cα ( t ) − cβ ( t ) ⎤⎥ = ⎡⎢ cα ( 0 ) − cβ ( 0 ) ⎤⎥
2 2 2 2
⎟ ⎜⎜
⎜ 0 − � ⎟ ⎝ cβ ( t ) ⎠ 2 ⎣ ⎦ 2⎣ ⎦
⎜ ⎟
⎝ 2⎠
Thus, the zcomponent of the spin does not change with time! This is perhaps a bit
surprising. We continue to compute the x and y components:
⎛ �⎞
⎜ 0 2 ⎟ ⎛ cα ( t ) ⎞ �
Î x ( t ) = ( cα ( t ) * cβ ( t ) *) ⎜ ⎟ ⎜⎜ ⎟⎟ = ⎡c ⎣ α ( t ) * cβ ( t ) + cβ ( t ) * cα ( t )⎤⎦
⎜ � 0 ⎟ ⎝ cβ ( t ) ⎠ 2
⎜ ⎟
⎝2 ⎠
�
= cα ( 0 ) cβ ( 0 ) ⎡⎣ e−iω0 t + e+iω0 t ⎦⎤ = �cα ( 0 ) cβ ( 0 ) cos ω� 0 t
� �
2
Thus, the x and y components oscillate with time at the shielded Larmor frequency ω� 0 . It is
convenient to define a magnetization vector that contains these three expectation values:
( )
�
M ( t ) ≡ Iˆx ( t ) Î y ( t ) Î z ( t )
It is fairly easy to see that the magnetization is precessing about the magnetic field: the
projection onto the magnetic field axis is constant, while the perpendicular motion is tracing
out a circular path. This is precisely the behavior one would expect from a classical magnetic
moment in a magnetic field. In this case, the magnetic field would exert a torque on the
magnetic moment according to:
�
dM ( t ) � �
= M ( t ) × γ B eff
dt
where we note that the magnetic moment feels the shielded magnetic field Beff. This gives us
three differential equations for the components of the magnetization, called Bloch
Equations:
dM x ( t )
= γ ⎡⎣ M y ( t ) Bzeff − M z ( t ) Byeff ⎤⎦ = γ M y ( t ) Bzeff
dt
dM y ( t )
= −γ ⎡⎣ M x ( t ) Bzeff − M z ( t ) Bxeff ⎤⎦ = −γ M x ( t ) Bzeff
dt
dM z ( t )
= γ ⎡⎣ M x ( t ) Byeff − M y ( t ) Bxeff ⎦⎤ = 0
dt
Where we have noted that only the zcomponent of the magnetic field is nonzero. Further,
it is easy to see by substitution that our quantum mechanical predictions for Î x (t ) and
Î y (t ) satisfy the equations above for M x ( t ) and M y ( t ) , respectively (try it and see). Thus,
we find that the quantum evolution of the average spin exactly follows the classical
equations of motion! We find comfort in this conclusion, because it is usually much easier to
think in terms of classical properties whenever possible, giving us a very nice semiclassical
way of interpreting NMR.
This rather surprising result turns out to be true for a single spin evolving in an arbitrary
time depenedent magnetic field Beff(t). To prove this, we have to use Ehrenfest’s theorem,
which states that for an arbitrary operator O, the time dependent average value of O
satisfies
d i
Ô ( t ) = ⎡ Hˆ , Oˆ ⎤ ( t )
dt � ⎣ ⎦
What does this gain us? Well, with this result in hand it is relatively easy to derive the
correct differential equations for our favorite time dependent magnetic field:
�
B eff ( t ) = − Bzeff − Bx cos (ω t )
This is the magnetic field we apply in an NMR experiment and being able to visualize the
dynamics will help us understand how the experiment works. It is relatively straightforward
to work out the associated Bloch Equations for this magnetic field. They are:
dM x ( t )
= γ M y ( t ) Bzeff
dt
dM y ( t )
= −γ ⎡⎣ M x ( t ) Bzeff + M z ( t ) Bx cos (ω t )⎤⎦
dt
dM z ( t )
= −γ M y ( t ) Bx cos (ωt )
dt
These equations can actually be solved analytically to obtain the magnetization as a function
of time. From these equations we obtain the picture below:
ω � ω0 ω = ω0 ω � ω0
Here, we are plotting the magnetization as a function of time for various choices of the
frequency of the oscillating magnetic field component. If our field oscillates too quickly
(first case) then the magnetization just sees the average field and noting interesting
happens – we just get precession about the average field. If the oscillating field is too slow,
the magnetization oscillates around the instantaneous field and we get a sort of hulahoop
motion of the magnetization. However, if we hit the frequency just right (middle) we can get
the magnetization to invert – to go from “up” to “down”. Thus, we see that the absorption
condition in NMR is associated with flipping the magnetization of the system.
Now, we note that at resonance, with the field on continuously, the spin will actually flip from
“up” to “down” and back to “up” and back to “down”… as a function of time. It is this
oscillation that shows up in our NMR spectrum. However, it is possible to turn the
oscillating field on and off as a function of time. Thus, for example, if we kept the field on
for exactly π/ω0 then the system would only have time to flip one time – all the up spins would
be converted to down and vice versa. Such a pulsed magnetic field is called an inversion
pulse, for obvious reasons. Meanwhile, if we kept the field on for exactly π/2ω0 we could
drive all the magnetization into the xy plane. This is called a π/2 pulse. Further, we note
these pulses only work if we are on resonance with a particular proton’s Larmor frequency;
from the above figure it is clear that if we are off resonance, we can’t get the spins to flip.
Thus, one can imagine fairly complex sequences of inversion pulses and π/2 pulses applied at
various frequencies being used to isolate different couplings within a complicated molecule
(like a protein). Thus, it should not be surprising that cutting edge NMR experiments are all
timeresolved in order to extract the maximum information from the molecule.