Surveys
and
Monographs
Volume 72
Jacobi Operators
and
Completely Integrable
Nonlinear Lattices
Gerald Teschl
University of Vienna
Note: The AMS has granted the permission
to post this online edition! This version is for
personal use only! If you like this book and
want to support the idea of online versions,
please consider buying this book:
http://www.ams.org/bookstoregetitem?
item=surv72
American Mathematical Society
Editorial Board
Georgia M. Benkart Michael Loss
Peter Landweber Tudor Stefan Ratiu, Chair
1991 Mathematics subject classication. 39Axx, 47B39, 58F07
Abstract. This book is intended to serve both as an introduction and a reference to
spectral and inverse spectral theory of Jacobi operators (i.e., second order symmetric
dierence operators) and applications of these theories to the Toda and Kacvan Moerbeke
hierarchy.
Starting from second order dierence equations we move on to selfadjoint operators
and develop discrete WeylTitchmarshKodaira theory, covering all classical aspects like
Weyl mfunctions, spectral functions, the moment problem, inverse spectral theory, and
uniqueness results. Next, we investigate some more advanced topics like locating the
essential, absolutely continuous, and discrete spectrum, subordinacy, oscillation theory,
trace formulas, random operators, almost periodic operators, (quasi)periodic operators,
scattering theory, and spectral deformations.
Then, the Lax approach is used to introduce the Toda hierarchy and its modied
counterpart, the Kacvan Moerbeke hierarchy. Uniqueness and existence theorems for the
initial value problem, solutions in terms of Riemann theta functions, the inverse scattering
transform, B acklund transformations, and soliton solutions are discussed.
Corrections and complements to this book are available from:
http://www.mat.univie.ac.at/gerald/ftp/bookjac/
Typeset by A
M
SL
A
T
E
X, PicT
E
X, and Makeindex. Version: November 7, 2012.
Library of Congress CataloginginPublication Data
Teschl, Gerald, 1970
Jacobi Operators and Completely Integrable Nonlinear Lattices / Gerald Teschl.
p. cm. (Mathematical surveys and monographs, ISSN 00765376 ; V. 72)
Includes bibliographical references and index.
ISBN 0821819402
1. Jacobi operators. 2. Dierentiable dynamical systems. I. Title. II. Series: Math
emtical surveys and monographs ; no. 72.
QA329.2.T47 1999 9939165
515.7242dc21 CIP
Copying and reprinting. Individual readers of this publication, and nonprot libraries
acting for them, are permitted to make fair use of the material, such as to copy a chapter
for use in teaching or research. Permission is granted to quote brief passages from this
publication in reviews, provided the customary acknowledgement of the source is given.
Republication, systematic copying, or multiple reproduction of any material in this
publication (including abstracts) is permitted only under license from the American Math
ematical Society. Requests for such permissions should be addressed to the Assistant to
the Publisher, American Mathematical Society, P.O. Box 6248, Providence, Rhode Island
029406248. Requests can also be made by email to reprintpermission@ams.org.
c 2000 by the American Mathematical Society. All rights reserved.
The American Mathematical Society retains all rights
except those granted too the United States Government.
To Susanne
Contents
Preface xiii
Part 1. Jacobi Operators
Chapter 1. Jacobi operators 3
1.1. General properties 3
1.2. Jacobi operators 13
1.3. A simple example 19
1.4. General second order dierence expressions 20
1.5. The innite harmonic crystal in one dimension 22
Chapter 2. Foundations of spectral theory for Jacobi operators 27
2.1. Weyl mfunctions 27
2.2. Properties of solutions 29
2.3. Positive solutions 34
2.4. Weyl circles 37
2.5. Canonical forms of Jacobi operators and the moment problem 40
2.6. Some remarks on unbounded operators 47
2.7. Inverse spectral theory 54
Chapter 3. Qualitative theory of spectra 59
3.1. Locating the spectrum and spectral multiplicities 59
3.2. Locating the essential spectrum 62
3.3. Locating the absolutely continuous spectrum 65
3.4. A bound on the number of eigenvalues 72
Chapter 4. Oscillation theory 75
4.1. Pr ufer variables and Sturms separation theorem 75
ix
x Contents
4.2. Classical oscillation theory 80
4.3. Renormalized oscillation theory 83
Chapter 5. Random Jacobi operators 87
5.1. Random Jacobi operators 87
5.2. The Lyapunov exponent and the density of states 91
5.3. Almost periodic Jacobi operators 100
Chapter 6. Trace formulas 105
6.1. Asymptotic expansions 105
6.2. General trace formulas and xi functions 109
Chapter 7. Jacobi operators with periodic coecients 115
7.1. Floquet theory 115
7.2. Connections with the spectra of nite Jacobi operators 119
7.3. Polynomial identities 123
7.4. Two examples: period one and two 124
7.5. Perturbations of periodic operators 126
Chapter 8. Reectionless Jacobi operators 133
8.1. Spectral analysis and trace formulas 133
8.2. Isospectral operators 140
8.3. The nitegap case 142
8.4. Further spectral interpretation 150
Chapter 9. Quasiperiodic Jacobi operators and Riemann theta functions 153
9.1. Riemann surfaces 153
9.2. Solutions in terms of theta functions 155
9.3. The elliptic case, genus one 163
9.4. Some illustrations of the RiemannRoch theorem 165
Chapter 10. Scattering theory 167
10.1. Transformation operators 167
10.2. The scattering matrix 171
10.3. The GelfandLevitanMarchenko equations 175
10.4. Inverse scattering theory 180
Chapter 11. Spectral deformations Commutation methods 187
11.1. Commuting rst order dierence expressions 187
11.2. The single commutation method 189
11.3. Iteration of the single commutation method 193
11.4. Application of the single commutation method 196
11.5. A formal second commutation 198
Contents xi
11.6. The double commutation method 200
11.7. Iteration of the double commutation method 206
11.8. The Dirichlet deformation method 209
Notes on literature 217
Part 2. Completely Integrable Nonlinear Lattices
Chapter 12. The Toda system 223
12.1. The Toda lattice 223
12.2. Lax pairs, the Toda hierarchy, and hyperelliptic curves 226
12.3. Stationary solutions 233
12.4. Time evolution of associated quantities 236
Chapter 13. The initial value problem for the Toda system 241
13.1. Finitegap solutions of the Toda hierarchy 241
13.2. Quasiperiodic nitegap solutions and the timedependent
BakerAkhiezer function 247
13.3. A simple example continued 250
13.4. Inverse scattering transform 251
13.5. Some additions in case of the Toda lattice 254
13.6. The elliptic case continued 256
Chapter 14. The Kacvan Moerbeke system 257
14.1. The Kacvan Moerbeke hierarchy and its relation to the Toda
hierarchy 257
14.2. Kac and van Moerbekes original equations 263
14.3. Spectral theory for supersymmetric Diractype dierence operators 264
14.4. Associated solutions 265
14.5. Nsoliton solutions on arbitrary background 267
Notes on literature 273
Appendix A. Compact Riemann surfaces a review 275
A.1. Basic notation 275
A.2. Abelian dierentials 277
A.3. Divisors and the RiemannRoch theorem 279
A.4. Jacobian variety and Abels map 283
A.5. Riemanns theta function 286
A.6. The zeros of the Riemann theta function 288
A.7. Hyperelliptic Riemann surfaces 291
Appendix B. Herglotz functions 299
Appendix C. Jacobi Dierence Equations with Mathematica 309
xii Contents
C.1. The package DiEqs and rst order dierence equations 309
C.2. The package JacDEqs and Jacobi dierence equations 313
C.3. Simple properties of Jacobi dierence equations 315
C.4. Orthogonal Polynomials 319
C.5. Recursions 321
C.6. Commutation methods 322
C.7. Toda lattice 328
C.8. Kacvan Moerbeke lattice 331
Bibliography 335
Glossary of notations 347
Index 353
Preface
The name of the game
Jacobi operators appear in a variety of applications. They can be viewed as
the discrete analogue of SturmLiouville operators and their investigation has many
similarities with SturmLiouville theory. Spectral and inverse spectral theory for
Jacobi operators plays a fundamental role in the investigation of completely inte
grable nonlinear lattices, in particular the Toda lattice and its modied counterpart,
the Kacvan Moerbeke lattice.
Why I have written this book
Whereas numerous books about SturmLiouville operators have been written,
only few on Jacobi operators exist. In particular, there is currently no monograph
available which covers all basic topics (like spectral and inverse spectral theory,
scattering theory, oscillation theory and positive solutions, (quasi)periodic opera
tors, spectral deformations, etc.) typically found in textbooks on SturmLiouville
operators.
In the case of the Toda lattice a textbook by M. Toda [230] exists, but none
of the recent advances in the theory of nonlinear lattices are covered there.
Audience and prerequisites
As audience I had researchers in mind. This book can be used to get acquainted
with selected topics as well as to look up specic results. Nevertheless, no previous
knowledge on dierence equations is assumed and all results are derived in a self
contained manner. Hence the present book is accessible to graduate students as
well. Previous experience with SturmLiouville operators might be helpful but is
not necessary. Still, a solid working knowledge from other branches of mathematics
is needed. In particular, I have assumed that the reader is familiar with the theory
of (linear) selfadjoint operators in Hilbert spaces which can be found in (e.g.)
xiii
xiv Preface
[192] or [241]. This theory is heavily used in the rst part. In addition, the reader
might have to review material from complex analysis (see Appendix A and B) and
dierential equations on Banach manifolds (second part only) if (s)he feels (s)he
does not have the necessary background. However, this knowledge is mainly needed
for understanding proofs rather than the results themselves.
The style of this book
The style of this monograph is strongly inuenced by my personal bias. I
have striven to present an intuitive approach to each subject and to include the
simplest possible proof for every result. Most proofs are rather sketchy in character,
so that the main idea becomes clear instead of being drowned by technicalities.
Nevertheless, I have always tried to include enough information for the reader to
ll in the remaining details (her)himself if desired. To help researchers, using this
monograph as a reference, to quickly spot the result they are looking for, most
information is found in display style formulas.
The entire treatment is supposed to be mathematically rigorous. I have tried to
prove every statement I make and, in particular, these little obvious things, which
turn out less obvious once one tries to prove them. In this respect I had Marchenkos
monograph on SturmLiouville operators [167] and the one by Weidmann [241] on
functional analysis in mind.
Literature
The rst two chapters are of an introductory nature and collect some well
known folklore, the successive more advanced chapters are a synthesis of results
from research papers. In most cases I have rearranged the material, streamlined
proofs, and added further facts which are not published elsewhere. All results
appear without special attribution to who rst obtained them but there is a section
entitled Notes on literature in each part which contains references to the literature
plus hints for additional reading. The bibliography is selective and far from being
complete. It contains mainly references I (am aware of and which I) have actually
used in the process of writing this book.
Terminology and notation
For the most part, the terminology used agrees with generally accepted usage.
Whenever possible, I have tried to preserve original notation. Unfortunately I had
to break with this policy at various points, since I have given higher priority to
a consistent (and selfexplaining) notation throughout the entire monograph. A
glossary of notation can be found towards the end.
Contents
For convenience of the reader, I have split the material into two parts; one
on Jacobi operators and one on completely integrable lattices. In particular, the
second part is to a wide extent independent of the rst one and anybody interested
Preface xv
only in completely integrable lattices can move directly to the second part (after
browsing Chapter 1 to get acquainted with the notation).
Part I
Chapter 1 gives an introduction to the theory of second order dierence equa
tions and bounded Jacobi operators. All basic notations and properties are pre
sented here. In addition, this chapter provides several easy but extremely helpful
gadgets. We investigate the case of constant coecients and, as a motivation for
the reader, the innite harmonic crystal in one dimension is discussed.
Chapter 2 establishes the pillars of spectral and inverse spectral theory for
Jacobi operators. Here we develop what is known as discrete WeylTitchmarsh
Kodaira theory. Basic things like eigenfunction expansions, connections with the
moment problem, and important properties of solutions of the Jacobi equation are
shown in this chapter.
Chapter 3 considers qualitative theory of spectra. It is shown how the essen
tial, absolutely continuous, and point spectrum of specic Jacobi operators can
be located in some cases. The connection between existence of subordinate solu
tions and continuity of spectral measures is discussed. In addition, we investigate
under which conditions the number of discrete eigenvalues is nite.
Chapter 4 covers discrete SturmLiouville theory. Both classical oscillation and
renormalized oscillation theory are developed.
Chapter 5 gives an introduction to the theory of random Jacobi operators. Since
there are monographs (e.g., [40]) devoted entirely to this topic, only basic results
on the spectra and some applications to almost periodic operators are presented.
Chapter 6 deals with trace formulas and asymptotic expansions which play a
fundamental role in inverse spectral theory. In some sense this can be viewed as an
application of Kreins spectral shift theory to Jacobi operators. In particular, the
tools developed here will lead to a powerful reconstruction procedure from spectral
data for reectionless (e.g., periodic) operators in Chapter 8.
Chapter 7 considers the special class of operators with periodic coecients.
This class is of particular interest in the onedimensional crystal model and sev
eral profound results are obtained using Floquet theory. In addition, the case of
impurities in onedimensional crystals (i.e., perturbation of periodic operators) is
studied.
Chapter 8 again considers a special class of Jacobi operators, namely reec
tionless ones, which exhibit an algebraic structure similar to periodic operators.
Moreover, this class will show up again in Chapter 10 as the stationary solutions
of the Toda equations.
Chapter 9 shows how reectionless operators with no eigenvalues (which turn
out to be associated with quasiperiodic coecients) can be expressed in terms of
Riemann theta functions. These results will be used in Chapter 13 to compute
explicit formulas for solutions of the Toda equations in terms of Riemann theta
functions.
xvi Preface
Chapter 10 provides a comprehensive treatment of (inverse) scattering the
ory for Jacobi operators with constant background. All important objects like re
ection/transmission coecients, Jost solutions and GelfandLevitanMarchenko
equations are considered. Again this applies to impurities in onedimensional crys
tals. Furthermore, this chapter forms the main ingredient of the inverse scattering
transform for the Toda equations.
Chapter 11 tries to deform the spectra of Jacobi operators in certain ways. We
compute transformations which are isospectral and such which insert a nite num
ber of eigenvalues. The standard transformations like single, double, or Dirichlet
commutation methods are developed. These transformations can be used as pow
erful tools in inverse spectral theory and they allow us to compute new solutions
from old solutions of the Toda and Kacvan Moerbeke equations in Chapter 14.
Part II
Chapter 12 is the rst chapter on integrable lattices and introduces the Toda
system as hierarchy of evolution equations associated with the Jacobi operator via
the standard Lax approach. Moreover, the basic (global) existence and uniqueness
theorem for solutions of the initial value problem is proven. Finally, the stationary
hierarchy is investigated and the BurchnallChaundy polynomial computed.
Chapter 13 studies various aspects of the initial value problem. Explicit for
mulas in case of reectionless (e.g., (quasi)periodic) initial conditions are given in
terms of polynomials and Riemann theta functions. Moreover, the inverse scatter
ing transform is established.
The nal Chapter 14 introduces the Kac vanMoerbeke hierarchy as modied
counterpart of the Toda hierarchy. Again the Lax approach is used to establish
the basic (global) existence and uniqueness theorem for solutions of the initial
value problem. Finally, its connection with the Toda hierarchy via a Miuratype
transformation is studied and used to compute Nsoliton solutions on arbitrary
background.
Appendix
Appendix A reviews the theory of Riemann surfaces as needed in this mono
graph. While most of us will know Riemann surfaces from a basic course on complex
analysis or algebraic geometry, this will be mainly from an abstract viewpoint like
in [86] or [129], respectively. Here we will need a more computational approach
and I hope that the reader can extract this knowledge from Appendix A.
Appendix B compiles some relevant results from the theory of Herglotz func
tions and Borel measures. Since not everybody is familiar with them, they are
included for easy reference.
Appendix C shows how a program for symbolic computation, Mathematica
,
can be used to do some of the computations encountered during the main bulk.
While I dont believe that programs for symbolic computations are an indispens
able tool for doing research on Jacobi operators (or completely integrable lattices),
they are at least useful for checking formulas. Further information and Mathe
matica
p
(I, M) = f (I, M)[
nI
[f(n)[
p
< , 1 p < ,
nI
[f(n)[
p
_
1/p
, 1 p < , f
= sup
nI
[f(n)[,
makes
p
(I, M), 1 p , a Banach space as well.
3
4 1. Jacobi operators
Furthermore,
0
(I, M) denotes the set of sequences with only nitely many
values being nonzero and
p
nI
f(n), g(n)
M
,
f = f
2
=
_
f, f, f, g
2
(I, M). (1.4)
For what follows we will choose I = Z for simplicity. However, straightforward
modications can be made to accommodate the general case I Z.
During most of our investigations we will be concerned with dierence expres
sions, that is, endomorphisms of (Z);
(1.5)
R : (Z) (Z)
f Rf
(we reserve the name dierence operator for dierence expressions dened on a
subset of
2
(Z)). Any dierence expression R is uniquely determined by its corre
sponding matrix representation
(1.6) (R(m, n))
m,nZ
, R(m, n) = (R
n
)(m) =
m
, R
n
,
where
(1.7)
n
(m) =
m,n
=
_
0 n ,= m
1 n = m
is the canonical basis of (Z). The order of R is the smallest nonnegative integer
N = N
+
+N
.
If no such number exists, the order is innite.
We call R symmetric (resp. skewsymmetric) if R(m, n) = R(n, m) (resp.
R(m, n) = R(n, m)).
Maybe the simplest examples for a dierence expression are the shift expres
sions
(1.8) (S
kZ
R(., . +k)(S
+
)
k
, (S
)
1
= S
.
Here R(., . +k) denotes the multiplication expression with the sequence (R(n, n +
k))
nZ
, that is, R(., . +k) : f(n) R(n, n +k)f(n). In order to simplify notation
we agree to use the short cuts
(1.10) f
= S
f, f
++
= S
+
S
+
f, etc.,
1.1. General properties 5
whenever convenient. In connection with the dierence expression (1.5) we also
dene the diagonal, upper, and lower triangular parts of R as follows
(1.11) [R]
0
= R(., .), [R]
kN
R(., . k)(S
)
k
,
implying R = [R]
+
+ [R]
0
+ [R]
.
Having these preparations out of the way, we are ready to start our investigation
of second order symmetric dierence expressions. To set the stage, let a, b (Z, R)
be two realvalued sequences satisfying
(1.12) a(n) R0, b(n) R, n Z,
and introduce the corresponding second order, symmetric dierence expres
sion
(1.13)
: (Z) (Z)
f(n) a(n)f(n + 1) +a(n 1)f(n 1) +b(n)f(n)
.
It is associated with the tridiagonal matrix
(1.14)
_
_
_
_
_
_
_
_
.
.
.
.
.
.
.
.
.
a(n 2) b(n 1) a(n 1)
a(n 1) b(n) a(n)
a(n) b(n + 1) a(n + 1)
.
.
.
.
.
.
.
.
.
_
_
_
_
_
_
_
_
and will be our main object for the rest of this section and the tools derived here
even though simple will be indispensable for us.
Before going any further, I want to point out that there is a close connection
between second order, symmetric dierence expressions and second order, symmet
ric dierential expressions. This connection becomes more apparent if we use the
dierence expressions
(1.15) (f)(n) = f(n + 1) f(n), (
fg)(n) = f(n)(
f)(n) (1.17)
and the summation by parts formula (also known as Abel transform)
(1.18)
n
j=m
g(j)(f)(j) = g(n)f(n + 1) g(m1)f(m) +
n
j=m
(
g)(j)f(j).
6 1. Jacobi operators
Both are readily veried. Nevertheless, let me remark that ,
are no derivations
since they do not satisfy Leibnitz rule. This very often makes the discrete case
dierent (and sometimes also harder) from the continuous one. In particular, many
calculations become much messier and formulas longer.
There is much more to say about relations for the dierence expressions (1.15)
analogous to the ones for dierentiation. We refer the reader to, for instance, [4],
[87], or [147] and return to (1.13).
Associated with is the eigenvalue problem u = zu. The appropriate setting
for this eigenvalue problem is the Hilbert space
2
(Z). However, before we can
pursue the investigation of the eigenvalue problem in
2
(Z), we need to consider
the Jacobi dierence equation
(1.19) u = z u, u (Z), z C.
Using a(n) ,= 0 we see that a solution u is uniquely determined by the values u(n
0
)
and u(n
0
+1) at two consecutive points n
0
, n
0
+1 (you have to work much harder
to obtain the corresponding result for dierential equations). It follows, that there
are exactly two linearly independent solutions.
Combining (1.16) and the summation by parts formula yields Greens formula
(1.20)
n
j=m
_
f(g) (f)g
_
(j) = W
n
(f, g) W
m1
(f, g)
for f, g (Z), where we have introduced the (modied) Wronskian
(1.21) W
n
(f, g) = a(n)
_
f(n)g(n + 1) g(n)f(n + 1)
_
.
Greens formula will be the key to selfadjointness of the operator associated with
in the Hilbert space
2
(Z) (cf. Theorem 1.5) and the Wronskian is much more
than a suitable abbreviation as we will show next.
Evaluating (1.20) in the special case where f and g both solve (1.19) (with the
same parameter z) shows that the Wronskian is constant (i.e., does not depend on
n) in this case. (The index n will be omitted in this case.) Moreover, it is nonzero
if and only if f and g are linearly independent.
Since the (linear) space of solutions is two dimensional (as observed above) we
can pick two linearly independent solutions c, s of (1.19) and write any solution u
of (1.19) as a linear combination of these two solutions
(1.22) u(n) =
W(u, s)
W(c, s)
c(n)
W(u, c)
W(c, s)
s(n).
For this purpose it is convenient to introduce the following fundamental solutions
c, s (Z)
(1.23) c(z, ., n
0
) = z c(z, ., n
0
), s(z, ., n
0
) = z s(z, ., n
0
),
fullling the initial conditions
(1.24)
c(z, n
0
, n
0
) = 1, c(z, n
0
+ 1, n
0
) = 0,
s(z, n
0
, n
0
) = 0, s(z, n
0
+ 1, n
0
) = 1.
Most of the time the base point n
0
will be unessential and we will choose n
0
= 0
for simplicity. In particular, we agree to omit n
0
whenever it is 0, that is,
(1.25) c(z, n) = c(z, n, 0), s(z, n) = s(z, n, 0).
1.1. General properties 7
Since the Wronskian of c(z, ., n
0
) and s(z, ., n
0
) does not depend on n we can eval
uate it at n
0
(1.26) W(c(z, ., n
0
), s(z, ., n
0
)) = a(n
0
)
and consequently equation (1.22) simplies to
(1.27) u(n) = u(n
0
)c(z, n, n
0
) +u(n
0
+ 1)s(z, n, n
0
).
Sometimes a lot of things get more transparent if (1.19) is regarded from the
viewpoint of dynamical systems. If we introduce u = (u, u
+
) (Z, C
2
), then (1.19)
is equivalent to
(1.28) u(n + 1) = U(z, n + 1)u(n), u(n 1) = U(z, n)
1
u(n),
where U(z, .) is given by
U(z, n) =
1
a(n)
_
0 a(n)
a(n 1) z b(n)
_
,
U
1
(z, n) =
1
a(n 1)
_
z b(n) a(n)
a(n 1) 0
_
. (1.29)
The matrix U(z, n) is often referred to as transfer matrix. The corresponding
(nonautonomous) ow on (Z, C
2
) is given by the fundamental matrix
(z, n, n
0
) =
_
c(z, n, n
0
) s(z, n, n
0
)
c(z, n + 1, n
0
) s(z, n + 1, n
0
)
_
=
_
_
_
U(z, n) U(z, n
0
+ 1) n > n
0
1l n = n
0
U
1
(z, n + 1) U
1
(z, n
0
) n < n
0
. (1.30)
More explicitly, equation (1.27) is now equivalent to
(1.31)
_
u(n)
u(n + 1)
_
= (z, n, n
0
)
_
u(n
0
)
u(n
0
+ 1)
_
.
Using (1.31) we learn that (z, n, n
0
) satises the usual group law
(1.32) (z, n, n
0
) = (z, n, n
1
)(z, n
1
, n
0
), (z, n
0
, n
0
) = 1l
and constancy of the Wronskian (1.26) implies
(1.33) det (z, n, n
0
) =
a(n
0
)
a(n)
.
Let us use (z, n) = (z, n, 0) and dene the upper, lower Lyapunov exponents
(z) = limsup
n
1
[n[
ln (z, n, n
0
) = limsup
n
1
[n[
ln (z, n),
(z) = liminf
n
1
[n[
ln (z, n, n
0
) = liminf
n
1
[n[
ln (z, n). (1.34)
Here
(1.35)  = sup
uC
2
\0]
u
C
2
u
C
2
denotes the operator norm of . By virtue of (use (1.32))
(1.36) (z, n
0
)
1
(z, n) (z, n, n
0
) (z, n
0
)
1
(z, n)
8 1. Jacobi operators
the denition of
(z),
(z) 0 if
a(n) is bounded. In fact, since
(z) =
() =
().
Lemma 1.1. Suppose that [a(n)[ does not grow or decrease exponentially and that
[b(n)[ does not grow exponentially, that is,
(1.37) lim
n
1
[n[
ln [a(n)[ = 0, lim
n
1
[n[
ln(1 +[b(n)[) = 0.
If Hyp
() R
2
such that
v V
() lim
n
1
[n[
ln (, n)v =
(),
v , V
() lim
n
1
[n[
ln (, n)v =
(), (1.38)
respectively.
Proof. Set
(1.39) A(n) =
_
1 0
0 a(n)
_
and abbreviate
U(z, n) = 0
due to our assumption (1.37). Moreover,
(1.43)
min(1, a(n))
max(1, a(0))

(z, n)
(z, n)
max(1, a(n))
min(1, a(0))
and hence lim
n
[n[
1
ln 
(), v ,
V
() it is not hard to see that the lemma becomes false if a(n) is exponentially
decreasing.
For later use observe that
(1.44) (z, n, m)
1
=
a(n)
a(m)
J (z, n, m)
J
1
, J =
_
0 1
1 0
_
,
(where
m=n0+1
K(z, n, m)g(m), (1.48)
where
(1.49)
n1
j=n0
f(j) =
_
_
n1
j=n0
f(j) for n > n
0
0 for n = n
0
n01
j=n
f(j) for n < n
0
,
satises (1.46) and the initial conditions f(n
0
) = f
0
, f(n
0
+ 1) = f
1
as can be
checked directly. The summation kernel K(z, n, m) has the following properties:
K(z, n, n) = 0, K(z, n + 1, n) = a(n)
1
, K(z, n, m) = K(z, m, n), and
(1.50) K(z, n, m) =
u(z, m)v(z, n) u(z, n)v(z, m)
W(u(z), v(z))
for any pair u(z), v(z) of linearly independent solutions of u = zu.
Another useful result is the variation of constants formula. It says that if
one solution u of (1.19) with u(n) ,= 0 for all n Z is known, then a second (linearly
independent, W(u, v) = 1) solution of (1.19) is given by
(1.51) v(n) = u(n)
n1
j=n0
1
a(j)u(j)u(j + 1)
.
10 1. Jacobi operators
It can be veried directly as well.
Sometimes transformations can help to simplify a problem. The following two
are of particular interest to us. If u fullls (1.19) and u(n) ,= 0, then the sequence
(n) = u(n + 1)/u(n) satises the (discrete) Riccati equation
(1.52) a(n)(n) +
a(n 1)
(n 1)
= z b(n).
Conversely, if fullls (1.52), then the sequence
(1.53) u(n) =
n1
j=n0
(j) =
_
_
n1
j=n0
(j) for n > n
0
1 for n = n
0
n01
j=n
(j)
1
for n < n
0
,
fullls (1.19) and is normalized such that u(n
0
) = 1. In addition, we remark that
the sequence (n) might be written as nite continued fraction,
(1.54) a(n)(n) = z b(n)
a(n 1)
2
z b(n 1)
.
.
.
a(n
0
+ 1)
2
z b(n
0
+ 1)
a(n
0
)
(n
0
)
for n > n
0
and
(1.55) a(n)(n) =
a(n)
2
z b(n + 1)
.
.
.
a(n
0
1)
2
z b(n
0
) a(n
0
)(n
0
)
for n < n
0
.
If a is a sequence with a(n) ,= 0 and u fullls (1.19), then the sequence
(1.56) u(n) = u(n)
n1
j=n0
a(j),
fullls
(1.57)
a(n)
a(n)
u(n + 1) +a(n 1) a(n 1) u(n 1) +b(n) u(n) = z u(n).
Especially, taking a(n) = sgn(a(n)) (resp. a(n) = sgn(a(n))), we see that it is no
restriction to assume a(n) > 0 (resp. a(n) < 0) (compare also Lemma 1.6 below).
We conclude this section with a detailed investigation of the fundamental solu
tions c(z, n, n
0
) and s(z, n, n
0
). To begin with, we note (use induction) that both
c(z, n k, n) and s(z, n k, n), k 0, are polynomials of degree at most k with
respect to z. Hence we may set
(1.58) s(z, n k, n) =
k
j=0
s
j,k
(n)z
j
, c(z, n k, n) =
k
j=0
c
j,k
(n)z
j
.
1.1. General properties 11
Using the coecients s
j,k
(n) and c
j,k
(n) we can derive a neat expansion for
arbitrary dierence expressions. By (1.9) it suces to consider (S
)
k
.
Lemma 1.2. Any dierence expression R of order at most 2k +1 can be expressed
as
(1.59) R =
k
j=0
_
c
j
+s
j
S
+
_
j
, c
j
, s
j
(Z), k N
0
,
with c
j
= s
j
= 0 if and only if R = 0. In other words, the set
j
, S
+
jN0
forms
a basis for the space of all dierence expressions.
We have
(1.60) (S
)
k
=
k
j=0
_
c
j,k
+s
j,k
S
+
_
j
,
where s
j,k
(n) and c
j,k
(n) are dened in (1.58).
Proof. We rst prove (1.59) by induction on k. The case k = 0 is trivial. Since
the matrix element
k
(n, n k) =
k1
j=0
a(n j
0
1
) ,= 0 is nonzero we can choose
s
k
(n) = R(n, n +k +1)/
k
(n 1, n +k 1), c
k
(n) = R(n, n k)/
k
(n, n k) and
apply the induction hypothesis to R(c
k
s
k
S
+
)
k
. This proves (1.59). The rest
is immediate from
(1.61) (R(s(z, ., n))(n) =
k
j=0
s
j
(n)z
j
, (R(c(z, ., n))(n) =
k
j=0
c
j
(n)z
j
.
n1
j=n0+1
a(j)
=
z
k
k
j=1
p
n0,n
(j)z
kj
k
j=1
a(n
0
+j)
,
where k = n n
0
1 0 and
(1.66) p
n0,n
(j) =
tr(J
j
n0,n
)
j1
=1
p
n0,n
()tr(J
j
n0,n
)
j
, 1 j k.
To verify the rst equation, use that if z is a zero of s(., n, n
0
), then (s(z, n
0
+
1, n
0
), . . . , s(z, n1, n
0
)) is an eigenvector of (1.64) corresponding to the eigenvalue
z. Since the converse statement is also true, the polynomials (in z) s(z, n, n
0
) and
det(z J
n0,n
) only dier by a constant which can be deduced from (1.30). The
second is a wellknown property of characteristic polynomials (cf., e.g., [91]).
The rst few traces are given by
tr(J
n0,n0+k+1
) =
n0+k
j=n0+1
b(j),
tr(J
2
n0,n0+k+1
) =
n0+k
j=n0+1
b(j)
2
+ 2
n0+k1
j=n0+1
a(j)
2
,
tr(J
3
n0,n0+k+1
) =
n0+k
j=n0+1
b(j)
3
+ 3
n0+k1
j=n0+1
a(j)
2
(b(j) +b(j + 1)),
tr(J
4
n0,n0+k+1
) =
n0+k
j=n0+1
b(j)
4
4
n0+k1
j=n0+1
a(j)
2
_
b(j)
2
+b(j + 1)b(j)
+b(j + 1)
2
+
a(j)
2
2
_
+ 4
n0+k2
j=n0+1
a(j + 1)
2
. (1.67)
1.2. Jacobi operators 13
An explicit calculation yields for n > n
0
+ 1
c(z, n, n
0
) =
a(n
0
)z
nn02
n1
j=n0+1
a(j)
_
_
1
1
z
n1
j=n0+2
b(j) +O(
1
z
2
)
_
_
,
s(z, n, n
0
) =
z
nn01
n1
j=n0+1
a(j)
_
_
1
1
z
n1
j=n0+1
b(j) +O(
1
z
2
)
_
_
, (1.68)
and (using (1.63) and (1.62)) for n < n
0
c(z, n, n
0
) =
z
n0n
n01
j=n
a(j)
_
_
1
1
z
n0
j=n+1
b(j) +O(
1
z
2
)
_
_
,
s(z, n, n
0
) =
a(n
0
)z
n0n1
n01
j=n
a(j)
_
_
1
1
z
n01
j=n+1
b(j) +O(
1
z
2
)
_
_
. (1.69)
1.2. Jacobi operators
In this section we scrutinize the eigenvalue problem associated with (1.19) in the
Hilbert space
2
(Z).
Recall that the scalar product and norm is given by
(1.70) f, g =
nZ
f(n)g(n), f =
_
f, f, f, g
2
(Z),
where the bar denotes complex conjugation.
For simplicity we assume from now on (and for the rest of this monograph)
that a,b are bounded sequences.
Hypothesis H. 1.4. Suppose
(1.71) a, b
H, b
H and
(1.73) H 2a
+b
,
where H denotes the operator norm of H.
14 1. Jacobi operators
Proof. The fact that lim
n
W
n
(f, g) = 0, f, g
2
(Z), together with Greens
formula (1.20) shows that H is selfadjoint, that is,
(1.74) f, Hg = Hf, g, f, g
2
(Z).
For the rest consider a(n)
2
+a(n 1)
2
+b(n)
2
= H
n

2
H
2
and
(1.75) [f, Hf[ (2a
+b
)f
2
.
, b
by
(1.76) a
(n) = (n)a(n), b
(n) = b(n), n Z,
and the unitary involution U
by
(1.77) U
= U
1
=
_ n1
j=0
(j)
m,n
_
m,nZ
.
Let H be a Jacobi operator associated with the dierence expression (1.13). Then
H
dened as
(1.78) H
= U
HU
1
f)(n) = a
(n)f(n + 1) +a
(n 1)f(n 1) +b
(n)f(n)
and is unitarily equivalent to H.
Proof. Straightforward.
The next transformation is equally useful and will be referred to as reection
at n
0
. It shows how information obtained near one endpoint, say +, can be
transformed into information near the other, .
1.2. Jacobi operators 15
Lemma 1.7. Fix n
0
Z and consider the unitary involution
(1.80) (U
R
f)(n) = (U
1
R
f)(n) = f(2n
0
n)
or equivalently (U
R
f)(n
0
+k) = f(n
0
k). Then the operator
(1.81) H
R
= U
R
HU
1
R
,
is associated with the sequences
(1.82) a
R
(n
0
k 1) = a(n
0
+k), b
R
(n
0
k) = b(n
0
+k), k Z,
or equivalently a
R
(n) = a(2n
0
n 1), b
R
(n) = b(2n
0
n).
Proof. Again straightforward.
Associated with U
R
are the two orthogonal projections
(1.83) P
R
=
1
2
(1l U
R
), P
R
+P
+
R
= 1l, P
R
P
+
R
= P
+
R
P
R
= 0
and a corresponding splitting of H into two parts H = H
+
R
H
R
, where
(1.84) H
R
= P
+
R
HP
R
+P
R
HP
R
=
1
2
(H H
R
).
The symmetric part H
+
R
(resp. antisymmetric part H
R
) commutes (resp. anticom
mutes) with U
R
, that is, [U
R
, H
+
R
] = U
R
H
+
R
H
+
R
U
R
= 0 (resp. U
R
, H
R
=
U
R
H
R
+H
R
U
R
= 0). If H = H
R
we enter the realm of supersymmetric quantum
mechanics (cf., e.g., [225] and Section 14.3).
After these two transformations we will say a little more about the spectrum
(H) of H. More precisely, we will estimate the location of (H).
Lemma 1.8. Let
(1.85) c
(n), sup
nZ
c
+
(n)].
Proof. We will rst show that H is semibounded from above by sup c
+
. From
(1.16) we infer
f, Hf =
nZ
_
a(n)[f(n + 1) f(n)[
2
+ (a(n 1) +a(n) +b(n))[f(n)[
2
_
. (1.87)
By Lemma 1.6 we can rst choose a(n) > 0 to obtain
(1.88) f, Hf sup
nZ
c
+
(n) f
2
.
Similarly, choosing a(n) < 0 we see that H is semibounded from below by inf c
,
(1.89) f, Hf inf
nZ
c
(n) f
2
,
completing the proof.
16 1. Jacobi operators
We remark that these bounds are optimal in the sense that equality is attained
for (e.g.) a(n) = 1/2, b(n) = 0 (cf. Section 1.3).
We will not only consider H but also restrictions of H; partly because they are
of interest on their own, partly because their investigation gives information about
H.
To begin with, we dene the following restrictions H
,n0
of H to the subspaces
2
(n
0
, ),
H
+,n0
f(n) =
_
a(n
0
+ 1)f(n
0
+ 2) +b(n
0
+ 1)f(n
0
+ 1), n = n
0
+ 1
(f)(n), n > n
0
+ 1
,
H
,n0
f(n) =
_
a(n
0
2)f(n
0
2) +b(n
0
1)f(n
0
1), n = n
0
1
(f)(n), n < n
0
1
. (1.90)
In addition, we also dene for R
H
0
+,n0
= H
+,n0+1
, H
+,n0
= H
+,n0
a(n
0
)
1
n0+1
, .
n0+1
, ,= 0,
H
,n0
= H
,n0
, H
,n0
= H
,n0+1
a(n
0
)
n0
, .
n0
, ,= . (1.91)
All operators H
,n0
are bounded and selfadjoint.
Last, we dene the following nite restriction H
n1,n2
to the subspaces
2
(n
1
, n
2
)
(1.92) H
n1,n2
f(n) =
_
_
_
a(n
1
+ 1)f(n
1
+ 2) +b(n
1
+ 1)f(n
1
+ 1), n = n
1
+ 1
(f)(n), n
1
+ 1 < n < n
2
1
a(n
2
2)f(n
2
2) +b(n
2
1)f(n
2
1), n = n
2
1
.
The operator H
n1,n2
is clearly associated with the Jacobi matrix J
n1,n2
(cf. (1.64)).
Moreover, we set H
,
n1,n2
= H
n1,n2
, H
0,2
n1,n2
= H
,2
n1+1,n2
, and
H
1,2
n1,n2
= H
,2
n1,n2
a(n
1
)
1
1
n1+1
, .
n1+1
,
1
,= 0,
H
1,2
n1,n2
= H
1,
n1,n2+1
a(n
2
)
2
n2
, .
n2
,
2
,= . (1.93)
Remark 1.9. H
+,n0
can be associated with the following domain
(1.94) D(H
+,n0
) = f
2
(n
0
, )[ cos()f(n
0
) + sin()f(n
0
+ 1) = 0,
= cot() ,= 0, if one agrees that only points with n > n
0
are of signicance and
that the last point is only added as a dummy variable so that one does not have
to specify an extra expression for (f)(n
0
+ 1). In particular, the case =
(i.e., corresponding to the boundary condition f(n
0
) = 0) will be referred to as
Dirichlet boundary condition at n
0
. Analogously for H
,n0
, H
1,2
n1,n2
.
One of the most important objects in spectral theory is the resolvent (Hz)
1
,
z (H), of H. Here (H) = C(H) denotes the resolvent set of H. The matrix
elements of (H z)
1
are called Green function
(1.95) G(z, n, m) =
n
, (H z)
1
m
, z (H).
Clearly,
(1.96) ( z)G(z, ., m) =
m
(.), G(z, m, n) = G(z, n, m)
and
(1.97) ((H z)
1
f)(n) =
mZ
G(z, n, m)f(m), f
2
(Z), z (H).
1.2. Jacobi operators 17
We will derive an explicit formula for G(z, n, m) in a moment. Before that we need
to construct solutions u
0
(.) = G(z, ., 0), z (H).
By construction u fullls (1.19) only for n > 0 and n < 0. But if we take u(z, 2),
u(z, 1) as initial condition we can obtain a solution u
(z, n) of u = zu on the
whole of (Z) which coincides with u(z, n) for n < 0. Hence u
(z) satises u
(z)
(z), u
+
(z))
_
u
+
(z, n)u
(z, m) for m n
u
+
(z, m)u
(z, n) for n m
,
z (H). Indeed, since the right hand side of (1.99) satises (1.96) and is square
summable with respect to n, it must be the Green function of H.
For later use we also introduce the convenient abbreviations
g(z, n) = G(z, n, n) =
u
+
(z, n)u
(z, n)
W(u
(z), u
+
(z))
,
h(z, n) = 2a(n)G(z, n, n + 1) 1
=
a(n)(u
+
(z, n)u
(z, n + 1) +u
+
(z, n)u
(z, n + 1))
W(u
(z), u
+
(z))
. (1.100)
Note that for n m we have
G(z, n, m) = g(z, n
0
)c(z, n, n
0
)c(z, m, n
0
)
+g(z, n
0
+ 1)s(z, n, n
0
)s(z, m, n
0
)
+h(z, n
0
)
c(z, n, n
0
)s(z, m, n
0
) +c(z, m, n
0
)s(z, n, n
0
)
2a(n
0
)
c(z, n, n
0
)s(z, m, n
0
) c(z, m, n
0
)s(z, n, n
0
)
2a(n
0
)
. (1.101)
Similar results hold for the restrictions: Let
(1.102) s
(z, n, n
0
) = sin()c(z, n, n
0
) cos()s(z, n, n
0
)
with = cot() (i.e., the sequence s
(z, n, n
0
) fullls the boundary condition
cos()s
(z, n
0
, n
0
) + sin()s
(z, n
0
+ 1, n
0
) = 0). Then we obtain for the resol
vent of H
,n0
(1.103) ((H
,n0
z)
1
u)(n) =
m
>
<
n0
G
,n0
(z, m, n)u(m), z (H
,n0
),
where
(1.104) G
,n0
(z, m, n) =
1
W(s
(z), u
(z))
_
s
(z, n, n
0
)u
(z, m) for m
n
s
(z, m, n
0
)u
(z, n) for n
m
(use (H
,n0
z)
1
to show the existence of u
(z, .) for z (H
,n0
)).
18 1. Jacobi operators
Remark 1.10. The solutions being square summable near (resp. satisfying the
boundary condition cos()f(n
0
) + sin()f(n
0
+ 1) = 0) are unique up to constant
multiples since the Wronskian of two such solutions vanishes (evaluate it at
(resp. n
0
)). This implies that the point spectrum of H, H
,n0
is always simple.
In addition to H
,n0
we will be interested in the following direct sums of these
operators
(1.105) H
n0
= H
,n0
H
+,n0
,
in the Hilbert space f
2
(Z)[ cos()f(n
0
) + sin()f(n
0
+ 1) = 0. The reason
why H
n0
is of interest to us follows from the close spectral relation to H as can be
seen from their resolvents (resp. Green functions)
G
n0
(z, n, m) = G(z, n, m)
G(z, n, n
0
)G(z, n
0
, m)
G(z, n
0
, n
0
)
,
G
n0
(z, n, m) = G(z, n, m)
(z, n
0
)
1
(G(z, n, n
0
+ 1) +G(z, n, n
0
))
(G(z, n
0
+ 1, m) +G(z, n
0
, m)), R, (1.106)
where
(z, n) =
(u
+
(z, n + 1) +u
+
(z, n))(u
(z, n + 1) +u
(z, n))
W(u
(z), u
+
(z))
= g(z, n + 1) +
a(n)
h(z, n) +
2
g(z, n). (1.107)
Remark 1.11. The operator H
n0
is equivalently given by
(1.108) H
n0
= (1l P
n0
)H(1l P
n0
)
in the Hilbert space (1l P
n0
)
2
(Z) = f
2
(Z)[
n0
, f = 0, where P
n0
denotes
the orthogonal projection onto the onedimensional subspace spanned by
n0
=
cos()
n0
+ sin()
n0+1
, = cot(), [0, ) in
2
(Z).
Finally, we derive some interesting dierence equations for g(z, n) to be used
in Section 6.1.
Lemma 1.12. Let u, v be two solutions of (1.19). Then g(n) = u(n)v(n) satises
(1.109)
(a
+
)
2
g
++
a
2
g
z b
+
+
a
2
g
+
(a
)
2
g
z b
= (z b
+
)g
+
(z b)g,
and
(1.110)
_
a
2
g
+
(a
)
2
g
+ (z b)
2
g
_
2
= (z b)
2
_
W(u, v)
2
+ 4a
2
gg
+
_
.
Proof. First we calculate (using (1.19))
(1.111) a
2
g
+
(a
)
2
g
= (z b)
2
g +a(z b)(uv
+
+u
+
v).
Adding (z b)
2
g and taking squares yields the second equation. Dividing both
sides by z b and adding the equations corresponding to n and n + 1 yields the
rst.
1.3. A simple example 19
Remark 1.13. There exists a similar equation for
)
2
(
)
+
(a
A)
2
(
+B
2
_
2
= (A
B)
2
_
(
A
a
W(u, v))
2
+ 4(a
+
)
2
)
+
_
, (1.112)
with
A = a +(z b
+
) +
2
a
+
,
B = a
(z b
+
) +((z b
+
)(z b) +a
+
a
a
2
)
+
2
a
+
(z b). (1.113)
It can be veried by a long and tedious (but straightforward) calculation.
1.3. A simple example
We have been talking about Jacobi operators for quite some time now, but we
have not seen a single example yet. Well, here is one, the free Jacobi operator
H
0
associated with constant sequences a(n) = a, b(n) = b. The transformation
z 2az +b reduces this problem to the one with a
0
(n) = 1/2, b
0
(n) = 0. Thus we
will consider the equation
(1.114)
1
2
_
u(n + 1) +u(n 1)
_
= zu(n).
Without restriction we choose n
0
= 0 throughout this section (note that we have
s(z, n, n
0
) = s(z, n n
0
), etc.) and omit n
0
in all formulas. By inspection (try the
ansatz u(n) = k
n
) u
(z, n) = (z R
1/2
2
(z))
n
,
where R
1/2
2
(z) =
z 1
z = [
, u
+
) = R
1/2
2
(z) we need a second solution for z
2
= 1, which is given
by s(1, n) = (1)
n+1
n. For the fundamental solutions we obtain
s(z, n) =
(z +R
1/2
2
(z))
n
(z R
1/2
2
(z))
n
2R
1/2
2
(z)
,
c(z, n) =
s(z, n 1)
s(z, 1)
= s(z, n 1). (1.117)
Notice that s(z, n) = (1)
n+1
s(z, n). For n > 0 we have the following expansion
s(z, n) =
[[n/2]]
j=0
_
n
2j + 1
_
(z
2
1)
j
z
n2j1
=
[[n/2]]
k=0
_
_
(1)
k
[[n/2]]
j=k
_
n
2j + 1
__
j
k
_
_
_
z
n2k1
, (1.118)
20 1. Jacobi operators
where [[x]] = supn Z[ n < x. It is easily seen that we have H
0
 = 1 and
further that
(1.119) (H
0
) = [1, 1].
For example, use unitarity of the Fourier transform
(1.120)
U :
2
(Z) L
2
(, )
u(n)
nZ
u(n)e
inx
.
which maps H
0
to the multiplication operator by cos(x).
The Green function of H
0
explicitly reads (z C[1, 1])
(1.121) G
0
(z, m, n) =
(z +R
1/2
2
(z))
]mn]
R
1/2
2
(z)
.
In particular, we have
g
0
(z, n) =
1
R
1/2
2
(z)
= 2
j=0
_
2j
j
_
1
(2z)
2j+1
h
0
(z, n) =
z
R
1/2
2
(z)
=
j=0
_
2j
j
_
1
(2z)
2j
. (1.122)
Note that it is sometimes convenient to set k = z + R
1/2
2
(z) (and conversely z =
1
2
(k +k
1
)), or k = exp(i) (and conversely z = cos()) implying
(1.123) u
(z, n) = k
n
= e
in
.
The map z k = z + R
1/2
2
(z) is a holomorphic mapping from the set
+
(C )[1, 1] to the unit disk z C[ [z[ < 1. In addition, viewed as a map
on the Riemann surface of R
1/2
2
(z), it provides an explicit isomorphism between
the Riemann surface of R
1/2
2
(z) and the Riemann sphere C .
1.4. General second order dierence expressions
We consider the dierence expression
(1.124) f(n) =
1
w(n)
_
f(n + 1) +f(n 1) +d(n)f(n)
_
,
where w(n) > 0, d(n) R, and (w(n)w(n + 1))
1
, w(n)
1
d(n) are bounded
sequences. It gives rise to an operator
H, called Helmholtz operator, in the
weighted Hilbert space
2
(Z; w) with scalar product
(1.125) f, g =
nZ
w(n)f(n)g(n), f, g
2
(Z; w).
Greens formula (1.20) holds with little modications and
H is easily seen to be
bounded and selfadjoint. There is an interesting connection between Jacobi and
Helmholtz operators stated in the next theorem.
1.4. General second order dierence expressions 21
Theorem 1.14. Let H be the Jacobi operator associated with the sequences a(n) >
0, b(n) and let
H be the Helmholtz operator associated with the sequences w(n) > 0,
d(n). If we relate these sequences by
(1.126)
w(2m) = w(0)
m1
j=0
a(2j)
2
a(2j + 1)
2
w(2m+ 1) =
1
a(2m)
2
w(2m)
, d(n) = w(n)b(n),
respectively
(1.127) a(n) =
1
_
w(n)w(n + 1)
, b(n) =
d(n)
w(n)
,
then the operators H and
H are unitarily equivalent, that is, H = U
HU
1
, where
U is the unitary transformation
(1.128)
U :
2
(Z; w)
2
(Z)
u(n)
_
w(n)u(n)
.
Proof. Straightforward.
Remark 1.15. (i). The most general threeterm recurrence relation
(1.129) f(n) = a(n)f(n + 1) +
j=n0
a(j)
c(j + 1)
,
c(n) = w(n) a(n) = w(n + 1) c(n + 1), d(n) = w(n)
b(n). (1.131)
Let
H be the selfadjoint operator associated with in
2
(Z; w). Then the unitary
operator
(1.132)
U :
2
(Z; w)
2
(Z)
u(n)
_
w(n)u(n)
transforms
H into a Jacobi operator H = U
HU
1
in
2
(Z) associated with the
sequences
a(n) =
c(n)
_
w(n)w(n + 1)
= sgn( a(n))
_
a(n) c(n + 1),
b(n) =
d(n)
w(n)
=
b(n). (1.133)
22 1. Jacobi operators
In addition, the Wronskians are related by
c(n)
_
f(n)g(n + 1) f(n + 1)g(n)
_
=
a(n)
_
(Uf)(n)(Ug)(n + 1) (Uf)(n + 1)(Ug)(n)
_
. (1.134)
(ii). Let c(n) > 0 be given. Dening
(1.135)
w(2m) =
m1
j=0
_
a(2j)c(2j + 1)
c(2j)a(2j + 1)
_
2
w(2m+ 1) =
c(2m)
2
a(2m)
2
w(2m)
, d(n) = w(n)b(n),
the transformation U maps H to an operator
H = UHU
1
associated with the
dierence expression (1.130).
1.5. The innite harmonic crystal in one dimension
Finally, I want to say something about how Jacobi operators arise in applications.
Despite the variety of possible applications of dierence equations we will focus
on only one model from solid state physics: the innite harmonic crystal in one
dimension. Hence we consider a linear chain of particles with harmonic nearest
neighbor interaction. If x(n, t) denotes the deviation of the nth particle from its
equilibrium position, the equations of motion read
m(n)
d
2
dt
2
x(n, t) = k(n)(x(n + 1, t) x(n, t)) +k(n 1)(x(n 1, t) x(n, t))
= (
@
@
@
@
@
@

x(n,t)
This model is only valid as long as the relative displacement is not too large (i.e.,
at least smaller than the distance of the particles in the equilibrium position).
Moreover, from a physical viewpoint it is natural to assume k, m, m
1
(Z, R)
and k(n) ,= 0. Introducing conjugate coordinates
(1.137) p(n, t) = m(n)
d
dt
x(n, t), q(n, t) = x(n, t),
the system (1.136) can be written as Hamiltonian system with Hamiltonian given
by
(1.138) 1(p, q) =
nZ
_
p(n)
2
2m(n)
+
k(n)
2
(q(n + 1) q(n))
2
_
.
1.5. The innite harmonic crystal in one dimension 23
Since the total energy of the system is supposed to be nite, a natural phase space
for this system is (p, q)
2
(Z, R
2
) with symplectic form
(1.139) ((p
1
, q
1
), (p
2
, q
2
)) =
nZ
_
p
1
(n)q
2
(n) p
2
(n)q
1
(n)
_
.
Using the symplectic transform
(1.140) (p, q) (v, u) = (
p
m
,
mq)
we get a new Hamiltonian
(1.141)
1(v, u) =
1
2
nZ
_
v(n)
2
+ 2a(n)u(n)u(n + 1) +b(n)u(n)
2
_
,
where
(1.142) a(n) =
k(n)
_
m(n)m(n + 1)
, b(n) =
k(n) +k(n 1)
m(n)
.
The corresponding equations of evolution read
d
dt
u(n, t) =
1(v, u)
v(n, t)
= v(n, t),
d
dt
v(n, t) =
1(v, u)
u(n, t)
= Hu(n, t), (1.143)
where H is our usual Jacobi operator associated with the sequences (1.142). Equiv
alently we have
(1.144)
d
2
dt
2
u(n, t) = Hu(n, t).
Since this system is linear, standard theory implies
u(n, t) = cos(t
H)u(n, 0) +
sin(t
H)
H
v(n, 0),
v(n, t) = cos(t
H)v(n, 0)
sin(t
H)
H
Hu(n, 0), (1.145)
where
(1.146) cos(t
H) =
=0
(1)
t
2
(2)!
H
,
sin(t
H)
H
=
=0
(1)
t
2+1
(2 + 1)!
H
.
In particular, introducing
(1.147) S(, n) = (c(, n), s(, n))
and expanding the initial conditions in terms of eigenfunctions (cf. Section 2.5,
equation (2.133))
(1.148) u() =
nZ
u(n, 0)S(, n), v() =
nZ
v(n, 0)S(, n)
24 1. Jacobi operators
we infer
u(n, t) =
_
(H)
(u() cos(t
) +v()
sin(t
)S(, n)d(),
v(n, t) =
_
(H)
(v() cos(t
) u()
sin(t
Hv(0), (1.150)
since [ cos(t
)[ 1, [
sin(t
)
t
[ 1 for t R, 0.
Given this setting two questions naturally arise. Firstly, given m(n), k(n) what
can be said about the characteristic frequencies of the crystal? Spectral theory for
H deals with this problem. Secondly, given the set of characteristic frequencies
(i.e., the spectrum of H), is it possible to reconstruct m(n), k(n)? This question is
equivalent to inverse spectral theory for H once we establish how to reconstruct k,
m from a, b. This will be done next.
Note that u(n) =
_
m(n) > 0 solves u = 0. Hence, if we assume k > 0, then
H 0 by Corollary 11.2. In particular,
(1.151) (H) [0, 2(k
m
1

)]
and 0 is in the essential spectrum of H by Lemma 3.8. Moreover, since
(1.152)
nN
1
a(n)u(n)u(n + 1)
=
nN
1
k(n)
=
(recall k
is unitary equivalent
to H (and so are
H
and H
).
From a physical viewpoint the case of a crystal with N atoms in the base cell is
of particular interest, that is, k and m are assumed to be periodic with period N.
In this case, the same is true for a, b and we are led to the study of periodic Jacobi
operators. The next step is to consider impurities in this crystal. If such impurities
appear only local, we are led to scattering theory with periodic background (to be
dealt with in Chapter 7). If they are randomly distributed over the crystal we are
led to random operators (to be dealt with in Chapter 5).
1.5. The innite harmonic crystal in one dimension 25
But for the moment let us say a little more about the simplest case where m
and k are both constants (i.e., N = 1). After the transformation t
_
m
2k
t we can
even assume m = 2k = 1, that is,
(1.154)
d
2
dt
2
x(n, t) =
1
2
(x(n + 1, t) +x(n 1, t)) x(n, t).
The socalled plane wave solutions are given by
(1.155) u
(n, t) = e
i(n ()t)
,
where the wavelength
1
and the frequency are connected by the dispersion
relation
(1.156) () =
_
1 cos() =
2[ sin(
2
)[.
Since u
nZ
u(n, 0)e
in
, v() =
nZ
v(n, 0)e
in
we obtain
(1.158) u(n, t) =
1
2
_
_
u() cos(()t) +v()
sin(()t)
()
_
e
in
d.
Or, equivalently,
(1.159) u(n, t) =
mZ
c
nm
(t)u(m, 0) +s
nm
(t)v(m, 0),
where
c
n
(t) =
1
2
_
cos(()t)e
in
d = J
2]n]
(
2t),
s
n
(t) =
1
2
_
sin(()t)
()
e
in
d =
_
t
0
c
n
(s)ds
=
t
2]n]+1
2
]n]
([n[ + 1)!
1
F
2(
2[n[ + 1
2
; (
2[n[ + 3
2
, 2[n[ + 1);
t
2
2
). (1.160)
Here J
n
(x),
p
F
q
(u; v; x) denote the Bessel and generalized hypergeometric func
tions, respectively. From this form one can deduce that a localized wave (say
compactly supported at t = 0) will spread as t increases (cf. [194], Corollary to
Theorem XI.14). This phenomenon is due to the fact that dierent plane waves
travel with dierent speed and is called dispersion.
You might want to observe the following fact: if the coupling constant k(n
0
)
between the n
0
th and (n
0
+ 1)th particle is zero (i.e., no interaction between
these particles), then the chain separates into two parts (cf. the discussion after
Theorem 1.5 and note that k(n
0
) = 0 implies a(n
0
) = 0 for the corresponding
Jacobi operator).
We will encounter chains of particles again in Section 12.1. However, there we
will consider a certain nonlinear interaction.
Chapter 2
Foundations of spectral
theory for Jacobi operators
The theory presented in this chapter is the discrete analog of what is known as Weyl
TitchmarshKodaira theory for SturmLiouville operators. The discrete version has
the advantage of being less technical and more transparent.
Again, the present chapter is of fundamental importance and the tools devel
oped here are the pillars of spectral theory for Jacobi operators.
2.1. Weyl mfunctions
In this section we will introduce and investigate Weyl mfunctions. Rather than
the classical approach of Weyl (cf. Section 2.4) we advocate a dierent one which
is more natural in the discrete case.
As in the previous chapter, u
(z, n
0
) =
n01
, (H
,n0
z)
1
n01
= G
,n0
(z, n
0
1, n
0
1).
By virtue of (1.104) we also have the more explicit form
(2.2) m
+
(z, n
0
) =
u
+
(z, n
0
+ 1)
a(n
0
)u
+
(z, n
0
)
, m
(z, n
0
) =
u
(z, n
0
1)
a(n
0
1)u
(z, n
0
)
.
The base point n
0
is of no importance in what follows and we will only consider
m
(z) = m
)
and that it satises
(2.3) m
(z) = m
(z), [m
(z)[ (H
z)
1

1
[Im(z)[
.
Moreover, m
(z) is a Herglotz function (i.e., it maps the upper half plane into
itself, cf. Appendix B). In fact, this is a simple consequence of the rst resolvent
27
28 2. Foundations of spectral theory
identity
Im(m
(z)) = Im(z)
1
, (H
z)
1
(H
z)
1
= Im(z)(H
z)
1
1

2
. (2.4)
Hence by Theorem B.2, m
(z) =
_
R
d
()
z
, z CR,
where
=
_
(,]
d
() =
1
lim
0
lim
0
+
_
Im(m
(x + i))dx.
Here we have normalized
() = 0,
< (H
).
Let P
(H
(z) =
1
, (H
z)
1
1
=
_
R
d
1
, P
(,]
(H
)
1
z
.
Thus we see that d
= d
1
, P
(,]
(H
)
1
is the spectral measure of H
() =
1
, (H
+
)
1
.
Moreover, there is a close connection between the socalled moment problem (i.e.,
determining d
from
d
. Indeed, since m
,0
= 1, m
,1
= b(1), m
,2
= a(1
0
1
)
2
+b(1)
2
, etc., we
infer
(2.9) b(1) = m
,1
, a(1
0
1
)
2
= m
,2
(m
,1
)
2
, etc. .
We will consider this topic in Section 2.5.
You might have noticed, that m
(z, n) = u
(z, n
0
)
n1
j=n0
(a(j
0
1
)m
(z, j))
1
and that m
+
(z, n) satises the following recurrence relation
(2.11) a(n
0
1
)
2
m
(z, n) +
1
m
(z, n 1)
+z b(n) = 0.
The functions m
(z, n) =
u
(z, n + 1)
a(n)u
(z, n)
, m
(z) = m
(z, 0)
is often more convenient than the original one. The connection is given by
(2.13) m
+
(z, n) = m
+
(z, n), m
(z, n) =
a(n)
2
m
(z, n) z +b(n)
a(n 1)
2
(note also m
(z, n+1)
1
= a(n)
2
m
=
a(0)
2
a(1)
2
d
.
You might want to note that m
(z) =
z b(n)
a(0)
2
+
_
R
d
()
z
.
Finally, we introduce the Weyl mfunctions m
,n
.
They are dened analogously to m
+,n
in
terms of H
+,n
suggests that m
+
(z, n), ,= 0, can be expressed in terms of m
+
(z, n).
Using (1.91) and the second resolvent identity we obtain
(2.16) m
+
(z, n) =
n+1
, (H
+,n
z)
1
n+1
=
m
+
(z, n)
a(n)m
+
(z, n)
.
Similarly, the Weyl mfunctions m
,n
, ,= , can be
expressed in terms of m
(z, n + 1),
(2.17) m
(z, n) =
n
, (H
,n
z)
1
n
=
m
(z, n + 1)
1 a(n)m
(z, n + 1)
.
2.2. Properties of solutions
The aim of the present section is to establish some fundamental properties of special
solutions of (1.19) which will be indispensable later on.
As an application of Weyl mfunctions we rst derive some additional properties
of the solutions u
(z, n) = a(0)u
(z, 0)
_
a(0)
1
c(z, n) m
(z)s(z, n)
_
,
30 2. Foundations of spectral theory
where the constant (with respect to n) a(0)u
(z, 0) = 1 (bearing in mind that c(z, n), s(z, n) are polynomials with respect
to z), we infer for instance (using (2.4))
(2.19) u
(z, n) = u
(z, n).
Moreover, u
). But we can do
even better. If is an isolated eigenvalue of H
, then m
(z, 0) = (z ) we
can remove the singularity at z = . In summary,
Lemma 2.2. The solution u
), respectively. If we choose
(2.20) u
(z, n) =
c(z, n)
a(0)
m
(z)s(z, n),
then u
), u
(z, .) , 0, and u
(z, .) =
u
(z, n).
Moreover, the sums
(2.21)
j=n
u
+
(z
1
, j)u
+
(z
2
, j),
n
j=
u
(z
1
, j)u
(z
2
, j)
are holomorphic with respect to z
1
(resp. z
2
) provided u
(z
1
, .) (resp. u
(z
2
, .)) are.
Proof. Only the last assertion has not been shown yet. It suces to prove the
claim for one n, say n = 1. Without restriction we suppose u
+
(z, 0) = a(0)
1
(if
u
+
(z, 0) = 0, shift everything by one). Then u
+
(z, n) = (H
+
z)
1
1
(n), n 1,
and hence
(2.22)
j=1
u
+
(z
1
, j)u
+
(z
2
, j) =
1
, (H
+
z
1
)
1
(H
+
z
2
)
1
2
(N)
.
The remaining case is similar.
Remark 2.3. If (
0
,
1
) (H), we can dene u
(, n), [
0
,
1
]. Indeed, by
(2.20) it suces to prove that m
(),
(2.23) m
t
() =
1
, (H
)
2
1
< 0, (
0
,
1
)
(compare also equation (B.18)). Here the prime denotes the derivative with respect
to . However, u
(
0,1
, n) might not be square summable near in general.
In addition, the discrete eigenvalues of H
+
are the zeros of
a(0)
m
+
() (see
(2.16)) and hence decreasing as a function of . Similarly, the discrete eigenvalues
of H
j=m
[u(z, j)[
2
,
where [.]
n
denotes the Weyl bracket,
(2.25) [u(z)]
n
=
W
n
(u(z), u(z))
2iIm(z)
= a(n)
Im(u(z, n)u(z, n + 1))
Im(z)
.
Especially for u
(z)]
n
=
_
j=n+1
[u
+
(z, j)[
2
j=
[u
(z, j)[
2
and for s(z, n)
(2.27) [s(z)]
n
=
_
j=0
[s(z, j)[
2
, n 0
0
j=n+1
[s(z, j)[
2
, n < 0
.
Moreover, let u(z, n), v(z, n) be solutions of (1.19). If u(z, n) is dierentiable
with respect to z, we obtain
(2.28) ( z)u
t
(z, n) = u(z, n)
(here the prime denotes the derivative with respect to z) and using Greens formula
(1.20) we infer
(2.29)
n
j=m
v(z, j)u(z, j) = W
n
(v(z), u
t
(z)) W
m1
(v(z), u
t
(z)).
Even more interesting is the following result.
Lemma 2.4. Let u
(z),
d
dz
u
(z)) =
_
j=n+1
u
+
(z, j)
2
n
j=
u
(z, j)
2
.
Proof. Greens formula (1.20) implies
(2.31) W
n
(u
+
(z), u
+
( z)) = (z z)
j=n+1
u
+
(z, j)u
+
( z, j)
32 2. Foundations of spectral theory
and furthermore,
W
n
(u
+
(z), u
t
+
(z)) = lim
zz
W
n(u
+
(z),
u
+
(z) u
+
( z)
z z
)
=
j=n+1
u
+
(z, j)
2
. (2.32)
(
0
), u
+
(
0
)) = 0 and hence u
(
0
, n) dier only by a (nonzero) constant
multiple. Moreover,
d
dz
W(u
(z), u
+
(z))
z=0
= W
n
(u
(
0
), u
t
+
(
0
)) +W
n
(u
t
(
0
), u
+
(
0
))
=
jZ
u
(
0
, j)u
+
(
0
, j) (2.33)
by the lemma and hence
(2.34) G(z, n, m) =
P(
0
, n, m)
z
0
+O(z
0
)
0
,
where
(2.35) P(
0
, n, m) =
u
(
0
, n)u
(
0
, m)
jZ
u
(
0
, j)
2
.
Similarly, for H
we obtain
(2.36) lim
z
(z )G
(z, n, m) =
s(, n)s(, m)
jN
s(, j)
2
.
Thus the poles of the kernel of the resolvent at isolated eigenvalues are simple.
Moreover, the negative residue equals the kernel of the projector onto the corre
sponding eigenspace.
As a second application we show monotonicity of G(z, n, n) with respect to z
in a spectral gap. Dierentiating (1.99), a straightforward calculation shows
(2.37) G
t
(z, n, n) =
u
+
(z, n)
2
W
n
(u
(z), u
(z)) u
(z, n)
2
W
n
(u
+
(z), u
+
(z))
W(u
(z), u
+
(z))
2
,
which is positive for z R(H) by Lemma 2.4. The same result also follows from
G
t
(z, n, m) =
d
dz
n
, (H z)
1
m
=
n
, (H z)
2
jZ
G(z, n, j)G(z, j, m). (2.38)
Finally, let us investigate the qualitative behavior of the solutions u
(z) more
closely.
2.2. Properties of solutions 33
Lemma 2.5. Suppose z C
ess
(H
). Then we can nd C
(z, n)[ C
exp(
n), n N.
For
we can choose
(2.40)
= ln
_
1 + (1 )
sup
R]
dist((H
), z)
2 sup
nN
[a(n)[
_
, > 0
(the corresponding constant C
will depend on ).
Proof. By the denition of H
(H
+
z)
1
P
n
=
1
, (P
H
+
P
z)
1
=
1
, (H
+
z +Q
)
1
n
, (2.41)
where
(2.42) (P
u)(n) = e
n
u(n)
and
(2.43) (Q
u)(n) = a(n)(e
1)u(n 1).
Moreover, choosing = ln(1 + (1 )/(2 sup
nN
[a(n)[)) we have
(2.44) Q
 2 sup
nN
[a(n)[(e
1) = (1 ).
Using the rst resolvent equation
(H
+
z +Q
)
1
 (H
+
z)
1
 +(H
+
z)
1

Q
(H
+
z +Q
)
1
 (2.45)
and (H
+
z)
1
 =
1
implies the desired result
(2.46) [e
(n1)
G
+
(z, 1, n)[ (H
+
z +Q
)
1

1
This result also shows that for any solution u(z, n) which is not a multiple of
u
const
[a(n)[
e
n
since otherwise the Wronskian of u and u
() = 0
ess
(H
)
since
() >
, ,
ess
(H
).
34 2. Foundations of spectral theory
2.3. Positive solutions
In this section we want to investigate solutions of u = u with (H). We will
assume
(2.49) a(n) < 0
throughout this section. As a warmup we consider the case < (H).
Lemma 2.6. Suppose a(n) < 0 and let < (H). Then we can assume
(2.50) u
(, n) > 0, n Z,
and we have
(2.51) (n n
0
)s(, n, n
0
) > 0, n Zn
0
.
Proof. From (H ) > 0 one infers (H
+,n
) > 0 and hence
(2.52) 0 <
n+1
, (H
+,n
)
1
n+1
=
u
+
(, n + 1)
a(n)u
+
(, n)
,
showing that u
+
() can be chosen to be positive. Furthermore, for n > n
0
we
obtain
(2.53) 0 <
n
, (H
+,n0
)
1
n
=
u
+
(, n)s(, n, n
0
)
a(n
0
)u
+
(, n
0
)
,
implying s(, n, n
0
) > 0 for n > n
0
. The remaining case is similar.
The general case (H) requires an additional investigation. We rst prove
that (2.51) also holds for (H) with the aid of the following lemma.
Lemma 2.7. Suppose a(n) < 0 and let u , 0 solve u = u. Then u(n) 0,
n n
0
(resp. n n
0
), implies u(n) > 0, n > n
0
(resp. n < n
0
). Similarly,
u(n) 0, n Z, implies u > 0, n Z.
Proof. Fix n > n
0
(resp. n < n
0
). Then, u(n) = 0 implies u(n 1) > 0 (since
u cannot vanish at two consecutive points) contradicting 0 = (b(n) )u(n) =
a(n)u(n + 1) a(n 1)u(n 1) > 0.
The desired result now follows from
(2.54) s(, n, n
0
) = lim
0
s( , n, n
0
) 0, n > n
0
,
and the lemma. In addition, we note
(2.55) b(n) = a(n)s(, n + 1, n 1) > 0, (H).
The following corollary is simple but powerful.
Corollary 2.8. Suppose u
j
(, n), j = 1, 2, are two solutions of u = u, (H),
with u
1
(, n
0
) = u
2
(, n
0
) for some n
0
Z. Then if
(2.56) (n n
0
)(u
1
(, n) u
2
(, n)) > 0 (resp. < 0)
holds for one n Zn
0
, then it holds for all. If u
1
(, n) = u
2
(, n) for one
n Zn
0
, then u
1
and u
2
are equal.
Proof. Use u
1
(, n) u
2
(, n) = c s(, n, n
0
) for some c R.
2.3. Positive solutions 35
In particular, this says that for (H) solutions u(, n) can change sign
(resp. vanish) at most once (since s(z, n, n
0
) does).
For the sequence
(2.57) u
m
(, n) =
s(, n, m)
s(, 0, m)
, m N,
this corollary implies that
(2.58)
m
() = u
m
(, 1) =
s(, 1, m)
s(, 0, m)
is increasing with m since we have u
m+1
(, m) > 0 = u
m
(, m). Next, since
a(0)s(, 1, m) +a(1)s(, 1, m) = ( b(0))s(, 0, m) implies
(2.59)
m
() <
b(0)
a(0)
,
we can dene
(2.60)
+
() = lim
m
m
(), u
+
(, n) = lim
m
u
m
(, n) = c(, n) +
+
()s(, n).
By construction we have u
+
(, n) > u
m
(, n), n N, implying u
+
(, n) > 0, n N.
For n < 0 we have at least u
+
(, n) 0 since u
m
(, n) > 0 and thus u
+
(, n) > 0,
n Z, by Lemma 2.7.
Let u(, n) be a solution with u(, 0) = 1, u(, 1) = (). Then, by the above
analysis, we infer that u(, n) > 0, n N, is equivalent to ()
+
() and hence
u(, n) u
+
(, n), n N (with equality holding if and only if u() = u
+
()). In
this sense u
+
() is the minimal positive solution (also principal or recessive
solution) near +. In particular, since every solution can change sign at most
once, we infer that if there is a square summable solution near +, then it must
be equal to u
+
(, n) (up to a constant multiple), justifying our notation.
Moreover, if u() is dierent from u
+
(), then constancy of the Wronskian
(2.61)
u
+
(, n)
u(, n)
u
+
(, n + 1)
u(, n + 1)
=
W(u(), u
+
())
a(n)u(, n)u(, n + 1)
together with W(u(), u
+
()) = a(0)(
+
() ()) > 0 shows that the sequence
u
+
(, n)/u(, n) is decreasing for all u() ,= u
+
() if and only if u
+
() is minimal.
Moreover, we claim
(2.62) lim
n
u
+
(, n)
u(, n)
= 0.
In fact, suppose lim
n
u
+
(, n)/u(, n) = > 0. Then u
+
(, n) > u(, n), n
N, and hence u
= (u
+
(, n)u(, n))/(1) > 0, n N. But u
+
(, n) < u
(, n)
implies u(, n) < u
+
(, n), a contradiction.
Conversely, (2.62) for one u() uniquely characterizes u
+
(, n) since (2.62)
remains valid if we replace u() by any positive linear combination of u() and
u
+
().
Summing up (2.61) shows that
(2.63)
j=0
1
a(j)u(, j)u(, j + 1)
=
1
a(0)(
+
() ())
< .
36 2. Foundations of spectral theory
Moreover, the sequence
(2.64) v(, n) = u(, n)
j=n
1
a(j)u(, j)u(, j + 1)
> 0
solves u = u (cf. (1.51)) and equals u
+
(, n) up to a constant multiple since
lim
n
v(, n)/u(, n) = 0.
By reection we obtain a corresponding minimal positive solution u
(, n) near
. Let us summarize some of the results obtained thus far.
Lemma 2.9. Suppose a(n) < 0, (H) and let u(, n) be a solution with
u(, n) > 0, n 0. Then the following conditions are equivalent.
(i). u(, n) is minimal near .
(ii). We have
(2.65)
u(, n)
v(, n)
v(, 0)
u(, 0)
, n 0,
for any solution v(, n) with v(, n) > 0, n 0.
(iii). We have
(2.66) lim
n
u(, n)
v(, n)
= 0.
for one solution v(, n) with v(, n) > 0, n 0.
(iv). We have
(2.67)
jN
1
a(j)u(j)u(j + 1)
= .
Recall that minimality says that for a solution u(, n) with u(, 0) = 1, u(, 1) =
() to be positive on N, we need ()
+
(). Similarly, for u(, n) to be pos
itive on N we need ()
+
() ()
(, n) +
1 +
2
u
+
(, n), [1, 1].
Two cases may occur
(i). u
(, n), u
+
(, n) are linearly dependent (i.e.,
+
() =
(, n), u
+
(, n) are linearly independent and
(2.69) u
(, n) =
1 +
2
u
+
(, n) +
1
2
u
(, n),
is positive if and only if [1, 1]. In this case H is called subcritical
.
If H > (H), then H is always subcritical by Lemma 2.6. To emphasize
this situation one sometimes calls H supercritical if H > (H).
2.4. Weyl circles 37
In case (ii) it is not hard to show using (2.62) that for two positive solutions
u
j
(, n), j = 1, 2, we have
(2.70) u
(, n) =
1 +
2
u
1
(, n) +
1
2
u
2
(, n) > 0 [1, 1],
if and only if the solutions u
1,2
equal u
up to constant multiples.
Remark 2.10. Assuming a(n) < 0, the requirement H 0 is also necessary
for a positive solution to exist. In fact, any positive solution can be used to factor
H = A
(z, .), c
(z, 0) = sin(), s
(z, 1) = cos(),
c
(z, 0) = cos(), c
(z, 1) = sin(),
where = cot(). Clearly,
W(c
(z), s
(
j
(N), n) (since s
N
(z) = sgn(N)
W
N
(s
(
j
(N)), c
(z))
a(0)W
N
(s
(
j
(N)), s
(z))
, N Z0
(independent of the eigenvalue chosen), then the solution
(2.74) u
N
(z, n) = a(0)
1
c
(z, n) sgn(N) m
N
(z)s
(z, n)
satises the boundary condition at N, that is, W
N
(s
(
1
), u
N
(z)) = 0. The func
tion m
N
(z) is rational (w.r.t. z) and has poles at the eigenvalues
j
(N). In par
ticular, they are the analogs of the Weyl mfunctions for nite Jacobi operators.
Hence it suggests itself to consider the limit N , where our hope is to obtain
the Weyl mfunctions for the Jacobi operators H
.
We x
0
R and set
(2.75)
N
=
s
(
0
, N + 1)
s
(
0
, N)
38 2. Foundations of spectral theory
implying
0
=
j
(N) for one j. It will turn out that the above choice for
N
is not
really necessary since m
N
(z) will, as a consequence of boundedness of H, converge
for arbitrary sequences
N
. However, the above choice also works in the case of
unbounded operators.
Before we turn to the limit, let us derive some additional properties for nite
N. The limits
lim
zj(N)
W
N
(s
(
j
(N)), c
(z)) = a(0),
lim
zj(N)
W
N
(s
(
j
(N)), s
(z))
z
j
(N)
= W
N
(s
(
j
(N)), s
t
(
j
(N))) (2.76)
imply that all poles of m
N
(z, ) are simple. Using (2.29) to evaluate (2.76) one
infers that the negative residue at
j
(N) is given by
(2.77)
N
(
j
(N)) =
_
N
0
n=
1
N+1
s
(
j
(N), n)
2
_
1
, N
>
<
0.
The numbers
N
(
j
(N)) are called norming constants. Hence one gets
(2.78) m
N
(z) =
j=1
N
(
j
(N))
z
j
(N)
+
_
1
a(0)
,
1
,=
zb(
1
0
)
a(0)
2
,
1
=
, N
>
<
0,
and with the help of (2.24) we obtain
(2.79)
N
0
n=
1
N+1
[u
N
(z, n)[
2
=
Im( m
N
(z))
Im(z)
, N
>
<
0,
that is, m
N
(z) are Herglotz functions.
Now we turn to the limits N . Fix z CR. Observe that if
1
varies
in R, then
N
takes all values in R
N times. Rewriting (2.73) shows that
(2.80) m
N
(z) =
sgn(N)
a(0)
c
(z, N)
N
+c
(z, N + 1)
s
(z, N)
N
+s
(z, N + 1)
is a Mobius transformation and hence the values of m
N
(z) for dierent
N
R lie on a circle (also called Weyl circle) in the complex plane (note that
z R would correspond to the degenerate circle R). The center of this circle
is
(2.81) c
N
= sgn(N)
W
N
(c
(z), s
(z))
2a(0)Im(z)[s
(z)]
N
and the radius is
(2.82) r
N
=
W(c
(z), s
(z))
2a(0)Im(z)[s
(z)]
N
=
1
2[Im(z)[s
(z)]
N
[
.
Using
(2.83) [a(0)
1
c
(z) sgn(N) ms
(z)]
N
= [s
(z)]
N([ mc
N
[
2
r
2
N
)
2.4. Weyl circles 39
we see that this circle is equivalently characterized by the equation
(2.84) m C[[a(0)
1
c
(z) + sgn(N) ms
(z)]
N
= 0.
Since [.]
N
, N > 0, is decreasing with respect to N, the circle corresponding to N+1
lies inside the circle corresponding to N. Hence these circles tend to a limit point
for any sequence (
N
R )
NN
since
(2.85) lim
N
[s
(z)]
N
=
n=0
[s
(z, n)[
2
=
(otherwise H
+
would have a nonreal eigenvalue). Similarly for N < 0. Thus the
pointwise convergence of m
N
(z) is clear and we may dene
(2.86) m
(z) = lim
N
m
N
(z).
Moreover, the above sequences are locally bounded in z (x N and take all circles
corresponding to a (suciently small) neighborhood of any point z and note that
all following circles lie inside the ones corresponding to N) and by Vitalis theorem
([229], p. 168) they converge uniformly on every compact set in C
= z C[
Im(z) > 0, implying that m
.
Remark 2.11. (i). Since m
N
(z) converges for arbitrary choice of the sequence
N
we even have
(2.87) m
(z) = lim
N
c
(z, N)
a(0)s
(z, N)
.
Moreover, this approach is related to (2.60). Using (1.62), (1.63) shows
m
() =
c(, m)/s(, m) and establishes the equivalence.
(ii). That the Weyl circles converge to a point is a consequence of the boundedness
of a(n), b(n). In the general case the limit could be a circle or a point (independent
of z CR). Accordingly one says that is limit circle or limit point at . (See
Section 2.6 for further information on unbounded operators.)
As anticipated by our notation, m
(z, n) = a(0)
1
c
(z, n) m
(z)s
(z, n)
is square summable near . Since m
+
lies in the interior of all Weyl circles the
limit lim
N
[u
+
(z)]
N
0 must exist and hence u
+
2
+
(Z) by (2.24). Moreover,
u
+
2
+
(Z) implies [u
+
]
= 0. Similarly for u
n=
1
[u
(z, n)[
2
=
Im( m
(z))
Im(z)
,
implies that m
(z, 0) = a(0)
1
cos() m
(z) =
1
a(0)
+
_
R
d
()
z
, ,=
0
.
40 2. Foundations of spectral theory
This nally establishes the connection
(2.91) m
(z) = m
(z) =
u
(z, 1)
a(0)u
(z, 0)
as expected. Furthermore, m
1
is unique up to a constant) by
(2.92) m
1
(z) =
1
a(0)
a(0) cos(
2
1
) m
2
(z) sin(
2
1
)
a(0) sin(
2
1
) m
2
(z) cos(
2
1
)
,
where
1,2
= cot(
1,2
). Specializing to the case
1
= ,
2
= we infer
(2.93) m
+
(z, n) =
m
+
(z) +a(0)
1
a(0)m
+
(z)
, m
(z, n) =
m
(z, 1) +a(0)
1
1 a(0)m
(z, 1)
which should be compared with (2.16), (2.17), respectively.
2.5. Canonical forms of Jacobi operators and the
moment problem
The aim of this section is to derive canonical forms for H, H
j=0
s
j,n
(0)H
j
+
1
=
n
,
implying that
1
is a cyclic vector for H
+
. We recall the measure
d
+
() = d
1
, P
(,]
(H
+
)
1
(2.95)
and consider the Hilbert space L
2
(R, d
+
). Since d
+
is supported on (H
+
) this
space is the same as the space L
2
((H
+
), d
+
). The scalar product is given by
(2.96) f, g
L
2 =
_
R
f()g() d
+
().
If f, g are polynomials we can evaluate their scalar product without even knowing
d
+
() since
(2.97) f, g
L
2 = f(H
+
)
1
, g(H
+
)
1
.
Applying this relation in the special case f() = s(, m), g() = s(, n), we obtain
from equation (2.94) that the polynomials s(z, n), n N, are orthogonal with
respect to this scalar product, that is,
(2.98) s(, m), s(, n)
L
2 =
_
R
s(, m)s(, n) d
+
() =
m,n
.
We will see in Theorem 4.5 that s(, n) has n1 distinct real roots which interlace
the roots of s(, n + 1).
2.5. Canonical forms and the moment problem 41
Now consider the following transformation U from the set
0
(N) onto the set
of all polynomials (eigenfunction expansion)
(Uf)() =
n=1
f(n)s(, n),
(U
1
F)(n) =
_
R
s(, n)F()d
+
(). (2.99)
A simple calculation for F() = (Uf)() shows that U is unitary,
(2.100)
n=1
[f(n)[
2
=
_
R
[F()[
2
d
+
().
This leads us to the following result.
Theorem 2.12. The unitary transformation
(2.101)
U :
2
(N) L
2
(R, d
+
)
f(n)
n=1
f(n)s(, n)
(where the sum is to be understood as norm limit) maps the operator H
+
to the
multiplication operator by . More explicitly,
(2.102) H
+
=
U
1
H
U,
where
(2.103)
H : L
2
(R, d
+
) L
2
(R, d
+
)
F() F()
.
Proof. Since d
+
is compactly supported the set of all polynomials is dense in
L
2
(R, d
+
) (Lemma B.1) and U extends to the unitary transformation
U. The rest
follows from
HF() =
UH
+
U
1
F() =
UH
+
_
R
s(, n)F()d
+
()
=
U
_
R
s(, n)F()d
+
() = F(). (2.104)
pp
(H
+
) = R[ is a growth point of
+,pp
,
ac
(H
+
) = R[ is a growth point of
+,ac
,
sc
(H
+
) = R[ is a growth point of
+,sc
. (2.106)
42 2. Foundations of spectral theory
Recall that
pp
(H
+
) is in general not equal to the point spectrum
p
(H
+
) (i.e.,
the set of eigenvalues of H
+
). However, we have at least
(2.107)
pp
(H
+
) =
p
(H
+
),
where the bar denotes closure.
An additional decomposition in continuous and singular part with respect to
the Hausdor measure dh
c
(H
+
) = R[ is a growth point of
+,c
,
s
(H
+
) = R[ is a growth point of
+,s
.
They will be used in Section 3.3.
Finally, we show how a
2
, b can be reconstructed from the measure
+
. In fact,
even the moments m
+,j
, j N, are sucient for this task. This is generally known
as (Hamburger) moment problem.
Suppose we have a given sequence m
j
, j N
0
, such that
(2.109) C(k) = det
_
_
_
_
_
m
0
m
1
m
k1
m
1
m
2
m
k
.
.
.
.
.
.
.
.
.
.
.
.
m
k1
m
k
m
2k2
_
_
_
_
_
> 0, k N.
Without restriction we will assume m
0
= 1. Using this we can dene a sesquilinear
form on the set of polynomials as follows
(2.110) P(), Q()
L
2 =
j,k=0
m
j+k
p
j
q
k
,
where P(z) =
j=0
p
j
z
j
, Q(z) =
j=0
q
j
z
j
(note that all sums are nite). The
polynomials
(2.111) s(z, k) =
1
_
C(k 1)C(k)
det
_
_
_
_
_
_
_
m
0
m
1
m
k1
m
1
m
2
m
k
.
.
.
.
.
.
.
.
.
.
.
.
m
k2
m
k1
m
2k3
1 z z
k1
_
_
_
_
_
_
_
, k N
(set C(0) = 1), form a basis for the set of polynomials which is immediate from
(2.112) s(z, k) =
C(k 1)
C(k)
_
z
k1
+
D(k 1)
C(k 1)
z
k2
+O(z
k3
)
_
,
where D(0) = 0, D(1) = m
1
, and
(2.113) D(k) = det
_
_
_
_
_
m
0
m
1
m
k2
m
k
m
1
m
2
m
k1
m
k+1
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
m
k1
m
k
m
2k3
m
2k1
_
_
_
_
_
, k N.
2.5. Canonical forms and the moment problem 43
Moreover, this basis is orthonormal, that is,
(2.114) s(, j), s(, k)
L
2 =
j,k
,
since
s(, k),
j
=
1
_
C(k 1)C(k)
det
_
_
_
_
_
_
_
m
0
m
1
m
k1
m
1
m
2
m
k
.
.
.
.
.
.
.
.
.
.
.
.
m
k2
m
k1
m
2k3
m
j
m
j+1
m
j+k1
_
_
_
_
_
_
_
=
_
0, 0 j k 2
_
C(k)
C(k1)
, j = k 1
. (2.115)
In particular, the sesquilinear form (2.110) is positive denite and hence an inner
product (note that C(k) > 0 is also necessary for this).
Expanding the polynomial zs(z, k) in terms of s(z, j), j N, we infer
zs(z, k) =
k+1
j=0
s(, j), s(, k)
L
2s(z, j)
= a(k)s(z, k + 1) +b(k)s(z, k) +a(k 1)s(z, k 1) (2.116)
(set s(z, 0) = 0) with
(2.117) a(k) = s(, k + 1), s(, k)
L
2, b(k) = s(, k), s(, k)
L
2, k N.
In addition, comparing powers of z in (2.116) shows
(2.118) a(k) =
_
C(k 1)C(k + 1)
C(k)
, b(k) =
D(k)
C(k)
D(k 1)
C(k 1)
.
In terms of our original setting this says that given the measure d
+
(or its
moments, m
+,j
, j N) we can compute s(, n), n N, via orthonormalization of
the set
n
, n N
0
. This xes s(, n) up to a sign if we require s(, n) realvalued.
Then we can compute a(n), b(n) as above (up to the sign of a(n) which changes if
we change the sign of s(, n)). Summarizing, d
+
uniquely determines a(n)
2
and
b(n). Since knowing d
+
() is equivalent to knowing m
+
(z), the same is true for
m
+
(z) (compare also the proof of Theorem 2.29). In fact, we have an even stronger
result.
Theorem 2.13. Suppose that the bounded measure d
+
is not supported on a nite
set. Then there exists a unique bounded Jacobi operator H
+
having d
+
as spectral
measure.
Proof. We have already seen that the necessary and sucient condition for our
reconstruction procedure to work is that the sesquilinear form generated by the
moments m
j
of d
+
is positive denite. Pick any nonzero polynomial P(). Due
to our assumption we can nd > 0 and an interval I such that
+
(I) ,= 0 and
P()
2
, I. Hence P(), P()
+
(I).
As a consequence we can dene a(n), b(n), s(, n), and the unitary transform
U as before. By construction H
+
=
U
1
H
U is a bounded Jacobi operator associ
ated with a(n), b(n). That
+
is the spectral measure of H
+
follows from (using
44 2. Foundations of spectral theory
(
U
1
)() = 1)
(2.119)
1
, P
(H
+
)
1
=
U
1
, P
(
H)
U
1
=
_
R
()d
+
()
for any Borel set R.
If d
+
is supported on N points, the reconstruction procedure will break down
after N steps (i.e., C(N + 1) = 0) and we get a nite Jacobi matrix with d
+
as
spectral measure.
We also remark
(2.120) c(z, n) = a(0)
_
R
s(z, n) s(, n)
z
d
+
(), n N,
since one easily veries
+
c(z, n) = zc(z, n) a(0)
0
(n), n N (use (2.98) with
m = 1). Moreover, this implies
(2.121) u
+
(z, n) =
c(z, n)
a(0)
m
+
(z)s(z, n) =
_
R
s(, n)
z
d
+
()
and it is not hard to verify
(2.122) G
+
(z, n, m) =
_
R
s(, n)s(, m)
z
d
+
().
The Jacobi operator H can be treated along the same lines. Since we essentially
repeat the analysis of H
+
we will be more sketchy.
Consider the vector valued polynomials
(2.123) S(z, n) =
_
c(z, n), s(z, n)
_
.
The analog of (2.94) reads
(2.124) s(H, n)
1
+c(H, n)
0
=
n
.
This is obvious for n = 0, 1 and the rest follows from induction upon applying H
to (2.124). We introduce the spectral measures
(2.125) d
i,j
(.) = d
i
, P
(,]
(H)
j
,
and the (hermitian) matrix valued measure
(2.126) d =
_
d
0,0
d
0,1
d
1,0
d
1,1
_
.
The diagonal part consists of positive measures and the odiagonal part can be
written as the dierence of two positive measures
(2.127) d
0,1
() = d
1,0
() = d
0,1,+
() d
0,1,
(),
where
d
0,1,+
() =
1
2
d(
0
+
1
), P
(,]
(H)(
0
+
1
),
d
0,1,
() =
1
2
(d
0
, P
(,]
(H)
0
+d
1
, P
(,]
(H)
1
). (2.128)
2.5. Canonical forms and the moment problem 45
Moreover, d is a positive matrix measure and we have a corresponding Hilbert
space L
2
(R, C
2
, d) with scalar product given by
(2.129) F, G
L
2 =
1
i,j=0
_
R
F
i
()G
j
()d
i,j
()
_
R
F()G()d()
and if F, G are vector valued polynomials, then
(2.130) F, G
L
2 = F
0
(H)
0
+F
1
(H)
1
, G
0
(H)
0
+G
1
(H)
1
.
By (2.124) the vector valued polynomials S(, n) are orthogonal with respect to d,
(2.131) S(., m), S(., n)
L
2 =
m,n
.
The formulas analogous to (2.117) then read
(2.132) a(n) = S(, n + 1), S(, n)
L
2, b(n) = S(, n), S(, n)
L
2, n Z.
Next, we consider the following transformation U from the set
0
(Z) onto the
set of vectorvalued polynomials (eigenfunction expansion)
(Uf)() =
nZ
f(n)S(, n),
(U
1
F)(n) =
_
R
S(, n)F()d(). (2.133)
Again a simple calculation for F() = (Uf)() shows that U is unitary,
(2.134)
nZ
[f(n)[
2
=
_
R
F()F()d().
Extending U to a unitary transformation
U we obtain as in the case of H
+
the
following
Theorem 2.14. The unitary transformation
(2.135)
U :
2
(Z) L
2
(R, C
2
, d)
f(n)
n=
f(n)S(, n)
(where the sum is to be understood as norm limit) maps the operator H to the
multiplication operator by , that is,
(2.136)
H =
UH
U
1
,
where
(2.137)
H : L
2
(R, C
2
, d) L
2
(R, C
2
, d)
F() F()
.
For the Green function of H we obtain
(2.138) G(z, n, m) =
_
R
S(, n)S(, m)
z
d().
By Lemma B.13, in order to characterize the spectrum of H one only needs to
consider the trace d
tr
of d given by
(2.139) d
tr
= d
0,0
+d
1,1
.
Let the Lebesgue decomposition (cf. (2.105)) of d
tr
be given by
(2.140) d
tr
= d
tr
pp
+d
tr
ac
+d
tr
sc
,
46 2. Foundations of spectral theory
then the pure point, absolutely continuous, and singular continuous spec
tra of H are given by
(H) = R[ is a growth point of
tr
,
pp
(H) = R[ is a growth point of
tr
pp
,
ac
(H) = R[ is a growth point of
tr
ac
,
sc
(H) = R[ is a growth point of
tr
sc
. (2.141)
The Weyl matrix M(z) is dened as
M(z) =
d()
z
1
2a(0)
_
0 1
1 0
_
=
_
g(z, 0)
h(z,0)
2a(0)
h(z,0)
2a(0)
g(z, 1)
_
, z C(H). (2.142)
Explicit evaluation yields
(2.143) M(z) =
1
m
+
(z) + m
(z)
_
1
a(0)
2
m+(z) m(z)
2a(0)
m+(z) m(z)
2a(0)
m
+
(z) m
(z)
_
,
and the determinant reads
(2.144) det M(z) =
1
4a(0)
2
.
In terms of the original Weyl mfunctions we obtain
g(z, 0) =
1
z b(0) a(0)
2
m
+
(z) a(1)
2
m
(z)
,
h(z, 0) = (z b(0) +a(0)
2
m
+
(z) a(1)
2
m
(z, n), s
(z, n) =
_
c
(z, n), s
(z, n)
_
= U
S(z, n),
where U
=
_
cos sin
sin cos
_
.
Hence all objects need to be rotated by the angle . For instance, introducing
M
(z) = U
M(z)U
1
we infer
M
(z) =
1
m
+
(z) + m
(z)
_
_
1
a(0)
2
m
+
(z) m
(z)
2a(0)
m
+
(z) m
(z)
2a(0)
m
+
(z) m
(z)
_
_
. (2.148)
Note also
(2.149)
1
a(0)
2
( m
+
(z) + m
(z))
= sin
2
()
(z, 0).
2.6. Some remarks on unbounded operators 47
If we restrict ourselves to the absolutely continuous part of the spectrum, we
can do even more. We abbreviate m
() = lim
0
m
())
a(0)
2
[ m
+
() + m
()[
2
d,
d
0,1,ac
() =
Im( m
+
())Re( m
()) Re( m
+
())Im( m
())
a(0)[ m
+
() + m
()[
2
d,
d
1,1,ac
() =
Im([ m
()[
2
m
+
() +[ m
()[
2
m
+
())
[ m
+
() + m
()[
2
d. (2.150)
Choosing the new basis
(2.151)
_
u
+
(, n)
u
(, n)
_
= V ()
_
c(, n)
s(, n)
_
with
(2.152) V () =
1
m
+
() + m
()
_
1
a(0)
m
+
()
1
a(0)
m
()
_
we get
(2.153) (V
1
())
d
ac
() V
1
() =
1
_
Im( m
()) 0
0 Im( m
+
())
_
d.
Note that V () is not unitary. We will show how to diagonalize all of d in Sec
tion 3.1.
2.6. Some remarks on unbounded operators
In this section we temporarily drop the boundedness assumption on the coecients
a, b. This renders H unbounded and implies that we are no longer able to dene H
on all of
2
(Z). Nevertheless we can dene the minimal and maximal operator
allied with as follows
(2.154)
H
min
: D(H
min
)
2
(Z)
f f
,
H
max
: D(H
max
)
2
(Z)
f f
,
with
(2.155) D(H
min
) =
0
(Z), D(H
max
) = f
2
(Z)[f
2
(Z).
By Greens formula (1.20) we have H
min
= H
max
and
(2.156)
H
max
= H
min
: D(H
max
)
2
(Z)
f f
,
with
(2.157) D(H
max
) = f D(H
max
)[ lim
n
W
n
(f, g) = 0, g D(H
max
).
Here H
.
48 2. Foundations of spectral theory
Since we might have H
max
,= H
max
, H
max
might not be selfadjoint in general.
The key to this problem will be the limit point (l.p.), limit circle (l.c.) classi
cation alluded to in Remark 2.11 (ii). To make things precise, we call l.p. at
if s(z
0
, .) ,
2
(N) for some z
0
CR. Otherwise is called l.c. at .
In order to draw some rst consequences from this denition we note that
all considerations of Section 2.4 do not use boundedness of a, b except for (2.85).
However, if
nN
s(z
0
, n) < (considering only = for simplicity), then the
circle corresponding to m
N
(z
0
) converges to a circle instead of a point as N .
If m
(z
0
) is dened to be any point on this limiting circle, everything else remains
unchanged. In particular, u
+
(z
0
, .)
2
(N) and s(z
0
, .)
2
(N) shows that every
solution of u = z
0
u is in
2
(N) if is l.c. at +.
This enables us to reveal the connections between the l.p. / l.c. characterization
and the selfadjointness of unbounded Jacobi operators. We rst consider H
.
We recall that H
min,+
is symmetric and that (H
min,+
)
= H
max,+
. We need
to investigate the deciency indices d
= dimKer(H
max,+
z
), z
, of
H
min,+
. They are independent of z
= d
+
) since H
min,+
is real (cf. [192], [241], Chapter 8). Thus, to compute d
it suces to consider
Ker(H
max,+
z
0
). Since any element of Ker(H
max,+
z
0
) is a multiple of s(z
0
) we
infer d
= d
+
= 0 if s(z
0
) ,
2
(N) and d
= d
+
= 1 if s(z
0
)
2
(N). This shows
that H
max,+
is selfadjoint if and only if is l.p. at +. Moreover, s(z
0
)
2
(N)
implies s(z)
2
(N) for all z CR since d
z z
0
a(0)
n
j=n0+1
(c(z
0
, n)s(z
0
, j) c(z
0
, j)s(z
0
, n))u(j). (2.158)
By assumption, there exists a constant M 0 such that
(2.159)
j=n0+1
[c(z
0
, j)[
2
M,
j=n0+1
[s(z
0
, j)[
2
M.
2.6. Some remarks on unbounded operators 49
Invoking the CauchySchwarz inequality we obtain the estimate (n > n
0
)
j=n0+1
(c(z
0
, n)s(z
0
, j) c(z
0
, j)s(z
0
, n))u(j)
j=n0+1
[c(z
0
, n)s(z
0
, j) c(z
0
, j)s(z
0
, n)[
2
n
j=n0+1
[u(j)[
2
M
_
[c(z
0
, n)[
2
+[s(z
0
, n)[
2
_
n
j=n0+1
[u(j)[
2
. (2.160)
Since n
0
is arbitrary we may choose n
0
in (2.159) so large, that we have 4a(0)
1
[z
z
0
[M
2
1. Again using CauchySchwarz
n
j=n0+1
[u(j)[
2
([u(n
0
)[
2
+[u(n
0
+ 1)[
2
)
M +
2[z z
0
[
a(0)
M
2
n
j=n0+1
[u(j)[
2
([u(n
0
)[
2
+[u(n
0
+ 1)[
2
)
M +
1
2
n
j=n0+1
[u(j)[
2
, (2.161)
where
M is chosen such that
j=n0+1
[c(z
0
, j, n
0
)[
2
M,
j=n0+1
[s(z
0
, j, n
0
)[
2
M holds. Solving for the left hand side nishes the proof,
(2.162)
n
j=n0+1
[u(j)[
2
2([u(n
0
)[
2
+[u(n
0
+ 1)[
2
)
M.
nN
1
[a(n)[
2
[a(0)[
nN
[c(z, n)[
2
nN
[s(z, n)[
2
.
This shows that a sucient condition for to be l.p. at + is
(2.165)
nN
1
[a(n)[
= .
50 2. Foundations of spectral theory
The remaining question is: What happens in the l.c. case? Obviously we need
some suitable boundary conditions. The boundary condition
(2.166) BC
n0,
(f) = cos()f(n
0
) + sin()f(n
0
+ 1) = 0
of Remark 1.9 makes no sense if n
0
= . However, (2.166) can be written as
W
n0
(v, f) = 0, where v is any sequence satisfying BC
n0,
(v) = 0 and [v(n
0
)[ +
[v(n
0
+ 1)[ , = 0. Moreover, all dierent boundary conditions can be obtained by
picking v such that W
n0
(v, v) = 0 and W
n0
(v, f) ,= 0 for some f. This latter
characterization of the boundary condition may be generalized.
We dene the set of boundary conditions for at + by
(2.167) BC
+
() = v D(H
max,+
)[ W
+
(v, v) = 0, W
+
(v, f) ,= 0 for some
f D(H
max,+
) if is l.c. at +.
Observe that the rst requirement holds if v is real. The second is void if is l.p.
(at +). Otherwise, if is l.c., there is at least one real v for which it holds (if
not, (iii) of Lemma 2.16 implies that is l.p.). Two sequences v
1,2
BC
+
() are
called equivalent if W
+
(v
1
, v
2
) = 0.
Lemma 2.17. Let v BC
+
() and set
(2.168) D
+
(v) = f D(H
max,+
)[W
+
(v, f) = 0.
Then
(i) W
+
(v, f) = 0 W
+
(v, f) = 0,
(ii) W
+
(g, f) = 0 for f, g D
+
(v).
Moreover, W
+
(v
1
, v
2
) = 0 is equivalent to D
+
(v
1
) = D
+
(v
2
).
Proof. For all f
1
, . . . , f
4
D(H
max
) we can take the limits n +in the Pl ucker
identity
(2.169) W
n
(f
1
, f
2
)W
n
(f
3
, f
4
) +W
n
(f
1
, f
3
)W
n
(f
4
, f
2
) +W
n
(f
1
, f
4
)W
n
(f
2
, f
3
) = 0.
Now choose f
1
= v, f
2
=
f, f
3
= v, f
4
= f to conclude W
+
(v, f) = 0 implies
W
+
(v, f) = 0. Then choose f
1
= v, f
2
=
f, f
3
= f, f
4
= g to show (ii). The last
assertion follows from (ii) upon choosing v = v
1
, g = v
2
.
Combining this lemma with Greens formula (1.20) shows
Theorem 2.18. Choose v BC
+
(), then
(2.170) H
+
: D
+
(v)
2
(N)
f f
is a selfadjoint extension of H
min,+
.
In the l.p. case the boundary condition W
+
(v, f) = 0 is of course always
fullled and thus D
+
(v) = D(H
max,+
) for any v BC
+
().
Clearly, we can also dene selfadjoint operators H
n0,+
and H
n0,+
correspond
ing to H
+
as we did in Section 1.2 for the bounded case.
Now, that we have found selfadjoint extensions, let us come back to the Weyl
mfunctions. We x v(n) = s
+
(z) =
1
a(0)
lim
n+
W
n
(s
(), c
(z))
W
n
(s
(), s
(z))
2.6. Some remarks on unbounded operators 51
lies on the limiting circle. This is clear if is l.p. at +. Otherwise, l.c. at
+ implies c
(z), s
(z) D(H
max,+
) and both Wronskians converge to a limit
as pointed out earlier in this section. Moreover, if W
n
(s
(), s
(z)) = 0, then
s
(z) D(H
+
) and hence z (H
+
). In particular, W
n
(s
(), s
(z)) ,= 0 for
z CR and we can call m
+
(z) the Weyl mfunction of H
+
.
In addition, the function
(2.172) u
+
(z, n) =
c
(z, n)
a(0)
m
+
(z)s
(z, n)
is in
2
+
(Z) and satises the boundary condition
(2.173) W
+
(s
(), u
(z)) = 0.
The boundary condition uniquely characterizes u
+
(z, n) up to a constant in the l.c.
case.
We have seen that the question of selfadjointness is simple if is l.p.. One the
other hand, the spectrum gets simple if is l.c..
Lemma 2.19. If is l.c. at +, then the resolvent of H
+
is a HilbertSchmidt
operator. In particular, this implies that H
+
has purely discrete spectrum, (H
+
) =
d
(H
+
), and
(2.174)
d
(H+)
1
1 +
2
< .
Proof. The result is a consequence of the following estimate
(n,m)Z
2
[G
+
(z, m, n)[
2
=
1
[W[
2
nZ
_
[u
+
(z, n)[
2
m<n
[s(z, m)[
2
+[s(z, n)[
2
mn
[u
+
(z, m)[
2
_
2
[W[
2
u
+
(z)
2
s(z)
2
,
where W = W(s(z), u
+
(z)).
Our next goal is to nd a good parameterization of all selfadjoint extensions
if is l.c. at +.
First of all note that any real solution of u = u, R, is in BC
+
() (since
W(u, u) ,= 0 for any linearly independent solution of u = u). Now x
(2.175) v
1
)/a(0).
Lemma 2.20. All selfadjoint extensions of H
min,+
correspond to some v
with
unique [0, ).
Proof. Let H
+
be a selfadjoint extension of H
min,+
and
0
(H
+
) be an eigen
value with corresponding eigenfunction s(
0
, n). Using Greens formula (1.20) with
f = s(
0
), g D(H
+
) we see D(H
+
) D
+
(s(
0
)) and hence D(H
+
) = D
+
(s(
0
))
by maximality of selfadjoint operators. Let [0, ) be the unique value for
which
(2.176) W
+
(v
, s(
0
)) = cos()W
+
(c(0), s(
0
))+sin()W
+
(s(0), s(
0
)) = 0.
52 2. Foundations of spectral theory
Then D
+
(s(
0
)) = D
+
(v
).
Now we turn back again to operators on
2
(Z). We use that corresponding
denitions and results hold for the other endpoint as well. With this in mind
we have
Theorem 2.21. Choose v
BC
(v
, f) = 0
is selfadjoint.
Again, if is l.p. at , the corresponding boundary condition is void and
can be omitted. We also note that if is l.c. at both , then we have not found
all selfadjoint extensions since we only consider separated boundary conditions
(i.e., one for each endpoint) and not coupled ones which connect the behavior of
functions at and +.
As before we have
Lemma 2.22. If is l.c. at both , then the resolvent of H is a HilbertSchmidt
operator. In particular, the spectrum of H is purely discrete.
Most results found for bounded Jacobi operators still hold with minor modi
cations. One result that requires some changes is Theorem 2.13.
Theorem 2.23. A measure d
+
which is not supported on a nite set is the spectral
measure of a unique Jacobi operator H
+
if and only if the set of polynomials is dense
in L
2
(R, d
+
).
Proof. If the set of polynomials is dense in L
2
(R, d
+
) we can use the same proof
as in Theorem 2.13 to show existence of a unique Jacobi operator with d
+
as
spectral measure.
Conversely, let H
+
be given, and let U be a unitary transform mapping H
+
to multiplication by in L
2
(R, d
+
) (which exists by the spectral theorem). Then
[(U
1
)()[
2
= 1 since
(2.178)
_
d
+
=
1
, P
(H
+
)
1
= U
1
, P
()U
1
=
_
[(U
1
)()[
2
d
+
()
for any Borel set R. So, by another unitary transformation, we can assume
(U
1
)() = 1. And since the span of (H
+
)
j
1
, j N
0
, is dense in
2
(Z), so is the
span of (U(H
+
)
j
1
)() =
j
in L
2
(R, d
+
).
A measure d
+
which is not supported on a nite set and for which the set of
polynomials is dense in L
2
(R, d
+
) will be called Jacobi measure.
There are some interesting consequences for the moment problem.
Lemma 2.24. A set m
+,j
jN
forms the moments of a Jacobi measure if and
only if (2.109) holds.
Moreover,
(2.179) supp(
+
) [R, R] [m
+,j
[ R
j
, j N.
2.6. Some remarks on unbounded operators 53
Proof. If (2.109) holds we get sequences a(n), b(n) by (2.118). The spectral mea
sure of any selfadjoint extension has m
+,j
as moments.
If supp(
+
) [R, R], then [m
+,j
[
_
[[
j
d
+
() R
j
_
d
+
() = R
j
. Con
versely, if supp(
+
) , [R, R], then there is an > 0 such that C
=
+
([[[ >
R +) > 0 and hence [m
+,2j
[
_
]]>R+
2j
d
+
() C
(R +)
2j
.
Finally, let us look at uniqueness.
Theorem 2.25. A measure is uniquely determined by its moments if and only if
the associated Jacobi dierence expression (dened via (2.118)) is l.p. at +.
Proof. Our assumption implies that H
min,+
is essentially selfadjoint and hence
(H
min,+
z)D(H
min,+
) is dense in
2
(N) for any z C
. Denote by H
0
the set
of polynomials on R and by H the closure of H
0
with respect to the scalar product
(2.110). Then (z)H
0
is dense in H and hence there is a sequence of polynomials
P
z,n
(), z C
L
2 = lim
n
1, P
z,n
jN
is l.p. at +.
Proof. Our estimate implies that e
iz
L
1
(R, d) for [Im(z)[ < 1/R. Hence the
Fourier transform
(2.183)
_
R
e
iz
d() =
j=0
m
+,j
(iz)
j
j!
is holomorphic in the strip [Im(z)[ < 1/R. This shows that the Fourier transform
is uniquely determined by the moments and so is the Borel transform and hence
the measure (see (B.9)).
54 2. Foundations of spectral theory
2.7. Inverse spectral theory
In this section we present a simple recursive method of reconstructing the sequences
a
2
, b when the Weyl matrix (cf. (2.142))
(2.184) M(z, n) =
_
g(z, n)
h(z,n)
2a(n)
h(z,n)
2a(n)
g(z, n + 1)
_
, z C(H),
is known for one xed n Z. As a consequence, we are led to several uniqueness
results.
By virtue of the Neumann series for the resolvent of H we infer (cf. (6.2) below
and Section 6.1 for more details)
g(z, n) =
1
z
b(n)
z
2
+O(
1
z
3
),
h(z, n) = 1
2a(n)
2
z
2
+O(
1
z
3
). (2.185)
Hence a(n)
2
, b(n) can be easily recovered as follows
b(n) = lim
z
z(1 +zg(z, n)),
a(n)
2
=
1
2
lim
z
z
2
(1 +h(z, n)). (2.186)
Furthermore, we have the useful identities (use (1.100))
(2.187) 4a(n)
2
g(z, n)g(z, n + 1) = h(z, n)
2
1
and
(2.188) h(z, n + 1) +h(z, n) = 2(z b(n + 1))g(z, n + 1),
which show that g(z, n) and h(z, n) together with a(n)
2
and b(n) can be determined
recursively if, say, g(z, n
0
) and h(z, n
0
) are given.
In addition, we infer that a(n)
2
, g(z, n), g(z, n+1) determine h(z, n) up to one
sign,
(2.189) h(z, n) =
_
1 + 4a(n)
2
g(z, n)g(z, n + 1)
_
1/2
,
since h(z, n) is holomorphic with respect to z C(H). The remaining sign can
be determined from the asymptotic behavior h(z, n) = 1 +O(z
2
).
Hence we have proved the important result that M(z, n
0
) determines the se
quences a
2
, b. In fact, we have proved the slightly stronger result:
Theorem 2.27. One of the following set of data
(i) g(., n
0
) and h(., n
0
)
(ii) g(., n
0
+ 1) and h(., n
0
)
(iii) g(., n
0
), g(., n
0
+ 1), and a(n
0
)
2
for one xed n
0
Z uniquely determines the sequences a
2
and b.
Remark 2.28. (i) Let me emphasize that the two diagonal elements g(z, n
0
) and
g(z, n
0
+ 1) alone plus a(n
0
)
2
are sucient to reconstruct a(n)
2
, b(n). This is in
contradistinction to the case of onedimensional Schrodinger operators, where the
diagonal elements of the Weyl matrix determine the potential only up to reection.
2.7. Inverse spectral theory 55
You might wonder how the Weyl matrix of the operator H
R
associated with
the (at n
0
) reected coecients a
R
, b
R
(cf. Lemma 1.7) look like. Since reection
at n
0
exchanges m
(z, n
0
) (i.e., m
R,
(z, n
0
) = m
(z, n
0
)) we infer
g
R
(z, n
0
) = g(z, n
0
),
h
R
(z, n
0
) = h(z, n
0
) + 2(z b(n
0
))g(z, n
0
),
g
R
(z, n
0
+ 1) =
a(n
0
)
2
a(n
0
1)
2
g(z, n
0
+ 1) +
z b(n
0
)
a(n
0
1)
2
_
h(z, n
0
)
+ (z b(n
0
))g(z, n
0
)
_
, (2.190)
in obvious notation.
(ii) Remark 6.3(ii) will show that the sign of a(n) cannot be determined from either
g(z, n
0
), h(z, n
0
), or g(z, n
0
+ 1).
The odiagonal Green function can be recovered as follows
(2.191) G(z, n +k, n) = g(z, n)
n+k1
j=n
1 +h(z, j)
2a(j)g(z, j)
, k > 0,
and we remark
a(n)
2
g(z, n + 1) a(n 1)
2
g(z, n 1) + (z b(n))
2
g(z, n)
= (z b(n))h(z, n). (2.192)
A similar procedure works for H
+
. The asymptotic expansion
(2.193) m
+
(z, n) =
1
z
b(n + 1)
z
2
a(n + 1)
2
+b(n + 1)
2
z
3
+O(z
4
)
shows that a(n + 1)
2
, b(n + 1) can be recovered from m
+
(z, n). In addition, (2.11)
shows that m
+
(z, n
0
) determines a(n)
2
, b(n), m
+
(z, n), n > n
0
. Similarly, (by
reection) m
(z, n
0
) determines a(n1)
2
, b(n), m
(z, n
0
) determine a(n)
2
, b(n) except for a(n
0
1)
2
, a(n
0
)
2
, b(n
0
). However,
since a(n
0
1)
2
, a(n
0
)
2
, b(n
0
), and m
(z, n
0
) can be computed from m
(z, n
0
) we
conclude:
Theorem 2.29. The quantities m
+
(z, n
0
) and m
(z, n
0
) uniquely determine a(n)
2
and b(n) for all n Z.
Next, we recall the function
(z, n) =
a(n)
1 +
2
z
b(n + 1) + 2a(n) +
2
b(n)
z
2
+O(
1
z
3
).
Our goal is to prove
Theorem 2.30. Let
1
,
2
R with
1
,=
2
be given. Then
j
(., n
0
),
j = 1, 2, for one xed n
0
Z uniquely determine a(n)
2
, b(n) for all n Z (set
1
)
2
,
g(z, n + 1) =
2
2
1
(z, n) +
2
1
2
(z, n) + 2
1
2
R(z)
(
2
1
)
2
,
h(z, n) =
2
1
(z, n) +
1
2
(z, n) + (
1
+
2
)R(z)
(2a(n))
1
(
2
1
)
2
, (2.195)
where R(z) is the branch of
(2.196) R(z) =
_
(
2
1
)
2
4a(n)
2
+
1
(z, n)
2
(z, n)
_
1/2
=
1
+
2
2a(n)
+O(
1
z
),
which is holomorphic for z C(H) and has asymptotic behavior as indicated. If
one of the numbers
1
,
2
equals , one has to replace all formulas by their limit
using g(z, n) = lim
(z, n).
Proof. Clearly, if (
1
,
2
) ,= (0, ), (, 0), we can determine a(n) from equation
(2.194). Hence by Theorem 2.27 it suces to show (2.195). Since the rst equation
follows from (2.187) and the other two, it remains to establish the last two equations
in (2.195). For this we prove that the system
(2.197) (g
+
)
2
+ 2
j
2a(n)
hg
+
+
2
j
4a(n)
2
(h
2
1) = g
+
j
(z, n), j = 1, 2,
has a unique solution (g
+
, h) = (g(z, n + 1), h(z, n)) for [z[ large enough which is
holomorphic with respect to z and satises the asymptotic requirements (2.185).
We rst consider the case
j
,= 0, . Changing to new variables (x
1
, x
2
), x
j
=
(2a(n)/
j
)g
+
+h, our system reads
(2.198) x
2
j
1 =
1
2
j
2a(n)
j
(z, n)
1
(x
1
x
2
), j = 1, 2.
Picking [z[ large enough we can assume
j
(z, n) ,= 0 and the solution set of the new
system is given by the intersection of two parabolas. In particular, (2.197) has at
most four solutions. Two of them are clearly g
+
= 0, h = 1. But they do not have
the correct asymptotic behavior and hence are of no interest to us. The remaining
two solutions are given by the last two equations of (2.195) with the branch of R(z)
arbitrarily. However, we only get correct asymptotics (g
+
= z
1
+ O(z
2
) resp.
h = 1 + O(z
2
)) if we x the branch as in (2.196). This shows that g(z, n + 1),
h(z, n) can be reconstructed from
1
,
2
and we are done. The remaining cases
can be treated similarly.
Finally, we want to give an alternative characterization of the sequences g(z),
h(z) respectively m
(z) +
1
m
(z)
= (z b),
m
+
(z) = z
1
+O(z
2
)
m
(z) = a
2
z +O(z
0
)
.
Then (i) (resp. (ii)) is necessary and sucient for g(z), h(z) (resp. m
(z)) to be
the corresponding coecients of the Weyl Mmatrix (resp. the Weyl mfunctions)
of H.
Moreover, if g(z), h(z) and m
(z, n)
, g(z, n + 1) =
m
+
(z, n) m
(z, n)
m
+
(z, n) + m
(z, n)
,
h(z, n) =
m
+
(z, n) m
(z, n)
m
+
(z, n) + m
(z, n)
, (2.201)
respectively
(2.202) m
(z, n) =
1 h(z, n)
2a(n)
2
g(z, n)
=
2g(z, n + 1)
1 h(z, n)
,
then one condition implies the other.
Proof. Necessity has been established earlier in this section. The relation between
the two sets of data is straightforward and hence it suces to consider (e.g.) m
(z).
Since m
(z) are both meromorphic near they are uniquely determined by their
Laurent expansion around . But the coecients of the Laurent expansion can
be determined uniquely using the recursion relations for m
i
m
+
(
0
+ i) > 0.
59
60 3. Qualitative theory of spectra
Now let us relate the spectrum to the asymptotic behavior of s(, n). Let us
rst show that s(, n) does not grow too fast for in the spectrum. More precisely,
we have
Lemma 3.1. Let f
2
(N), then the set
(3.3) S
f,+
() = R[f(n)s(, n)
2
(N),
is of full H
+
spectral measure, that is,
+
(S
f,+
()) = 1. The same is true for the
set
(3.4) S
+
() = R[ liminf
n
1
n
n
m=1
[s(, m)[
2
< .
Proof. Since the polynomials s(, n) are orthonormal, one veries
(3.5) f
2
=
_
R
nN
[f(n)s(, n)[
2
d
+
().
Hence
nN
[f(n)s(, n)[
2
< a.e. with respect to d
+
. Similarly, by Fatous
lemma we have
(3.6)
_
R
liminf
n
1
n
n
m=1
[s(, m)[
2
d
+
() < liminf
n
_
R
1
n
n
m=1
[s(, m)[
2
d
+
() = 1,
which implies the second claim.
Choosing f(n) = n
1/2
, > 0, we see for example
(3.7) sup
nN
[s(, n)[
n
1/2+
< , a.e. d
+
.
As a consequence we obtain
Theorem 3.2. Let > 0, then
(H
+
) = S
+
() =
_
R[ sup
nN
[s(, n)[
n
1/2+
<
_
=
_
R[ limsup
n
1
n
ln(1 +[s(, n)[) = 0
_
. (3.8)
The spectrum of H
+
is simple.
Proof. Denote the last two sets (without closure) in (3.8) by S
1
, S
2
. Then we need
to show (H
+
) = S
+
= S
1
= S
2
. By equation (2.47) we have S
+
, S
1
S
2
(H
+
)
and since S
+,1,2
are all of full spectral measure, the claim follows as in the proof of
(B.27).
That the spectrum of H
+
is simple is immediate from Theorem 2.12.
Now let us turn to H. Here we have (H) if and only if g(z, 0) +g(z, 1) is
not holomorphic in z near and
(3.9) (H) = R[ lim
0
Im(g( + i, 0) +g( + i, 1)) > 0.
To obtain again a characterization in terms of solutions, we need the analog of
Lemma 3.1.
3.1. Locating the spectrum and spectral multiplicities 61
Lemma 3.3. Let f
2
(Z), then the set
(3.10) S
f
() = R[f(n)s
(, n)
2
(Z) for some = ()
is of full H spectral measure, that is,
tr
(S
f
()) = 1. The same is true for the set
(3.11) S() = R[ liminf
n
1
n
n
m=1
[s
(, m)[
2
< for some = ().
Proof. We use U() from (B.62) which diagonalizes d and change to the new pair
of solutions
(3.12) U(, n) = U()S(, n).
Now, proceeding as in the proof of Lemma 3.1, we obtain
(3.13) f =
_
R
_
r
1
()
nZ
[f(n)U
1
(z, n)[
2
+r
2
()
nZ
[f(n)U
2
(z, n)[
2
_
d
tr
().
Hence for a.e. with respect to d
tr
we have
nZ
r
i
()[f(n)U
i
(, n)[
2
< for
i = 1, 2. Since r
1
() +r
2
() = 1, r
i
()U
i
(, n) is nonzero for one i. The real part is
the required solution. The second part follows from Fatous lemma as in the proof
of the previous lemma.
As before, choosing f(n) = (1 + [n[)
1/2
, > 0, we obtain the following
result.
Theorem 3.4. Let > 0, then
(H) = S() =
_
R[ sup
nZ
[s
(, n)[
n
1/2+
< for some = ()
_
=
_
R[ limsup
n
1
n
ln(1 +[s
i
(g( + i, 0) +g( + i, 1)) > 0.
At each discrete eigenvalue
d
(H) either m
+
(z, n) and m
(z, n) + m
+
(z, n) = 0.
(ii). It is interesting to know which parts of the spectrum of H can be read o
from d
0,0
alone, that is, from g(z, 0). Since g(z, 0) = a(0)
2
( m
(z) + m
+
(z))
1
,
we know
ess
(H
)
ess
(H
+
) and it will follow from Lemma 3.7 that g(z, 0) de
termines the essential spectrum of H. Moreover, if
d
(H), we either have
u
(, 0), u
+
(, 0) ,= 0 and hence m
(z, n) + m
+
(z, n) = 0 (implying a pole of
g(z, 0) at z = ) or u
(, 0), u
+
(, 0) = 0 and hence poles for both m
(z, n) (im
plying a zero of g(z, 0) at z = ). Summarizing, g(z, n) determines (H) except for
the discrete eigenvalues which are also (discrete) simultaneous eigenvalues of H
and H
+
.
62 3. Qualitative theory of spectra
To say more about the spectral multiplicity of H, we need to investigate the
determinant of R = d/d
tr
. By (B.61) we infer
(3.16) R
i,j
() = lim
0
Im(M
i,j
( + i))
Im(g( + i, 0) +g( + i, 1))
.
Hence, using (2.143) we see
det R() =
lim
0
Im( m
+
( + i))Im( m
( + i))
a(0)
2
[ m
+
( + i) + m
( + i)[
2
1
Im(g( + i, 0) +g( + i, 1))
2
(3.17)
Observe that the rst factor is bounded by (2a(0))
2
. Now Lemma B.14 immedi
ately gives
Lemma 3.6. The singular spectrum of H has spectral multiplicity one. The ab
solutely continuous spectrum of H has multiplicity two on
ac
(H
+
)
ac
(H
) and
multiplicity one on
ac
(H)(
ac
(H
+
)
ac
(H
)).
Proof. Recall from Appendix B that the set
(3.18) M
s
= R[ lim
0
Im(g( + i, 0) +g( + i, 1)) =
is a minimal support for the singular spectrum (singular continuous plus pure point
spectrum) of H. Hence the singular part of H is given by H
s
= H
pp
H
sc
=
P
Ms
(H)H and the absolutely continuous part is given by H
ac
= (1 P
Ms
(H))H.
So we see that the singular part has multiplicity one by Lemma B.14.
For the absolutely continuous part use that
(3.19) M
ac,
= R[0 < lim
0
Im( m
( + i)) <
are minimal supports for the absolutely continuous spectrum of H
. Again the
remaining result follows from Lemma B.14.
3.2. Locating the essential spectrum
The objective of this section is to locate the essential spectrum (which by the
way is always nonempty since H is bounded) for certain classes of Jacobi operators.
Since our operators have spectral multiplicity at most two, the essential spectrum
is the set of all accumulation points of the spectrum.
We rst show that it suces to consider H
. In fact, even H
+
is enough since
the corresponding results for H
n0
into
2
(Z) by dening (e.g.) H
n0
u(n
0
) = 0, we see from
(1.106) that the kernel of the resolvent corresponding to this new operator
H
n0
is
given by
(3.20)
G
n0
(z, n, m) = G(z, n, m)
G(z, n, n
0
)G(z, n
0
, m)
G(z, n
0
, n
0
)
1
z
n0
(n)
n0
(m).
Hence
H
n0
is a rank two resolvent perturbation of H implying
ess
(H) =
ess
(
H
n0
).
However, the last term in (3.20) is clearly articial and it is more natural to consider
H
n0
as rank one perturbation.
3.2. Locating the essential spectrum 63
Lemma 3.7. For any Jacobi operator H we have
(3.21)
ess
(H) =
ess
(H
0
) =
ess
(H
)
ess
(H
+
)
and, in addition,
ess
(H
) =
ess
(H
,n0
) for any n
0
Z, R .
Moreover, for
0
<
1
we have
(3.22)
dimRan P
(0,1)
(H
n0
) 1 dimRan P
(0,1)
(H) dimRan P
(0,1)
(H
n0
) + 1.
Proof. The rst part follows since nite rank perturbations have equivalent essen
tial spectra.
For the second claim, there is nothing to prove if (
0
,
1
)
ess
(H) ,= . If this
is not the case, let us view g(., n
0
) as a function from (
0
,
1
) R
= S
1
.
Then g(, n
0
) = implies that is a simple eigenvalue of H but not of H
n0
. On
the other hand, if g(, n
0
) = 0, then either is a single eigenvalue of H
n0
(but not
of H) or a double eigenvalue of H
n0
and a single eigenvalue of H. Combining this
observation with the fact that g(., n
0
) is monotone (cf. equation (2.37)) nishes the
proof.
Note that (H
+
) (H
+
) =
ess
(H
+
) for ,= .
The following lemma shows that is in the essential spectrum of H if there
exists a non square summable solution which does not grow exponentially. This is
a strengthening of (2.48).
Lemma 3.8. Suppose there is a solution of u = u such that
(3.23) lim
k
1
n
k
ln(u(2n
k
)
2
+u(2n
k
1)
2
) = 0
for some increasing sequence 0 < n
k
< n
k+1
. Then
ess
(H
+
). Similarly for
H
.
Proof. Normalizing u(0)
2
+ u(1)
2
= 1 we have u(n) = s
(, n) (up to sign).
Altering b(1) we can assume = and shifting b(n) it is no restriction to set
= 0 (i.e., u(n) = s(0, n)). We will try to construct a Weyl sequence (cf. [192],
Thm. VII.12) as follows. Set
(3.24) u
N
(n) =
_
s(0,n)
N
j=1
s(0,j)
2
, n N
0 , n > N
.
Then u
N
 = 1 and
H
+
u
N
(n) =
1
_
N
j=1
s(0, j)
2
_
a(N)s(0, N)
N
(n)
+ (a(N 1)s(0, N 1) +b(N)s(0, N))
N+1
(n)
_
. (3.25)
Hence
(3.26) H
+
u
2N

2
const
f(N)
N
j=1
f(j)
,
64 3. Qualitative theory of spectra
where f(n) = s(0, 2n)
2
+s(0, 2n1)
2
. If the right hand side does not tend to zero,
we must have
(3.27) f(n)
n
j=1
f(j), > 0,
and hence f(n) (1 )
1n
. But this contradicts our assumption (3.23)
(3.28) 0 = lim
k
1
n
k
ln f(n
k
) lim
k
1 n
k
n
k
ln(1 ) = ln(1 ) > 0.
Hence H
+
u
2N
 0 implying 0 (H
+
) by Weyls theorem. If 0 were a discrete
eigenvalue, s(0, n) would decrease exponentially by Lemma 2.5 and the limit in
(3.23) would be negative. So 0
ess
(H
+
) as claimed.
Next, we note that the essential spectrum depends only on the asymptotic
behavior of the sequences a(n) and b(n).
Lemma 3.9. Let H
1,2
be two Jacobi operators and suppose
(3.29) lim
n
([a
2
(n)[ [a
1
(n)[) = lim
n
(b
2
(n) b
1
(n)) = 0.
Then
(3.30)
ess
(H
1,+
) =
ess
(H
2,+
).
Proof. Clearly we can assume a
1,2
(n) > 0. The claim now follows easily from the
fact that, by assumption, H
2,+
H
1,+
is compact (approximate H
2,+
H
1,+
by
nite matrices and use that norm limits of compact operators are compact).
Finally, let us have a closer look at the location of the essential spectrum. We
abbreviate
(3.31) c
= liminf
n
c
(n), c
+
= limsup
n
c
+
(n).
Lemma 3.10. We have
(3.33)
ess
(H
+
) [c
, c
+
].
If, in addition, a(n), b(n) are slowly oscillating, that is,
(3.34) lim
n
([a(n + 1)[ [a(n)[) = 0, lim
n
(b(n + 1) b(n)) = 0,
then we have
(3.35)
ess
(H
+
) = [c
, c
+
].
Proof. Clearly we can assume a(n) > 0. But we may even assume c
+
(n)
c
+
and c
(n) c
. In fact, consider
H
+
associated with the sequences a(n) =
mina(n), c
+
a(n1)b(n), c
a(n1)+b(n),
b(n) = b(n). Since a(n) a(n) =
max0, c
+
(n) c
+
, c
(n) we infer
ess
(H
+
) =
ess
(
H
+
) from Lemma 3.9.
3.3. Locating the absolutely continuous spectrum 65
For the rst claim we use
f, H
+
f = a(0)[f(1)[
2
+
n=1
_
a(n)[f(n + 1) f(n)[
2
+ (a(n 1) +a(n) +b(n))[f(n)[
2
_
(3.36)
and proceed as in Lemma 1.8, proving (3.33).
Suppose (3.34) holds and x (c
, c
+
). To prove (3.35) we need to show
ess
(H
+
). By Weyls criterion (cf. [192], Thm. VII.12) it suces to exhibit an
orthonormal sequence u
N
(N) with (H
+
)u
N
 0 as N . For each
N N we can nd n
N
> N(N 1)/2 such that
(i). there is a
N
[c
(n
N
), c
+
(n
N
)] such that [
N
[ 1/
N and
(ii). [a(n) a
N
[ + [a(n 1) a
N
[ + [b(n) b
N
[ 1/
N for n
N
n n
N
+ N,
where a
N
= (a(n
N
) +a(n
N
1))/2), b
N
= b(n
N
).
If we set
(3.37) k
N
=
N
b
N
+
_
(
N
b
N
)
2
4a
2
N
2a
N
,
we have [k
N
[ = 1 since [
N
b
N
[ 2a
N
. Now dene u
N
(n) = (k
N
)
n
/
N for
n
N
n < n
N
+N and u
N
(n) = 0 otherwise. Then u
N
 = 1 and u
N
(n)u
M
(n) = 0,
N ,= M. Moreover, we have
(H
+
)u
N
 (H
N,+
N
)u
N
 +(
N
)u
N
 +(H
+
H
N,+
)u
N

4a
N
+
1
N
+
1
N
0 (3.38)
as N , where in the last estimate we have used
N
(k
N
)
n
=
N
(k
N
)
n
, (i), and
(ii), respectively. This proves
ess
(H
+
) and hence (3.35).
3.3. Locating the absolutely continuous spectrum
In this section we will try to locate the absolutely continuous spectrum of H in
some situations. We will rst investigate the half line operator H
+
whose spectral
properties are related to the boundary behavior of m
+
(+i) as 0 (see Appen
dix B). Our strategy in this section will be to relate this boundary behavior to the
behavior of solutions of u = u.
But rst we start with a preliminary lemma which shows that the absolutely
continuous spectrum of H is completely determined by those of H
and H
+
.
Lemma 3.11. For any n
0
Z, R, the absolutely continuous part of the
spectral measures corresponding to H
and H
,n0
, respectively, are equivalent. For
xed n
0
and dierent , the singular parts are mutually disjoint. In particular,
(3.39)
ac
(H
) =
ac
(H
,n0
).
The same is true for the trace measure d
tr
corresponding to H and d
+
+d
. In
particular,
(3.40)
ac
(H) =
ac
(H
)
ac
(H
+
).
66 3. Qualitative theory of spectra
In addition, let us abbreviate g(, n) = limsup
0
Im(g( + i, n)), m
(, n) =
limsup
0
Im(m
(, n
0
)) (3.41)
are minimal supports for the absolutely continuous spectrum of H.
Proof. Using (2.16) and (i) of Theorem B.8 we infer
(3.42) d
+,n0,ac
() =
2
[ a(n
0
)m
+
(, n
0
)[
2
d
+,n0,ac
()
and since H
+,n0
= H
0
+,n0+1
the equivalence of the absolutely continuous parts of
the spectral measures follows for the + sign. That the singular parts are disjoint
follows from (2.92) since m
1
+
() = implies m
2
+
() ,= for all
2
,=
1
. The
part is similar.
It remains to establish the claims for H. It is no restriction to set n
0
= 0.
By Theorem B.8 (ii) the set L of all R for which both limits lim
0
m
( +
i) exist and are nite is of full Lebesgue measure. For L we can compute
lim
0
M(+i) from (2.143). Hence we obtain M
ac,1
L = M
ac,2
L = M
ac,3
L.
Moreover, M
ac,1
L = [0 < Im(g(, 0) + g(, 1)) and the claim is immediate
from Lemma B.7 and Lemma B.12.
In particular, this says that if limsup
0
Im(m
+
( + i)) < , (
0
,
1
),
then the spectrum of H
+
is purely absolutely continuous in (
0
,
1
). Note that we
could replace m
(z, n) by m
(
+
) of
d
+,c
can be found by investigating the boundary values of m
+
( + i) as 0.
We set
(3.43) f
(1,n)
=
_
n
j=1
[f(j)[
2
and extend this denition for real n 1 by linear interpolation of f
2
(1,n)
. That
is, we have f
2
(1,n)
=
_
n
0
[f([[x]] + 1)[
2
dx, where [[.]] denotes the integer part (see
(4.14)).
Lemma 3.12. Let = a(0)(2s()
(1,n)
c()
(1,n)
)
1
, then
(3.44) 5
24 [a(0)m
+
( + i)[
s()
(1,n)
c()
(1,n)
5 +
24.
3.3. Locating the absolutely continuous spectrum 67
Proof. Let n N, then by (2.158)
u
+
( + i, n) =
c(, n)
a(0)
m
+
( + i)s(, n)
i
a(0)
n
j=1
(c(, n)s(, j) c(, j)s(, n))u
+
( + i, j). (3.45)
Hence one obtains after a little calculation (as in the proof of Lemma 2.15)
u
+
( + i)
(1,n)

c()
a(0)
m
+
( + i)s()
(1,n)
2s()
(1,n)

c()
a(0)

(1,n)
u
+
( + i)
(1,n)
. (3.46)
Using the denition of and (2.89) we obtain

c()
a(0)
m
+
( + i)s()
2
(1,n)
4u
+
( + i)
2
(1,n)
4u
+
( + i)
2
(1,)
=
4
Im(m
+
( + i))
8s()
(1,n)

c()
a(0)

(1,n)
Im(m
+
( + i)). (3.47)
Combining this estimate with
(3.48) 
c()
a(0)
m
+
( + i)s()
2
(1,n)
_

c()
a(0)

(1,n)
[m
+
( + i)[s()
(1,n)
_
2
shows (1 x)
2
8x, where x = [a(0)m
+
( + i)[s()
(1,n)
c()
1
(1,n)
.
We now introduce the concept of subordinacy, [0, 1]. A nonzero solu
tion u of u = zu is called subordinate at +with respect to another solution
v if
(3.49) liminf
n+
u
2
(1,n)
v
(1,n)
= 0.
Moreover, u is called subordinate at + if (3.49) holds for for any linearly
independent solution v.
An
1
subordinate solution is also an
2
subordinate solution for any
1
2
and if there is no
2
subordinate solution, then there is also no
1
subordinate
solution.
For example, if u
2
(
+
) if s() is not 
subordinate with respect to c(). If [0, 1), the converse is also true.
In particular, if no subordinate solution at + exists for any (
1
,
2
),
then
+
is purely continuous on (
1
,
2
), that is, [
1
,
2
]
c
(H
+
) and (
1
,
2
)
s
(H
+
) = , (0, 1].
The case = 1 deserves special attention, since 1continuous is the same as
absolutely continuous. We will call u subordinate with respect to v if
(3.50) lim
n+
u
(1,n)
v
(1,n)
= 0.
We use the ordinary limit (instead of liminf) in the case = 1 for convenience
later on. We will infer from Lemma 3.14 below that, for our purpose, this makes
no dierence.
It is easy to see that if u is subordinate with respect to v, then it is subordinate
with respect to any linearly independent solution. In particular, a subordinate
solution is unique up to a constant. Moreover, if a solution u of u = u, R, is
subordinate, then it is real up to a constant, since both the real and the imaginary
part are subordinate. For z CR we know that there is always a subordinate
solution near , namely u
() is subordinate or m
+
( + i0) = /a(0), we can reduce it to the rst
case using the transform b(1) b(1) a(0)/ (see (2.16)).
We are interested in N
. In particular,
(3.53)
ac
(H
) = N
()
ess
.
Proof. By Theorem B.8 (ii) we may assume m
+
(+i0) exists and is nite almost
everywhere. But for those values we have 0 < Im(m
) B
f,
()
ess
.
Choosing f(n) = (1 +[n[)
1/2
, > 0, we even see
(3.56)
ac
(H
) R[
() = 0
ess
.
In addition, let
(3.57)
B
() = R[ liminf
n
1
n
n
m=1
[u(m)[
2
< for all solutions of u = u,
then B
() N
() and
(3.58)
ac
(H
) = B
()
ess
.
Proof. We only consider the + sign. By Lemma 3.1 the sets
(3.59) S
f,
() = R[f(n)s
(, n)
2
(N)
are of full H
f,+
()S
f,+
() if ,=
. Moreover, since d
+,ac
and d
+,ac
are equivalent, we infer
+,ac
(RS
f,+
()) =
+,ac
(RS
f,+
()) = 0. Thus S
f,+
() and S
f,+
() are both supports for d
+,ac
and
so is their intersection.
The second assertion follows since S
f,+
() implies
+
() = 0 for our
particular choice of f.
Finally, as before B
() is a support for
ac
(H
() N
(). Suppose B
+
() is such
that a subordinate solution u exists, that is, lim
n+
u
(1,n)
/v
(1,n)
= 0. By
assumption, there is a subsequence n
j
such that n
1
j
v
(1,nj)
is bounded. Hence
n
1
j
u
(1,nj)
must tend to zero. But this contradicts constancy of the Wronskian
(3.60) 0 < [W(u, v)[ =
1
n
j
nj
m=1
[W
n
(u, v)[ const
u
(1,nj+1)
n
j
v
(1,nj+1)
n
j
0
(use CauchySchwarz). The case B
() is similar.
Now, we want to extend our results to operators H on
2
(Z). The spectrum
of H is related to the boundary behavior of g( + i, n), n = 0, 1, which again is
related to the boundary behavior of m
( + i) by equation (2.201).
Lemma 3.17. Suppose liminf
0
Im(m
(+i)[ <
for (
1
,
2
), = or = +. Then [
1
,
2
]
ac
(H) and (
1
,
2
)
s
(H) =
. Here
s
(H) =
sc
(H)
pp
(H).
Proof. Consider (e.g.) = +. Our assumption implies that [
1
,
2
]
ess
(H
+
)
ess
(H). Moreover, using (2.201) one easily shows
(3.61) limsup
0
Im(g( + i, n)) < , n = 0, 1.
70 3. Qualitative theory of spectra
Together with Lemma B.7 this implies that the singular spectrum of H is not
supported on (
1
,
2
).
This last result is the key to our next theorem.
Theorem 3.18. On every subinterval (
1
,
2
) N
is
purely absolutely continuous in the sense that
(3.62) [
1
,
2
]
ac
(H
), (
1
,
2
)
s
(H
) = .
In addition, the spectrum of H is purely absolutely continuous on every subin
terval (
1
,
2
) N
() N
+
(), that is,
(3.63) [
1
,
2
]
ac
(H), (
1
,
2
)
s
(H) = .
Proof. Clearly Lemma 3.12 (and all other considerations) holds for any other
boundary condition . Hence only the claim concerning H needs to be proven.
If N
+
(), we know limsup
0
[ m
+
( + i)[ = C() < and we need to show
liminf
0
Im( m
+
( + i)) > 0. Suppose that we can nd a sequence
n
0 such
that Im( m
+
( + i
n
)) 0 as n . Now set m
+
( + i
n
) = x
n
+ iy
n
and
n
= a(0)x
n
. Since limsup [x
n
[ < we can pass to a subsequence such that x
n
converges. This implies (x
n
, y
n
,
n
) (x
0
, 0,
0
= a(0)x
0
) and by (2.92)
[ m
0
+
( + i
n
)[ =
1
[a(0)[
1 +a(0)
0
(x
n
+ iy
n
)
0
a(0)(x
n
+ iy
n
)
1
[a(0)[
[1 +
0
n
[
_
(
0
n
)
2
+y
2
n
(3.64)
contradicting our assumption. The case N
() is similar.
The relation between the absolutely continuous spectrum and (non)existence
of subordinate solutions is often referred to as principle of subordinacy.
Finally, let us show how these results can be applied. Note that B
() if
all solutions of u = u are bounded near .
Theorem 3.19. Suppose the sequences a(n) > 0, b(n) satisfy
(3.65)
nZ
[a(n + 1) a(n)[ +[b(n + 1) b(n)[ <
and set
(3.66) lim
n
a(n) = a
, lim
n
b(n) = b
.
Then for any (2a
+b
, 2a
+b
ess
(H
) =
ac
(H
) = [2a
+b
, 2a
+b
],
sc
(H
) = ,
pp
(H
) (, 2a
+b
] [2a
+b
, ). (3.67)
Moreover,
ess
(H) =
ac
(H) =
ac
(H
+
)
ac
(H
),
sc
(H) = ,
pp
(H) R
ess
(H). (3.68)
3.3. Locating the absolutely continuous spectrum 71
Proof. After the transformation z 2a
+
z +b
+
we can assume a
+
= 1/2, b
+
= 0.
Consider the quantity
(3.69) K(n) = a(n)(u(n + 1)
2
+u(n)
2
) + (b(n) )u(n)u(n + 1),
where u is a solution of u = u. For (1, 1) we can nd N N, > 0 such
that
(3.70)
b(n)
2a(n)
1 , n N.
Whence
K(n)
a(n)
u(n + 1)
2
+u(n)
2
(1 )2u(n)u(n + 1)
= (u(n + 1)
2
+u(n)
2
) + (1 )(u(n + 1) u(n))
2
(u(n + 1)
2
+u(n)
2
). (3.71)
Thus to show u bounded near + it suces to show this result for K. One
computes
K(n + 1) K(n) =
a(n + 1) a(n)
a(n + 1)
(K(n) + (a(n + 1) +a(n))u(n + 1)
2
)
+
a(n)(b(n + 1) b(n))
a(n + 1)
u(n)u(n + 1) (3.72)
which implies using (3.71)
(3.73) K(n + 1) (1 +C(n))K(n), C(n)
1
(Z),
where
(3.74) C(n) =
[a(n + 1) a(n)[
a(n + 1)
(1 +
a(n + 1) +a(n)
a(n)
) +
[b(n + 1) b(n)[
2a(n + 1)
.
Hence
(3.75) K(n) K(n
0
)
m=n0
(1 +C(m)), n n
0
,
is bounded. A similar argument shows u bounded near . The rest follows from
Lemma 3.10 and Theorem 3.18.
Note that (3.65) does not exclude eigenvalues at the boundary of the essential
spectrum. For example, taking
(3.76) a(n) =
1
2
, b(n) =
2 3n
2
4 +n
4
,
there is an eigenvalue at 1 with corresponding eigenfunction u
(1, n) = (1 +n
2
)
1
.
The reason is that the rst moment
nZ
[nb(n)[ is not nite. We will learn more
about this case in Chapter 10.
72 3. Qualitative theory of spectra
3.4. A bound on the number of eigenvalues
In this section we will derive a BirmanSchwinger type bound on the number of
eigenvalues below the essential spectrum of certain Jacobi operators.
Theorem 3.20. Let H
0
, H be operators in a Hilbert space H. Suppose H
0
0
and
0
ess
(H
0
). Let H be such that HH
0
is HilbertSchmidt, then, for <
0
we have
(3.77) dimRan P
(,)
(H) tr((H H
0
)(H
0
)
1
)
2
.
Similarly, if H
0
0
and
0
ess
(H
0
). Then
(3.78) dimRan P
(,)
(H) tr((H H
0
)( H
0
)
1
)
2
for >
0
.
Proof. Since H H
0
is compact we have
ess
(H) =
ess
(H
0
). Using the same
notation as in [195], Theorem XIII.1 we infer as in the proof of the proposition
after Theorem XIII.2 that
n
(H
0
+ (H H
0
)) is monotone nonincreasing with
respect to [0, ). Now we use that
(3.79) (H
0
+(H H
0
))u = u, u H, <
0
,
is equivalent to
(3.80) v = K
v, v H,
where K
=
_
(H
0
)
1
(H
0
H)
_
(H
0
)
1
and u =
_
(H
0
)
1
v. Notice
that K
=
tr((H
0
H)(H)
1
)
2
< . Denote the nonzero eigenvalues of K
by 1/
m
. Then
each eigenvalue
n
(H
0
+(HH
0
)) with
n
(H) < passes when =
m
(0, 1).
Hence
(3.81) dimRan P
(,)
(H)
nN
1
2
n
= tr K
2
nZ
2[a(n) a
0
(n)[
_
g
0
(, n + 1)
_
g
0
(, n)
+[b(n) b
0
(n)[ g
0
(, n)
_
2
(3.85)
for <
0
H
0
, >
0
H
0
, respectively.
Proof. The kernel of (H H
0
)(H
0
)
1
is given by
K(, n, m) = (a(n) a
0
(n))G(, n + 1, m) + (b(n) b
0
(n))G(, n, m)
+ (a(n 1) a
0
(n 1))G(, n 1, m) (3.86)
which can be estimated by
K(, n, m) = [a(n) a
0
(n)[
_
g
0
(, n + 1)
_
g
0
(, m)
+[b(n) b
0
(n)[
_
g
0
(, n)
_
g
0
(, m)
+[a(n 1) a
0
(n 1)[
_
g
0
(, n 1)
_
g
0
(, m). (3.87)
Now
tr((H H
0
)( H
0
)
1
)
2
=
n,mZ
K(, n, m)K(, m, n)
n,mZ
K(, n, m)
K(, m, n)
=
n,mZ
K(, n, n)
K(, m, m) =
_
nZ
K(, n, n)
_
2
(3.88)
concludes the proof.
Applications of these results will be given in Section 10.2.
Chapter 4
Oscillation theory
There is a close connection between the number of sign ips of solutions of Jacobi
equations and the spectra of the corresponding Jacobi operators. The investigation
of this interconnection was started by Sturm (in the case of dierential equations)
and is now known as oscillation theory.
4.1. Pr ufer variables and Sturms separation
theorem
In this section we will study oscillation properties of solutions of (1.19). To be
more precise, we are interested in the number of sign ips of solutions. It will be
convenient to assume
(4.1) a(n) < 0, b(n) R.
In the sequel, a solution of (1.19) will always mean a realvalued, nonzero
solution of (1.19). Given a solution u(, .) of u = u, R, we introduce Pr ufer
variables
u
(, .),
u
(, .) via
u(, n) =
u
(, n) sin
u
(, n),
u(, n + 1) =
u
(, n) cos
u
(, n). (4.2)
Notice that the Pr ufer angle
u
(, n) is only dened up to an additive integer
multiple of 2 (which depends on n).
Inserting (4.2) into ( )u = 0 yields
(4.3) a(n) cot
u
(, n) +a(n 1) tan
u
(, n 1) = b(n)
and
(4.4)
u
(, n) sin
u
(, n) =
u
(, n 1) cos
u
(, n 1).
Equation (4.3) is a discrete Riccati equation (cf. (1.52)) for cot
u
(n) and (4.4) can
be solved if
u
(n) is known provided it is replaced by
(4.5) a(n)
u
(, n) = a(n 1)
u
(, n 1) = 0
75
76 4. Oscillation theory
if sin
u
(, n) = cos
u
(, n 1) = 0 (use u = u and (4.8) below). Explicitly,
(4.6)
u
(, n) =
u
(, 0)
n1
m=0
cos
u
(, m)
sin
u
(, m+ 1)
,
where cos
u
(, m)/ sin
u
(, m+1) has to be replaced by a(m+1)/a(m) whenever
sin
u
(, m+ 1) = 0.
The Wronskian of two solutions u
i
(
i
, n), i = 1, 2, reads
(4.7) W
n
(u
1
(
1
), u
2
(
2
)) = a(n)
u1
(
1
, n)
u2
(
2
, n) sin(
u1
(
1
, n)
u2
(
2
, n)).
The next lemma considers zeros of solutions and their Wronskians more closely.
In particular, we will show that a solutions (resp. their Wronskians) must switch
sign at a zero.
Lemma 4.1. Let u
1,2
be solutions of u
1,2
=
1,2
u
1,2
corresponding to
1
,=
2
,
respectively. Then
(4.8) u
1
(n) = 0 u
1
(n 1)u
1
(n + 1) < 0.
Moreover, suppose W
n
(u
1
, u
2
) = 0 but W
n1
(u
1
, u
2
)W
n+1
(u
1
, u
2
) ,= 0, then
(4.9) W
n1
(u
1
, u
2
)W
n+1
(u
1
, u
2
) < 0.
Otherwise, if W
n
(u
1
, u
2
) = W
n+1
(u
1
, u
2
) = 0, then necessarily
(4.10) u
1
(n + 1) = u
2
(n + 1) = 0 and W
n1
(u
1
, u
2
)W
n+2
(u
1
, u
2
) < 0.
Proof. The fact u
1
(n) = 0 implies u
1
(n 1)u
1
(n + 1) ,= 0 (otherwise u
1
vanishes
identically) and a(n)u
1
(n + 1) = a(n 1)u
1
(n 1) (from u
1
= u
1
) shows
u
1
(n 1)u
1
(n + 1) < 0.
Next, from (1.20) we infer
(4.11) W
n+1
(u
1
, u
2
) W
n
(u
1
, u
2
) = (
2
1
)u
1
(n + 1)u
2
(n + 1).
And W
n
(u
1
, u
2
) = 0 is equivalent to u
1
(n) = c u
2
(n), u
1
(n + 1) = c u
2
(n + 1) for
some c ,= 0. Hence applying the above formula gives
(4.12) W
n1
(u
1
, u
2
)W
n+1
(u
1
, u
2
) = c
2
(
2
1
)
2
u
1
(n)
2
u
1
(n + 1)
2
,
proving the rst claim. If W
n
(u
1
, u
2
), W
n+1
(u
1
, u
2
) are both zero, we must have
u
1
(n + 1) = u
2
(n + 1) = 0 and as before W
n1
(u
1
, u
2
)W
n+2
(u
1
, u
2
) = (
2
1
)
2
u
1
(n)u
1
(n + 2)u
2
(n)u
2
(n + 2). Hence the claim follows from the rst part
(4.8).
We can make the Pr ufer angle
u
(, .) unique by xing, for instance,
u
(, 0)
and requiring
(4.13) [[
u
(, n)/]] [[
u
(, n + 1)/]] [[
u
(, n)/]] + 1,
where
(4.14) [[x]] = supn Z[ n < x.
Since our solutions u(, n) will usually depend continuously on , the same should
be true for
u
(, n). In particular, continuity in should be compatible with the
requirement (4.13).
4.1. Pr ufer variables and Sturms theorem 77
Lemma 4.2. Let R be an interval. Suppose u(, n) is continuous with respect
to and (4.13) holds for one
0
. Then it holds for all if we require
u
(., n) C(, R).
Proof. Fix n and set
(4.15)
u
(, n) = k +(),
u
(, n + 1) = k + (), k Z,
where () (0, ], () (0, 2]. If (4.13) should break down, then by continuity
we must have one of the following cases for some
1
. (i) (
1
) = 0 and
(
1
) (, 2), (ii) (
1
) = and (
1
) (0, ), (iii) (
1
) = 0 and (
1
)
(0, ), (iv) (
1
) = 2 and (
1
) (0, ). Abbreviate R = (
1
, n)(
1
, n + 1).
Case (i) implies 0 > sin((
1
)) = cos(k) sin(k +(
1
)) = R
1
u(
1
, n+1)
2
> 0,
contradicting (i). Case (ii) is similar. Case (iii) implies (
1
) = /2 and hence
1 = sin(k + /2) cos(k) = R
1
u(
1
, n)u(
1
, n + 2) contradicting (4.8). Again,
case (iv) is similar.
Let us call a point n Z a node of a solution u if either u(n) = 0 or
a(n)u(n)u(n + 1) > 0. Then, [[
u
(n)/]] = [[
u
(n + 1)/]] implies no node at n.
Conversely, if [[
u
(n +1)/]] = [[
u
(n)/]] +1, then n is a node by (4.8). Denote by
#(u) the total number of nodes of u and by #
(m,n)
(u) the number of nodes of u
between m and n. More precisely, we will say that a node n
0
of u lies between m
and n if either m < n
0
< n or if n
0
= m but u(m) ,= 0. Hence we conclude that
the Pr ufer angle of a solution counts the number of nodes.
Lemma 4.3. Let m < n. Then we have for any solution u
(4.16) #
(m,n)
(u) = [[
u
(n)/]] lim
0
[[
u
(m)/ +]]
and
(4.17) #(u) = lim
n
_
[[
u
(n)/]] [[
u
(n)/]]
_
.
Next, we prove the analog of Sturms separation theorem for dierential
equations.
Lemma 4.4. Let u
1,2
be solutions of u = u corresponding to
1
2
. Suppose
m < n are two consecutive points which are either nodes of u
1
or zeros of W
.
(u
1
, u
2
)
(the cases m = or n = + are allowed if u
1
and u
2
are both in
2
(Z)
and W
(u
1
, u
2
) = 0, respectively) such that u
1
has no further node between m
and n. Then u
2
has at least one node between m and n + 1. Moreover, suppose
m
1
< < m
k
are consecutive nodes of u
1
. Then u
2
has at least k 1 nodes
between m
1
and m
k
. Hence we even have
(4.18) #
(m,n)
(u
2
) #
(m,n)
(u
1
) 1.
Proof. Suppose u
2
has no node between m and n + 1. Hence we may assume
(perhaps after ipping signs) that u
1
(j) > 0 for m < j < n, u
1
(n) 0, and
u
2
(j) > 0 for m j n. Moreover, u
1
(m) 0, u
1
(n + 1) < 0, and u
2
(n + 1) 0
provided m, n are nite. By Greens formula (1.20)
(4.19) 0 (
2
1
)
n
j=m+1
u
1
(j)u
2
(j) = W
n
(u
1
, u
2
) W
m
(u
1
, u
2
).
78 4. Oscillation theory
Evaluating the Wronskians shows W
n
(u
1
, u
2
) < 0, W
m
(u
1
, u
2
) > 0, which is a
contradiction.
It remains to prove the last part. We will use induction on k. The case k = 1 is
trivial and k = 2 has already been proven. Denote the nodes of u
2
which are lower
or equal than m
k+1
by n
k
> n
k1
> . If n
k
> m
k
we are done since there are
k 1 nodes n such that m
1
n m
k
by induction hypothesis. Otherwise we can
nd k
0
, 0 k
0
k, such that m
j
= n
j
for 1 + k
0
j k. If k
0
= 0, we are done
and hence we can suppose k
0
1. By induction hypothesis it suces to show that
there are k k
0
nodes n of u
2
with m
k0
n m
k+1
. By assumption m
j
= n
j
,
1 +k
0
j k, are the only nodes n of u
2
such that m
k0
n m
k+1
. Abbreviate
m = m
k0
, n = m
k+1
and assume without restriction u
1
(m + 1) > 0, u
2
(m) > 0.
Since the nodes of u
1
and u
2
coincide we infer 0 <
n
j=m+1
u
1
(j)u
2
(j) and we can
proceed as in the rst part to obtain a contradiction.
We call oscillatory if one solution of u = 0 has an innite number of nodes.
In addition, we call oscillatory at if one solution of u = 0 has an innite
number of nodes near . We remark that if one solution of ( )u = 0 has
innitely many nodes, so has any other (corresponding to the same ) by (4.18).
Furthermore,
1
oscillatory implies
2
oscillatory for all
2
1
(again by
(4.18)).
Now we turn to the special solution s(, n) characterized by the initial condi
tions s(, 0) = 0, s(, 1) = 1. As in Lemma 2.4 we infer
(4.20) W
n
(s(), s
t
()) =
n
j=1
s(, j)
2
.
Here the prime denotes the derivative with respect to . Evaluating the above
equation using Pr ufer variables shows
(4.21)
t
s
(, n) =
1
a(n)
s
(, n)
2
n
j=1
s(, j)
2
.
In particular,
t
s
(, n) < 0 for n < 1,
t
s
(, 1) =
t
s
(, 0) = 0, and
t
s
(, n) > 0 for
n > 0. Equation (4.21) implies that nodes of s(, n) for n N move monotonically
to the left without colliding. In addition, since s(, n) cannot pick up nodes locally
by (4.8), all nodes must enter at and since
t
s
(, 0) = 0 they are trapped inside
(0, ).
We will normalize
s
(, 0) = 0 implying
s
(, 1) = /2. Since s(, n) is a
polynomial in , we easily infer s(, n
0
)
>
<
0 for xed n
0
>
<
0 and suciently
small (see (1.68) and (1.69)). This implies
(4.22) <
s
(, n
0
) < /2, n
0
< 1, 0 <
s
(, n
0
) < , n
0
1,
for xed n
0
and suciently small. Moreover, dividing (4.3) by and letting
using (4.22) shows
(4.23) lim
cot(
s
(, n))
1
=
1
a(n)
, n
+1
< 1
,
4.1. Pr ufer variables and Sturms theorem 79
and hence
(4.24)
s
(, n) =
_
2
a(n)
+o(
1
), n < 1
a(n)
+o(
1
), n 1
,
as .
Now what happens with
s
(, n) as increases? Suppose n 1 for simplicity.
We already know that for small,
s
(, n) starts near 0. Then it increases as
increases by (4.21). It crosses /2 at the rst zero of s(, n+1). Next it crosses at
the rst zero of s(, n). This process continues until it nally crosses n at the last
zero of s(, n + 1). Here we have assumed that s(, n), n N, has precisely n 1
distinct real zeros. However, this is easily seen to hold, since s(
0
, n) = 0 implies
that
0
is an eigenvalue for H
0,n
(see (1.65)). Hence
0
must be real and simple
since H
0,n
is selfadjoint and its spectrum is simple (cf. Remark 1.10), respectively.
Let us summarize these ndings. The interlacing of zeros is a classical result from
the theory orthogonal polynomials.
Theorem 4.5. The polynomial s(, n), n N, has n 1 real and distinct roots
denoted by
(4.25)
n
1
<
n
2
<
n
n1
.
The zeros of s(, n) and s(, n + 1) are interlacing, that is,
(4.26)
n+1
1
<
n
1
<
n+1
2
< <
n
n1
<
n+1
n
.
Moreover,
(4.27) (H
0,n
) =
n
j
n1
j=1
.
Proceeding as for (4.24) we also note
(4.28)
s
(, n) =
_
(2n+1)
2
+
a(n)
+o(
1
), n < 1
n
a(n)
+o(
1
), n 1
,
as +.
Analogously, let u
t
+
(, n) =
j=n+1
u
+
(, j)
2
a(n)
+
(, n)
2
< 0,
(, n) =
n
j=
u
(, j)
2
a(n)
(, n)
2
> 0, (4.29)
where we have abbreviated
u
=
,
u
=
.
Since H is bounded from below we can normalize
(4.30) 0 <
(, n) =
a(n)
+o(
1
),
+
(, n) =
2
a(n)
+o(
1
), n Z,
as .
80 4. Oscillation theory
4.2. Classical oscillation theory
Before we come to the rst applications we recall a lemma from functional analysis.
It will be one of our main ingredients in the following theorems.
Lemma 4.6. Let H be a bounded selfadjoint operator and
j
, 1 j k, be
linearly independent elements of a (separable) Hilbert space H.
(i). Let R. If
(4.32) , H < 
2
for any nonzero linear combination =
k
j=1
c
j
j
, then
(4.33) dimRan P
(,)
(H) k.
Similarly, , H > 
2
implies dimRan P
(,)
(H) k.
(ii). Let
1
<
2
. If
(4.34) (H
2
+
1
2
) <
2
1
2

for any nonzero linear combination =
k
j=1
c
j
j
, then
(4.35) dimRan P
(1,2)
(H) k.
Proof. (i). Let M = span
j
H. We claim dimP
(,)
(H)M = dimM = k.
For this it suces to show KerP
(,)
(H)[
M
= 0. Suppose P
(,)
(H) = 0,
,= 0. Then, abbreviating d
(x) = d, P
(,x)
(H), we see that for any nonzero
linear combination
, H =
_
R
xd
(x) =
_
[,)
xd
(x)
_
[,)
d
(x) = 
2
. (4.36)
This contradicts our assumption (4.32). (ii). Using the same notation as before we
need to show KerP
(1,2)
(H)[
M
= 0. If P
(1,2)
(H) = 0, ,= 0, then,
(H
2
+
1
2
)
2
=
_
R
(x
2
+
1
2
)
2
d
(x) =
_
x
2
d
(x +
2
+
1
2
)
(
2
1
)
2
4
_
(x +
2
+
1
2
) =
(
2
1
)
2
4

2
, (4.37)
where = (, (
2
1
)/2] [(
2
1
)/2, ). But this is a contradiction as
before.
Using this result we can now show that the number of nodes of s(, n) equals
the number of eigenvalues below . We begin with the case of nite and halfline
operators.
Theorem 4.7. Let R. Then we have
(4.38) dimRan P
(,)
(H
0,n
) = #
(0,n)
(s()), n > 1,
and
(4.39) dimRan P
(,)
(H
+
) = #
(0,+)
(s()).
The same theorem holds if + is replaced by .
4.2. Classical oscillation theory 81
Proof. We only carry out the proof for the plus sign (the other part following from
reection). By virtue of (4.21), (4.24), and Lemma 4.3 we infer
(4.40) dimRan P
(,)
(H
0,n
) = [[
s
(, n)/]] = #
(0,n)
(s()), n > 1,
since (H
0,n
) if and only if
s
(, n) = 0 mod . Let k = #(s()) if #(s()) <
, otherwise the following argument works for arbitrary k N. If we pick n so large
that k nodes of s() are to the left of n, we have k eigenvalues
1
< <
k
< of
H
0,n
. Taking an arbitrary linear combination (m) =
k
j=1
c
j
s(
j
, n), c
j
C, for
m < n and (m) = 0 for m n a straightforward calculation (using orthogonality
of s(
j
)) veries
(4.41) , H
+
< 
2
.
Invoking Lemma 4.6 shows
(4.42) dimRan P
(,)
(H
+
) k.
For the reversed inequality we can assume k = #(s()) < . Consider
H
0,n
=
H
0,n
1l on
2
(0, n)
2
(n 1, ). Since
H
0,n
H strongly as n this
implies ([116], Lemma 5.2)
(4.43) dimRan P
(,)
(H
+
) lim
n
dimRan P
(,)
(H
0,n
) = k
completing the proof.
Remark 4.8. (i). Consider the following example
a(n) =
1
2
, n N,
b(1) = 1, b(2) = b
2
, b(3) =
1
2
, b(n) = 0, n 4. (4.44)
The essential spectrum of H
+
is given by
ess
(H
+
) = [1, 1] and one might expect
that H
+
has no eigenvalues below the essential spectrum if b
2
. However,
since we have
(4.45) s(1, 0) = 0, s(1, 1) = 1, s(1, 2) = 0, s(1, n) = 1, n 3,
Theorem 4.7 shows that, independent of b
2
R, there is always precisely one
eigenvalue below the essential spectrum.
(ii). By a simple transformation we obtain the corresponding result for H
+,n0
,
,= 0,
(4.46) dimRan P
(,)
(H
+,n0
) = #
(0,+)
(s
(, ., n
0
)),
where s
(, ., n
0
) is the solution satisfying the boundary condition in (1.94) (see
(2.71)). Similar modications apply to Theorems 4.13, 4.16, and 4.17 below.
As a consequence of Theorem 4.7 we infer the following connection between
being oscillatory or not and the inmum of the essential spectrum.
Corollary 4.9. We have
(4.47) dimRan P
(,)
(H
) <
if and only if is nonoscillatory near and hence
(4.48) inf
ess
(H
) = inf R[ ( ) is oscillatory at .
82 4. Oscillation theory
Moreover, let
0
< <
k
< . . . be the eigenvalues of H
()).
Proof. Again it suces to prove the rst equality. By virtue of (4.29) and (4.31)
we infer
(4.50) dimRan P
(,)
(H
,n
) = [[
(, n)/]], n Z.
We rst want to show that [[
(, n)/]] = #
(,n)
(u
(
1
, n)/]] = k 1 for some
1
R
(saying that u
(
1
, n) > k for m n. Now pick
0
such that
(
0
, n) = k. Then
u
(
0
, .) has a node at n but no node between and n (by Lemma 4.3). Now
apply Lemma 4.4 to u
(
0
, .), u
(
1
, .) to obtain a contradiction. The rest follows
as in the proof of Theorem 4.7.
As before we obtain
Corollary 4.11. We have
(4.51) dimRan P
(,)
(H) <
if and only if is nonoscillatory and hence
(4.52) inf
ess
(H) = inf R[ ( ) is oscillatory.
Furthermore, let
0
< <
k
< . . . be the eigenvalues of H below the essential
spectrum of H. Then the eigenfunction corresponding to
k
has precisely k 1
nodes.
Remark 4.12. (i). Corresponding results for the projection P
(,)
(H) can be
obtained from P
(,)
(H) = P
(,)
(H). In fact, it suces to change the de
nition of a node according to u(n) = 0 or a(n)u(n)u(n+1) < 0 and P
(,)
(H) to
P
(,)
(H) in all results of this section.
(ii). Dening u
j
with
1
<
1
< <
k
<
2
. Again let (m) =
k
j=1
c
j
s(
j
, n),
c
j
C for m < n and (m) = 0 for n m be an arbitrary linear combination.
Then
(4.56) (H
+
2
+
1
2
) <
2
1
2

together with the Lemma 4.6 implies
(4.57) dimRan P
(1,2)
(H
+
) k.
To prove the second inequality we use
H
0,n
= H
0,n
2
1l H
+
strongly as n
and proceed as before
(4.58) dimRan P
(1,2)
(H
+
) liminf
n
P
[1,2)
(
H
0,n
) = k
since P
[1,2)
(
H
0,n
) = P
[1,2)
(H
0,n
).
4.3. Renormalized oscillation theory
The objective of this section is to look at the nodes of the Wronskian of two solutions
u
1,2
corresponding to
1,2
, respectively. We call n Z a node of the Wronskian if
W
n
(u
1
, u
2
) = 0 and W
n+1
(u
1
, u
2
) ,= 0 or if W
n
(u
1
, u
2
)W
n+1
(u
1
, u
2
) < 0. Again we
will say that a node n
0
of W(u
1
, u
2
) lies between m and n if either m < n
0
< n or
if n
0
= m but W
n0
(u
1
, u
2
) ,= 0. We abbreviate
(4.59)
u1,u2
(n) = (
u2
(n)
u1
(n)) mod 2.
and require
(4.60) [[
u1,u2
(n)/]] [[
u1,u2
(n + 1)/]] [[
u1,u2
(n)/]] + 1.
We will x
1
R and a corresponding solution u
1
and choose a second solution
u(, n) with [
1
,
2
]. Now let us consider
(4.61) W
n
(u
1
, u()) = a(n)
u1
(n)
u
(, n) sin(
u1,u
(, n))
as a function of [
1
,
2
]. As in the case of solutions, we rst show that the
normalization (4.60) is compatible with continuity in .
Lemma 4.14. Suppose
u1,u
(
1
, .) satises (4.60), then we have
(4.62)
u1,u
(, n) =
u
(, n)
u1
(n),
where
u
(, .),
u1
(.) both satisfy (4.13). That is,
u1,u
(., n) C([
1
,
2
], R) and
(4.60) holds for all
u1,u
(, .) with [
1
,
2
]. In particular, the second inequality
in (4.13) is attained if and only if n is a node of W
.
(u
1
, u()). Moreover, denote by
#
(m,n)
W(u
1
, u
2
) the total number of nodes of W
.
(u
1
, u
2
) between m and n. Then
(4.63) #
(m,n)
W(u
1
, u
2
) = [[
u1,u2
(n)/]] lim
0
[[
u1,u2
(m)/ +]]
and
#W(u
1
, u
2
) = #
(,)
W(u
1
, u
2
)
= lim
n
_
[[
u1,u2
(n)/]] [[
u1,u2
(n)/]]
_
. (4.64)
84 4. Oscillation theory
Proof. We x n and set
(4.65)
u1,u
(, n) = k +(),
u1,u
(, n + 1) = k + (),
where k Z, (
1
) (0, ] and (
1
) (0, 2]. Clearly (4.62) holds for =
1
since W
.
(u
1
, u(
1
)) is constant. If (4.60) should break down, we must have one
of the following cases for some
0
1
. (i) (
0
) = 0, (
0
) (, 2], or (ii)
(
0
) = , (
0
) (0, ], or (iii) (
0
) = 2, (
0
) (, ], or (iv) (
0
) = 0,
(
0
) (, ]. For notational convenience let us set = (
0
), = (
0
) and
u1
(n) =
1
(n),
u
(
0
, n) =
2
(n). Furthermore, we can assume
1,2
(n) = k
1,2
+
1,2
,
1,2
(n + 1) = k
1,2
+
1,2
with k
1,2
Z,
1,2
(0, ] and
1,2
(0, 2].
Suppose (i). Then
(4.66) W
n+1
(u
1
, u(
0
)) = (
0
1
)u
1
(n + 1)u(
0
, n + 1).
Inserting Pr ufer variables shows
(4.67) sin(
2
1
) = cos
2
(
1
) 0
for some > 0 since = 0 implies
1
=
2
. Moreover, k = (k
2
k
1
) mod 2 and
k + = (k
2
k
1
) +
2
1
implies = (
2
1
) mod 2. Hence we have
sin 0 and (, 2] implies = 2. But this says
1
=
2
= /2 and
1
=
2
= . Since we have at least (
2
) > 0 and hence
2
(
2
) > /2,
2
(
2
) > for > 0 suciently small. Thus from (
2
) (, 2) we get
(4.68) 0 > sin (
2
) = sin(
2
(
2
) ) > 0,
contradicting (i).
Suppose (ii). Again by (4.66) we have sin(
2
1
) 0 since
1
=
2
. But
now (k + 1) = (k
1
k
2
) mod 2. Furthermore, sin(
2
1
) = sin() 0 says
= since (0, ]. Again this implies
1
=
2
= /2 and
1
=
2
= . But
since () increases/decreases precisely if () increases/decreases for near
0
,
(4.60) stays valid.
Suppose (iii) or (iv). Then
(4.69) W
n
(u
1
, u(
0
)) = (
0
1
)u
1
(n + 1)u(
0
, n + 1).
Inserting Pr ufer variables gives
(4.70) sin(
2
1
) = sin(
1
) sin(
2
)
for some > 0. We rst assume
2
>
1
. In this case we infer k = (k
2
k
1
) mod 2
implying
2
1
= 0 mod 2 contradicting (4.70). Next assume
2
1
. Then
we obtain (k + 1) = (k
2
k
1
) mod 2 implying
2
1
= mod 2 and hence
sin(
2
1
) 0 from (4.70). Thus we get
1
=
2
= /2
1
=
2
= , and hence
1
= 0 mod 2 contradicting (iii), (iv). This settles (4.62).
Furthermore, if () (0, ], we have no node at n since () = implies
() = by (ii). Conversely, if () (, 2] we have a node at n since () = 2
is impossible by (iii). The rest being straightforward.
Equations (4.16), (4.62), and (4.63) imply
Corollary 4.15. Let
1
2
and suppose u
1,2
satisfy u
1,2
=
1,2
u
1,2
, respectively.
Then we have
(4.71) [#
(n,m)
W(u
1
, u
2
) (#
(n,m)
(u
2
) #
(n,m)
(u
1
))[ 2
4.3. Renormalized oscillation theory 85
Now we come to a renormalized version of Theorem 4.13. We rst need the
result for a nite interval.
Theorem 4.16. Fix n
1
< n
2
and
1
<
2
. Then
(4.72) dimRan P
(1,2)
(H
n1,n2
) = #
(n1,n2)
W(s(
1
, ., n
1
), s(
2
, ., n
2
)).
Proof. We abbreviate
(4.73) (, n) =
s(1,.,n1),s(,.,n2)
(n)
and normalize (perhaps after ipping the sign of s(
1
, ., n
1
)) (
1
, n) (0, ]. From
(4.21) we infer
(4.74) dimRan P
(1,2)
(H
n1,n2
) = lim
0
[[(
2
, n
1
)/ +]]
since (H
n1,n2
) is equivalent to (, n
1
) = 0 mod . Using (4.63) completes
the proof.
Theorem 4.17. Fix
1
<
2
. Then
(4.75) dimRan P
(1,2)
(H
+
) = #
(0,+)
W(s(
1
), s(
2
)).
The same theorem holds if + is replaced by .
Proof. Again, we only prove the result for H
+
. Set k = #
(0,)
W(s(
1
), s(
2
))
provided this number is nite and k N otherwise. We abbreviate
(4.76) (, n) =
s(1),s()
(n)
and normalize (
1
, n) = 0 implying (, n) > 0 for >
1
. Hence, if we choose
n so large that all k nodes are to the left of n, we have
(4.77) (, n) > k.
Thus we can nd
1
<
1
< <
k
<
2
with (
j
, n) = j. Now dene
(4.78)
j
(m) =
_
s(
j
, m)
j
s(
1
, m), m n
0, m n
,
where
j
,= 0 is chosen such that s(
j
, m) =
j
s(
1
, m) for m = n, n + 1. Further
more observe that
(4.79)
j
(m) =
_
j
s(
j
, m)
1
1
s(
1
, m), m n
0, m n
and that s(
1
, m), s(
j
, .), 1 j k, are orthogonal on 1, . . . , n. Next, let =
k
j=1
c
j
j
, c
j
C, be an arbitrary linear combination, then a short calculation
veries
(4.80) (H
+
2
+
1
2
) <
2
1
2
.
Invoking Lemma 4.6 gives
(4.81) dimRan P
(1,2)
(H
+
) k.
To prove the reversed inequality is only necessary if #
(0,)
W(s(
1
),s(
2
)) < .
In this case we look at H
,
0,n
with = s(
2
, n+1)/s(
2
, n). By Theorem 4.16 and
Remark 4.8 (ii) we have
(4.82) dimRan P
(1,2)
(
H
,
0,n
) = #
(0,n)
W(s(
1
), s(
2
)).
86 4. Oscillation theory
Now use strong convergence of
H
,
0,n
= H
,
0,n
1
1l to H
+
as n to obtain
(4.83) dimRan P
(1,2)
(H
+
) liminf
n
dimRan P
(1,2)
(
H
,
0,n
) = k
completing the proof.
As a consequence we obtain the analog of Corollary 4.9.
Corollary 4.18. Let u
1,2
satisfy u
1,2
=
1,2
u
1,2
. Then
(4.84) #
(0,)
W(u
1
, u
2
) < dimRan P
(1,2)
(H
) < .
Proof. By (4.18) and Corollary 4.15 we learn that #
(0,)
W(u
1
, u
2
) is nite if
and only if #
(0,)
W(s(
1
), s(
2
)) is nite.
Finally, we turn to our main result for Jacobi operators H on Z.
Theorem 4.19. Fix
1
<
2
and suppose [
1
,
2
]
ess
(H) = . Then
dimRan P
(1,2)
(H) = #W(u
(
1
), u
(
2
))
= #W(u
(
1
), u
(
2
)). (4.85)
Proof. Since the proof is similar to the proof of Theorem 4.17 we will only outline
the rst part. Let k = #W(u
+
(
1
), u
(
2
)) if this number is nite and k N
otherwise. Pick n > 0 so large that all zeros of the Wronskian are between n and
n. We abbreviate
(4.86) (, n) =
u+(1),u()
(n)
and normalize (
1
, n) [0, ) implying (, n) > 0 for >
1
. Hence, if we
choose n N so large that all k nodes are between n and n, we can assume
(4.87) (, n) > k.
Thus we can nd
1
<
1
< <
k
<
2
with (
j
, n) = 0 mod . Now dene
(4.88)
j
(m) =
_
u
j
, m) m n
j
u
+
(
1
, m) m n
,
where
j
,= 0 is chosen such that u
j
, m) =
j
u
+
(
1
, m) for m = n, n + 1. Now
proceed as in the previous theorems.
Again, we infer as a consequence.
Corollary 4.20. Let u
1,2
satisfy u
1,2
=
1,2
u
1,2
. Then
(4.89) #W(u
1
, u
2
) < dimRan P
(1,2)
(H) < .
Proof. Follows from Corollary 4.18 and dimRan P
(1,2)
(H) nite if and only if
both dimRan P
(1,2)
(H
) and dimRan P
(1,2)
(H
+
) nite (see (3.22)).
Remark 4.21. (i). Lemma 1.6 shows that all theorems remain valid if our hy
pothesis a(n) < 0 is replaced by a(n) ,= 0.
(ii). Dening u
f()d().
If you are not familiar with the basics of probability (respectively measure)
theory, you might want to refer to any textbook (e.g., [39]) rst.
We will choose
(5.1) =
Z
0
= = ((j))
jZ
[(j)
0
,
where
0
is a bounded Borel subset of (R0)R, and T is the algebra generated
by the cylinder sets (i.e., by sets of the form [(j) B
i
, 1 j n with
B
i
Borel subsets of
0
).
On we have a discrete invertible dynamical system, namely the shift
operators
(5.2) T
i
:
(j) (j i)
.
Since the shift of a cylinder set is again one, T(= T
1
) is measurable.
87
88 5. Random Jacobi operators
The measure is called invariant (or stationary) if (TF) = (F) for any
F T. Alternatively, T is called measure preserving. In addition, our dynamical
system T is called ergodic (with respect to ) if any shift invariant set F = TF
has probability (F) zero or one.
Given (, T, ) as above we can dene random variables
a :
(Z, R0)
(j) a
(n) =
1
(n)
b :
(Z, R0)
(j) b
(n) =
2
(n)
(5.3)
and a corresponding random Jacobi operator
(5.4) H
= a
S
+
+a
+b
.
Since we assumed
0
bounded, there is a constant C (independent of ) such that
H
0
. It is easy to see that is ergodic. This
is known as Anderson model.
Now what are the questions we should ask about H
= and
n
i
= f
1
(I
n
i
) = [
i
2
n
f <
i+1
2
n
,
= f
1
(). The sets
n
i
, i Z , are disjoint and (
n
i
) 0, 1 due to
ergodicity. Moreover, since
(5.5)
iZ]
(
n
i
) = (
_
iZ]
n
i
) = () = 1
there exists i
n
Z such that (
n
in
) = 1 and (
n
i
) = 0 , i ,= i
n
. And since
the nesting intervals I
n+1
in+1
I
n
in
must converge to a point x
0
=
nN
I
n
in
we have
(5.6) (f
1
(x
0
)) = (
nN
n
in
) = lim
n
(
n
in
) = 1.
= S
i
H
S
i
,
5.1. Random Jacobi operators 89
we see that H
T
i
and H
(H
)) = dimRanP
(H
).
Based on this observation we obtain the following theorem.
Theorem 5.3. There exists a set R such that
(5.9) (H
) = a.s. .
In addition
(5.10)
d
(H
) = a.s. .
Proof. We start by claiming that the function P
(H
) is weakly measurable
(i.e., f, P
(H
(H
(H
)) =
nZ
n
, P
(H
)
n
is an invariant (by (5.7)) random variable and hence a.s. constant by the lemma.
Moreover,
tr(P
(H
)) = E(tr(P
(H
)))
= E
_
nZ
n
, P
(H
)
n
_
=
nZ
E(
n
, P
(H
)
n
)
=
nZ
E(
0
, P
(H
T
n
)
0
) =
nZ
E(
0
, P
(H
)
0
) (5.12)
is zero or innity depending on E(
0
, P
(H
)
0
) = 0 or not. Here we have used
invariance of with respect to T
n
in the last equality.
For any pair of rational numbers (p, q) Q
2
we set d
(p,q)
= 0, if the almost
sure value of trP
(p,q)
(H
) is 0, , respectively. Set
(5.13)
=
(p,q)Q
2
(p,q)
,
(p,q)
= [trP
(p,q)
(H
) = d
(p,q)
.
Then (
) = 1 for (
(p,q)
) = 1 and the intersection is countable.
Now for
1
,
2
we claim (H
1
) = (H
2
). Indeed, if , (H
1
), there
exist two rational numbers
1
< <
2
such that
(5.14) 0 = dimRanP
(1,2)
(H
1
) = d
(1,2)
= dimRanP
(1,2)
(H
1
)
implying , (H
2
). Interchanging
1
and
2
nishes the rst part.
For the remaining one suppose
d
(H
). Then
(5.15) 0 < dimRanP
(1,2)
(H
) <
with
1
< <
2
suciently close to . This is impossible for
. So
d
(H
) = ,
.
90 5. Random Jacobi operators
Considering the corresponding projections onto the absolutely continuous, sin
gular continuous, and pure point subspaces we expect a similar result for the corre
sponding spectra. The tricky part is to establish measurability of these projections.
Theorem 5.4. There exist sets
ac
,
sc
, and
pp
such that
ac
(H
) =
ac
a.s. ,
sc
(H
) =
sc
a.s. ,
pp
(H
) =
pp
a.s. . (5.16)
Proof. Given the proof of the previous theorem, we only need to show that two
of the projections P
ac
(H), P
sc
(H), and P
pp
(H) are weakly measurable (for xed
, which is omitted here for notational simplicity). Indeed, we have P
ac
(H) =
P
ac
(H)P
N
P
c
(H)e
itH
_
_
_
_
_
= 0,
where
N
(n) = 1 if [n[ N and
N
(n) = 0 if [n[ > N. Hence
f, P
c
(H)g =
1
T
_
T
0
f, e
itH
(1l
N
+
N
)P
c
(H)e
itH
gdt
= lim
T
1
T
_
T
0
f, e
itH
(1l
N
)P
c
(H)e
itH
gdt
= lim
T
1
T
_
T
0
f, e
itH
(1l
N
)e
itH
gdt
lim
T
1
T
_
T
0
f, e
itH
(1l
N
)P
pp
(H)e
itH
gdt (5.18)
Now we estimate the last term using P
pp
(H)g =
j
g
j
, where g
j
are eigenfunctions
of H, that is, Hg
j
=
j
g
j
.
f, e
itH
(1l
N
)P
pp
(H)e
itH
g e
itH
f(1l
N
)e
itH
P
pp
(H)g
f
j
(1l
N
)e
itj
g
j
 f
j
(1l
N
)g
j
. (5.19)
Hence we have
(5.20) f, P
c
(H)g = lim
N
lim
T
1
T
_
T
0
f, e
itH
(1l
N
)e
itH
gdt
which is measurable since e
itH
=
j=0
(itH)
j
/j!.
Finally, P
s
(H) is measurable because of
(5.21) f, P
s
(H)g = lim
n
sup
I1,]I]<1/n
f, P
I
(H)g
(see Lemma B.6).
Since eigenvalues of H
)) = 0.
5.2. The Lyapunov exponent and the density of states 91
Proof. Since trP
]
(H
M, M b
M.
In particular, we have H
 3M.
We recall the denition of the Lyapunov exponent of H
(z) = lim
n
1
[n[
ln 
(z, n)
provided this limit exists. For its investigation we need some standard results from
ergodic theory.
A sequence of random variables f
n
: R is called a subadditive process
if
(5.24) f
m+n
() f
m
() +f
n
(T
m
),
where T is measure preserving. If equality holds, the process is called additive.
Then we have the subadditive ergodic theorem by Kingman [148] (see [214]
for a simple proof).
Theorem 5.7. If f
n
nN
is a subadditive process satisfying E([f
n
[) < , n N,
and inf
n
E(f
n
)/n > , then f
n
()/n converges almost surely and, if T is ergodic,
(5.25) lim
n
1
n
f
n
() = lim
n
1
n
E(f
n
) = inf
nN
1
n
E(f
n
) a.s. .
In the additive case we have f
n
() =
n1
m=0
f
1
(T
m
) and the theorem says
that lim
n
f
n
()/n = E(f
1
) almost surely if E([f
1
[) < . This is known as
Birkhos ergodic theorem.
Now we can prove the following result.
Theorem 5.8. For each z C there exists a number
(5.26) (z) = inf
nN
1
n
E(ln 
(z, n))
(independent of ) such that
(5.27) lim
n
1
[n[
ln 
(z, n +m, m)
(z, m)
f
n
(T
m
) +f
m
(). (5.28)
Moreover, for n > 0 (recall (1.33))
(5.29)
1
M

(z, n)
n
j=1
U
(z, 0, n)
1
)
= E(ln 
T
n
(z, n, 0)
1
) = E(ln 
(z, n, 0)
1
)
= E(ln a
(n) ln a
(0) + ln 
(z) =
(z) = (z)
almost surely.
We call (z) the Lyapunov exponent. It has several interesting properties. First
of all, by the considerations after (1.36) we have (z) 0. Before we can proceed,
we need to review the concept of a subharmonic function.
A measurable function f : C R is called submean if max(f, 0) is
locally integrable and
(5.32) f(z
0
)
1
2ir
2
_
]zz0]r
f(z) dz dz, r > 0.
It is called uppersemicontinuous (u.s.c.) if for any sequence z
n
z
0
we have
(5.33) limsup
znz0
f(z) f(z
0
).
If f is both, submean and u.s.c., it is called subharmonic. As an immediate
consequence we obtain that
(5.34) f(z
0
) = lim
r0
1
2ir
2
_
]zz0]r
f(z) dz dz
if f is subharmonic. The properties of being submean or subharmonic are preserved
by certain limiting operations.
Lemma 5.9. (i). Suppose f
n
are submean and f
n+1
f
n
. Then f(z) = inf
n
f
n
(z)
is submean. If, in addition, f
n
are subharmonic, then so is f.
(ii). Suppose f
n
are submean and sup
n
max(f
n
, 0) is locally integrable. Then
sup
n
f
n
and limsup
n
f
n
are submean.
5.2. The Lyapunov exponent and the density of states 93
Proof. (i). Switching to f
n
f
1
it is no restriction to assume f
n
0 and hence
the rst assertion follows from the monotone convergence theorem. The second one
follows since the inmum of u.s.c. functions is again u.s.c..
(ii). Since
sup
n
f
n
(z
0
)
1
2ir
2
sup
n
_
]zz0]r
f
n
(z) dz dz
1
2ir
2
_
]zz0]r
sup
n
f
n
(z) dz dz, (5.35)
we see that sup
n
f
n
is submean. Combining this result with the previous one and
limsup
n
f
n
= inf
N
sup
nN
f
n
nishes the proof.
As a rst observation note that the norm 
(z
0
, n) = 
1
2ir
2
_
]zz0]r
(z) of H
are submean.
Proof. By Fubinis theorem the function
(5.37) E(ln 
(z, n))
is submean. By the estimate (5.29) and the dominated convergence theorem it is
also continuous and hence subharmonic. Thus, (z) is the limit of a monotone not
increasing sequence of subharmonic functions
(5.38)
1
2
n
E(ln 
(z, 2
n
)) (z).
The rest follows from our previous lemma.
This result looks pretty technical at the rst sight, however, in connection with
(5.34) it says that in order to show (z) = (z) for some subharmonic function (z)
it suces to show equality for z CR. This will be used in what follows since the
case z CR is usually easier to handle as we will see below.
For z CR we know the asymptotic behavior of the solutions of
u = zu.
There are solutions u
,
(z, n) which decay exponentially as n and grow
exponentially in the other direction. Hence it is not hard to see that
(5.39) (z) = lim
n
1
2n
ln
_
[u
(z, n)[
2
+[u
(z, n + 1)[
2
_
for any solution u
) E(ln [m
,
(z)[) (5.40)
almost surely. In addition,
(5.41) (z) = E(ln a
) + ln [E
z[ Re
_
E
E0
()
z
d,
where () = E(
,
(, n)) and
,
(, n) are the functions associated with the
Weyl mfunctions m
,
(z, n) (see Theorem B.11).
Proof. We rst compute
lim
n
1
2n
ln
_
[u
,+
(z, n)[
2
+[u
,+
(z, n + 1)[
2
_
=
= lim
n
1
2n
_
ln(1 +[m
,+
(z, n)[
2
) + 2 ln [u
,+
(z, n + 1)[
_
= lim
n
1
n
n1
j=0
ln [a(j)m
,+
(z, j)[, (5.42)
where we have used (2.3). Since m
,+
(z, n) is measurable and satises (2.3) we can
invoke Birkhos ergodic theorem to see that the limit is E(ln a
)+E(ln [m
,+
(z)[).
Moreover, using (compare (6.56) below)
(5.43) ln m
,+
(z, n) = ln(E
,
z) +
_
E,
E,0
,+
(, n)
z
d,
and Fubinis theorem we see
(5.44) E(ln [m
,+
(z)[) = ln [E
z[ + Re
_
E
E0
E(
,+
(, n))
z
d.
Computing the remaining three limits in a similar fashion the result follows from
(5.39).
Note that (5.41) implies
(5.45) ( + i) = ( i), R.
Next we want to shed some additional light on the function () which appeared
in the last theorem. To do this we rst introduce the socalled density of states. We
consider the nite operators H
,0,n
with Dirichlet boundary conditions at 0 and n.
The integrated density of states for this operator is dened by
(5.46) N
n
() =
1
n
dimRanP
(,)
(H
,0,n+1
) =
1
n
trP
(,)
(H
,0,n+1
).
That is we take the number of eigenstates below and divide it by the total number
of eigenstates. In particular, N
n
() is zero below the spectrum of H
,0,n+1
, one
above the spectrum of H
,0,n+1
, and increasing in between. By Theorem 4.7 we
also have
(5.47) N
n
() =
1
n
#
(0,n+1)
(s
())
5.2. The Lyapunov exponent and the density of states 95
which is why N
n
() is also called rotation number.
Our hope is that N
n
() will tend to some limit as n .
If the process
(5.48) f
n
() = trP
(,)
(H
,0,n+1
)
were subadditive, the desired result would follow directly from Kingmans Theorem.
Unfortunately this is not true! To see this observe
(5.49) H
,0,n+m+1
= H
,0,m+1
H
T
m
,0,n+1
+M
,m
,
where M
,m
= a
(m)(
m
, .
m+1
+
m+1
, .
m
). In the simplest case m = n = 1
the above equation reads
(5.50)
_
b
1
a
1
a
1
b
2
_
=
_
b
1
0
0 0
_
_
0 0
0 b
2
_
+a
1
_
0 1
1 0
_
.
Let b
+
= max(b
1
, b
2
) and b
= min(b
1
, b
2
). Then the eigenvalues of this matrix are
b
c and b
+
+c for suitable c > 0. Hence
(5.51) f
2
() = f
1
() +f
1
(T
1
) +
_
_
0 , b
c
1 , b
c < b
0 , b
< b
+
1, b
+
< b
+
+c
0 , b
+
+c <
shows that neither f
n
() nor f
n
() is subadditive in general.
However, since M
,m
is a rank two perturbation, we have at least
(5.52) f
n+m
() f
m
() +f
n
(T
m
) + 2
and hence
f
n
() = f
n
() +2 is subadditive. Clearly f
n
()/n converges if and only
if
f
n
()/n converges in which case the limits are equal. Summarizing,
(5.53) lim
n
N
n
() = N()
almost surely. The quantity N() is called integrated density of states. Because
of (5.47) it is also known as rotation number. Some of its properties are collected
in the following lemma.
Lemma 5.12. There is a set
of probability one such that for continuous f,
(5.54) lim
n
_
f()dN
n
() =
_
f()dN() = E(
0
, f(H
)
0
),
.
In other words, dN
n
converges weakly to dN almost surely. Moreover, N() is
continuous and (5.54) also holds for f =
(,0)
.
Proof. Note that f E(
0
, f(H
)
0
) is a positive linear functional on the space
C([3M, 3M]) and hence by the RiezMarkov Theorem there is a (positive proba
bility) measure d
N such that
(5.55)
_
f()d
N() = E(
0
, f(H
)
0
).
96 5. Random Jacobi operators
We begin by showing that (5.54) holds with dN replaced by d
N. Let us choose
f() =
k
, k N
0
, rst and consider the process
(5.56) f
n
() = tr (H
,0,n+1
)
k
=
n
m=1
m
, (H
,0,n+1
)
k
m
.
Since each summand depends on n (unless k = 0, 1) it is not clear that this process
is (sub)additive. Hence we will try to replace it by
(5.57)
f
n
() = tr (H
)
k
=
n
m=1
m
, (H
)
k
m
,
which is additive. By Birkhos theorem we have lim
n
f
n
()/n =
_
k
d
N()
for in a set
k
of probability one. Now what about the error we make when
replacing f
n
by
f
n
? The matrix elements
m
, (H
)
k
m
and
m
, (H
)
k
m
only
dier for m = n, n 1, . . . , n n(k) and the dierence can be estimated using
a
M and [b
[ M. Hence we obtain
(5.58) [
f
n
() f
n
()[ C
k
(M)
and so f
n
()/n converges if and only if
f
n
()/n converges in which case the limits
are equal. Choosing
=
k=0
k
we see that (5.54) holds if f is a polynomial.
Now let f be continuous and choose a polynomial f
_
f()d
N()
_
f()dN
n
()
_
f
()dN
n
()
_
f
()d
N()
+
+
_
[f() f
()[dN
n
() +
_
[f() f
()[d
N(). (5.59)
Performing the limit n we see
(5.60) lim
n
_
f()d
N()
_
f()dN
n
()
and letting 0 establishes (5.54) for continuous f.
Let f
() =
2
/(
2
+
2
). Using dominated convergence we obtain
N(
0
) = lim
0
_
f
()d
N() = lim
0
E(
0
, f
(H
)
0
)
= E(lim
0
0
, f
(H
)
0
) = E(
0
, P
0]
(H
)
0
) = 0, (5.61)
where the last equality follows from Theorem 5.5. This shows that
N() is contin
uous.
To show (5.54) for f() =
(,0)
() we approximate f by continuous func
tions f
given by f
() = 1 (
0
)/ for
0
< <
0
+ and f
() = f()
else. The parameter is chosen such that
N((
0
,
0
)) /2 (this is possible
since
N is continuous). Now observe
(5.62)
lim
n
_
[f() f
()[dN
n
() lim
n
_
g
()dN
n
() =
_
g
()d
N()
2
,
where g
() = (1 [
0
[/)
(0,0+)
(), and proceed as before.
5.2. The Lyapunov exponent and the density of states 97
Equation (5.54) for f =
(,)
nally tells us N() =
N().
Remark 5.13. (i). The proof of the previous lemma shows that
(5.63)
N() = lim
n
1
n
tr(
(0,n+1)
P
(,)
(H
)) = lim
n
1
n
n
m=1
m
, P
(,)
(H
)
m
,
where the right hand side is an additive process.
(ii). We can replace H
,0,n+1
by H
,n1,n+1
, that is,
(5.64) N() = lim
n
1
2n + 1
trP
(,)
(H
,n1,n+1
)),
or we could choose dierent boundary conditions.
In particular,
(5.65) N() = lim
n
1
n
#
(0,n+1)
(u
())
for any real (nontrivial) solution
() = u
() (compare (4.18)).
Note that N() can also be viewed as the average of the spectral measures
d
,n,n
() since by (5.54) for f =
(,)
,
(5.66) N() = E(
n
, P
(,)
(H
)
n
) = E(
,n,n
((, ))).
In particular, we obtain the following corollary.
Corollary 5.14. We have (dN) = . Moreover, N() = 0 for E
0
and
N() = 1 for E
.
Another consequence is the Thouless formula.
Theorem 5.15. For any z C we have
(5.67) (z) = E(ln a
) +
_
E
E0
ln [ z[dN()
and N() = ().
Proof. Invoking (1.65) we get
(5.68)
1
n
ln [s
(, n)[ =
1
n
n
m=1
ln a
(m) +
_
ln [ z[dN
n
()
for z CR. Using (5.39) with u
= s
(z, 0)) =
_
dN()
z
,
then
(5.70)
( + i) =
_
dN(x)
(x )
2
2
= Im(g( + i)), [[ > 0.
98 5. Random Jacobi operators
In particular, the moments of dN are the expectations of the coecients of the
Laurent expansion of g
)
j
0
) =
_
j
dN().
Explicitly we have (
_
dN() = 1)
_
dN() = E(b
(0)),
_
2
dN() = 2E(a
(0)
2
) +E(b
(0)
2
), etc. . (5.72)
In addition, solving (2.11) for the
1
m,(z,n)
term, taking imaginary parts, logs,
and expectations we nd
(5.73) 2( + i) = E
_
ln(1 +
a
(
0
1
)Im(m
,
( + i))
)
_
, [[ > 0.
Finally, let us try to apply the results found in this section. To begin with, note
that we know
(z) =
() =
(z) =
= [(, ) B, B
) = 0 we obtain
from Fubinis theorem
(5.75) ( )(B) =
_
[B
[d() =
_
R
(B
)d = 0
that [B
(z) is submean
(z
0
)
1
2ir
2
_
]zz0]r
(z) dz dz
=
1
2ir
2
_
]zz0]r
(z) dz dz = (z
0
). (5.76)
Now we are ready to prove the following theorem which should be compared
with Theorem 3.16.
Theorem 5.17. The absolutely continuous spectrum of H
is given by
(5.77)
ac
= R[() = 0
ess
almost surely.
Proof. Combining Theorem 3.16 and the previous lemma we have immediately
ac
R[() = 0
ess
. The converse is a little harder.
5.2. The Lyapunov exponent and the density of states 99
Abbreviate N = R[() = 0. Using x(1 +x/2)
1
ln(1 +x), x 0, we
obtain from (5.73)
(5.78) E
_
n
,
( + i)
_
E
_
ln(1 +
n
,
( + i)
2
)
_
= 2( + i),
where n
,
(z) = a(
0
1
)Im(m
,
(z)) +
Im(z)
2
. Moreover, we have by (5.70) and
Fatous lemma
E
_
liminf
0
1
n
,
( + i)
_
liminf
0
E
_
1
n
,
( + i)
_
= 2 liminf
0
( + i)
= n
,
( + i) (see the notation from the previous
proof) and n = n
( + i) = g
( + i, 0)
1
. An elementary calculation using
Im(n) = n
+
+n
veries
_
1
n
+
+
1
n
_
(n
+
n
)
2
+ (Re(n))
2
[n[
2
=
_
1
n
+
+
1
n
__
1 4
n
+
n
[n[
2
_
=
1
n
+
+
1
n
+ 4Im(
1
n
). (5.81)
Invoking (5.70) and (5.78) we see
(5.82) E
__
1
n
+
+
1
n
_
(n
+
n
)
2
+ (Re(n))
2
[n[
2
_
4
_
( + i)
( + i)
_
.
Moreover, since for a.e. the limits n
,
(+i0) and n
pp
) I = .
Proof. By (5.80) and Lemma B.4 we see that g
( + i0, 0)
1
is holomorphic on
I and nonzero. So the same is true for g
(n), b
(n))
nZ
are determined by either
(a
(n), b
(n))
nn0
or (a
(n), b
(n))
nn0
for any n
0
Z.
Proof. It is no restriction to assume n
0
= 0. By (5.80) and Lemma B.4 we learn
that m
,+
(z) uniquely determines m
,
(z) and vice versa. Hence the claim follows
from Theorem 2.29.
Sequences (a
(n), b
(n))
nZ
of the above type are referred to as determin
istic sequences. Hence the theorem says that for nondeterministic sequences we
have
ac
= . In particular this holds for the Anderson model. An example for
deterministic sequences are almost periodic ones which will be discussed in the next
section.
5.3. Almost periodic Jacobi operators
Now we will consider the case where the coecients a, b are almost periodic se
quences. We review some basic facts rst.
Let c
0
= (a
0
, b
0
)
(Z, R
+
)
mZ
and call the closure the hull of c
0
, hull(c
0
) = (c
0
). The sequence c
0
is called
almost periodic if its hull is compact. Note that compactness implies that (H.5.1)
holds for almost periodic sequences.
For an almost periodic sequence c
0
we can dene a product
(5.83) : (c
0
) (c
0
) (c
0
)
(c
m
, c
n
) c
m
c
n
= c
m+n
,
which makes (c
0
) an abelian group. Moreover, this product is continuous since
c
n1
c
m1
c
n0
c
m0
 c
n1+m1
c
n1+m0
 +c
n1+m0
c
n0+m0

= c
m1
c
m0
 +c
n1
c
n0
, (5.84)
where we have used the triangle inequality and translation invariance of the norm
in
(Z)
F
d.
5.3. Almost periodic Jacobi operators 101
Then also satises (5.85) and hence must be equal to up to normalization
by uniqueness of the Haar measure. Hence (G) = (hull(c
0
))(G) and using
(F) = (hull(c
0
)) = (F) we obtain for G = F that (F) = (F)
2
implying
(F) 0, 1. Thus is ergodic.
As a consequence we note
Theorem 5.21. Let (a, b) be almost periodic. For all Jacobi operators correspond
ing to sequences in the hull of (a, b), the spectrum is the same and the discrete
spectrum is empty.
Proof. By Theorem 5.3 the claim holds for a set of full Haar measure. Since such
a set is dense we can approximate any point (a, b) in the hull by a sequence (a
n
, b
n
)
within this set in the sup norm. Hence the corresponding operators H
n
converge
to H in norm and the result follows.
Since the absolutely continuous part is not stable under small perturbations,
Theorem 5.4 cannot be directly extended to the the entire hull of an almost periodic
sequence.
Theorem 5.22. Let (a, b) be almost periodic. For all Jacobi operators correspond
ing to sequences in a subset of full Haar measure of hull((a, b)), the pure point,
singular continuous, absolutely continuous spectrum is the same, respectively.
One way of constructing examples of almost periodic operators is to choose
periodic functions p
1
(x), p
2
(x) (with period one) and consider
(5.88) (a
(n), b
(n)) = (p
1
(2
1
n +
1
), p
1
(2
2
n +
2
)), R
2
, [0, 2)
2
.
Note that such sequences are even quasi periodic by construction.
We will restrict our attention to the most prominent model where p
1
(x) = 1 and
p
2
(x) = cos(x). The corresponding operator H
(n), b
(n +N) = b
(n) and
(5.90) hull(1, b
) = (1, b
+2n
)
N
n=1
.
The underlying group is Z
N
and the Haar measure is the normalized counting
measure. This is a special case of periodic sequences and will be investigated in
Chapter 7.
The case we are interested in here is , Q. Then
(5.91) hull(1, b
) = (1, b
)[ [0, 2),
where the right hand side is compact as the continuous image of the circle. The
underlying group is S
1
= e
i
[ [0, 2) and the Haar measure is
d
2
. Note that
the shift satises S
m
b
= b
+2m
.
We start our investigation with a bound for the Lyapunov exponent from below.
Lemma 5.23. Suppose , Q, then the Lyapunov exponent of H
satises
(5.92) (z) ln
2
.
102 5. Random Jacobi operators
Proof. Observe
(5.93) b
(n) =
2
(e
2in
w + e
2in
w
1
), w = e
2i
,
and (cf. (1.29))
(5.94) U
(z, n) =
1
w
_
0 w
w zw
2
(e
2in
w
2
+ e
2in
)
_
.
Hence f
n
(w) = w
n
(z, n) = f
n
(w).
In particular, applying Theorem 5.17 we obtain
Theorem 5.24. If > 2 and , Q, then
ac
(H
) = for a.e. .
Moreover, we can even exclude point spectrum if is well approximated by
rational numbers and hence b
, bz
m=1
k
(z, n, m)(U(z, m) U
k
(z, m))(z, m1)
we see
(5.101) sup
]n]2N
k
(z, n)
k
(z, n) 2N
k
M(z)(M(z)
2
C
k
)
N
k
which tends to zero as k .
5.3. Almost periodic Jacobi operators 103
Abbreviate x
j
= u
k
(jN
k
) = 
k
(z, jN
k
)u(0). Periodicity of a
k
, b
k
implies
det
k
(z, N
k
) = 1 and
k
(z, jN
k
) =
k
(z, N
k
)
j
(compare also Section 7.1). Hence
using the characteristic equation (A
2
+ tr(A)A + det(A) = 0) and the triangle
inequality it is not hard to show (set A =
k
(z, N
k
)
1
) that x
1
+ x
1
and
x
2
+ x
1
1, where = tr
k
(z, N
k
). Hence max(x
2
+ x
2
, x
1
+ x
1
) 1
proving the claim.
In order for b
M
k
N
k
1
k
N
k
.
Theorem 5.26. Suppose is a Liouville number, then
pp
(H
) = .
Proof. For b
,k
(n) = cos(2
M
k
N
k
n +) we estimate
sup
]n]2N
k
[b
(n) b
,k
(n)[ = sup
]n]2N
k
[ cos(2n +) cos(2
M
k
N
k
n +)[
2 sup
]n]2N
k
[n[
M
k
N
k
4N
k
k
N
k
(5.103)
and the result follows from the previous lemma.
Combining Theorem 5.24 and Theorem 5.26 we obtain the
Corollary 5.27. The spectrum of H
(z, n).
Since both quantities are dened as expectations of resolvents
g(z, n) =
n
, (H z)
1
n
,
(z, n) = (1 +
2
)
n
, (H z)
1
n
a(n)
, (6.1)
all we have to do is invoking Neumanns expansion for the resolvent
(H z)
1
=
N1
j=0
H
j
z
j+1
+
1
z
N
H
N
(H z)
1
=
j=0
H
j
z
j+1
, [z[ > H. (6.2)
In summary, we obtain the following result as simple consequence (see also Sec
tion 2.1) of the spectral theorem.
Lemma 6.1. Suppose
2
(Z) with  = 1. Then
(6.3) g(z) = , (H z)
1
(),
105
106 6. Trace formulas
where d
() = d, P
(,]
(H) is the spectral measure of H associated to the
sequence . Moreover,
Im(g(z)) = Im(z)(H z)
1

2
,
g(z) = g(z), [g(z)[ (H z)
1

1
]Im(z)]
, (6.5)
and
(6.6) g(z) =
j=0
, H
j
z
j+1
.
We note that solving (1.107) for h(z, n) shows that h(z, n) can be written as
the dierence of two Herglotz functions.
Lemma 6.1 implies the following asymptotic expansions for g(z, n), h(z, n),
and
(z, n).
Theorem 6.2. The quantities g(z, n), h(z, n), and
j=0
g
j
(n)
z
j+1
, g
0
= 1,
h(z, n) = 1
j=0
h
j
(n)
z
j+1
, h
0
= 0,
(z, n) =
a(n)
j=0
j
(n)
z
j+1
,
0
= 1 +
2
. (6.7)
Moreover, the coecients are given by
g
j
(n) =
n
, H
j
n
,
h
j
(n) = 2a(n)
n+1
, H
j
n
,
j
(n) = (
n+1
+
n
), H
j
(
n+1
+
n
)
= g
j
(n + 1) +
a(n)
h
j
(n) +
2
g
j
(n), j N
0
. (6.8)
Remark 6.3. Using the unitary transform U
j
2
+bg
j
,
h
j+1
h
j+1
= 2
_
a
2
g
+
j
(a
)
2
g
j
_
+b
_
h
j
h
j
_
. (6.9)
Proof. The rst equation follows from
(6.10) g
j+1
(n) = H
n
, H
j
n
=
h
j
(n) +h
j
(n 1)
2
+b(n)g
j
(n)
6.1. Asymptotic expansions 107
using H
n
= a(n)
n+1
+a(n 1)
n1
+b(n)
n
. Similarly,
h
j+1
(n) = b(n)h
j
(n) + 2a(n)
2
g
j
(n + 1) + 2a(n 1)a(n)
n+1
, H
j
n1
= b(n + 1)h
j
(n) + 2a
2
g
j
(n) + 2a(n)a(n + 1)
n+2
, H
j
n
(6.11)
and eliminating
n+1
, H
j
n1
completes the proof.
The system (6.9) does not determine g
j
(n), h
j
(n) uniquely since it requires
solving a rst order recurrence relation at each step, producing an unknown sum
mation constant each time
h
j+1
(n) h
j+1
(n
0
) =
n
m=n0+1
_
2a(m)
2
g
j
(m+ 1) 2a(m1)
2
g
j
(m1)
+b(m)h
j
(m) b(m)h
j
(m1)
_
. (6.12)
To determine the constant observe that the right hand side must consists of two
summands, the rst one involving a(n + k), b(n + k) with [k[ < j and the second
one involving a(n
0
+k), b(n
0
+k) with [k[ < j. This follows since the left hand side
is of this form (use H
n
= a(n)
n+1
+ a(n 1)
n1
+ b(n)
n
and induction). In
particular, the rst summand must be equal to h
j+1
(n) and the second to h
j+1
(n
0
).
This determines the unknown constant.
Since this procedure is not very straightforward, it is desirable to avoid these
constants right from the outset. This can be done at the expense of giving up
linearity. If we compare powers of z in (2.187) we obtain
(6.13) h
j+1
= 2a
2
j
=0
g
j
g
+
1
2
j
=0
h
j
h
, j N,
which determines g
j
, h
j
recursively together with the rst equation of (6.9). Ex
plicitly we obtain
g
0
= 1, g
1
= b, g
2
= a
2
+ (a
)
2
+b
2
,
h
0
= 0, h
1
= 2a
2
, h
2
= 2a
2
(b
+
+b) (6.14)
and hence from (6.8)
0
= 1 +
2
,
1
= b
+
+ 2a +b
2
,
2
= (a
+
)
2
+a
2
+ (b
+
)
2
+ 2a(b
+
+b) + (a
2
+ (a
)
2
+b
2
)
2
. (6.15)
Remark 6.5. (i). Let us advocate another approach which produces a recursion
for g
j
only. Inserting the expansion (6.7) for g(z, n) into (1.110) and comparing
coecients of z
j
one infers
g
0
= 1, g
1
= b, g
2
= a
2
+ (a
)
2
+b
2
,
g
3
= a
2
(b
+
+ 2b) + (a
)
2
(2b +b
) +b
3
,
g
j+1
= 2bg
j
a
2
g
+
j1
+ (a
)
2
g
j1
b
2
g
j1
1
2
j1
=0
k
j1
k
+ 2a
2
_
j1
=0
g
j1
g
+
2b
j2
=0
g
j2
g
+
+b
2
j3
=0
g
j3
g
+
_
, (6.16)
108 6. Trace formulas
for j 3, where k
0
(n) = b(n) and
(6.17) k
j
= a
2
g
+
j1
(a
)
2
g
j1
+b
2
g
j1
2bg
j
+g
j+1
, j N.
If g
j
(n) is known, h
j
(n) can be determined using
(6.18) h
j+1
= bh
j
+g
j+2
2bg
j+1
+a
2
g
+
j
(a
)
2
g
j
+b
2
g
j
, j N
0
.
Equation (6.18) follows after inserting the rst equation of (6.9) into the second.
(ii). Analogously, one can get a recurrence relation for
j
using (1.112). Since this
approach gets too cumbersome we omit further details at this point.
Next, we turn to Weyl mfunctions. As before we obtain
Lemma 6.6. The quantities m
(z, n) =
j=0
m
,j
(n)
z
j+1
, m
,0
(n) = 1.
The coecients m
,j
(n) are given by
(6.20) m
,j
(n) =
n1
, (H
,n
)
j
n1
, j N,
and satisfy
m
,0
= 1, m
,1
= b
,
m
,j+1
= b
m
,j
+ (a
+
)
2
j1
=0
m
,j1
m
+
,
, j N. (6.21)
Finally, we consider the solutions
(z, n) = a(n) m
(z, n) =
_
a(n)
z
_
1
j=0
,j
(n)
z
j
.
The coecients
,j
(n) satisfy
+,0
= 1,
+,1
= b
+
,
+,j+1
= b
+
+,j
+ (a
+
)
2
j1
=0
+,j1
+
+,
, j 1, (6.23)
and
,0
= 1,
,1
= b,
,j+1
= b
,j
j
=1
,j+1
,
(a
)
2
1,j
, j 1. (6.24)
6.2. General trace formulas and xi functions 109
Proof. Existence is clear from
(6.25)
+
(z, n) = a(n)m
+
(z, n),
(z, n) =
z +b(n) +a(n 1)
2
m
(z, n)
a(n)
,
and the previous lemma. Moreover, inserting (6.22) into the Riccati equation (1.52)
produces the relations (6.23), (6.24) which determine the coecients
,j
uniquely.
Since
(z, .) =
z
a
_
1
b
z
(a
)
2
z
2
(a
)
2
b
z
3
(a
)
2
((a
)
2
+ (b
)
2
)
z
4
+O(
1
z
5
)
_
,
+
(z, .) =
a
z
_
1 +
b
+
z
+
(a
+
)
2
+ (b
+
)
2
z
2
+
(a
+
)
2
(2b
+
+b
++
) + (b
+
)
3
z
3
+
((a
+
)
2
+ (b
+
)
2
)
2
+ (a
+
)
2
((b
+
+b
++
)
2
+ (a
++
)
2
)
z
4
+O(
1
z
5
)
_
. (6.26)
This lemma has some interesting consequences for u
(z, n, n
0
) = c(z, n, n
0
)+
(z, n
0
)s(z, n, n
0
) are square summable near .
Moreover, we obtain
u
(z, n, n
0
) =
_
_
_
n1
j=n0
z
a(j)
_
_
_
_
_
_1
1
z
n1
j=n0
b(j) +O(
1
z
2
)
_
_
_,
u
+
(z, n, n
0
) =
_
_
_
n1
j=n0
a(j)
z
_
_
_
_
_
_1 +
1
z
n
j=n0+1
b(j) +O(
1
z
2
)
_
_
_. (6.27)
Remark 6.8. If we reect at a point n
0
(cf. (1.82)) we obtain from (6.8)
(6.28) g
R,j
(n
0
+k) = g
j
(n
0
k), h
R,j
(n
0
+k) = h
j
(n
0
k 1)
and thus
(6.29)
R,j
(n
0
+k) =
2
1/
j
(n
0
k).
In addition, since we have m
R,+
(z, n
0
+k) = m
(z, n
0
k) we infer
(6.30)
+,+2
(n
0
+k) = a
R
(n
0
k)
2
R,,
(n
0
k).
6.2. General trace formulas and xi functions
The objective of the present section is to develop trace formulas for arbitrary Jacobi
operators. Applications will be given at the end of Section 7.5 and in Section 8.1.
Our main ingredient will be Kreins spectral shift theory (for the special case
of rank one perturbations). Hence we will recall some basic results rst.
Consider the following rank one perturbation
(6.31) H
n,
= H +
n
, .
n
, 0,
110 6. Trace formulas
and abbreviate
(6.32) g
(z, n) =
n
, (H
n,
z)
1
n
.
By virtue of the second resolvent formula one infers
(6.33) (H
n,
z)
1
= (H z)
1
(H
n,
z)
1
n
, .(H z)
1
n
and hence
(6.34) g
(z, n) =
g(z, n)
1 +g(z, n)
.
Furthermore, we compute (H
n,
z)
1
n
= (1 +g
(z, n))(H z)
1
n
and thus
(6.35) (H
n,
z)
1
= (H z)
1
1 +g(z, n)
(H z)
1
n
, .(H z)
1
n
(a special case of Kreins resolvent formula). We have
ess
(H
n,
) =
ess
(H) and
an investigation of the discrete poles of (6.35) shows
(6.36)
d
(H
n,
) = R
ess
(H)[g(, n) =
1
.
Moreover, since g(, n) is increasing (w.r.t. ) in each spectral gap, there can be at
most one discrete eigenvalue of H
n,
in each spectral gap.
Computing traces we obtain
tr
_
(H
n,
z)
1
(H z)
1
_
=
1 +g(z, n)
n
, (H z)
2
=
d
dz
ln(1 +g(z, n)). (6.37)
Next, note that g(z, n) and hence 1 + g(z, n) is Herglotz. Hence we have the
following exponential representation (cf. Lemma B.12 (iii))
(6.38) 1 +g(z, n) = exp
_
_
R
(, n)
z
d
_
,
where
(6.39)
(, n) =
1
lim
0
arg
_
1 +g( + i, n)
_
.
Now we can rewrite (6.37) as
(6.40) tr
_
(H
n,
z)
1
(H z)
1
_
=
_
R
(, n)
( z)
2
d.
Moreover,
_
=
_
R
(, n)d.
However, (6.41) can be easily obtained rigorously by comparing coecients in (6.35)
and then taking traces of the resulting equations.
More general, Kreins result implies
(6.42) tr
_
f(H
n,
) f(H)
_
=
_
R
f
t
()
(, n)d
6.2. General trace formulas and xi functions 111
for continuously dierentiable f (f can be assumed compactly supported since only
values in the support of
(, n)
are called perturbation determinant and spectral shift function of the pair
(H, H
n,
), respectively.
Let us collect some parts of these results in our rst theorem.
Theorem 6.9. Let
(n) = tr
_
(H
n,
)
+1
H
+1
_
= ( + 1)
_
R
(, n)d,
with (cf. (6.8))
(6.44) b
(0)
(n) = , b
()
(n) = ( + 1)g
(n) +
j=1
g
j
(n)b
(j1)
(n), N.
Proof. The claim follows after expanding both sides of
(6.45) ln(1 +g(z, n)) =
_
R
(, n)
z
d
and comparing coecients using the following connections between the series of
g(z) and ln(1 +g(z)) (cf., e.g., [185]). Let g(z) have the Laurent expansion
(6.46) g(z) =
=1
g
as z . Then we have
(6.47) ln(1 +g(z)) =
=1
c
,
where
(6.48) c
1
= g
1
, c
= g
j=1
j
g
j
c
j
, 2.
_
R
(, n)d
2
.
Next, we want to investigate the case where the coupling constant tends to
. It is clear that H
,n
f will only converge if f(n) = 0. On the other hand,
(6.50) (H
n,
z)
1
(H z)
1
1
g(z, n)
(H z)
1
n
, .(H z)
1
n
converges (in norm) as . Comparison with (1.106) shows (H
n,
z)
1
(H
n
z)
1
if we embed (H
n
z)
1
into
2
(Z) using (H
n
z)
1
n
= 0, that
is, (H
n
z)
1
= (1l P
n
)(H z)
1
(1l P
n
) (cf. Remark 1.11). In particular,
(6.40) is still valid
(6.51) tr
_
(H
n
z)
1
(H z)
1
_
=
_
R
(, n)
( z)
2
d.
112 6. Trace formulas
Here the xi function (, n) is dened by
(6.52) g(z, n) = [g(i, n)[ exp
__
R
_
1
z
1 +
2
_
(, n)d
_
, z C(H),
respectively by
(6.53) (, n) =
1
lim
0
arg g( + i, n), arg(.) (, ].
Since (., n) is not compactly supported, the asymptotic expansions require a some
what more careful analysis. We abbreviate
(6.54) E
0
= inf (H), E
= sup (H),
and note that g(, n) > 0 for < E
0
, which follows from (H ) > 0 (implying
(H )
1
> 0). Similarly, g(, n) < 0 for > E
.
Using (6.55) together with the asymptotic behavior of g(., n) we infer
(6.56) g(z, n) =
1
E
z
exp
_
_
E
E0
(, n)d
z
_
.
As before we set H
n
n
= 0 implying
tr
_
f(H
n
) f(H)
_
=
_
R
f
t
()(, n)d
=
_
E
E0
f
t
()(, n)d f(E
) (6.57)
for continuously dierentiable f with compact support (again compact support is
no restriction since only values in [E
0
, E
] contribute).
Theorem 6.10. Let (, n) be dened as above. Then we have the following trace
formula
(6.58) b
()
(n) = tr(H
(H
n
)
) = E
_
E
E0
1
(, n)d,
where
b
(1)
(n) = b(n),
b
()
(n) = g
(n)
1
j=1
g
j
(n)b
(j)
(n), > 1. (6.59)
Proof. The claim follows after expanding both sides of
(6.60) ln
_
(E
z)g(z, n)
_
=
_
E
E0
(, n)d
z
and comparing coecients as before.
6.2. General trace formulas and xi functions 113
The special case = 1 of equation (6.58) reads
(6.61) b(n) = E
_
E
E0
(, n)d =
E
0
+E
2
+
1
2
_
E
E0
(1 2(, n))d.
Similarly, H
n0
can be obtained as the limit of the operator H +
n
, .
n
as
. We have
(6.62)
(z, n) =
a(n)
exp
_
_
R
(, n)d
z
_
, z C(H
n
), R0,
where
(6.63)
(, n) =
1
lim
0
arg
_
( + i, n)
_
=
_
0, a(n) < 0
1, a(n) > 0
and 0 sgn(a(n))
(, n) 1. The function
(, n) is compactly supported
and we get as before
Theorem 6.11. Let
(, n)d, N,
where
b
,(0)
(n) = 1 +
2
,
b
,()
(n) = ( + 1)
(n)
a(n)
j=1
j
(n)b
,(j1)
(n), N. (6.65)
Specializing to = 0 in (6.64) we obtain
(6.66) a(n) =
1
+
1
_
R
(, n)d.
Finally, observe that
(, n
0
) for two values of and one xed n
0
determines
the sequences a(n), b(n) for all n Z.
Lemma 6.12. Let
1
,=
2
R be given. Then (
j
,
j
(., n
0
)), j = 1, 2, for
one xed n
0
Z uniquely determines a(n)
2
, b(n) for all n Z.
Proof. If we know and
(z, n
0
), we know a(n
0
) by (6.66) and
(z, n
0
) by
(6.62). Hence the result follows from Theorem 2.30.
Chapter 7
Jacobi operators with
periodic coecients
Some of the most interesting Jacobi operators are those with periodic coecients a,
b. In this chapter we will develop Floquet theory to investigate this case. This will
allow us to give a complete characterization of the qualitative behavior of solutions
and hence also a complete spectral characterization. Especially, we will see that
all important quantities associated with periodic sequences have a simple polyno
mial dependence with respect to the spectral parameter z and can be computed
explicitly. In fact, this feature is shared by a much larger class of Jacobi operators
to be considered in the next chapter. In addition, this chapter also provides the
motivation for denitions being made there.
7.1. Floquet theory
For this entire chapter we will assume a, b to be periodic with period N (for N = 1
see Sections 1.3 and 7.4).
Hypothesis H. 7.1. Suppose there is an N N such that
(7.1) a(n +N) = a(n), b(n +N) = b(n).
In addition, we agree to abbreviate
(7.2) A =
N
j=1
a(n
0
+j) =
N
j=1
a(j), B =
N
j=1
b(n
0
+j) =
N
j=1
b(j).
We start by recalling the transfer matrix (z, n, n
0
) (cf. (1.30)) which transfers
initial conditions u(n
0
), u(n
0
+1) at n
0
of a solution u into the corresponding values
u(n), u(n+1) at n. It suggests itself to investigate what happens if we move on N
steps, that is, to look at the monodromy matrix
(7.3) M(z, n
0
) = (z, n
0
+N, n
0
).
115
116 7. Periodic Jacobi operators
A rst naive guess would be that all initial conditions return to their starting
values after N steps (i.e., M(z, n
0
) = 1l) and hence all solutions are periodic. How
ever, this is too much to hope for since it already fails for N = 1 (see Section 1.3).
On the other hand, since it does not matter whether we start our N steps at
n
0
, at n
0
+ N, or even n
0
+ N, Z, we infer that M(z, n
0
) is periodic, that is,
M(z, n
0
+ N) = M(z, n
0
), n
0
Z. Moreover, we even have (z, n
0
+ N, n
0
) =
M(z, n
0
)
. Thus (z, n, n
0
) exhibits an exponential behavior if we move on N
points in each step. If we factor out this exponential term, the remainder should
be periodic.
For this purpose we rewrite M(z, n
0
) a little bit. Using periodicity (7.1) we
obtain from (1.33)
(7.4) det M(z, n
0
) = 1
and hence we can nd a periodic matrix Q(z, n
0
) such that
(7.5) M(z, n
0
) = exp (iNQ(z, n
0
)), tr Q(z, n
0
) = 0.
Now we can write
(7.6) (z, n, n
0
) = P(z, n, n
0
) exp (i(n n
0
)Q(z, n
0
)), P(z, n
0
, n
0
) = 1l.
A simple calculation
P(z, n +N, n
0
) = (z, n +N, n
0
)M(z, n
0
)
1
exp ( i(n n
0
)Q(z, n
0
))
= (z, n +N, n
0
+N) exp ( i(n n
0
)Q(z, n
0
))
= (z, n, n
0
) exp ( i(n n
0
)Q(z, n
0
)) = P(z, n, n
0
) (7.7)
shows that P(z, n, n
0
) is indeed periodic as anticipated.
Next we want to turn this result into a corresponding result for the solutions
of u = zu. The key is clearly to investigate the Jordan canonical form of M(z, n
0
)
(resp. Q(z, n
0
)).
From (1.32) we see
M(z, n
1
) = (z, n
1
+N, n
0
+N)M(z, n
0
)(z, n
0
, n
1
)
= (z, n
1
, n
0
)M(z, n
0
)(z, n
1
, n
0
)
1
(7.8)
and hence the discriminant
(7.9) (z) =
1
2
tr M(z, n
0
) =
1
2
_
c(z, n
0
+N, n
0
) +s(z, n
0
+N + 1, n
0
)
_
,
plus the eigenvalues m
(z) = 1,
are independent of n
0
. The branch in the above root is xed as follows
(7.11) ((z)
2
1)
1/2
=
1
2A
2N1
j=0
_
z E
j
,
where (E
j
)
2N1
j=0
are the zeros of (z)
2
1 (cf. Appendix A.7). This denition
implies the following expansion for large z
(7.12) m
(z) = (2(z))
1
(1 +O(
1
z
2N
)) = (
A
z
N
)
1
(1
B
z
+O(
1
z
2
))
and will ensure [m
+
(z)[ 1, [m
(z)). If s(z, n
0
+N, n
0
) ,= 0 we set
(7.14) e
(z, n
0
) =
_
1
(z, n
0
)
_
,
where
(7.15)
(z, n) =
m
(z) c(z, n
0
+N, n
0
)
s(z, n
0
+N, n
0
)
=
c(z, n
0
+N + 1, n
0
)
m
(z) s(z, n
0
+N + 1, n
0
)
.
If s(, n
0
+N, n
0
) = 0, then s(, n, n
0
) is periodic with respect to n and s(, n
0
+
N, n
0
) = m
() and hence
(z, n
0
) tends to a nite
limit
(, n
0
) as z . Thus we can set e
(, n
0
) = (0, 1) and e
(, n
0
) =
(1,
(, n
0
)).
Now the matrix
(7.16) U(z, n
0
) = (e
+
(z, n
0
), e
(z, n
0
))
will transform M(z, n
0
) into
(7.17) U(z, n
0
)
1
M(z, n
0
)U(z, n
0
) =
_
m
+
(z) 0
0 m
(z)
_
.
Moreover, we have two corresponding solutions (Floquet solutions)
(7.18) u
(n, z) = e
1
(z, n
0
)c(z, n, n
0
) +e
2
(z, n
0
)s(z, n, n
0
),
satisfying
(7.19) u
(z, n +N) = m
(z)u
(z, n).
They are linearly independent and unique up to multiples.
Case 2). (z) = 1 (and hence m
+
(z) = m
(z, n) = p
(z, n)e
iq(z)n
, p
(z, n +N) = p
(z, n).
(ii). If (z) = 1, then either all solutions satisfy
(7.27) u(z, n +N) = u(z, n)
or there are two solutions satisfying
(7.28) u(z, n) = p(z, n), u(z, n) = p(z, n) +np(z, n)
with p(z, n +N) = p(z, n), p(z, n +N) = p(z, n).
If we normalize the Floquet solutions by u
(z, n
0
) = 1 they are called Floquet
functions
(z, n, n
0
) and a straightforward calculation yields
(z, n, n
0
) = c(z, n, n
0
) +
(z, n
0
)s(z, n, n
0
),
(z, n) =
(z, n + 1, n
0
)
(z, n, n
0
)
. (7.29)
Observe that
(z, n) is periodic,
(z, n+N) =
(z, n) =
s(z, n +N + 1, n)
s(z, n +N, n)
+O(z
12N
),
+
(z, n) =
c(z, n +N, n)
s(z, n +N, n)
+O(z
12N
). (7.30)
7.2. Connections with the spectra of nite Jacobi operators 119
7.2. Connections with the spectra of nite Jacobi
operators
In this section we want to gain further insight by considering nite Jacobi matrices.
The results will also shed some light on our trace formulas.
First, we look at the operators H
1,2
n1,n2
. We are mainly interested in the special
case
H
n0
= H
,
n0,n0+N
associated with the following Jacobi matrices (cf. (1.64))
n0
= J
n0,n0+N
,
H
0
n0
= J
n0+1,n0+N+1
,
n0
=
_
_
_
_
_
_
_
_
_
b
n0+1
a
n0+1
a
n0+1
b
n0+2
.
.
.
.
.
.
.
.
.
.
.
.
.
.
. b
n0+N1
a
n0+N1
a
n0+N1
b
n0+N
_
_
_
_
_
_
_
_
_
, (7.31)
(abbreviating a(n), b(n) as a
n
, b
n
) with
b
n0+1
= b
n0+1
a
n0
1
,
b
n0+N
= b
n0
a
n0
, and R0.
Second, we look at the operators
(7.32)
H
n0
=
_
_
_
_
_
_
_
_
_
b
n0+1
a
n0+1
e
i
a
n0
a
n0+1
b
n0+2
.
.
.
.
.
.
.
.
.
.
.
.
.
.
. b
n0+N1
a
n0+N1
e
i
a
n0
a
n0+N1
b
n0+N
_
_
_
_
_
_
_
_
_
,
with [0, 2). These latter operators are associated with the following coupled
boundary conditions ( [0, 2))
(7.33)
u(n
0
) exp(i) u(n
0
+N) = 0,
u(n
0
+ 1) exp(i) u(n
0
+ 1 +N) = 0.
The spectrum of
H
n0
is given by the zeros
j
(n
0
)
N
j=1
(where
N = N 1 for
= 0, and
N = N otherwise) of the characteristic equation
(7.34) det(z
H
n0
) =
j=1
(z
j
(n
0
)).
In the special case = we set
j
(n
0
) =
j
(n
0
) and consequently
0
j
(n
0
) =
j
(n
0
+ 1) by (7.31).
As in (1.65) we can give an alternate characterization of the eigenvalues
j
(n
0
).
Recall from Remark 1.9 that
H
n0
can be obtained from H by imposing boundary
conditions at n
0
, n
0
+N. Hence any eigenfunction must satisfy these boundary con
ditions. The solution satisfying the boundary condition at n
0
is (up to a constant)
given by
(7.35) s
(z, n, n
0
) = sin()c(z, n, n
0
) cos()s(z, n, n
0
), = cot().
120 7. Periodic Jacobi operators
It is an eigenfunction of
H
n0
if and only if it also satises the boundary condition
at n
0
+N. Thus we infer for R0
cos()s
(z, n
0
+N, n
0
) + sin()s
(z, n
0
+N + 1, n
0
)
=
( +
1
)
1
A
N
j=1
(z
j
(n
0
)), (7.36)
where the leading coecient follows from (1.68). For = 0, we have
c(z, n
0
+N + 1, n
0
) =
a(n
0
)
A
N1
j=1
(z
j
(n
0
+ 1)),
s(z, n
0
+N, n
0
) =
a(n
0
)
A
N1
j=1
(z
j
(n
0
)). (7.37)
Since the function satisfying the boundary condition at one side is unique (up to a
constant), the spectrum (
H
n0
) is simple (cf. Remark 1.10).
For later use note that standard linear algebra immediately yields the following
trace relations (recall (7.2))
N1
j=1
j
(n
0
) = tr
H
n0
= B b(n
0
),
N
j=1
j
(n
0
) = tr
H
n0
= B ( +
1
)a(n
0
). (7.38)
Higher powers can be computed from (1.67).
Next, we turn to
H
n0
. Again the spectrum is given by the zeros E
N
j=1
of
the characteristic equation
(7.39) det(z
H
n0
) =
N
j=1
(z E
j
).
Our notation suggests that (
H
n0
) is independent of n
0
. In fact, extending the
eigenfunctions of
H
n0
to a solution of (1.19) on Z, we get a solution with the
property u(n +N) = e
i
u(n). Conversely, any solution of this kind gives rise to an
eigenfunction of
H
n0
. In other words, the spectrum of
H
n0
can be characterized by
(7.40) (
H
n0
) (m
+
() e
i
)(m
() e
i
) = 0 () = cos(),
where the independence of n
0
is evident. Moreover, we have (
H
n0
) = (
H
n0
) (
is to be understood mod2) since the corresponding operators are antiunitarily
equivalent.
In the special case = 0, we set
H
0
n0
=
H
+
n0
,
H
n0
=
H
n0
and similarly for
the eigenvalues E
0
j
= E
+
j
, E
j
= E
j
.
By the above analysis, E
j
are the zeros of (z) 1 and we may write
(7.41) (z) 1 =
1
2A
N
j=1
(z E
j
).
7.2. Connections with the spectra of nite Jacobi operators 121
Since the N 1 zeros of d(z)/dz must interlace the N zeros of both (z) 1 we
infer
(7.42) E
1
< E
1
E
2
< E
2
E
3
< E
(1)
N1
N
< E
(1)
N
N
for sgn(A) = (1)
N
. The numbers in the above sequence are denoted by E
j
,
0 j 2N 1. A typical discriminant is depicted below.

6
E0 E1 E2 E3 E4 E5
()
1
+1
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
This also shows (using (7.42)) that for (0, )(, 2) the spectrum is simple.
The case when the spectrum of
H
n
is degenerated can be easily characterized.
Lemma 7.4. The following conditions for E R are equivalent.
(i). E is a double eigenvalue of
H
n
.
(ii). All solutions of u = Eu satisfy u(E, n +N) = u(E, n).
(iii). M(E, n) = 1l.
(iv). E = E
2j1
= E
2j
(
H
n
) for all R .
In particular, if one of these conditions holds for one n Z, it automatically
holds for all.
The trace relation for
H
n
reads
(7.43)
N
j=1
E
j
= tr
H
n
= B, or
2N1
j=0
E
j
= 2B.
It can be used to eliminate B in the previous two trace relations (7.38) which
produces the same trace relations already found in Section 6.2. This will become
clear in Chapter 8 (see (8.11)).
Now, let us use this information to investigate H. Our choice for the branch in
(7.11) implies [m
+
(z)[ < 1 for [z[ large (see also (7.12) and Appendix A.7). Thus
+
is square summable near + for [z[ large. But since
+
is holomorphic for
z CR it is square summable for all z CR (see Lemma 2.2). So we must have
[m
+
(z)[ < 1 for z CR and hence by continuity [m
+
(z)[ 1 for z C.
In particular, we can use
,
+
) =
2a(n
0
)((z)
2
1)
1/2
s(z, n
0
+N, n
0
)
= a(n
0
)
_
+
(z, n
0
)
(z, n
0
)
_
and if we evaluate (1.99) at m = n = n
0
(u
+
(z, n
0
)
(z, n
0
)
=
N1
j=1
(z
j
(n
0
))
2N1
j=0
_
z E
j
.
122 7. Periodic Jacobi operators
Since n
0
is arbitrary we must have
(7.46)
+
(z, n, n
0
)
(z, n, n
0
) =
a(n
0
)
a(n)
s(z, n +N, n)
s(z, n
0
+N, n
0
)
=
N1
j=1
z
j
(n)
z
j
(n
0
)
.
Similarly, we compute
(z, n) =
(
+
(z, n) +)(
(z, n) +)
a(n)(
+
(z, n)
(z, n))
=
(m
+
(z) sin
1
()s
(z) sin
1
()s
(z, n +N + 1, n)
2 sin
2
()a(n)((z)
2
1)
1/2
=
a(n)
N
j=1
(z
j
(n))
2N1
j=0
_
z E
j
, (7.47)
where we have used (7.15), det M(z, n) = 1, and m
+
(z) +m
(z) = 2(z).
By (2.141) and (7.45) the spectrum of H is given by
(7.48) (H) = R[[()[ 1 =
N1
_
j=0
[E
2j
, E
2j+1
]
and it is purely absolutely continuous (
p
(H) =
sc
(H) = ).
The sets
(7.49)
0
= (, E
0
],
j
= [E
2j1
, E
2j
],
N
= [E
2N1
, ),
1 j N 1, are called spectral gaps. In addition, since g(z, n) and
(z, n) are
Herglotz functions we infer (compare the argument in Section 8.1)
(7.50)
j
(n)
j
, 1 j N 1, R ,
and
(7.51)
N
(n)
N
for a(n) < 0,
N
(n)
0
for a(n) > 0.
The most important case is = ,
j
(n) =
j
(n) and the typical situation looks
as follows.
E0 E1
1(n)
E2 E3
2(n)
E4 E5
The fact that s
j
(n
0
), n, n
0
) satises the same boundary condition at n
0
and
n
0
+N implies that s
j
(n
0
), n, n
0
) is a Floquet solution. Hence we can associate
a sign
j
(n) with
j
(n) such that
(7.52) M(n,
j
(n))
_
sin()
cos()
_
= m
j
(n)
(
j
(n))
_
sin()
cos()
_
.
7.3. Polynomial identities 123
If = we abbreviate
j
(n) =
j
(n). Moreover, if
j
(n) does not coincide
with one of the band edges E
j
, then the corresponding Floquet solution is an
eigenfunction of H
j
(n),n
and consequently
(7.53) (H
n
) = (H)
j
(n)
N
j=1
.
This characterization of the pairs (
j
(n),
j
(n)) will be useful later on because it
does not rely on periodicity.
Remark 7.5. Another way to investigate periodic operators would be a (constant
ber) direct integral decomposition of H
(7.54)
2
(Z)
=
_
[0,2)
2
(n
0
, n
0
+N + 1)
d
2
, H
=
_
[0,2)
n0
d
2
,
compare [195], Chapter XIII.16. However, since this approach will not reveal any
new information for us, we will not investigate it further.
7.3. Polynomial identities
In the previous section we have seen that several objects exhibit a simple polynomial
structure with respect to z. The present section further pursues this observation.
We rst dene the following polynomials
G(z, n) =
A
a(n)
s(z, n +N, n) =
A
a(n 1)
c(z, n +N, n 1)
=
N1
j=1
(z
j
(n)),
H(z, n) = A(s(z, n +N + 1, n) c(z, n +N, n)),
R(z) = 4A
2
((z)
2
1) =
2N1
j=0
(z E
j
). (7.55)
In terms of these polynomials the monodromy matrix reads
(7.56) M(z, n) =
1
A
_
A(z)
1
2
H(z, n) a(n)G(z, n)
a(n)G(z, n + 1) A(z) +
1
2
H(z, n)
_
.
Using (7.56) and (7.52) we get
(7.57) H(n,
j
(n + 1)) = R
1/2
(
j
(n + 1)), H(n,
j
(n)) = R
1/2
(
j
(n)),
where we have set R
1/2
(z) = 2A((z)
2
1)
1/2
and used the convenient abbreviation
(7.58) R
1/2
(
j
(n)) =
j
(n)R
1/2
(
j
(n)),
j
(n) = (
j
(n),
j
(n)).
124 7. Periodic Jacobi operators
Thus Newtons interpolation formula yields
H(z, n) =
N1
j=1
R
1/2
(
j
(n))
N1
k,=j
z
k
(n)
j
(n)
k
(n)
+ (z b(n))G(z, n)
=
N1
j=1
R
1/2
(
j
(n + 1))
N1
k,=j
z
k
(n + 1)
j
(n + 1)
k
(n + 1)
+ (z b(n + 1))G(z, n + 1), (7.59)
where the factor in front of G(z, n), G(z, n + 1) follows after expanding both sides
and considering the highest two coecients. In addition, we obtain from (7.59)
H(z, n) +H(z, n 1) = 2(z b(n))G(z, n),
H(z, n) H(z, n 1) = 2
N1
j=1
R
1/2
(
j
(n))
N1
k,=j
z
k
(n)
j
(n)
k
(n)
. (7.60)
Next, from (7.56) and (7.60) we obtain
1
A
N1
j=1
R
1/2
(
j
(n))
N1
k,=j
z
k
(n)
j
(n)
k
(n)
= c(z, n +N 1, n 1) c(z, n +N, n)
= s(z, n +N + 1, n) s(z, n +N, n 1). (7.61)
Finally, we have
(7.62)
(z, n) =
H(z, n) R
1/2
(z)
2a(n)G(z, n)
=
2a(n)G(z, n + 1)
H(z, n) R
1/2
(z)
,
and thus
(7.63) 4a(n)
2
G(z, n)G(z, n + 1) = H(z, n)
2
R(z),
which is equivalent to det M(z, n) = 1.
Remark 7.6. In the case of closed gaps the factor
(7.64) Q(z) =
(z E
2j
)
cancels from all equations since
j
(n) = E
2j1
= E
2j
if j
t
. Here
(7.65)
t
= 1 j N 1[E
2j1
= E
2j
, = 1 j N 1[E
2j1
< E
2j
.
In particular, observe
_
z E
2j1
_
z E
2j
= z
j
(n) for j
t
.
7.4. Two examples: period one and two
In this section we consider the two simplest cases when the period is one or two.
We start with period one. We assume a(n) =
1
2
, b(n) = 0 and consider them
to be periodic with period N. By inspection, the Floquet functions (n
0
= 0) are
given by
(z, n) = u
(z, n) = (z R
1/2
2
(z))
n
,
(z) = z R
1/2
2
(z). (7.66)
7.4. Two examples: period one and two 125
Hence the discriminant and the Floquet multipliers read
(z) =
1
2
(s(z, N + 1) s(z, N 1)) =
1
2
(m
+
(z) +m
(z)),
m
(z))
2
=
1
4
((z +R
1/2
2
(z))
N
(z R
1/2
2
(z))
N
)
2
= (z
2
1)s(z, N)
2
. (7.68)
Therefore we have
(7.69) E
0
= 1, E
2j1
= E
2j
=
j
=
j
= cos(
j
N
), E
2N1
= 1,
1 j N 1, and all spectral gaps
j
, 1 j N 1 are closed. We can now
verify the rst relation in (7.38) directly and from the second we obtain
(7.70)
N
=
+
1
2
.
Which shows that
N
N
for < 0 and
N
0
for 0 < .
Now let us consider the case of period N = 2. In this case A = a(0)a(1) and
B = b(0) +b(1). Moreover, one computes
(7.71) (z) =
(z
B
2
)
2
C
2
2A
, C
2
= a(0)
2
+a(1)
2
+
(b(0) b(1))
2
4
and hence
R(z) = 4A
2
((z)
2
1) = ((z
B
2
)
2
C
2
)
2
4A
2
= (z E
0
)(z E
1
)(z E
2
)(z E
3
), (7.72)
where
(7.73) E
0,1
=
B
2
_
C
2
2[A[, E
2,3
=
B
2
+
_
C
2
2[A[.
Conversely, we have B = E
0
+ E
2
= E
1
+ E
3
, C
2
= ((E
0
E
2
)
2
+ (E
1
E
3
)
2
)/8,
and [A[ = ((E
0
E
2
)
2
(E
1
E
3
)
2
)/16. Our trace formulas (7.38) produce
(7.74)
1
(0) = b(1),
1
(1) = b(0),
1
(n) = B ( +
1
)a(n)
and we observe (
1
(0)) = (
1
(1)) =
(1(0)1(1))
2
4C
2
8A
= (
1
). A short
computation shows
(7.75) m
1(0)
(
1
) =
a(1)
a(0)
, m
1(1)
(
1
) =
a(0)
a(1)
,
where m
(
1
) = m
(
1
(0)) = m
(
1
(1)). Moreover, we have
1
(0) =
1
(1) =
sgn([a(0)[ [a(1)[) and a(0) = a(1) implies (
1
)
2
= 1. Solving m
(
1
) for a(n)
shows
126 7. Periodic Jacobi operators
a(0)
2
= Am
1(0)
(
1
) = A((
1
)
1
(0)((
1
)
2
1)
1/2
),
a(1)
2
= Am
1(0)
(
1
) = A((
1
) +
1
(0)((
1
)
2
1)
1/2
). (7.76)
In particular, this shows that the coecients a(n), b(n) are expressible in terms of
E
j
, 0 j 3, and (e.g.)
1
(0),
1
(0). This result will be generalized in Section 8.1.
7.5. Perturbations of periodic operators
In this section we are going to study shortrange perturbations H of periodic oper
ators H
p
associated with sequences a, b satisfying a(n) a
p
(n) and b(n) b
p
(n)
as [n[ . Our main hypothesis for this section reads
Hypothesis H. 7.7. Suppose a
p
, b
p
are given periodic sequences and H
p
is the
corresponding Jacobi operator. Let H be a perturbation of H
p
such that
(7.77)
nZ
[n(a(n) a
p
(n))[ < ,
nZ
[n(b(n) b
p
(n))[ < .
We start with two preliminary lemmas.
Lemma 7.8. Consider the Volterra sum equation
(7.78) f(n) = g(n) +
m=n+1
K(n, m)f(m).
Suppose there is a sequence
K(n, m) such that
(7.79) [K(n, m)[
K(n, m),
K(n + 1, m)
K(n, m),
K(n, .)
1
(0, ).
Then, for given g
(0, ), fullling
the estimate
(7.80) [f(n)[
_
sup
m>n
[g(m)[
_
exp
_
m=n+1
K(n, m)
_
.
If g(n) and K(n, m) depend continuously (resp. holomorphically) on a parameter
and if
K does not depend on this parameter, then the same is true for f(n).
Proof. Using the standard iteration trick
(7.81) f
0
(n) = g(n), f
j+1
(n) =
m=n+1
K(n, m)f
j
(m),
we see that the solution is formally given by
(7.82) f(n) =
j=0
f
j
(n).
To prove uniqueness and existence it remains to show that this iteration converges.
We claim
(7.83) [f
j
(n)[
sup
m>n
[g(m)[
j!
_
m=n+1
K(n, m)
_
j
, j 1,
7.5. Perturbations of periodic operators 127
which follows from
[f
j+1
(n)[
m=n+1
K(n, m)
sup
l>m
[g(l)[
j!
_
l=m+1
K(m, l)
_
j
sup
m>n
[g(m)[
j!
m=n+1
K(n, m)
_
l=m+1
K(n, l)
_
j
sup
m>n
[g(m)[
(j + 1)!
m=n+1
_
_
l=m
K(n, l)
_
j+1
l=m+1
K(n, l)
_
j+1
_
=
sup
m>n
[g(m)[
(j + 1)!
_
l=n+1
K(n, l)
_
j+1
,
where we have used (s =
K(n, m) 0, S =
=m+1
K(n, ) 0)
(7.84) (S +s)
j+1
S
j+1
= sS
j
j
=0
(1 +
s
S
)
(j + 1)sS
j
, j N.
This settles the iteration and the estimate (7.80). The rest follows from uniform
convergence of the series (7.82).
Remark 7.9. A similar result holds for equations of the type
(7.85) f(n) = g(n) +
n1
m=
K(n, m)f(m).
Lemma 7.10. Assume (H.7.7). Then there exist solutions u
(z, .), z C, of
u = zu satisfying
(7.86) lim
n
[m
p
(z)
n/N
(u
(z, n) u
p,
(z, n))[ = 0,
where u
p,
(z, .), m
p
(z), and N are the Floquet solutions, Floquet multipliers, and
period of H
p
, respectively (cf. (7.19)). In addition, u
(z, n) =
a
p
(n
0
1
)
a(n
0
1
)
u
p,
(z, n)
n1
m=
n+1
a
p
(n
0
1
)
a(n
0
1
)
K(z, n, m)u
(z, m),
where
K(z, n, m) =
((
p
)u
p,
(z))(m)u
p,+
(z, n) u
p,
(z, n)((
p
)u
p,+
(z))(m)
W
p
(u
p,
(z), u
p,+
(z))
=
s
p
(z, n, m+ 1)
a
p
(m+ 1)
(a(m) a
p
(m)) +
s
p
(z, n, m)
a
p
(m)
(b(m) b
p
(m))
+
s
p
(z, n, m1)
a
p
(m1)
(a(m1) a
p
(m1)) (7.88)
128 7. Periodic Jacobi operators
(W
p
(., ..) denotes the Wronskian formed with a
p
rather than a).
Proof. We only prove the claim for u
+
(z, .), the one for u
p
(z)
n/N
u
+
(z, .) we get a sequence which is
bounded near +.
Suppose u
+
(z, .) satises (disregarding summability for a moment)
u
+
(z, n) =
a
p
(n)
a(n)
u
p,+
(z, n)
+
m=n+1
a
p
(n)
a(n)
m
p
(z)
(nm)/N
K(z, n, m) u
+
(z, m). (7.89)
Then u
+
(z, .) fullls u = zu and (7.86). Hence, if we can apply Lemma 7.8, we
are done. To do this, we need an estimate for K(z, n, m) or, equivalently, for the
growth rate of the transfer matrix (z, n, m). By (7.6) it suces to show
(7.90) (m
p
(z)M
p
(z, m))
n
 const(z, m) n, n N.
Abbreviate M(z) = m
p
(z)M
p
(z, m) and note that the eigenvalues of M(z) are 1
and m
p
(z)
2
. After performing the unitary transformation U(z) = (e(z), e
(z)),
where e(z) is the normalized eigenvector of M(z) corresponding to the eigenvalue
1 and e
p
(z)
2
_
.
Using
(7.92)
M(z)
n
=
_
1 (z)
n1
j=0
m
p
(z)
2j
0 m
p
(z)
2n
_
, n N,
and [m
p
(z)[ 1 we see 
M(z)
n
 max1, [(z)[n (1 + [(z)[)n, which is the
desired estimate. Since e(z) and (z) are continuous with respect to z, the constant
in (7.90) can be chosen independent of z as long as z varies in compacts.
Theorem 7.11. Suppose (H.7.7) holds. Then we have
ess
(H) = (H
p
), the point
spectrum of H is nite and conned to the spectral gaps of H
p
, that is,
p
(H)
R(H
p
). Furthermore, the essential spectrum of H is purely absolutely continuous.
Proof. That
ess
(H) =
ess
(H
p
) follows from Lemma 3.9. To prove the remaining
claims we use the solutions u
(, .) of u = u for (H
p
) found in Lemma 7.10.
Since u
(, .), (H
p
) are bounded and do not vanish near , there are
no eigenvalues in the essential spectrum of H and invoking Theorem 3.18 shows
that the essential spectrum of H is purely absolutely continuous. Moreover, (7.86)
with = E
0
implies that H E
0
is nonoscillatory since we can assume (perhaps
after ipping signs) u
p,
(E
0
, n) > 0, n Z, and by Corollary 4.11 there are
only nitely many eigenvalues below E
0
. Similarly, (using Remark 4.12) there are
only nitely many eigenvalues above E
2N+1
. Applying Corollary 4.20 in each gap
(E
2j1
, E
2j
), 1 j N, shows that the number of eigenvalues in each gap is nite
as well.
These results enable us to dene what is known as scattering theory for the
pair (H, H
p
), where H is a Jacobi operator satisfying (H.7.7).
7.5. Perturbations of periodic operators 129
Since we are most of the time interested in the case z (H
p
), we shall
normalize u
p,
(, 0) = 1 for (H
p
). In particular, note that we have u
p,
() =
u
p,
(), where the bar denotes complex conjugation. Since one computes
(7.93) W(u
(), u
()) = W
p
(u
p,
(), u
p,
()) =
2i sin(q()N)
s
p
(, N)
, (H
p
),
(s
p
(, n) is the solution of
p
u = zu corresponding to the initial condition s
p
(, 0) =
0, s
p
(, 1) = 1) we conclude that u
(), u
(, n) = ()u
(, n) +
()u
(, n), (H
p
),
where
() =
W(u
(), u
())
W(u
(), u
())
=
s
p
(, N)
2i sin(q()N)
W(u
(), u
+
()),
(z) =
W(u
(), u
())
W(u
(), u
())
=
s
p
(, N)
2i sin(q()N)
W(u
(), u
()). (7.95)
The function () can be dened for all CE
p,j
. Moreover, the Pl ucker
identity (2.169) with f
1
= u
, f
2
= u
, f
3
= u
, f
4
= u
, implies
(7.96) [()[
2
= 1 +[
()[
2
and
() =
().
Using (7.87) one can also show
(7.97)
W(u
(), u
+
()) = W
p
(u
p,
(), u
p,+
()) +
nZ
u
(, n)((
p
)u
p,
())(n)
and
(7.98) W(u
(), u
()) =
nZ
u
(, n)((
p
)u
p,
())(n).
We now dene the scattering matrix
(7.99) S() =
_
T() R
()
R
+
() T()
_
, (H
p
),
of the pair (H, H
p
), where T() = ()
1
and R
() = ()
1
()[
2
= 1
and T()R
+
() = T()R
().
The quantities T() and R
(, n) =
_
_
_
T()u
p,
(, n), n
u
p,
(, n) +R
()u
p,
(, n), n
, (H
p
).
If we regard u
p,
(, n) as incoming plain wave packet, then T()u
p,
(, n) and
R
()u
p,
(, n) are the transmitted and reected packets respectively.
130 7. Periodic Jacobi operators
The quantities T() and R
(z) = m
(z, 0)
as follows
T() =
u
(, 0)
u
(, 0)
2iIm( m
())
m
() + m
+
()
,
R
() =
u
(, 0)
u
(, 0)
m
() + m
()
m
() + m
+
()
, (H
p
). (7.101)
Here we have set m
() = lim
0
m
(, n)u
+
(, n)
W(u
(), u
+
())
= T()
s
p
(, N)
2i sin(q()N)
u
(, n)u
+
(, n)
=
s
p
(, N)
2i sin(q()N)
[u
(, n)[
2
_
1 +R
()
u
(, n)
u
(, n)
_
. (7.102)
We can now give a rst application of the trace formulas derived in Section 6.2.
Denote by E
0
E
1
< E
2
< E
2M
E
2M+1
the band edges of H, where
equality holds if and only if E
2j
= E
2j+1
is an eigenvalue of H. That is,
(7.103) (H) =
M1
_
j=0
[E
2j
, E
2j+1
].
Furthermore, dene the number
j
(n) associated with each spectral gap
j
=
(E
2j1
, E
2j
) by (compare Section 8.1 for further details)
(7.104)
j
(n) = supE
2j1
j
[g(, n) < 0
j
, 1 j M.
Then we infer
(, n) =
1
2
(E0,E)
() +
1
2
M
j=1
_
(E2j1,j(n))
()
(j(n),E2j)
()
_
+
(E,)
() +
1
arg
_
1 +R
()
u
(, n)
u
(, n)
_
(Hp)
() (7.105)
since we have
(7.106) (, n) =
1
2
+
1
arg
_
1 +R
()
u
(, n)
u
(, n)
_
, (H
p
).
Hence we obtain from (6.58)
b
()
(n) =
1
2
2M+1
j=0
E
j
M1
j=1
j
(n)
_
(Hp)
1
arg
_
1 +R
()
u
(, n)
u
(, n)
_
d. (7.107)
Remark 7.12. If H is reectionless, that is R
(, n)
_
= U()
_
c(, n)
s(, n)
_
,
where
(7.109) U() =
_
u
+
(, 0) u
+
(, 0) m
+
()
u
(, 0) u
(, 0) m
()
_
.
By the considerations at the end of Section 2.5 this choice of basis will diagonalize
the matrix measure in the eigenfunction expansion. A short calculation (using
(7.93)) shows
(7.110) [u
(, 0)[
2
=
sin(q()N)
a(0)
2
s
p
(, N)Im( m
())
, (H
p
),
that u
(, n) are not correctly normalized. However, this can be taken into account
easily. Transforming the spectral measure to this new basis yields (use (7.101))
d
ac
() = (U
1
())
d
ac
() U
1
()
=
s
p
(, N)
4 sin(q()N)
[T()[
2
_
1 0
0 1
_
(Hp)
()d. (7.111)
For the pure point part we can choose (e.g.)
(7.112) d
pp
() =
Ejp(H)
_
+
(E
j
) 0
0 0
_
d( E
j
),
where
+
(E
j
) is the norming constant corresponding to the eigenvalue E
j
, that is,
(7.113)
+
(E
j
) =
_
nZ
[u
+
(E
j
, n)[
2
_
1
and () = 0 for 0 respectively () = 1 for > 0.
Transforming (2.131) shows that U(, n) are orthogonal with respect to d =
d
ac
+d
pp
,
U(, m), U(, n)
L
2 =
1
i,j=0
_
R
U
i
(, m)U
j
(, n)d
i,j
()
_
R
U(, m)U(, n)d () =
m,n
. (7.114)
Moreover, we obtain a unitary transformation
U :
2
(Z) L
2
(R, d ) dened by
(
Uf)() =
nZ
f(n)U(, n),
(
U
1
F)(n) =
_
R
U(, n)F()d (), (7.115)
which maps the operator H to the multiplication operator by ,
(7.116)
UH
U
1
=
H,
132 7. Periodic Jacobi operators
where
(7.117)
HF() = F(), F() L
2
(R, d ).
Chapter 8
Reectionless Jacobi
operators
In this chapter we are going to look at a class of operators for which the trace
formulas of Section 6.2 become particularly simple. Based on this fact we are able
to give a detailed spectral and inverse spectral analysis.
8.1. Spectral analysis and trace formulas
In this section we discuss the direct spectral problem for a certain class of reec
tionless bounded Jacobi operators.
A Jacobi operator H is called reectionless if for all n Z,
(8.1) (, n) =
1
2
for a.e.
ess
(H),
where (, n) is the xifunction of H introduced in (6.53).
For instance, periodic operators and operators with purely discrete spectrum
are special cases of reectionless operators. The following result further illustrates
the reectionless condition (8.1).
Lemma 8.1. Suppose (H). Then the following conditions are equivalent.
(i). For all n Z, (, n) =
1
2
for a.e. .
(ii). For some n
0
Z, n
1
Zn
0
, n
0
+ 1,
(, n
0
) = (, n
0
+ 1) = (, n
1
) =
1
2
for a.e .
(iii). For some n
0
Z,
m
+
( + i0, n
0
) = m
( + i0, n
0
) for a.e. ,
or equivalently a(n
0
)
2
m
+
( + i0, n
0
) = a(n
0
1)
2
m
( + i0, n
0
) +b(n
0
).
(iv) For some n
0
Z,
Re(g(, n
0
)) = Re(h(, n
0
)) = 0 for a.e. .
133
134 8. Reectionless Jacobi operators
(v) For some n
0
Z,
m
+
( + i0, n
0
) = m
( + i0, n
0
) for a.e. .
(vi). For some n
0
Z and u
(z, n) as in (2.20),
u
+
( + i0, n
0
) = u
( + i0, n
0
) for a.e. .
Proof. Without restriction we choose n
0
= 0. Moreover, for a.e. we can
assume that both m
( + i0)[
2
+ [ m
+
( + i0)[
2
> 0,
s(, n
1
)c(, n
1
) ,= 0. Hence we x such a and abbreviate m
= m
(+i0). Using
(2.20) we see that the requirement (, n) =
1
2
, that is, g(+i0, n) = g( + i0, n),
is equivalent to
Re
_
( m
+ m
+
)(c(, n)
2
+a(n)( m
m
+
)s(, n)c(, n))
a(n)
2
( m
m
+
)s(, n)
2
_
= 0. (8.2)
Clearly (i) (ii). In order to prove that (ii) (iii) we rst pick n = 0 in
(8.2) yielding Re( m
) = Re( m
+
). Furthermore choosing n = 1 shows Im( m
) =
Im( m
+
) or Re( m
) = Re( m
+
) = 0. In the latter case, Re( m
) = Re( m
+
) = 0, we
obtain for n = n
1
that Im( m
+ m
+
)Im( m
m
+
) = 0 and since Im( m
) > 0 again
Im( m
) = Im( m
+
). This completes (ii) (iii). The case (iii) (i) is obvious
from (8.2) and the rest is evident from (2.20) and (2.92).
The last condition (vi) implies vanishing of the reection coecients R
() (cf.
(7.99)), explaining the term reectionless (see also (7.106)). Moreover, condition
(iii) allows us to strengthen Theorem 2.29 for reectionless Jacobi operators.
Lemma 8.2. Suppose H is reectionless and
ess
(H) is of positive Lebesgue mea
sure, then H is uniquely determined by one of the Weyl mfunctions m
+
(z, n
0
) or
m
(z, n
0
). The result still holds if the reectionless condition only holds for a set
of positive Lebesgue measure.
Proof. Combine Lemma 8.1 (iii) and Theorem B.8 (iii).
Since the spectrum of H is closed, it can be written as the complement of a
countable union of disjoint open intervals, that is,
(8.3) (H) = = R
_
jJ0]
j
,
where J N, J
0
= J 0,
0
= (, E
0
),
= (E
, ),
E
0
E
2j1
< E
2j
E
,
j
= (E
2j1
, E
2j
), j J, (8.4)
< E
0
< E
< ,
j
k
= for j ,= k.
We emphasize that the notation employed in (8.4) implies that E
2j
= E
2k+1
for
some k J
0
whenever E
2j
d
(H).
Next, we turn to Dirichlet eigenvalues associated with each spectral gap
j
.
Formally
j
(n) is the zero of g(z, n) in
j
. Unfortunately, this denition causes
8.1. Spectral analysis and trace formulas 135
trouble since
j
could lie at a boundary point of
j
where g(., n) is no longer
holomorphic. Hence we use the more sophisticated one
(8.5)
j
(n) = supE
2j1
j
[g(, n) < 0
j
, j J.
Strict monotonicity of g(, n) with respect to
j
(cf. (2.37)) then yields
(8.6)
g(, n) < 0, (E
2j1
,
j
(n)),
g(, n) > 0, (
j
(n), E
2j
),
j J.
Moreover, let R
ess
(H), then g(, n) = 0 implies u
(, n) = 0 or u
+
(, n) = 0
(if
d
(H) we have u
(, n) = u
+
(, n) = 0). Thus
d
(H
n
) and hence we
infer
(8.7) (H
n
) =
ess
(H)
j
(n)
jJ
.
However, observe that
j
(n) is not necessarily an eigenvalue of H
n
unless
j
(n) ,
ess
(H).
Equation (8.6) shows that (, n) = 1, (E
2j1
,
j
(n)), and (, n) = 0,
(
j
(n), E
2j
), j J. Thus we know (, n) for R
ess
(H) and if we assume
(, n) = 1/2 for
ess
(H) we obtain (see also Section 6.2)
Lemma 8.3. Let H be reectionless. Then
g(z, n) =
1
z E
exp
_
_
E
E0
(, n)d
z
_
=
1
z E
0
z E
jJ
z
j
(n)
_
z E
2j1
_
z E
2j
. (8.8)
In particular, denoting by
(E0,)
() +
(E,)
()
_
+
1
2
jJ
_
(E2j1,)
() +
(E2j,)
() 2
(j(n),)
()
_
=
1
2
(E0,E)
() +
1
2
jJ
_
(E2j1,j(n))
()
(j(n),E2j)
()
_
+
(E,)
() for a.e. R. (8.9)
In the case of two gaps the situation is depicted below.

6
E0 E1
1(n)
E2 E3
2(n)
E4 E5
(,n)
1/2
1
0
+E
jJ
(E
2j1
+E
2j
2
j
(n)
)
_
136 8. Reectionless Jacobi operators
and in the special case = 1
(8.11) b(n) =
1
2
_
E
0
+E
jJ
(E
2j1
+E
2j
2
j
(n))
_
.
Now we consider the inverse spectral problem associated with reectionless op
erators more closely. To reconstruct H we need (e.g.) m
j
(n) we only know g(z, n) and thus only a(n)
2
( m
(z, n)+ m
+
(z, n)) = g(z, n)
1
.
Thus it suces to split g(z, n)
1
into a(n)
2
m
+
(z, n) and a(n)
2
m
(z, n) = a(n
0
1
)
2
_
R
d
,n
()
z
.
Invoking Theorem 2.29 will then show that H is uniquely determined.
The process of splitting up d
n
can be reduced to splitting up the pure point,
absolutely continuous, and singularly continuous part, respectively.
We start with the pure point part and introduce
(8.15)
p
(H
n
) =
j
(n)
j
J
= R[
n
() > 0,
J N.
Note that
p
(H
n
) is known from d
n
. Even though the two sets
j
(n)
j
J
and
j
(n)
jJ
are closely related, unfortunately neither
j
(n)
j
J
j
(n)
jJ
nor
j
(n)
jJ
j
(n)
j
J
is true in general. We have observed before that
j
might
not be an eigenvalue if it lies at the boundary of its gap. On the other hand, an
accumulation point of the
j
s might be an eigenvalue (see Remark 8.11 (i) below).
Moreover, we introduce the numbers
(8.16)
R
j
(n) = lim
0
ig(
j
(n) + i, n)
1
=
n
(
j
(n)) > 0,
and
(8.17)
j
(n) = lim
0
h(
j
(n) + i, n) =
+,n
(
j
(n))
,n
(
j
(n))
+,n
(
j
(n)) +
,n
(
j
(n))
[1, 1].
Equivalently, we have (use (2.201))
(8.18) m
(z, n) =
1
j
(n)
2
R
j
(n)a(n)
2
z
j
(n)
+O(z
j
(n))
0
.
In particular,
j
(n) is either 1 (depending on whether
j
is an eigenvalue of H
+,n
or H
,n
) or in (1, +1) (if
j
is an eigenvalue of both H
,n
and hence also of H).
8.1. Spectral analysis and trace formulas 137
The numbers
R
j
(n) can be evaluated using (8.8)
R
j
(n)
2
= (
j
(n) E
0
)(
j
(n) E
)(
j
(n) E
2j1
)(
j
(n) E
2j
)
kJ\j]
(
j
(n) E
2k1
)(
j
(n) E
2k
)
(
j
(n)
k
(n))
2
. (8.19)
If
j
=
k
= E
2k
= E
2j1
for some k (resp.
j
=
k
= E
2k1
= E
2j
) the
vanishing factors (
j
k
)
2
in the denominator and (
j
E
2j1
)(
j
E
2k
) (resp.
(
j
E
2k1
)(
j
E
2j
)) in the numerator have to be omitted.
By now we know how to split up the pure point and the absolutely continuous
(which is being taken care of by the reectionless condition) part of d
n
, that is,
(8.20) d
n,,pp
(
j
(n)) =
1
j
(n)
2
d
n,pp
(
j
(n)), d
n,,ac
=
1
2
d
n,ac
.
Since we do not want to bother with the singularly continuous part, we will assume
sc
(H
n
) = , that is,
n,sc
= 0, for simplicity. In this case the spectral data E
j
, j
J 0, , plus
j
(n
0
), j J, plus
j
(n
0
), j
J, for one xed n
0
Z are minimal
and uniquely determine a(n)
2
, b(n). Without this assumption we would need to
introduce (n, .) :
sc
(H
n
) [1, 1] such that d
n,,sc
() =
1(n,)
2
d
n,sc
().
The corresponding modications are straightforward.
Now let us focus on the actual reconstruction of a(n)
2
, b(n) from given spectral
data as above and present an explicit expression of a(n)
2
, b(n) in terms of the
spectral data.
As a preparation we observe that d
n,+
d
n,
consists only of a pure point
part. In fact, due to the reectionless condition, the absolutely continuous part
cancels and the singularly continuous part is absent by assumption. As a conse
quence a(n)
2
( m
(z, n) m
+
(z, n)) is meromorphic and we obtain an interesting
expression for h(z, n) (cf. (2.201)).
Lemma 8.4. Suppose H is reectionless and
sc
(H
n
) = . Then we have
(8.21) h(z, n) = g(z, n)
_
z b(n) +
J
j
(n)
R
j
(n)
z
j
(n)
_
.
Proof. Since the (signed) measure d
n,+
d
n,
is pure point, the integral in the
corresponding Herglotz representation can be easily computed.
Our idea for reconstructing a(n), b(n) is to express the moments of m
(z, n) in
terms of the spectral data. Once this is done, the rest follows from (2.118). Since
we already know g(z, n) and h(z, n) in terms of the spectral data, it suces to
relate these quantities to m
(z, n) = z b(n)
_
1
g(z,n)
h(z,n)
g(z,n)
=
j=0
c
,j
(n)
z
j+1
, (8.22)
138 8. Reectionless Jacobi operators
where the coecients c
,j
(n) are to be determined. From (8.21) we infer
(8.23) c
,
(n) =
J
j
(n)
R
j
(n)
j
(n)
, N
0
.
Moreover, c
+,
(n) can be determined from (cf. (6.56))
(8.24)
1
g(z, n)
= z exp
_
=1
b
()
(n)
z
_
,
which implies
c
+,2
(n) = 1,
c
+,2
(n) =
1
j=1
c
+,j2
(n)b
(j)
(n), N. (8.25)
Thus c
+,
are expressed in terms of E
j
,
j
(n) (by (8.10)). Here c
+,2
(n) and
c
+,1
(n) have been introduced for notational convenience only.
In particular, combining the case = 0 with our previous results we obtain
(8.26) a(n
0
1
)
2
=
b
(2)
(n) b(n)
2
4
J
j
(n)
2
R
j
(n).
Similarly, for = 1,
b(n 1) =
1
a(n
0
1
)
2
_
2b
(3)
(n) 3b(n)b
(2)
(n) +b(n)
3
12
J
j
(n)
2
R
j
(n)
j
(n)
_
. (8.27)
However, these formulas are only the tip of the iceberg. Combining
c
,
(n) =
_
R
_
a(n)
2
d
+,n
() a(n 1)
2
d
,n
()
_
= a(n)
2
m
+,
(n) a(n 1)
2
m
,
(n), N
0
, (8.28)
with (2.118) we obtain the main result of this section.
Theorem 8.5. Let H be a given bounded reectionless Jacobi operator. Suppose
the singular continuous spectrum of H
n
is empty and the spectral data (E
j
), (
j
),
and (
j
) corresponding to H (as above) are given for one xed n Z. Then the
sequences a
2
, b can be expressed explicitly in terms of the spectral data as follows
a(n k
0
1
)
2
=
C
,n
(k + 1)C
,n
(k 1)
C
,n
(k)
2
,
b(n k) =
D
,n
(k)
C
,n
(k)
D
,n
(k 1)
C
,n
(k 1)
, k N, (8.29)
where C
,n
(k) and D
,n
(k) are dened as in (2.109) and (2.113) using m
,
(n) =
c
+,
(n)c
,
(n)
2a(n
0
1
)
. The quantities a(n)
2
, a(n1)
2
, and c
,
(n) have to be expressed in
terms of the spectral data using (8.26), (8.25), and (8.23), (8.10), respectively.
In addition, we get the neat
8.1. Spectral analysis and trace formulas 139
Corollary 8.6. Let H be a reectionless Jacobi operator with spectrum consisting
of only one band, that is (H) = [E
0
, E
E
0
)
2
16
, b(n) =
E
0
+E
2
.
If J is nite, that is, H has only nitely many spectral gaps, then
j
(n)
j
j
(n)
jJ
and we can forget about the s. This case will be investigated more
closely in Section 8.3.
Finally, we turn to general eigenvalues associated with H
n
. Associated with
each spectral gap
j
we set
(8.31)
j
(n) = supE
2j1
j
[
(, n) < 0
j
, j J.
The strict monotonicity of
(, n) with respect to
j
, j J
0
, that
is,
(8.32)
d
d
(, n) = (1 +
2
)
n
, (H )
2
n
,
j
,
then yields
(8.33)
(, n) < 0, (E
2j1
,
j
(n)),
(, n) > 0, (
j
(n), E
2j
),
j J.
Since
=
_
supE
(resp. E
0
). Summarizing,
(, n) is given by
(, n) =
1
2
(E0,E)
() +
1
2
jJ
_
(E2j1,
j
(n))
()
(
j
(n),E2j)
()
_
+
(E,
)
(), a(n) < 0, (8.35)
and
(, n) =
1
2
(E0,E)
() +
1
2
jJ
_
(E2j1,
j
(n))
()
(
j
(n),E2j)
()
_
,E0)
(), a(n) > 0. (8.36)
Thus we have for ,= 0, ,
(8.37)
(z, n) =
a(n)
z
(n)
z E
0
z E
jJ
z
j
(n)
_
z E
2j1
_
z E
2j
,
and we remark that the numbers
j
(n) are related to the spectrum of H
n
as follows
(8.38) (H
n
) =
ess
(H)
j
(n)
jJ]
.
Again observe that
j
(n) is not necessarily an eigenvalue of H
n
unless
j
(n) ,
ess
(H).
140 8. Reectionless Jacobi operators
Evaluation of (6.58) shows
b
,()
(n) =
2a(n)
_
E
+1
0
+E
+1
(n)
+1
+
jJ
(E
+1
2j1
+E
+1
2j
2
j
(n)
+1
)
_
(8.39)
and in the special case = 0,
(8.40) a(n) =
1
2( +
1
)
_
E
0
+E
(n) +
jJ
(E
2j1
+E
2j
2
j
(n))
_
.
8.2. Isospectral operators
In this section we will use the results of the previous section to characterize the
isospectral class of certain reectionless Jacobi operators.
Before we can do this, we need a condition which ensures
sc
(H
n
) = .
Hypothesis H. 8.7. Suppose is innite and the set of all accumulation points
of the set of band edges E
2j1
, E
2j
jJ
is countable.
We will denote the set of all accumulation points of the set of band edges by
(8.41)
E
j
J
,
J N.
Let us rst note some consequences of this hypothesis.
Lemma 8.8. Condition (H.8.7) holds if and only if (H) is a countable union of
disjoint closed intervals (which may degenerate to points).
Explicitly, we have
(8.42) (H) =
_
_
jJ0
j
_
_
_
j
j
_
,
where
(8.43)
j
= [E
2j
, E
(r)
2j
], j J
0
,
j
= [
E
j
,
E
(r)
j
], j
J,
with
(8.44) x
(r)
= infE
j
[x < E
j
or j = E
j
jJ]
E
j
J
.
Proof. Denote the right hand side of (8.42) by . Since x
(r)
= inf (H)[x <
we see (H). To show R (H) pick x , . There are three cases x
(r)
=
E
2j
, x
(r)
= E
2j1
, or x
(r)
=
E
j
. In the rst case we have x (E
2j1
, E
2j
) (H)
which is what we want. In the second case we set E = supE
2k
[E
2k
E
2j1
. By
minimality of x
(r)
= E
2j1
there cannot be an E
2k
in [x, E
2j1
) and we either have
E
2j1
= E
2k
, reducing it to the rst case, or E x and hence x [E, E
2j1
]
contradicting case two. Case three is similar to case two. Thus (H) = . The
intervals in (8.42) are disjoint unless E
2j
=
E
k
in which case the corresponding
intervals are equal. Hence after throwing out the superuous intervals we obtain a
disjoint union.
Conversely, let (H) =
j
J
[
E
2j
,
E
2j+1
] be a countable union of disjoint closed
intervals. Then we have E
j
E
j
. and the claim follows since
E
j
is closed.
j
the singularly continuous part must be
zero (since it is supported on a countable set). The considerations of the previous
section indicate that we can choose a number
j
(n) for each spectral gap and a
number
j
for each Dirichlet eigenvalue. Since the set of all Dirichlet eigenvalues
is contained in the set
j
(n)
jJ
E
j
J
we will specify
j
(n),
j
(n) for each
j
(n),
E
j
, j J,
J, respectively. However, the previous section shows that
j
(n)
(resp.
j
(n)) is not needed if
j
(n) (resp.
E
j
) is not in
p
(H
n
), we will get rid of
all superuous s by an equivalence relation later on.
We rst introduce the following hypothesis.
Hypothesis H.8.9. Suppose the set
__
(
j
,
j
)
_
jJ
, (
j
)
j
J
_
satises the following
requirements.
(i). (
j
,
j
) [E
2j1
, E
2j
] [1, 1], j J, and
j
[1, 1], j
J.
(ii). If
j
(E
2j1
, E
2j
) then
j
1.
(iii). If
j
= E
2j1
= E
2k
(resp.
j
= E
2j
= E
2k1
) for some j, k J then
(
j
,
j
) = (
k
,
k
) and
j
(1, 1).
(iv). If
j
=
E
k
for some j J, k
J we have
j
=
k
.
(v). For both , at least one of the sets j J[
j
< 1 or j
J[
j
< 1 is
innite.
Given Hypothesis (H.8.9) we dene the set
(8.45) T
0
() =
_
__
(
j
,
j
)
_
jJ
, (
j
)
j
J
_
[(H.8.9) holds
_
.
Furthermore, for each element in T
0
() we can dene g(z) as in (8.8) and R
j
,
R
j
,
j J,
J as in (8.16) with
j
(n) replaced by
j
,
E
j
, respectively. We will set
(8.46)
__
(
j
,
1,j
)
_
jJ
, (
1,j
)
j
J
_
=
__
(
j
,
2,j
)
_
jJ
, (
2,j
)
j
J
_
if
1,j
=
2,j
for all j j J[R
j
> 0 and if
1,j
=
2,j
for all j j
J[
R
j
> 0.
The set T
0
() modulo this equivalence relation will be denoted by T().
For any set satisfying (H.8.7) the isospectral set of reectionless Jacobi op
erators H in
2
(Z) is denoted by
(8.47) Iso
R
() = reectionless Jacobi operators H on
2
(Z)[(H) = , a > 0.
With each H Iso
R
() we can associate an element in T() by xing a base point
n Z. In fact, dening
j
(n) as in (8.5) and
j
(n),
j
(n) as in (8.17) with
j
(n)
replaced by
j
,
E
j
, respectively, we obtain a mapping
(8.48) I
n
: Iso
R
() T() .
Theorem 8.10. Suppose satises (H.8.7) and x n Z. Then the map I
n
is a
bijection.
Proof. It suces to show that I
n
is surjective (since Dirichlet data uniquely de
termine an operator). Let an element of T() be given. Dene g(z, n), d
n
,
R
j
(n),
R
j
(n) as in (8.8), (8.12), (8.13), respectively. Since d
n,pp
is supported
142 8. Reectionless Jacobi operators
on
j
jJ
E
j
J
(as noted before), we can split it up according to
d
n,,pp
(
j
) =
1
j
2
d
n,pp
(
j
),
d
n,,pp
(
E
j
(n)) =
1
j
2
d
n,pp
(
E
j
). (8.49)
Similarly, the absolutely continuous part splits up according to d
n,,ac
=
1
2
d
n,ac
.
Due to (H.8.9) this splitting is welldened and by virtue of (8.14) and Theorem 2.13
this denes an operator H. It remains to show (H) = . By Remark 3.5(ii) it
suces to consider all points
j
for which both m
(z, 0) =
_
g(z, 0)
1
(
R)(z
E)
1
_
1
, 0.
Then g
= lim
0
ig
(
E + i, 0)
1
= ,
one obtains the following case distinctions ( the sigma corresponding to
E).
(1). = 0, then
E ,
p
(H
,
).
(2). > 0, 1, then
E
p
(H
,
),
E
j0
,
p
(H
).
(3). > 0, (1, 1), then
E
p
(H
,
)
p
(H
).
(ii). Let us further comment on (H.8.9). If (ii) should not hold, then
j
would be
an eigenvalue of the operator obtained from I
1
n
(by (8.50)). Similarly, if we drop
condition (iii), then E
2j
=
j
might not be an eigenvalue.
8.3. The nitegap case
In this section we will prove some additional results in the case of reectionless
Jacobi operators with nitely many gaps. This class is closely connected to the
Toda hierarchy (to be studied in Chapter 12) and hence deserves special attention.
Given (E
j
)
2r+1
j=0
and ((
j
(n
0
),
j
(n
0
)))
r
j=1
for one xed n
0
we will reconstruct the
sequences a, b recursively. We will derive all results without using information from
the previous two sections.
8.3. The nitegap case 143
Our approach will be modeled after the information gained in the previous
chapter about periodic operators (in particular Section 7.3). However, note that
for periodic sequences (with no closed gaps) (E
j
)
2r+1
j=0
is already too much ((E
j
)
r+1
j=1
for some plus A would be enough). If H has no eigenvalues, that is E
2j
< E
2j+1
,
the constructed sequences will turn out to be quasiperiodic and to be expressible
in terms of Riemann theta functions. This case will be investigated more closely in
Section 9.2.
Let r N
0
and
(8.54) E
0
E
1
< E
2
E
3
< E
4
< E
2r1
E
2r
< E
2r+1
be given real numbers. Here we have excluded the case E
2j
= E
2j+1
since the
corresponding factors would cancel from all equations. In addition, we will assume
that E
2j
< E
2j+1
for at least one j.
Dene R
1/2
2r+2
(z) as
(8.55) R
1/2
2r+2
(z) =
2r+1
j=0
_
z E
j
,
where
. is dened as in (1.116). It is most natural to consider R
1/2
2r+2
(z) as a
function on the associated Riemann surface M which is the same as the Riemann
surface of
(8.56)
R
1/2
2g+2
(z) =
j
_
z E
2j
_
z E
2j+1
,
where = j[0 j r, E
2j
< E
2j+1
, g = [[ 1 0. As a set, M is given by all
pairs
(8.57) p = (z,
R
1/2
2g+2
(z)), z C,
plus two points at innity,
R
1/2
2g+2
the evaluation map, that is,
(8.58) (p) = z,
R
1/2
2g+2
(p) =
R
1/2
2g+2
(z), p = (z,
R
1/2
2g+2
(z)).
The points (E
2j
, 0), (E
2j+1
, 0)[j M are called branch points and g is called
the genus of M. In addition, we dene
(8.59) R
1/2
2r+2
(p) =
R
1/2
2g+2
(p)
j,
((p) E
2j
)
and p
= (z,
R
1/2
2g+2
(z))
= (z,
R
1/2
2g+2
(z)),
j
(n
0
) (E
2j1
, E
2j
) and
j
(n
0
) 1 or
j
(n
0
) E
2j1
, E
2j
and
j
(n
0
) (1, 1). (8.61)
If E
2j2
< E
2j1
=
j
(n
0
) (resp. E
2j
< E
2j+1
=
j
(n
0
)), then
j
(n
0
) is superu
ous and we will set
j
(n
0
) = 0 in this case.
(ii).
(8.62)
E
2j2
= E
2j1
=
j
(n
0
)
j1
(n
0
) =
j
(n
0
)
E
2j
= E
2j+1
=
j
(n
0
)
j+1
(n
0
) =
j
(n
0
)
,
for 1 j r. In particular,
1
(n
0
) ,= E
0
and
r
(n
0
) ,= E
2r+1
.
In the notation of the previous section we have
(8.63) T() =
_
(
j
(n
0
))
r
j=1
[(H.8.12) holds
_
with
(8.64) =
r
_
j=0
[E
2j
, E
2j+1
]
and Theorem 8.10 is valid.
We require (H.8.12) (ii) without loss of generality since otherwise the construc
tion below would give
j
(n) =
j
(n
0
) for all n and the factor z
j
(n) would
cancel with the factor
_
z E
2j2
_
z E
2j1
(resp.
_
z E
2j
_
z E
2j+1
) from
all equations. For convenience we set
(8.65) R
1/2
2r+2
(
j
(n
0
)) =
j
(n
0
)R
1/2
2r+2
(
j
(n
0
)), R
2r+2
(z) = (R
1/2
2r+2
(z))
2
.
To begin with, we dene the polynomial associated with the Dirichlet eigenval
ues
(8.66) G
r
(z, n
0
) =
r
j=1
(z
j
(n
0
))
and the residues
(8.67) R
j
(n
0
) = lim
zj(n0)
(z
j
(n
0
))
R
1/2
2r+2
(z)
G
r
(z, n
0
)
.
If
j
(n
0
) ,=
j1
(n
0
),
j+1
(n
0
) we have
(8.68) R
j
(n
0
) =
R
1/2
2r+2
(
j
(n
0
))
r
k,=j
(
j
(n
0
)
k
(n
0
))
0.
If
j
(n
0
) =
j1
(n
0
) or
j
(n
0
) =
j+1
(n
0
) the vanishing factors in the numerator
and denominator have to be omitted. For notational convenience we set in addition
(8.69)
R
j
(n
0
) =
_
j
(n
0
)R
j
(n
0
),
j
(n
0
) ,=
j1
(n
0
),
j+1
(n
0
)
j(n0)
2
R
j
(n
0
),
j
(n
0
) =
j1
(n
0
) or
j+1
(n
0
)
.
The factor 1/2 in the case where two Dirichlet eigenvalues coincide will ensure that
we do not count the corresponding residue twice.
8.3. The nitegap case 145
Using this notation we dene a second polynomial
H
r+1
(z, n
0
) =
r
j=1
R
j
(n
0
)
r
k=1
k=j
(z
k
(n
0
)) + (z b(n
0
))G
r
(z, n
0
)
= G
r
(z, n
0
)
_
z b(n
0
) +
r
j=1
R
j
(n
0
)
z
j
(n
0
)
_
, (8.70)
where
(8.71) b(n
0
) =
1
2
2r+1
j=0
E
j
r
j=1
j
(n
0
).
This denition implies
(8.72) lim
zj(n0)
H
r+1
(z, n
0
)
R
1/2
2r+2
(z)
=
j
(n
0
), 1 j r.
Our next task is to dene
j
(n
0
+ 1). This will be done using a procedure
motivated by the approach of Section 2.7. The last equation (8.72) shows that we
can dene a polynomial G
r
(z, n
0
+ 1) by (compare (2.187))
(8.73) 4a(n
0
)
2
G
r
(z, n
0
)G
r
(z, n
0
+ 1) = H
r+1
(z, n
0
)
2
R
2r+2
(z).
As we want the highest coecient of G
r
(z, n
0
+1) to be one, we have introduced a
nonzero constant a(n
0
)
2
,= 0.
First of all, note that G
r
(z, n
0
+ 1) does not vanish identically (H
2
r+1
is non
negative and R
2r+2
is not since E
2j
< E
2j+1
for at least one j) and that it is a
polynomial of degree at most r (since the highest two powers of z on the right
cancel due to the denition of b(n
0
)).
We claim that there are at least two zeros of G
r
(., n
0
)G
r
(., n
0
+1) in the interval
[E
2j1
, E
2j
]. Since
4a(n
0
)
2
G
r
(E
2j1
, n
0
)G
r
(E
2j1
, n
0
+ 1) H
r+1
(E
2j1
, n
0
)
2
0 and
4a(n
0
)
2
G
r
(E
2j
, n
0
)G
r
(E
2j
, n
0
+ 1) H
r+1
(E
2j
, n
0
)
2
0, (8.74)
this is clear if
j
(n
0
) (E
2j1
, E
2j
). If
j
(n
0
) E
2j1
, E
2j
, say
j
(n
0
) =
E
2j1
, we have to consider two cases. First, let E
2j2
< E
2j1
, then R
2r+2
(.)
has a rst order zero near E
2j1
and hence H
r+1
(z, n
0
)
2
R
2r+2
(z) < 0 for
z (E
2j1
, E
2j1
+ ) and some > 0 suciently small. Implying one zero in
(E
2j1
, E
2j
]. Second, let E
2j2
= E
2j1
, then the same conclusion applies because
of (8.72) and
j
(n
0
) (1, 1).
Since G
r
(z, n
0
+ 1) is of degree at most r we conclude that it must be of the
form
(8.75) G
r
(z, n
0
+ 1) =
r
j=1
(z
j
(n
0
+ 1)),
with
(8.76)
j
(n
0
+ 1) [E
2j1
, E
2j
].
Observe that E
2j2
= E
2j1
=
j
(n
0
) (resp. E
2j
= E
2j+1
=
j
(n
0
)) implies
E
2j1
,=
j
(n
0
+ 1) (resp. E
2j
,=
j
(n
0
+ 1)). Moreover, if E
2j2
= E
2j1
=
146 8. Reectionless Jacobi operators
j
(n
0
+ 1), then
j
(n
0
+ 1) =
j1
(n
0
+ 1) and if E
2j
= E
2j+1
=
j
(n
0
+ 1), then
j
(n
0
+ 1) =
j+1
(n
0
+ 1) (since the lefthand side of (8.73) has a double zero in
this case). In addition,
1
(n
0
+1) ,= E
0
(resp.
r
(n
0
+1) ,= E
2r+1
) since otherwise
we must have E
0
= E
1
implying that the lefthand side of (8.73) has a single zero
and the righthand side a double zero. It remains to dene
(8.77)
j
(n
0
+ 1) = lim
zj(n0+1)
H
r+1
(z, n
0
)
R
1/2
2r+2
(z)
, 1 j r.
Clearly, if H
r+1
(
j
(n
0
+ 1), n
0
) ,= 0 we have
j
(n
0
+ 1) 1 by (8.73). If
H
r+1
(
j
(n
0
+ 1), n
0
) = 0 we have
j
(n
0
+ 1) E
2j1
, E
2j
, say E
2j1
. Then, if
E
2j2
< E
2j1
we get
j
(n
0
+1) = 0 and if E
2j2
= E
2j1
we obtain
j
(n
0
+1)
(1, 1) since (8.73) is negative near E
2j1
.
Thus we have constructed the list (
j
(n
0
+ 1))
r
j=1
from the list (
j
(n
0
))
r
j=1
such that (
j
(n
0
+ 1))
r
j=1
satises (H.8.12).
The coecient a(n
0
)
2
can be calculated by expanding both sides of (8.73) (using
(8.66), (8.70), and (8.55)). Comparing the coecient of z
2g
shows
(8.78) a(n
0
)
2
=
1
2
r
j=1
R
j
(n
0
) +
b
(2)
(n
0
) b(n
0
)
2
4
> 0,
where
(8.79) b
(2)
(n) =
1
2
2r+1
j=0
E
2
j
r
j=1
j
(n)
2
.
That a(n
0
)
2
is positive follows from a(n
0
)
2
G
r
(E
2r+1
, n
0
)G
r
(E
2r+1
, n
0
+1) 0. In
addition, we obtain
H
r+1
(z, n
0
) =
r
j=1
R
j
(n
0
+ 1)
r
k,=j
(z
k
(n
0
+ 1))
+ (z b(n
0
+ 1))G
r
(z, n
0
+ 1), (8.80)
implying
(8.81) a(n
0
)
2
=
1
2
r
j=1
R
j
(n
0
+ 1) +
b
(2)
(n
0
+ 1) b(n
0
+ 1)
2
4
> 0,
and
(8.82) H
r+1
(z, n
0
+ 1) = H
r+1
(z, n
0
) + 2(z b(n
0
+ 1))G
r
(z, n
0
+ 1).
As (8.70) is symmetric in G
r
(z, n
0
+ 1) and G
r
(z, n
0
) we can even reverse this
process using (8.80) as denition for H
r+1
(which only diers from (8.70) by
the sign in front of R
1/2
2r+2
(.)) and dening
j
(n
0
) with reversed sign, that is,
j
(n
0
) = ((n
0
), lim
zj(n0)
H
r+1
(z, n
0
)/R
1/2
2r+2
(z)). Thus given (
j
(n
0
))
r
j=1
we
can recursively dene the set (
j
(n))
r
j=1
for all n Z. In summary we have proven
the following theorem.
Theorem 8.13. Given the list (
j
(n
0
))
r
j=1
we can dene the list (
j
(n))
r
j=1
for all
n Z, which again fullls (H.8.12). In addition, we get two sequences a(n), b(n)
8.3. The nitegap case 147
dened as
a(n)
2
=
1
2
r
j=1
R
j
(n) +
1
8
2r+1
j=0
E
2
j
1
4
r
j=1
j
(n)
2
b(n)
2
4
> 0,
b(n) =
1
2
2r+1
j=0
E
j
r
j=1
j
(n), (8.83)
(cf (8.69)). The sign of a(n) can be chosen freely. The sequences a(n), b(n) are real
and bounded.
Proof. It remains to show boundedness of a(n) and b(n). This is clear for b(n) by
virtue of (8.83). Combining (8.78) and (8.81)
(8.84) a(n)
2
+a(n 1)
2
=
b
(2)
(n) b(n)
2
2
we see that a(n) is bounded as well.
Remark 8.14. (i). The question when a(n)
2
, b(n) are periodic will be answered
in Theorem 9.6.
(ii). Suppose
j1
,=
j
,=
j+1
for a moment. Using the residue theorem we may
write
(8.85)
j
(n + 1) +
j
(n) =
1
2i
_
j
z
2H
r+1
(z, n)H
t
r+1
(z, n) R
t
2r+2
(z)
H
r+1
(z, n)
2
R
2r+2
(z)
dz,
(here the prime denotes the derivative with respect to z) where
j
is a closed path
in the complex plane encircling (once) the interval [E
2j1
, E
2j
] counterclockwise.
Further we may also write
(8.86)
j
(n + 1)
j
(n) =
1
2i
_
j
zd ln
G
r
(z, n + 1)
G
r
(z, n)
.
Let me point out another interesting observation. The procedure for cal
culating
j
(n) from
j
(n
0
) with n > n
0
diers from that with n < n
0
only
by two signs! One in the denition of H
r+1
and one in the denition of
j
.
These cancel in each intermediate step but not in the rst and not in the last.
This means calculating
j
(n
0
+ k) from
j
(n
0
) yields the same result as cal
culating
j
(n
0
k) from
j
(n
0
)
= (
j
(n
0
),
j
(n
0
))
= (
j
(n
0
),
j
(n
0
)).
Lemma 8.15. Denote the sequences calculated from
j
(n
0
)
r
j=1
by a
R
, b
R
. Then
we have a
R
(n
0
+k) = a(n
0
k 1), b
R
(n
0
+k) = b(n
0
k) (cf. Lemma 1.7).
In addition, if all
j
(n
0
) coincide with band edges, that is if
j
(n
0
)
r
j=1
E
2j1
, E
2j
r
j=1
, then we have
j
(n
0
+ k) =
j
(n
0
k)
, 1 j r, and hence
a(n
0
+k) = a(n
0
k 1), b(n
0
+k) = b(n
0
k).
Proof. Our observation says in formulas that
R,j
(n
0
+ k) =
j
(n
0
k)
. From
this the result for b
R
is clear from (8.83). For the corresponding result for a
R
use
(8.78) and (8.81).
148 8. Reectionless Jacobi operators
Our next task is to investigate the spectrum of the Jacobi operator H associated
with the sequences a, b. To do this we need solutions of the Jacobi and Riccati
equations rst.
Using our polynomials we dene two meromorphic functions on M,
(8.87) (p, n) =
H
r+1
(p, n) +R
1/2
2r+2
(p)
2a(n)G
r
(p, n)
=
2a(n)G
r
(p, n + 1)
H
r+1
(p, n) R
1/2
2r+2
(p)
and, the BakerAkhiezer function,
(8.88) (p, n, n
0
) =
n1
j=n0
(p, j).
Notice
(8.89) (p, n)
2
=
G
r
(p, n + 1)
G
r
(p, n)
H
r+1
(p, n) +R
1/2
2r+2
(p)
H
r+1
(p, n) R
1/2
2r+2
(p)
and hence
(8.90) (p, n, n
0
)
2
=
G
r
(p, n)
G
r
(p, n
0
)
n1
j=n0
H
r+1
(p, j) +R
1/2
2r+2
(p)
H
r+1
(p, j) R
1/2
2r+2
(p)
.
Theorem 8.16. The function (p, n) fullls
(8.91) a(n)(p, n) +
a(n 1)
(p, n 1)
= (p) b(n),
and hence (p, n, n
0
) satises
a(n)(p, n + 1, n
0
) +a(n 1)(p, n 1, n
0
) +b(n)(p, n, n
0
)
= (p)(p, n, n
0
). (8.92)
Proof. A straightforward calculation (using (8.82)) yields
a(n)(p, n) +
a(n 1)
(p, n 1)
=
H
r+1
(p, n) +R
1/2
2r+2
(p)
2G
r
(p, n)
+
H
r+1
(p, n 1) R
1/2
2r+2
(p)
2G
r
(p, n)
= (p) b(n), (8.93)
and the rest follows from
(8.94) (p, n) =
(p, n + 1, n
0
)
(p, n, n
0
)
.
With
, ) (cf.
Appendix A.7), that is,
(8.95)
(z, n) =
H
r+1
(z, n) R
1/2
2r+2
(z)
2a(n)G
r
(z, n)
, etc..
Now Lemma 6.7 together with
(z, n) = (a(n)/z)
1
(1 +O(z
1
)) implies
8.3. The nitegap case 149
Theorem 8.17. The functions
(z, n, n
0
) are square summable near and the
Green function of H is given by (recall the notation from (1.100))
g(z, n) =
G
r
(z, n)
R
1/2
2r+2
(z)
, z (H),
h(z, n) =
H
r+1
(z, n)
R
1/2
2r+2
(z)
, z (H). (8.96)
Hence the spectrum of H reads
(H) =
r
j=0
[E
2j
, E
2j+1
],
ac
(H) =
j
[E
j
, E
j+1
],
sc
(H) = ,
p
(H) = E
j
[j , . (8.97)
In addition, if
j
(n) ,=
j
(n
0
), we have
j(n)
(
j
(n), n, n
0
) = 0 if
j
(n)
2
= 1 and
(
j
(n), n, n
0
) =
+
(
j
(n), n, n
0
) = 0 if
j
(n)
2
< 1.
The facts concerning the spectrum of H can be read o from the following
lemma:
Lemma 8.18. The Weyl Mmatrix is given by
(8.98) M(z) =
1
2a(0)R
1/2
2r+2
(z)
_
2a(0)G
r
(z, 0) H
r+1
(z, 0)
H
r+1
(z, 0) 2a(0)G
r
(z, 1)
_
,
and the associated matrix valued measure d by
d
0,0
() =
G
r
(, 0)
iR
1/2
2r+2
()
ac(H)
()d +
Ejp(H)
P(E
j
, 0, 0)d( E
j
),
d
0,1
() =
H
r+1
(, 0)
2ia(0)R
1/2
2r+2
()
ac(H)
()d +
Ejp(H)
P(E
j
, 0, 1)d( E
j
),
d
1,1
() =
G
r
(, 1)
iR
1/2
2r+2
()
ac(H)
()d +
Ejp(H)
P(E
j
, 1, 1)d( E
j
),
(8.99)
where P(E
j
, n, m) is the residue of G(z, n, m) at z = E
j
(cf. (2.34)). Here () =
0 for 0 and () = 1 for > 0.
Proof. The Mmatrix easily follows from (8.96) and d
0,0
, d
0,1
, d
1,1
are imme
diate from (B.31) and (B.41) (cf. Remark 3.5(i)).
We remark that
j
(n
0
) in the case of
j
(n) = E
2j2
= E
2j1
can be alterna
tively interpreted in terms of Weyl mfunctions
(8.100)
a(n 1)
2
a(n)
2
lim
zj(n)
m
+,n
(z)
m
,n
(z)
= lim
zj(n)
+
(z, n)
(z, n)
=
1 +
j
(n)
1
j
(n)
> 0.
Last, we collect some useful relations. First note
a(n)
2
G
r
(z, n + 1) a(n 1)
2
G
r
(z, n 1) + (z b(n))
2
G
r
(z, n)
= (z b(n))H
r+1
(z, n). (8.101)
150 8. Reectionless Jacobi operators
Further, from (8.87) we get
(8.102) (p, n)(p
, n) =
G
r
(p, n + 1)
G
r
(p, n)
,
and hence
(8.103) (p, n, n
0
)(p
, n, n
0
) =
G
r
(p, n)
G
r
(p, n
0
)
.
Moreover,
(p, n) (p
, n) =
R
1/2
2r+2
(p)
a(n)G
r
(p, n)
,
(p, n) +(p
, n) =
H
r+1
(p, n)
a(n)G
r
(p, n)
. (8.104)
8.4. Further spectral interpretation
In this section we will try to relate (
j
(n))
r
j=1
to the spectra of the operators H
n
associated with a, b. In fact we will even be a bit more general.
We rst dene a polynomial K
r+1
(z, n) for R by
(p, n) + =
H
r+1
(p, n) + 2a(n)G
r
(p, n) +R
1/2
2r+2
(p)
2a(n)G
r
(p, n)
=
2a(n)K
r+1
(p, n)
H
r+1
(p, n) + 2a(n)G
r
(p, n) R
1/2
2r+2
(p)
. (8.105)
Thus we have
K
r+1
(z, n) = G
r
(z, n + 1) +
a(n)
H
r+1
(z, n) +
2
G
r
(z, n)
=
a(n)
r+1
j=1
(z
j
(n)), R0, (8.106)
where (
j
(n))
r
j=1
, with r = r if = 0 and r = r + 1 otherwise, are the zeros
of K
r+1
(z, n). Since we have K
r+1
(z, n) G
r
(z, n + 1) as 0 we conclude
j
(n) =
j
(n +1) +O() for 1 j r and
r+1
(n) = a(n)
1
+b(n +1) +O().
Also note that the analogue of (8.96) reads (z (H))
K
r+1
(z, n)
R
1/2
2r+2
(z)
=
(
+
(z, n + 1) +
+
(z, n))(
(z, n + 1) +
(z, n))
W(
(z),
+
(z))
=
on n
0
(which cancels).
As in the last section we can use
(8.108) (H
r+1
(z, n) + 2a(n)G
r
(z, n))
2
R
2r+2
(z) = 4a(n)
2
K
r+1
(z, n)G
r
(z, n)
to show
(8.109)
j
(n) [E
2j1
, E
2j
], 1 j r,
8.4. Further spectral interpretation 151
and ( R0)
(8.110)
r+1
(n) E
2r+1
for a(n) < 0,
r+1
(n) E
0
for a(n) > 0,
(whether
r+1
(n) E
2r+1
or
r+1
(n) E
0
follows from the behavior as 0
together with a continuity argument). Evaluating (8.108) at z =
j
(n) yields
(8.111) R
2r+2
(
j
(n)) = (H
r+1
(
j
(n), n) + 2a(n)G
r
(
j
(n), n))
2
,
and hence we dene
(8.112)
j
(n) = lim
z
j
(n)
H
r+1
(z, n) + 2a(n)G
r
(z, n)
R
1/2
2r+2
(z)
.
This implies (use (8.106))
R
1/2
2r+2
(
j
(n)) = H
r+1
(
j
(n), n) 2a(n)G
r
(
j
(n), n)
= H
r+1
(
j
(n), n) + 2a(n)
1
G
r
(
j
(n), n + 1). (8.113)
For the sake of completeness we also state the trace formulas in the nitegap
case.
Lemma 8.19. The zeros
j
(n) of K
r+1
(z, n) fulll for R (cf. (6.64))
(8.114) b
,()
(n) =
2a(n)
_
2r+1
j=0
E
+1
j
2
r+1
j=1
j
(n)
+1
_
.
Especially for = 0
(8.115) a(n) =
1
2( +
1
)
_
2r+1
j=0
E
j
2
r+1
j=1
j
(n)
_
.
We further compute
(8.116) ((p, n) +)((n, p
) +) =
K
r+1
(p, n)
G
r
(p, n)
,
and
a(n + 1)
2
K
r+1
(z, n + 1) = a(n)
2
G
r
(z, n) +b
(n + 1)H
r+1
(z, n)
+b
(n + 1)
2
G
r
(z, n + 1), (8.117)
where b
n
.
Theorem 8.20. Suppose R . Then
(H
n
) = (H)
j
(n)
r
j=1
,
ac
(H
n
) =
ac
(H),
sc
(H
n
) = , (8.118)
where r = r for = 0, and r = r +1 otherwise. Whether
j
(n) is an eigenvalue
of H
+,n
or H
,n
depends on whether
j
(n) = +1 or
j
(n) = 1, respectively.
If
j
(n) (1, 1), then
j
(n
0
) is an eigenvalue of both H
,n0
if
j
(n)
p
(H)
and no eigenvalue for both otherwise. Moreover, for all
1,2
R we have
1
j
(n) ,=
2
j
(n) provided
1
,=
2
.
152 8. Reectionless Jacobi operators
Proof. The claims concerning the spectra are immediate by looking at Weyl m
functions and applying Lemma B.7 and the following considerations. It remains
to show
1
j
(n) ,=
2
j
(n) provided
1
,=
2
. Suppose the contrary. Then we have
1
= (
2
j
(n), n) = (
1
j
(n), n) =
2
a contradiction. Nevertheless we remark
that the case
1
j
(n) =
2
j
(n) (but
1
j
(n) ,=
2
j
(n)) can occur.
Next, we will investigate how
j
(n) varies with . Therefore we use (8.106) to
calculate
(8.119)
r+1
(z, n)
z=
j
(n)
in two ways
a(n)
r+1
k,=j
(
j
(n)
k
(n))
j
(n)
=
1
a(n)
_
H
r+1
(
j
(n), n) + 2a(n)G
r
(
j
(n), n)
_
. (8.120)
This yields
Lemma 8.21. The variation of the zeros
j
(n) of K
r+1
(z, n) with respect to
R0 is described by
(8.121)
j
(n) =
R
1/2
2r+2
(
j
(n))
r+1
k,=j
(
j
(n)
k
(n))
, 1 j r + 1.
If
j
(n) =
j+1
(n) the vanishing factors in the numerator and denominator have
to be omitted.
Finally, we note that the connection with (8.23) is given by
(8.122) c
,
(n) =
r
j=1
j
(n)
R
j
(n), N.
Chapter 9
Quasiperiodic Jacobi
operators and Riemann
theta functions
In the previous chapter we have seen how Riemann surfaces arise naturally in the
investigation of reectionless nitegap sequences. However, there we have only
used them as a convenient abbreviation. In this chapter we want to use Riemann
surfaces as an independent tool which will help us to gain further insight. For
simplicity we will only consider the case of nonsingular surfaces, that is, r = g in
(8.59) (i.e.,
R
2g+2
(z) = R
2r+2
(z)). The corresponding sequences will turn out to
be expressible in terms of Riemann theta functions and are hence quasiperiodic.
The present chapter is entirely independent of Chapter 8 in the sense that
one could take Theorem 9.2 as a denition for the BakerAkhiezer function. But
without our knowledge from Chapter 7 and 8 there is no reason why this object
should be of any interest. Nevertheless we will show how all major results from
Chapter 8 can be obtained using only methods from this chapter.
9.1. Riemann surfaces
The purpose of this section is mainly to clarify notation. For further information
and references we refer the reader to Appendix A.
We consider the Riemann surface M associated with the following function
(9.1) R
1/2
2g+2
(z), R
2g+2
(z) =
2g+1
j=0
(z E
j
), E
0
< E
1
< < E
2g+1
,
g N. M is a compact, hyperelliptic Riemann surface of genus g. Again we will
choose R
1/2
2g+2
(z) as the xed branch
(9.2) R
1/2
2g+2
(z) =
2g+1
j=0
_
z E
j
,
153
154 9. Quasiperiodic Jacobi operators
where
. is deed as in (1.116).
A point of M is denoted by
(9.3) p = (z, R
1/2
2g+2
(z)), z C, or p =
and M comes with three holomorphic maps on it. The involution map
(9.4)
: M M
(z, R
1/2
2g+2
(z)) (z, R
1/2
2g+2
(z))
= (z, R
1/2
2g+2
(z))
,
the projection
(9.5)
: M C
(z, R
1/2
2g+2
(z)) z
,
and the evaluation map
(9.6)
R
1/2
2g+2
: M C
(z, R
1/2
2g+2
(z)) R
1/2
2g+2
(z)
.
The sets
= (z, R
1/2
2g+2
(z))[z C(H) are called upper, lower sheet, respec
tively.
Clearly, the eigenvalues
j
(n) can be viewed as points on our Riemann surface
j
(n) = (
j
(n),
j
(n)R
1/2
2g+2
(
j
(n))) M,
R
1/2
2g+2
(
j
(n)) =
j
(n)R
1/2
2g+2
(
j
(n)). (9.7)
The divisor associated with the eigenvalues
j
(n) is dened by
(9.8) T
(n)
(
j
(n)) = 1, 1 j g,
and 0 otherwise, where g = g + 1 if R0 and g = g if 0, . We have
(9.9) deg(T
(n)
) = g
and since
p
(H
n
) is simple, T
(n)
is nonspecial,
(9.10) i(T
(n)
) = 0,
by Lemma A.20. We set T
= T
(0)
.
We proceed with considering the function (p, n)+ with R0. Equation
(8.105) shows that there is a simple pole at
(n)
T
(n)
T
, R0.
If we want to include the case = 0 we have to add an additional eigenvalue at
+
(the eigenvalue
g+1
(n) disappears at innity)
(9.12) ((p, n)) = T
(n+1)
T
(n)
T
+T
+
.
9.2. Solutions in terms of theta functions 155
In principal, some of the poles and zeros might still cancel, but Theorem 8.20 shows
that this does not happen.
By taking Abels map
p0
on both sides we get ( R0)
p0
(T
(n)
) =
p0
(T
(n)
+T
),
p0
(T
(n+1)
+T
+
) =
p0
(T
(n)
+T
). (9.13)
This shows that
p0
(T
(n)
) is independent of R0 which can also be shown
using (8.121) (compare (13.52)).
Combining the last two equations yields for arbitrary
(9.14)
p0
(T
(n+1)
+T
+
) =
p0
(T
(n)
+T
).
A result we will soon investigate further.
Next, we are interested in the poles and zeros of the BakerAkhiezer function.
Because of (8.88) the divisor of (p, n) = (p, n, 0) reads
(9.15) ((p, n)) = T
(n)
T
+n(T
+
T
)
and by taking Abels map on both sides we obtain
(9.16)
p0
(T
(n)
) =
p0
(T
) nA
(
+
).
9.2. Solutions in terms of theta functions
We rst x the representatives of a canonical homology basis a
j
, b
j
g
j=1
on M.
We require a
j
to surround the points E
2j1
, E
2j
(changing sheets twice) and b
j
to
surround E
0
, E
2j1
on the upper sheet. The corresponding canonical basis
j
g
j=1
for 1
1
(M) (the space of holomorphic dierentials) has the following representation
(9.17) =
g
j=1
c(j)
j1
d
R
1/2
2g+2
and we have
(9.18)
_
aj
k
=
j,k
,
_
bj
k
=
j,k
.
To every compact Riemann surface of genus g N belongs a Riemann theta
function (cf. Sections A.5 and A.6)
(9.19) (z) =
mZ
g
exp 2i
_
m, z +
m, m
2
_
, z C
g
,
with z, z =
g
j=1
z
j
z
j
and the matrix of bperiods of the dierentials (9.17). It
has the following fundamental properties
(z) = (z),
(z +m+ n) = exp 2i
_
n, z
n, n
2
_
(z), n, m Z
g
. (9.20)
We introduce
(9.21) z(p, n) =
A
p0
(p)
p0
(T
(n)
)
p0
C
g
, z(n) = z(
+
, n),
156 9. Quasiperiodic Jacobi operators
where
p0
is the vector of Riemann constants
(9.22)
p0,j
=
1
g
k=1
j,k
2
, p
0
= (E
0
, 0),
and A
p0
(
p0
) is Abels map (for divisors). The hat indicates that we regard it as a
(singlevalued) map from
M (the fundamental polygon associated with M) to C
g
.
z(p, n) is independent of the base point p
0
and we note that z(
+
, n) = z(
, n+1)
mod Z
g
. We recall that the function (z(p, n)) has zeros precisely at the points
j
(n), 1 j g, that is,
(9.23) (z(p, n)) = 0 p
j
(n)
g
j=1
.
Since divisors of the type T
with
j
[E
2j1
, E
2j
] are of special importance
to us, we want to tell whether a given divisor T M
g
(M
g
, the gth symmetric
power of M) is of the above type or not by looking at
p0
(T).
Lemma 9.1. Introduce two manifolds
M
D
=
g
j=1
1
([E
2j1
, E
2j
])
g
1
S
1
,
J
D
= [A] J(M)[ [i Im(A+
p0
)] = [0]
g
1
S
1
, (9.24)
where [.] denotes the equivalence classes in the Jacobian variety J(M). Then the
mapping
(9.25)
p0
: M
D
J
D
p0
(T
)
is an isomorphism.
Proof. Let
j
[E
2j1
, E
2j
] and observe
(9.26) i Im
_
A
p0,k
(
j
)
_
=
j,k
2
.
This can be shown as in Section A.7 by taking all integrals as limits along the real
axis (we may intersect some bcycles though we are required to stay in
M, but we
only add acycles which are real). Thus
p0
(M
D
) J
D
. Since divisors in M
D
are
nonspecial by Lemma A.20 we see that
p0
is an immersion by (A.86) and injective
by Lemma A.9. Moreover,
p0
(M
D
) is open (since holomorphic maps are open)
and compact (since M
D
is). So
p0
(M
D
) is both open and closed, and hence must
be the whole of J
D
by connectedness.
As a consequence we note that when n changes continuously to n+1 in (9.16),
j
(n) changes continuously to
j
(n + 1), always remaining in its gap. Thus we
obtain
(9.27) (z(p, m))(z(p, n)) > 0
for (p) (, E
0
) (E
2g+1
, ) and m, n Z. Moreover, since the set of
all possible Dirichlet eigenvalues M
D
is compact, we infer that there are positive
constants C
1
C
2
such that
(9.28) C
1
[(z(n))[ C
2
.
9.2. Solutions in terms of theta functions 157
Finally, we recall
+,
, the normalized abelian dierential of the third kind
with simple poles at
g
j=1
(
j
)
R
1/2
2g+2
d,
where the constants
j
have to be determined from the normalization
(9.30)
_
aj
+,
= 2
E2j
_
E2j1
g
k=1
(z
k
)
R
1/2
2g+2
(z)
dz = 0,
which shows
j
(E
2j1
, E
2j
). With these preparations we are now able to express
the BakerAkhiezer function (p, n) in terms of theta functions.
Theorem 9.2. The BakerAkhiezer function is given by
(9.31) (p, n) = C(n, 0)
(z(p, n))
(z(p, 0))
exp
_
n
_
p
p0
+,
_
,
and is given by
(9.32) (p, n) = C(n)
(z(p, n + 1))
(z(p, n))
exp
_
_
p
p0
+,
_
.
The hat indicates that we require the path of integration to lie in
M. The base point
p
0
has been chosen to be (E
0
, 0) (any other branch point would do as well). The
constants C(n) and C(n, n
0
) are real and read
(9.33) C(n)
2
=
(z(n 1))
(z(n + 1))
, C(n, m) =
n1
j=m
C(j).
Proof. The function is at least welldened on
M since changing the path in
M
can only change the values of the integral in the exponential by 2i. But by
(9.34)
_
aj
+,
= 0,
the bperiods are given by (cf. (A.21))
(9.35)
_
bj
+,
= 2i
A
,j
(
+
),
which shows together with (9.20) that it extends to a meromorphic function on M.
Since is xed by its poles and zeros up to a constant (depending on n) it remains
to determine this constant. This can be done using (
+
, n)(
, n) = 1, which
follows from (8.103). From Section A.7 we recall
A
p0
(
+
) =
A
p0
(
) mod Z
g
and similarly we obtain
_
p
p0
+,
=
_
p
p0
+,
for (p) < E
0
. This justies
our choice for C(n, 0)
2
. By (9.27), C(n, m)
2
is positive and the formula for now
follows from (p, n) = (n + 1, p)/(n, p).
The sign of C(n) is related to that of a(n) as will be shown in Theorem 9.4
below. Next, let us prove some basic properties of and .
158 9. Quasiperiodic Jacobi operators
Lemma 9.3. Set (H) =
g
j=0
[E
2j
, E
2j+1
]. The functions and have the
following properties
(i). and hence is real for (p) R(H).
(ii). Let z R, then we have
(9.36) [
mZ
g
exp 2i
_
m+
1
2
n, z +
1
2
n +
1
2
m+
1
2
n, (m+
1
2
n)
_
,
z C
g
, which are clearly real for z R
g
. Inserting
(z(
k
, n)) = exp 2i
_
1
8
(k), (k) +
1
2
(k), Re(z(
k
, n))
_
_
0
(k)
_
(Re(z(
k
, n))) (9.40)
into shows that the complex factors cancel since C(n) R and z
j
(
k
, n + 1) =
z
j
(
k
, n) +
A
,j
(
+
). The argument for > E
2g+1
and < E
0
is even easier
since Im(z( , n)) = Im(
_
p0
+,
) = 0 in this case.
Now we turn to the second claim. Since the theta functions are quasiperiodic
we only have to investigate the exponential factor. For the asymptotic behavior we
need to know Re(
_
p
p0
+,
). We rst assume p
+
, (p) = R, and choose
as integration path the lift of the straight line from E
0
+ i to + i and take the
limit 0 (we may intersect as many bcycles as we like since all aperiods are
zero). Hence
(9.41)
_
p
p0
+,
=
_
E0
g
j=1
(
j
)d
R
1/2
2g+2
.
Clearly, (9.41) is negative for < E
0
. For [E
0
, E
1
], (9.41) is purely imaginary
and hence the real part is 0. At E
1
our real part starts to decrease from zero until
it hits its minimum at
1
. Then it increases until it reaches 0, which must be at
E
2
by (9.30). Proceeding like this we get the desired behavior of k(z). For p
+,
).
9.2. Solutions in terms of theta functions 159
Next we are interested in expanding near
, z)
induced by the projection on the upper/lower sheet
). We start with
(9.42) exp
_
p
_
p0
+,
_
=
_
a
z
_
1
_
1 +
b
z
+
c
z
2
+O(
1
z
3
)
_
1
, p
,
where a,
exp
_
_
E0
g
j=1
(z
j
)
R
1/2
2g+2
(z)
dz
_
> 0
and for
b we get from (9.30)
(9.44)
b =
1
2
2g+1
j=0
E
j
g
j=1
j
.
For the theta functions we get near
+
(cf. (A.119) and dont confuse z and
z)
(z(p, n + 1))
(z(p, n))
=
(z(n + 1))
(z(n))
_
1 +
1
z
g
j=1
B
j
(n + 1)
+
1
2z
2
_
_
g
j=1
B
j
(n + 1)
_
2
+
g
j=1
_
2g+1
k=0
E
k
2
+
c
j
(g 1)
c
j
(g)
_
B
j
(n + 1)
+
g
j=1
g
k=1
BB
j,k
(n + 1)
_
+O(
1
z
3
)
_
, (9.45)
with
B
j
(n) = c
j
(g)
w
j
ln
_
(w +z(n))
(w +z(n 1))
_
w=0
,
BB
j,k
(n) = c
j
(g)c
k
(g)
2
w
j
w
k
ln
_
(w +z(n))
(w +z(n 1))
_
w=0
. (9.46)
Near
(z, n) =
_
a(n)
z
_
1
_
1
b(n +
1
0
)
z
+O(
1
z
2
)
_
,
160 9. Quasiperiodic Jacobi operators
where the sequences a(n), b(n) are given by
a(n)
2
= a
2
(z(n + 1))(z(n 1))
(z(n))
2
,
b(n) =
b +
g
j=1
c
j
(g)
w
j
ln
_
(w +z(n))
(w +z(n 1))
_
w=0
. (9.48)
The sings of a(n) and C(n) must be opposite
(9.49) a(n) = a C(n)
(z(n + 1))
(z(n))
.
The sequences a(n), b(n) are obviously quasiperiodic with g periods (i.e., their
Fourier exponents possess a gterm integral basis).
We remark that we have
(9.50)
j
(n
0
)
g
j=1
E
2j1
, E
2j
g
j=1
[z(p, n
0
+k)] = [z(p
, n
0
k)].
Inserting this into (9.48) yields again a(n
0
+k) = a(n
0
k1), b(n
0
+k) = b(n
0
k)
(compare Lemma 8.15).
The expansion coecients a,
b can be interpreted as averages of our coecients
a(n) and b(n) as follows.
Lemma 9.5. Suppose a(n) > 0 and (p) (, E
0
), then
(9.51) lim
N
N
_
(p, N) = exp
_
p
_
p0
+,
_
.
In particular, we obtain
(9.52) a = lim
N
N
_
N
j=1
a(j),
b = lim
N
1
N
N
j=1
b(j),
that is, a is the geometric mean of the as and
b is the arithmetic mean of the bs.
Proof. The rst claim follows from lim
N
N
_
[(z(N))[ = 1 (cf. (9.28)). The
second follows by comparing asymptotic coecients with (9.47) respectively (6.27).
+,
=
2
N
E2j
_
E2j1
t
(z)dz
((z)
2
1)
1/2
=
2
N
ln [(z) + ((z)
2
1)
1/2
[
E2j
E2j1
= 0. (9.56)
For the bperiods we get
_
bj
+,
=
2
N
j1
k=0
E
2k+1
_
E
2k
t
(z) dz
((z)
2
1)
1/2
=
2i
N
j1
k=0
(1)
gk
E
2k+1
_
E
2k
t
(z) dz
_
1 (z)
2
=
2i
N
j1
k=0
(1)
gk
arcsin((z))
E
2k+1
E
2k
=
2i
N
j. (9.57)
Which proves the assertion, since we infer 2N
A
p0
(
+
) Z
g
.
Further we obtain in the periodic case for the Floquet multipliers
(9.58) m(p) = (p, N) = (1)
N
sgn(A) exp
_
N
_
p
p0
+,
_
and a =
N
_
[A[,
b =
B
N
.
Next we will show that the BakerAkhiezer function is even xed by its (nite)
poles and its behavior at innity
Lemma 9.7. Let T
M
g
be nonspecial (i.e., i(T
) = 0). Then there is a
meromorphic function (n, .) on M fullling
(9.59) ((n)) T
+n(T
+
T
).
Moreover, (n, .) is unique up to a multiple depending only on n and the divisor
T
zer
(n) M
g
dened by
(9.60) ((n)) = T
zer
(n) T
+n(T
+
T
),
is also nonspecial.
Proof. Abbreviate T
n
= T
n(T
+
T
) = 0 and
hence r(T
1
) 1 (otherwise, if r(T
1
) > 1, a proper linear combination of two
elements would lie in /(T
+ T
)
(9.61) a(n)(p, n) +
a(n 1)
(p, n 1)
+b(n) (p) = O(
1
z
)
or equivalently
(9.62) a(n)(p, n + 1) +a(n 1)(p, n 1) ((p) b(n))(p, n) = z
n
O(
1
z
).
The function on the left fullls the requirements for a BakerAkhiezer function
(9.59) and T
+
+ T
is part of T
zer
(n). Since T
zer
(n) is nonspecial, this is
impossible (by Lemma A.20) unless the left hand side is zero. This is clearly an
independent proof that (p, n) satises (p, n) = (p)(p, n).
Remark 9.8. From (9.61) we also obtain a dierent representation of b(n). If we
choose p = p
0
it reads
(9.63) b(n) = E
0
+ a
(z(n 1))
(z(n))
(z(p
0
, n + 1))
(z(p
0
, n))
+ a
(z(n))
(z(n 1))
(z(p
0
, n 1))
(z(p
0
, n))
.
We note that (8.101)(8.104) can be easily veried independently by considering
poles, zeros, and the behavior at
and comparing the second term in the expansion we get our trace relation
(9.64) b(n) =
1
2
2g+1
j=0
E
j
g
j=1
j
(n).
Upon evaluating the integral
(9.65) I =
1
2i
_
M
(.)d ln((z(., n))),
we get from the residue theorem
(9.66) I =
g
j=1
j
(n) +
p]
res
p((.)d ln((z(., n))),
and by direct calculation (using (A.83))
(9.67) I =
g
j=1
_
aj
j
.
9.3. The elliptic case, genus one 163
Which yields the relation
(9.68) b(n) =
1
2
2g+1
j=0
E
j
g
j=1
_
aj
j
+
g
j=1
c
j
(g)
w
j
ln
_
(w +z(n))
(w +z(n 1))
_
w=0
,
and thus
(9.69)
g
j=1
_
aj
j
=
g
j=1
j
.
9.3. The elliptic case, genus one
In this section we will show how the formulas of the last section can be expressed
in terms of Jacobis elliptic functions if g = 1. All integrals are evaluated using the
tables in [38].
Suppose E
0
< E
1
< E
2
< E
3
are given and R
1/2
4
(z) is dened as usual. We
introduce
(9.70) k =
_
E
2
E
1
E
3
E
1
E
3
E
0
E
2
E
0
(0, 1), k
t
=
_
E
3
E
2
E
3
E
1
E
1
E
0
E
2
E
0
(0, 1),
such that k
2
+k
t
2
= 1, and
(9.71) u(z) =
_
E
3
E
1
E
3
E
0
E
0
z
E
1
z
, C =
2
_
(E
3
E
1
)(E
2
E
0
)
.
Furthermore, we will use the following notation for Jacobis elliptic integrals of the
rst
(9.72)
_
z
0
dx
_
(1 x
2
)(1 k
2
x
2
)
= F(z, k),
second
(9.73)
_
z
0
_
1 x
2
1 k
2
x
2
dx = E(z, k),
and third kind
(9.74)
_
z
0
dx
(1
2
x
2
)
_
(1 x
2
)(1 k
2
x
2
)
= (z,
2
, k).
We set F(1, k) = K(k), E(1, k) = E(k), and (1,
2
, k) = (
2
, k), where k
(0, 1),
2
R and all roots are assumed to be positive for x (0, 1). We remark
that (z,
2
, k) has simple poles at z
2
=
2
and that E(z, k) has a simple pole at
.
First we determine , the abelian dierential of the rst kind (in the charts
(
, z)).
(9.75) = c(1)
d
R
1/2
4
= c(1)
dz
R
1/2
4
(z)
.
The normalization ([38], 254.00)
(9.76)
_
a1
= 2c(1)
_
E2
E1
dz
R
1/2
4
(z)
= 2c(1)CK(k) = 1
164 9. Quasiperiodic Jacobi operators
yields
(9.77) c(1) =
1
2CK(k)
.
In addition, we obtain ([38], 252.00)
(9.78)
1,1
=
_
b1
= 2c(1)
_
E1
E0
dx
R
1/2
4
(x)
= 2ic(1)CK(k
t
) = i
K(k
t
)
K(k)
.
Next we are interested in Abels map A
p0
, p
0
= (E
0
, 0). Let p = (z, R
1/2
4
(z)),
then we have ([38], 252.00)
(9.79) A
p0
(p) = [
_
p
p0
] = [c(1)
_
z
E0
dx
R
1/2
4
(x)
] = [
F(u(z), k)
2K(k)
]
and especially for p =
+
(9.80) A
p0
(
+
) = [
F(
_
E3E1
E3E0
, k)
2K(k)
].
The vector of Riemann constants is independent of the base point p
0
and clearly
given by
(9.81) = [
1
1,1
2
].
Now we turn to the Dirichlet eigenvalue
1
(n). We recall
(9.82) A
p0
(
1
(n)) = A
p0
(
1
(0)) 2nA
p0
(
+
).
If we set A
p0
(
1
(0)) = [ +
1,1
2
], where [0, 1), we obtain
(9.83) A
p0
(
1
(n)) = [
1,1
2
+
1
(n)
_
1(n)
E1
c(1)dx
R
1/2
4
(x)
] = [ +
1,1
2
n
F(
_
E3E1
E3E0
, k)
K(k)
],
with
1
(n) = (
1
(n),
1
(n)R
1/2
4
(
1
(n))). Evaluation of the integral yields ([38],
254.00)
(9.84)
_
1(n)
E1
c(1)dx
R
1/2
4
(x)
=
F(
_
E2E0
E2E1
1(n)E1
1(n)E0
, k)
2K(k)
and hence we infer
(9.85)
1
(n) = E
1
1
E2E1
E2E0
E0
E1
sn
2
(2K(k) 2nF(
_
E3E1
E3E0
, k))
1
E2E1
E2E0
sn
2
(2K(k) 2nF(
_
E3E1
E3E0
, k))
,
where sn(z) is Jacobis elliptic function (i.e., the inverse of F(z, k)). The sheet can
be read o from the sign of sn(2K(k) 2nF(. . . )).
9.4. Some illustrations of the RiemannRoch theorem 165
Now we search a more explicit expression for the dierential
,+
. We rst
determine the constant
1
from ([38], 254.00, 254.10, 340.01)
_
a1
,+
= 2
_
E2
E1
x
1
R
1/2
4
(x)
dx
= 2C
_
(E
1
E
0
)(
E
2
E
1
E
2
E
0
, k) + (E
0
1
)K(k)
_
= 0, (9.86)
which implies
(9.87)
1
= E
0
+
(E
1
E
0
)
K(k)
(
E
2
E
1
E
2
E
0
, k).
The bperiod is given by (cf. (9.35))
(9.88)
_
b1
,+
= 2
_
E1
E0
x
1
R
1/2
4
(x)
dx = 2i
_
1
F(
_
E3E1
E3E0
, k)
K(k)
_
.
And for p = (z, R
1/2
4
(z)) near p
0
we have ([38], 251.00, 251.10, 340.01)
_
p
p0
,+
=
_
z
E0
x
1
R
1/2
4
(x)
dx
= C(E
1
E
0
)
_
(1
(
E2E1
E2E0
, k)
K(k)
)F(u(z), k)
(u(z),
E
3
E
0
E
3
E
1
, k)
_
. (9.89)
To complete this section we recall that Riemanns theta function can be expressed
as a Jacobi theta function, namely
(9.90) (z) =
3
(z) =
nZ
e
i(2nz +
1,1
n
2
)
.
9.4. Some illustrations of the RiemannRoch
theorem
In this section we give some illustrations of the RiemannRoch theorem (Theo
rem A.2) in connection with hyperelliptic Riemann surfaces not branched at innity
and obtain a basis for the vector space /(nT
T
), n N
0
.
Theorem 9.9. Assume T
0 to be nonspecial (i.e., i(T
) = 0) and of degree g.
For n N
0
, a basis for the vector space /(nT
T
) is given by,
_j1
=0
(., )
_
n
j=0
, (9.91)
or equivalently, by
/(nT
T
) = span (., j)
n
j=0
. (9.92)
166 9. Quasiperiodic Jacobi operators
Proof. Since the above functions are linearly independent elements of the vector
space /(nT
T
) it remains to show that they are maximal. From i(T
) =
0 = i(nT
+T
) and the RiemannRoch theorem we obtain
r(nT
T
) = 1 +n, (9.93)
which proves the theorem.
Chapter 10
Scattering theory
In Section 7.5 we have considered perturbations of periodic Jacobi operators, that
is, we have started developing scattering theory with periodic background. In
this chapter we will investigate sequences a(n) > 0, b(n) which look asymptotically
like the free ones (compare Section 1.3).
10.1. Transformation operators
To begin with, we state the main hypothesis which will be assumed to hold through
out this chapter.
Hypothesis H. 10.1. Suppose a, b (Z, R) satisfy a > 0 and
(10.1) n(1 2a(n))
1
(Z), nb(n)
1
(Z).
In addition, it will turn out convenient to introduce
(10.2) A
(n) =
n1
m=
2a(m), A
+
(n) =
m=n
2a(m), A = A
(n)A
+
(n).
From Section 1.3 we know that it is better to consider the free problem in terms of
k = z
z 1
(k, n) of ( (k +k
1
)/2)f = 0 which asymptotically
look like the free ones, that is,
(10.3) f
(k, n) = k
n
(1 +o(1)) as n ,
respectively, lim
n
f
(k, n)k
n
= 1.
They are called Jost solutions and their properties are summarized in the
following theorem.
Theorem 10.2. Let a(n), b(n) satisfy (H.10.1), then there exist Jost solutions
f
(k, n) of
(10.4) f
(k, n) =
k +k
1
2
f
(k, n)k
n
= 1.
Moreover
f(k, n) = A
(n)k
n
f
p
(z) =
k
1
, s
p
(z, n, m) =
k
k
2
1
(k
nm
k
mn
) (see Section 1.3) and compute the limit
lim
k0
k
nm
K(z, n, m) = 2a(m) 1.
The previous theorem shows that we may set
(10.6) f
+
(k, n) =
k
n
A
+
(n)
_
1 +
j=1
K
+,j
(n)k
j
_
, [k[ 1,
where K
+,j
(n) are the Fourier coecients of
f
+
(k, n) given by
(10.7) K
+,j
(n) =
1
2i
_
]k]=1
f
+
(k, n)k
j
dk
k
, j N.
Taking the limit n in this formula (using the estimate (7.80) and dominated
convergence) shows lim
n
K
+,j
(n) = 0, j N.
Inserting (10.6) into (10.4) (set K
+
0
(n) = 1, K
+
j
(n) = 0, j < 0) yields
K
+,j+1
(n) K
+,j+1
(n 1) = 2b(n)K
+,j
(n) K
+,j1
(n)
+ 4a(n)
2
K
+,j1
(n + 1). (10.8)
Summing up (with respect to n) using lim
n
K
+,j
(n) = 0, j N, we obtain
K
+,1
(n) =
m=n+1
2b(m),
K
+,2
(n) =
m=n+1
_
2b(m)K
+,1
(m) + (4a(m)
2
1)
_
,
K
+,j+1
(n) = K
+,j1
(n + 1)
m=n+1
_
2b(m)K
+,j
(m)
+ (4a(m)
2
1)K
+,j1
(m+ 1)
_
. (10.9)
A straightforward induction argument shows
(10.10) [K
+,j
(n)[ D
+,j
(n)C
+
(n + [[
j
2
]] + 1), j N,
where
(10.11) C
+
(n) =
m=n
c(m), D
+,m
(n) =
m1
j=1
(1 +C
+
(n +j)),
10.1. Transformation operators 169
and c(n) = 2[b(n)[ +[4a(n)
2
1[. It is important to note that C
+
1
+
(Z). In fact,
summation by parts shows
(10.12)
m=n
mc(m) = (n 1)C
+
(n) +
m=n
C
+
(m)
since
(10.13) lim
n
nC
+
(n + 1) lim
n
m=n+1
mc(m) = 0.
Moreover, 1 D
+,m
(n) D
+
(n) = lim
j
D
+,j
(n) and lim
n
D
+
(n) = 1.
This gives also an estimate for
[K
+,j+1
(n) K
+,j+1
(n 1)[ 2[b(n)[[K
+,j
(n)[ +[4a(n)
2
1[[K
+,j1
(n)[
+[K
+,j1
(n + 1) K
+,j1
(n)[
c(n)D
+
(n)C
+
(n + [[
j
2
]] + 1) +[K
+,j1
(n + 1) K
+,j1
(n)[. (10.14)
Using induction we obtain
(10.15) [K
+,j
(n) K
+,j
(n 1)[ D
+
(n)C
+
(n)C
+
(n +[[
j 1
2
]] +1) +c(n +[[
j
2
]]).
In addition, the estimate (10.10) for K
+,j
(n) gives us an estimate on the rate
of convergence in (10.3)
(10.16) [f
+
(k, n)A(n)k
n
1[
_
_
[k[
1 [k[
D
+
(n)C
+
(n + 1), [k[ < 1
2[k[D
+
(n)
m=n
C
+
(m), [k[ 1
.
Moreover, since K
+,j
(n) C
(1 + C
)
j1
, C
= lim
n
C
+
(n), we can
take the limits (K
0
= 1)
K
1
= lim
n
K
+,1
(n) = 2
mZ
b(m),
K
2
= lim
n
K
+,2
(n) =
mZ
_
2b(m)K
+,1
(m) + (4a(m)
2
1)
_
K
j+1
= lim
n
K
+,j+1
(n) = K
j1
mZ
_
2b(m)K
+,j
(m)
+ (4a(m)
2
1)K
+,j1
(m+ 1)
_
. (10.17)
(It will become clear later why we have dropped the + subscript.)
Associated with K
+
(n, n +j) = K
+,j
(n) is the operator
(10.18) (/
+
f)(n) =
1
A
+
(n)
_
f(n) +
m=n+1
K
+
(n, m)f(m)
_
, f
+
(Z),
which acts as a transformation operator for the pair
0
, .
Theorem 10.3. Let
0
be the free Jacobi dierence expression. Then we have
(10.19) /
+
f = /
+
0
f, f
+
(Z).
170 10. Scattering theory
Proof. A straightforward calculation shows that /
+
= /
+
0
is equivalent to
(10.8).
By our estimate (10.10), /
+
maps
p
+
(Z), p [1, ) , into itself. More
over, by Lemma 7.8, /
+
is an automorphism of
+
(Z). Considering f
0
(Z) we
see that its inverse /
+
is given by
(10.20) (/
+
f)(n) = A
+
(n)
_
f(n) +
m=n+1
L
+
(n, m)
m1
j=n
2a(j)
f(m)
_
, f
+
(Z),
with
L
+
(n, m) = K
+
(n, m)
m1
=n+1
K
+
(n, )L
+
(, m)
= K
+
(n, m)
m1
=n+1
L
+
(n, )K
+
(, m), m > n.
From /
+
=
0
/
+
we infer that the coecients L
+,j
(n) = L
+
(n, n +j) satisfy
L
+,j+1
(n) L
+,j+1
(n 1) = 2b(n +j)L
+,j
(n) 4a(n +j 1)
2
L
+,j1
(n)
+L
+,j1
(n + 1) (10.21)
as well as an estimate of type (10.10). In particular, /
+
is an automorphism of
p
+
(Z), p [1, ) .
Similarly, we set
(10.22) f
(k, n) =
k
n
A
(n)
_
1 +
j=1
K
,j
(n)k
j
_
, [k[ 1,
which yields after inserting it into (1.19) (set K
,0
(n) = 1, K
,j
(n) = 0, j < 0)
K
,j+1
(n + 1) K
,j+1
(n) = 2b(n)K
,j
(n) +K
,j1
(n)
4a(n 1)
2
K
,j1
(n 1). (10.23)
As before we obtain (lim
n
K
,j
(n) = 0, j N)
K
,1
(n) =
n1
m=
2b(m),
K
,2
(n) =
n1
m=
2b(m)K
,1
(m) (4a(m1)
2
1),
K
,j+1
(n) = K
,j1
(n)
n1
m=
_
2b(m)K
,j
(m)
(4a(m1)
2
1)K
,j1
(m1)
_
. (10.24)
Furthermore, K
,j
(n) = K
f)(n) =
1
A
(n)
_
f(n) +
n1
m=
K
(n, m)f(m)
_
, f
(Z),
10.2. The scattering matrix 171
with inverse /
.
10.2. The scattering matrix
Our next objective is to introduce the scattering matrix. For [k[ = 1, k
2
,= 1, the
Jost solutions f
(k, n), f
(k
1
, n) are linearly independent since we have
(10.26) W(f
(k), f
(k
1
)) =
1 k
2
2k
.
In particular, we may set
(10.27) f
(k, n) = (k)f
(k
1
, n) +
(k)f
(k), f
(k))
W(f
(k), f
(k
1
))
=
2k
1 k
2
W(f
+
(k), f
(k)),
(k) =
W(f
(k), f
(k
1
))
W(f
(k), f
(k
1
))
=
2k
1 k
2
W(f
(k), f
(k
1
)). (10.28)
Note that we have
(10.29) (k) = (k),
(k) =
(k),
+
(k
1
) =
(k),
and k = k
1
for [k[ = 1. Moreover, the Pl ucker identity (2.169) with f
1
= f
,
f
2
= f
, f
3
= f
, f
4
= f
, implies
(10.30) [(k)[
2
= 1 +[
(k)[
2
, [
+
(k)[ = [
(k)[.
The function (k) can be dened for all [k[ 1. Moreover, the power series of (k)
follows formally from
(k) =
2k
1 k
2
lim
n
a(n)(f
+
(k, n)f
(k, n + 1) f
+
(k, n + 1)f
(k, n))
=
1
A
lim
n
(k, n) =
1
A
j=0
K
j
k
j
, (10.31)
where K
j
= lim
n
K
,j
(n) from (10.17). To make this precise we need to show
that we can interchange limit and summation in the last line. This follows from
[K
,j
(n)[ C
(1 +C
)
j1
at least for [k[ < (1 +C
)
1
. Hence it follows for all
[k[ < 1 since (k) is holomorphic for [k[ < 1.
The function (k) has even more interesting properties. To reveal its connection
with Kreins spectral shift theory, we compute its derivative. By virtue of (2.29)
we obtain
d
dk
W(f
+
(k), f
(k)) = W
m
(f
+
(k),
d
dk
f
(k)) +W
n
(
d
dk
f
+
(k), f
(k))
1 k
2
2k
2
n
j=m+1
f
+
(k, j)f
(k, n) =
n
k
f
(n, k) + k
n
o(1), n , and
f
(k, n) = k
n
((k) +o(1)), n , respectively. Hence we see
W
m
(f
+
(k),
d
dk
f
(k)) = m
1 k
2
2k
2
(k)
1
2k
2
(k) +o(1), m ,
W
n
(
d
dk
f
+
(k), f
(k)) = n
1 k
2
2k
2
(k)
1
2
(k) +o(1), n + (10.33)
and using this to evaluate the limits m , n shows
(10.34)
d
dk
(k) =
1
k
nZ
(f
+
(k, n)f
(k, n) (k)).
Next, observe that, due to (H.10.1), HH
0
is trace class. Indeed, the multiplication
operators by b(n), 2a(n) 1 are trace class and trace class operators form an ideal
in the Banach space of bounded linear operators. So we can rewrite this equation
in the form
d
dz
ln (k(z)) =
jZ
(G(z, n, n) G
0
(z, n, n))
= tr((H z)
1
(H
0
z)
1
), (10.35)
that is, A(k) (since A(0) = 1) is the perturbation determinant of the pair
H, H
0
. This implies ([155], Theorem 1)
(10.36) (k(z)) =
1
A
exp
_
_
R
()d
z
_
,
where
(10.37)
() =
1
lim
0
arg (k( + i)) [1, 1]
is of compact support. However, the reader should be warned that (k(z)) is not
Herglotz since we do not have 0
() 1 in general.
In addition, Kreins result implies
(10.38) tr(H
(H
0
)
) =
_
R
()d.
The spectrum of H can be easily characterized upon specializing to the case
a
p
(n) = 1/2, b
p
(n) = 0 in Theorem 7.11.
Theorem 10.4. Suppose (H.10.1). Then we have
ess
(H) =
ac
(H) = [1, 1],
p
(H) = =
k +k
1
2
[ [[ > 1, (k) = 0. (10.39)
Moreover, the essential spectrum of H is purely absolutely continuous and the point
spectrum of H is nite.
Hence we can denote the number of eigenvalues by N N and set
(10.40)
p
(H) =
j
=
k
j
+k
1
j
2
1jN
.
Note that because of (10.30) all zeros of of (k) lie inside the unit circle. Moreover,
there can be no nonreal zeros (otherwise H would have a nonreal eigenvalue).
10.2. The scattering matrix 173
(Note also (0) = A
1
> 0 by (10.31).) Using Theorem 3.20 we can even provide
a bound on the number of eigenvalues.
Theorem 10.5. Suppose (H.10.1). Then
(10.41) dimRan P
(1,)
(H) 1 +
_
nZ
c
(n)
_
2
,
where
(10.42) c
(n) =
_
n(n + 1)[1 2a(n)[ +[n[ max(b(n), 0).
Proof. We only prove the rst bound, the second being similar. Upon replacing
b(n) by min(b(n), 0) we can assume b(n) 0 (cf. [195], Theorem XIII.10). Our
goal is to apply Theorem 3.20. Since the limit of G
0
(, n, m) as 1 does not
exist, we will consider H
and use
dimRan P
(,1)
(H) 1 + dimRan P
(,1)
(H
+
)
+ dimRan P
(,1)
(H
). (10.43)
Where the last claim follows from the fact that between two eigenvalues of H, there
can be at most one Dirichlet eigenvalue (see Lemma 3.7). Hence we need to consider
H
0,+
. From (0 < m n, < 1)
(10.44) G
0,+
(, n, m) = 2k
1 k
2m
1 k
2
k
nm
(k =
2
1 (1, 0)) we infer
(10.45) [G
0,+
(, n, m)[ 2 min(m, n) 2
mn, 1.
Evaluating Theorem 3.20 as in Corollary 3.21 gives
(10.46) dimRan P
(,1)
(H
+
)
_
n=1
c
+
(n)
_
2
.
The bound for H
(k)
R
+
(k) T(k)
_
, [k[ = 1,
where T(k) = (k)
1
and R
(k) = (k)
1
(k)[
2
= 1, T(k)R
+
(k) +T(k)R
(k) = 0,
and T(k) = T(k), R
(k) = R
(k).
174 10. Scattering theory
The quantity T(k) is called transmission coecient and R
(k)[
2
C
2
[1 k
2
[
2
,
where we can choose C = 2 max
]k]=1
[W(f
+
(k), f
(k)[
2
) is integrable over the unit circle.
We proceed with a closer investigation of T(k). Using the Laurent expansion
for (k) one computes
(10.51) T(k) = A
_
1 +k
mZ
2b(m) +O(k
2
)
_
.
The only poles of T(k) are at k
j
, where
j
= (k
j
+ k
1
j
)/2
p
(H). The residue
can be computed from (10.34) which yields (cf. (2.33))
(10.52) res
kj
T(k) =
k
j
+,j
j
= k
j
,j
,
where
(10.53)
1
,j
=
mZ
[f
(m, k
j
)[
2
, f
(n, k
j
) =
j
f
+
(n, k
j
).
The constants
,j
are usually referred to as norming constants. The sets
(10.54) S
(H) = R
(k), [k[ = 1; (k
j
,
,j
), 1 j N
are called left/right scattering data for H, respectively. We will show that they
determine the sequences a, b uniquely in the following sections. As a rst step
toward this result we note
Lemma 10.7. Either one of the sets S
j=1
[k
j
[(k k
1
j
)
k k
j
_
exp
_
1
4i
_
]]=1
ln(1 [R
()[
2
)
+k
k
d
_
,
(for [k[ < 1) and
(10.56)
R
(k)
R
+
(k)
=
T(k)
T(k)
,
+,j
,j
=
_
res
kj
T(k)
k
j
_
2
.
Proof. It suces to prove the formula for T(k) since evaluating the residues pro
vides
j
and thus
,j
=
2
j
,j
. For this purpose consider
(10.57) t(k) =
N
j=1
k k
j
kk
j
1
T(k)
which is holomorphic for [k[ < 1 and has neither poles nor zeros inside the unit
disk. In addition, t(k) is continuous for [k[ 1 and has at most simple zeros at
k = 1. Applying the residue theorem in a circle with radius r < 1 we obtain
1
2i
_
]]=r
ln(t())
+k
k
d
= 2 ln(t(k)) ln(t(0)),
1
2i
_
]]=r
1
ln(t(
1
))
+k
k
d
= ln(t(0)). (10.58)
10.3. The GelfandLevitanMarchenko equations 175
Letting r 1 (dominated convergence), using t(k) = t(k
1
), [k[ = 1, and adding
up the resulting equations yields
(10.59)
1
2i
_
]]=1
ln([t()[
2
)
+k
k
d
_
.
This proves the claim upon observing [t(k)[ = [T(k)[, [k[ = 1, and T(0) > 0.
Comparing coecients in the asymptotic expansions of (10.55) and (10.36)
gives another trace formula
(10.61)
K
m
=
1
_
0
ln(1 [R
(e
i
, t)[
2
) cos(m)d +
N
j=1
k
m
j
k
m
j
m
,
where
K
m
= K
m
m1
j=1
j
m
K
mj
K
j
are the expansion coecients of
(10.62) ln (k) = ln A+
j=1
K
j
k
j
.
Moreover, expanding ln (k(z)) one can express the traces tr(H
(H
0
)
) in terms
of the coecients K
m
, for instance,
tr
_
H (H
0
)
_
=
1
2
K
1
,
tr
_
H
2
(H
0
)
2
_
=
1
16
(2K
2
+K
2
1
),
etc. . (10.63)
10.3. The GelfandLevitanMarchenko equations
Now we want to set up a procedure which allows the reconstruction of H from its
scattering data S
(n, m).
In addition, the decay condition (H.10.1) is reected by the estimates (10.10)
and (10.15). These turn out to be equivalent to the corresponding estimates for
K
(H).
Since K
+
(n, m) are essentially the Fourier coecients of f
+
(k, n) we compute
the Fourier coecients of (10.27). Therefore we multiply
(10.64) T(k)f
(k, n) =
_
f
+
(k
1
, n) +R
+
(k)f
+
(k, n)
_
176 10. Scattering theory
by (2i)
1
A
+
(n)k
n+j1
(j N
0
) and integrate around the unit circle. We rst
evaluate the right hand side
A
+
(n)
2i
_
]k]=1
f
+
(k
1
, n)k
n+j
dk
k
= K
+
(n, n +j),
A
+
(n)
2i
_
]k]=1
R
+
(k)f
+
(k, n)k
n+j
dk
k
=
m=0
K
+
(n, n +m)
F
+
(2n +m+j),
(10.65)
where
(10.66)
F
+
(n) =
1
2i
_
]k]=1
R
+
(k)k
n
dk
k
2
(Z, R)
are the Fourier coecients of R
+
(k
1
). By Parsevals identity we have
(10.67)
nZ
F
+
(n)
2
=
1
2i
_
]k]=1
[R
+
(k)[
2
dk
k
< 1.
Next we evaluate the left hand side using the residue theorem (take a contour inside
the unit disk enclosing all poles and let this contour approach the unit circle)
A
+
(n)
2i
_
]k]=1
T(k)f
(k, n)k
n+j
dk
k
= A
+
(n)
N
=1
k
+,
(k
, n)k
n+j1
+A
+
(n)
T(0)
A
(n)
0
(j)
= A
+
(n)
N
=1
+,
f
+
(k
, n)k
n+j
+A
+
(n)
2
0
(j)
=
m=0
K
+
(n, n +m)
N
=1
+,
k
2n+m+j
+A
+
(n)
2
0
(j). (10.68)
Dening the GelfandLevitanMarchenko operator
(10.69) T
+
n
f(j) =
m=0
F
+
(2n +m+j)f(m), f
2
(N
0
),
where
(10.70) F
+
(n) =
F
+
(n) +
N
=1
+,
k
n
,
leads us to the GelfandLevitanMarchenko equation
(10.71) (1l +T
+
n
)K
+
(n, n +.) = A
+
(n)
2
0
.
Similarly, multiplying
(10.72) T(k)f
+
(k, n) =
_
f
(k
1
, n) +R
(k)f
(k, n)
_
by (2i)
1
A
(n)k
n+j1
(j N
0
) and integrating around the unit circle we get
for K
(n, m)
(10.73) (1l +T
n
)K
(n, n +.) = A
(n)
2
0
,
10.3. The GelfandLevitanMarchenko equations 177
where
(10.74) T
n
f(j) =
m=0
F
(2n mj)f(m), f
2
(N
0
),
and
(10.75) F
(n) =
1
2i
_
]k]=1
R
(k)k
n
dk
k
+
N
=1
,
k
n
.
Our next objective is to consider the GelfandLevitanMarchenko operator T
+
n
a little closer. The structure of the GelfandLevitanMarchenko equation
K
+
(n, n +j) +F
+
(2n +j) +
=1
F
+
(2n + +j)K
+
(n, n +)
= A
+
(n)
0
(j) (10.76)
suggests that the estimates (10.10) and (10.15) for K
+
(n, n+j) should imply similar
estimates for F
+
(2n +j). This will be shown rst.
As a warmup we prove a discrete Gronwalltype inequality.
Lemma 10.8. Suppose C
1
, C
2
, f
+
(Z, [0, )), C
3
1
+
(Z, [0, )), C
1
C
3
< 1,
and
(10.77) f(n) C
1
(n)
_
C
2
(n) +
j=n
C
3
(j)f(j)
_
.
Then we have
(10.78) f(n) C
1
(n)
_
C
2
(n) +
j=n
C
1
(j)C
2
(j)C
3
(j)
j
=n
(1 C
1
()C
3
())
_
.
Proof. Using g(n) =
j=n
C
3
(j)f(j) one computes
(10.79) g(n) g(n + 1) C
3
(n)C
1
(n)(C
2
(n) +g(n))
and hence (solving for g(n))
(10.80) g(n)
j=n
C
1
(j)C
2
(j)C
3
(j)
j
=n
(1 C
1
()C
3
())
.
The rest follows from f(n) C
1
(n)(C
2
(n) +g(n)).
178 10. Scattering theory
Solving the GelfandLevitanMarchenko equation (10.76) for F
j
(n) = [F
+
(2n+
j)[, j = 1, 2, and using the estimate (10.10) we obtain
F(n) D
+
(n)
_
2C
+
(n + 1) +
=1
([F
+
(2n + + 1)[ +[F
+
(2n + + 2)[)
C
+
(n + [[
2
]] + 1)
_
D
+
(n)
_
2C
+
(n + 1) +
=0
[F
+
(2n + + 2)[C
+
(n +k + 1)
+ 2
=1
[F
+
(2n + + 1)[C
+
(n +k) +
=1
[F
+
(2n + + 2)[C
+
(n +k)
2D
+
(n)
_
C
+
(n + 1) +
=n
C
+
()F()
_
, (10.81)
where F(n) = F
1
(n) +F
2
(n). And hence by the lemma
[F
+
(2n + 1)[ +[F
+
(2n + 2)[ 2D
+
(n)C
+
(n + 1)
_
1 +
j=n
2D
+
(j)C
+
(j)
j
=n
(1 2D
+
()C
+
())
_
. (10.82)
for n large enough. In summary we have
(10.83) [F
+
(2n +j)[ O(1)C
+
(n + [[
j
2
]] + 1),
where O(1) are terms of order zero as n . Next we want to derive an estimate
for F
+
(2 +j), j = 1, 2. Set G
j
(n) = [F
+
(2n + j) F
+
(2n + j + 2)[, j = 1, 2,
then
G
1
(n) [K
+,1
(n) K
1,+
(n + 1)[
+
j=1
[F
+
(2n +j + 1)K
+,j
(n) F
+
(2n +j + 3)K
+,j
(n + 1)[
c(n + 1) +
j=1
[F
+
(2n +j + 1)[[K
+,j
(n) K
+,j
(n + 1)[
+
j=1
[F
+
(2n +j + 1) F
+
(2n +j + 3)[[K
+,j
(n + 1)[
c(n + 1)
+O(1)
j=1
C
+
(n + [[
j
2
]] + 1)(C
+
(n + 1)C
+
(n + [[
j+1
2
]]) +c(n + [[
j
2
]]))
+O(1)
j=1
[F
+
(2n +j + 1) F
+
(2n +j + 3)[C
+
(n + [[
j
2
]] + 2)
O(1)
_
c(n) +
j=n
(G
1
(j) +G
2
(j))C
+
(j + 1)
_
, (10.84)
10.3. The GelfandLevitanMarchenko equations 179
where c(n) = c(n + 1) +C
+
(n)C
+
(n + 1). Similarly, we obtain
(10.85) G
2
(n) O(1)
_
c(n) +
j=n
(G
1
(j) +G
2
(j))C
+
(j + 1)
_
.
Adding both equations gives an estimate for G(n) = G
1
(n) + G
2
(n) and again by
the lemma
(10.86) G(n) O(1)
_
c(n) +
j=n
c(j)C
+
(j + 1)
_
.
As a consequence we infer
(10.87) nG(n) O(1)
_
n c(n) +C
+
(n)
j=n
j c(j)
_
1
(N).
Similar results hold for F
(H).
Hypothesis H. 10.9. The scattering data
(10.88) S
(H) = R
(k), [k[ = 1; k
j
,
,j
, 1 j N
satisfy the following conditions.
(i). The consistency condition
(10.89)
R
(k)
R
+
(k)
=
T(k)
T(k)
,
+,j
,j
=
_
res
kj
T(k)
k
j
_
2
,
where T(k) is dened using (10.55).
(ii). The reection coecients R
(k) = R
(k), C
2
[1 k
2
[
2
1 [R
(k)[
2
for some C > 0. The Fourier coecients
F
of R
(k
1
) satisfy
(10.91)
j=1
j[
(j)
F
(j 2)[ < .
(iii). The values k
j
(1, 0) (0, 1), 1 j N, are distinct and the norming
constants
,j
, 1 j N, are positive.
Finally, we collect some properties of the operator T
+
n
.
Theorem 10.10. Fix n Z and consider T
+
n
:
p
(N
0
)
p
(N
0
), p [1, ),
as above. Then T
+
n
is a compact operator with norm
(10.92) T
+
n

p
m=2n
[F
+
(m)[.
For p = 2, T
+
n
is a selfadjoint trace class operator satisfying
(10.93) 1l +T
+
n
n
> 0, lim
n
n
= 1.
180 10. Scattering theory
The trace of T
+
n
is given by
(10.94) tr(T
+
n
) =
j=0
F
+
(2n + 2j) +
N
=1
+,
k
2n
1 k
.
Proof. The inequality for the norm of T
+
n
is a special case of Youngs inequality.
Moreover, cutting o F
+
after nitely many terms gives a sequences of nite rank
operators which converge to T
+
n
in norm; hence T
+
n
is compact.
Let f
2
(N
0
) and abbreviate
f(k) =
j=0
f(j)k
j
. Setting f(j) = 0 for j < 0
we obtain
(10.95)
j=0
f(j)T
+
n
f(j) =
1
2i
_
]k]=1
R
+
(k)k
2n
[
f(k)[
2
dk
k
+
N
=1
+,
k
2n
f(k
)[
2
from the convolution formula. Since R
+
(k) = R
+
(k) the integral over the imaginary
part vanishes and the real part can be replaced by
Re(R
+
(k)k
2n
) =
1
2
_
[1 +R
+
(k)k
2n
[
2
1 [R
+
(k)k
2n
[
2
_
=
1
2
_
[1 +R
+
(k)k
2n
[
2
+[T(k)[
2
_
1 (10.96)
(remember [T(k)[
2
+[R
+
(k)k
2n
[
2
= 1). This eventually yields the identity
j=0
f(j)(1l +T
+
n
)f(j) =
N
=1
+,
k
2n
f(k
)[
2
+
1
4i
_
]k]=1
_
[1 +R
+
(k)k
2n
[
2
+[T(k)[
2
_
[
f(k)[
2
dk
k
, (10.97)
which establishes 1l+T
+
n
0. In addition, by virtue of [1+R
+
(k)k
2n
[
2
+[T(k)[
2
> 0
(a.e.), 1 is no eigenvalue and thus 1l + T
+
n
n
for some
n
> 0. That
n
1
follows from T
+
n
 0.
To see that T
+
n
is trace class we use the splitting T
+
n
=
T
+
n
+
N
=1
T
+,
n
according to (10.70). The operators
T
+,
n
are positive and trace class. The operator
T
+
n
is given by multiplication with k
2n
R
+
(k) in Fourier space and hence is trace
class since [R
+
(k)[ is integrable.
A similar result holds for T
n
.
10.4. Inverse scattering theory
In this section we want to invert the process of scattering theory. Clearly, if
S
+
(H) (and thus T
+
n
) is given, we can use the GelfandLevitanMarchenko equa
tion (10.71) to reconstruct a(n), b(n) from T
+
n
a(n)
2
=
1
4
0
, (1l +T
+
n
)
1
0
, (1l +T
+
n+1
)
1
,
b(n) =
1
2
_
1
, (1l +T
+
n
)
1
0
, (1l +T
+
n
)
1
1
, (1l +T
+
n1
)
1
0
, (1l +T
+
n1
)
1
_
. (10.98)
10.4. Inverse scattering theory 181
In other words, the scattering data S
+
(H) uniquely determine a, b. Since T
+
n
is
trace class we can use Kramers rule to express the above scalar products. If we
delete the rst row and rst column in the matrix representation of 1l+T
+
n
we obtain
1l +T
+
n+1
. If we delete the rst row and second column in the matrix representation
of 1l +T
+
n
we obtain an operator 1l +(
+
n
. By Kramers rule we have
0
, (1l +T
+
n
)
1
0
=
det(1l +T
+
n+1
)
det(1l +T
+
n
)
,
1
, (1l +T
+
n
)
1
0
=
det(1l +(
+
n
)
det(1l +T
+
n
)
, (10.99)
where the determinants have to be interpreted as Fredholm determinants.
These formulas can even be used for practical computations. Let T
N
n
, (
N
n
be the
nite rank operators obtained from T
+
n
, (
+
n
, respectively, by setting F
+
(2n +j) =
0 for j > N. Then we have T
N
n
T
+
n
, (
N
n
(
+
n
in norm and hence also
(1l +T
N
n
)
1
(1l +T
+
n
)
1
as N . Furthermore,
0
, (1l +T
+
n
)
1
0
= lim
N
0
, (1l +T
N
n
)
1
= lim
N
det(1l +T
N
n+1
)
det(1l +T
N
n
)
,
1
, (1l +T
+
n
)
1
0
= lim
N
1
, (1l +T
N
n
)
1
= lim
N
det(1l +(
N
n
)
det(1l +T
N
n
)
, (10.100)
where
T
N
n
=
_
F
+
(2n +r +s), r +s N
0, r +s > N
_
0r,s
,
(
N
n
=
_
_
F
+
(2n +s + 1)
0,s
, r = 0
F
+
(2n +r +s + 2), r > 0, r +s N
0, r +s > N
_
_
0r,s
. (10.101)
For this procedure it is clearly interesting to know when given sets S
are the
scattering data of some operator H. In the previous section we have seen that
Hypothesis 10.9 is necessary. In this section we will show that it is also sucient
for S
1
(N). In addition, f
+
(k, n) = A
+
(n)
1
m=n
K
+
(n, m)k
m
, where K
+
(n, m) =
A
+
(n)
2
m
, (1l + T
+
n
)
1
0
and A
+
(n)
2
=
0
, (1l + T
+
n
)
1
0
, satises
+
f =
k+k
1
2
f.
Proof. We abbreviate
(10.102) C(n) =
j=0
[F
+
(n +j) F
+
(n +j + 2)[, C
1
(n) =
j=n
C(j)
182 10. Scattering theory
and observe C(n + 1) C(n) and
[F
+
(n)[ +[F
+
(n + 1)[
j=0
[F
+
(n + 2j) F
+
(n + 2j + 2)[
+
j=0
[F
+
(n + 2j + 1) F
+
(n + 2j + 3)[ = C(n). (10.103)
Moreover, let
K
+,j
(n) =
j
, (1l +T
+
n
)
1
0
0,j
= A
+
(n)
2
K
+
(n, n+j)
0,j
.
Then
(10.104)
K
+,j
(n) +F
+
(2n +j) +
=0
K
+,
(n)F
+
(2n + +j) = 0
or equivalently
K
+,j
(n) = (1l +T
+
n
)
1
F
+
(2n +j). Using
(10.105)
=0
[
K
+,
(n)[ N(n)C
1
(2n), N(n) = (1l +T
+
n
)
1

1
,
we obtain
[
K
+,j
(n)[ C(2n +j) +C(2n +j)
=0
[
K
+,
(n)[
C(2n +j)(1 +N(n)C
1
(2n)). (10.106)
Taking dierences in (10.104) we see
[
K
+,j
(n)
K
+,j
(n + 1)[ [F
+
(2n +j) F
+
(2n +j + 2)[
+
=0
_
[F
+
(2n +j +) F
+
(2n +j + + 2)[[
K
+,
(n)[
+[F
+
(2n +j + + 2)[[
K
+,
(n)
K
+,
(n + 1)[
_
(10.107)
Using
(1l +T
+
n
)(
K
+,j
(n)
K
+,j
(n + 1)) = F
+
(2n +j + 2) F
+
(2n +j)
+
=0
(F
+
(2n +j + + 2) F
+
(2n +j +))
K
+,
(n + 1) (10.108)
and hence
j=0
[
K
+,j
(n)
K
+,j
(n + 1)[ N(n)
j=1
_
[F
+
(2n +j + 2) F
+
(2n +j)[
+
=0
[F
+
(2n +j + + 2) F
+
(2n +j +)[[
K
+,
(n + 1)[
_
O(1)C(2n) (10.109)
we nally infer
[
K
+,j
(n)
K
+,j
(n + 1)[ [F
+
(2n +j) F
+
(2n +j + 2)[
+O(1)C(2n)C(2n +j). (10.110)
Thus n[
K
+,j
(n)
K
+,j
(n+1)[
1
(N) and the same is true for 4a
+
(n)
2
1, b
+
(n).
10.4. Inverse scattering theory 183
Next we turn to f
+
(k, n). Abbreviate
(K)(n, m) = 4a
+
(n)
2
K(n + 1, m) +K(n 1, m) + 2b
+
(n)K(n, m)
K(n, m+ 1) K(n, m1), (10.111)
then f
+
(k, n) satises
+
f =
k+k
1
2
f if K
+
= 0 (cf. (10.8)). We have K
+
(n, n) =
1 and K
+
(n, n + 1) =
m=n+1
2b
+
(m) by construction. Moreover, using the
GelfandLevitanMarchenko equation
(10.112) K
+
(n, m) = F
+
(n +m)
=n+1
K
+
(n, )F
+
( +m)
one checks
(10.113) (K
+
)(n, m) =
=n+1
(K
+
)(n, )F
+
( +m), m > n + 1.
By Lemma 7.8 this equation has only the trivial solution K
+
= 0 and hence the
proof is complete.
A corresponding result holds for given data S
, b
(k, n) = f
(k
1
, n)+
R
(k)f
m=0
K
+,m
(n)
F
+
(2n +m+j)
implying
(10.115)
j=
+,j
(n)k
nj
= R
+
(k)f
+
(k, n).
Using the GelfandLevitanMarchenko equation we further infer
+,j
(n) = A
+
(n)
0
(j)
1
A
+
(n)
_
K
+,j
(n) +
m=0
K
+,m
(n)
N
=0
+,
k
2n+m+j
_
= A
+
(n)
0
(j)
K
+,j
(n)
A
+
(n)
N
=1
+,
f
+
(k
, n)k
n+j
. (10.116)
184 10. Scattering theory
Next, computing the Fourier series we see
R
+
(k)f
+
(k, n) =
j=1
+,j
(n)k
n+j
+A
+
(n)k
n
f
+
(k
1
, n)
=1
+,
f
+
(k
, n)
k
n
k
n
1 k
k
1
(10.117)
and hence
(10.118) T(k)h
(k, n) = f
+
(k
1
, n) +R
+
(k)f
+
(k, n),
where
(10.119) h
(k, n) =
k
n
T(k)
_
A
+
(n) +
j=1
+,j
(n)k
j
=1
+,
f
+
(k
, n)
k
n
k
k k
_
.
Similarly, one obtains
(10.120) h
+
(k, n) =
k
n
T(k)
_
A
(n) +
j=1
,j
(n)k
j
=1
,
f
(k
, n)
k
n
k
k k
_
,
where
(10.121)
,j
(n) =
1
A
(n)
m=0
K
,m
(n)
F
+
(2n mj).
In particular, we see that k
n
h
(k
, n) = (res
kj
T(k))
1
,
k
(k
, n)
and (use also [T(k)[
2
+[R
(k)[
2
= 1)
(10.123) T(k)f
(k, n) = h
(k
1
, n) +R
(k)h
(k, n).
Eliminating (e.g.) R
(k
1
, n) +
R
(k)f
(k, n) we see
f
(k, n)h
(k, n) f
(k, n)h
(k, n) =
T(k)
_
h
(k, n)h
+
(k, n) f
(k, n)f
+
(k, n)
_
= G(k, n), [k[ = 1. (10.124)
Equation (10.122) shows that all poles of G(k, n) =
jZ
G
j
k
j
inside the unit circle
are removable and hence G
j
= 0, j < 0. Moreover, by G(k, n) = G(k, n) we see
G
j
= G
j
and thus G(k, n) = 0. If we compute 0 = lim
k0
h
(k, n)h
+
(k, n)
f
(k, n)f
+
(k, n) we infer A
(n)A
+
(n) = T(0) and thus a
(n) = a
+
(n) a(n).
Our next aim is to show h
(k, n) = f
jZ
H
j
k
j
=
h
(k, n)
f
(k, n)
=
f
+
(k, n)
h
+
(k, n)
, [k[ 1,
which is meromorphic inside the unit disk with H
0
= 1. However, if we choose
n large enough, then k
n
f
(k, n) = f
(k, n)
2
(k, n) and f
+
(k, n),
h
and H
+
, respectively.
Since the Weyl mfunctions of H
and H
+
are equal for some n, we conclude
H
= H
+
H (i.e., b
(n) = b
+
(n) b(n)).
Up to this point we have constructed an operator H which has the correct
reection coecients and the correct eigenvalues. To nish the proof we need to
show that our operator also has the correct norming constants. This follows from
(cf. (10.34))
(10.126)
nZ
f
(k
, n)f
+
(k
, n) = k
(res
kj
T(k))
1
and (10.122).
Chapter 11
Spectral deformations
Commutation methods
11.1. Commuting rst order dierence expressions
The idea of this section is to factor a given Jacobi operator into the product of
two simpler operators, that is, two operators associated with rst order dierence
expressions. Since it is not clear how to do this, we will rst investigate the other
direction, that is, we take a rst order dierence expression and multiply it with
its adjoint. The main reason why we take the adjoint as second operator is, since
this choice ensures selfadjointness of the resulting product (i.e., we get a Jacobi
operator). In addition, changing the order of multiplication produces a second
operator whose spectral properties are closely related to those of the rst. In fact,
we have the wellknown result by von Neumann (see, e.g., [241], Theorem 5.39 and
[54], Theorem 1).
Theorem 11.1. Suppose A is a bounded operator with adjoint A
. Then AA
and
A
and A
A for z C((A
A) 0) are related by
(AA
z)
1
=
1
z
_
A(A
Az)
1
A
1
_
,
(A
Az)
1
=
1
z
_
A
(AA
z)
1
A1
_
. (11.1)
Let
o
,
e
be two realvalued sequences fullling
(11.2)
o
,
e
(Z),
o
(n),
e
(n) > 0.
187
188 11. Spectral deformations
The second requirement is for convenience only (it will imply a
j
(n) < 0 below).
Now consider the (bounded) operators
(11.3)
A
2
(Z)
2
(Z)
f(n)
o
(n)f(n + 1) +
e
(n)f(n),
A
2
(Z)
2
(Z)
f(n)
o
(n 1)f(n 1) +
e
(n)f(n),
where A
A and H
2
= AA
(Z) and a
j
(n) < 0, b
j
(n) > 0. Next, observe that
the quantities
(11.6)
1
(n) =
e
(n)
o
(n)
,
2
(n) =
o
(n)
e
(n + 1)
satisfy
(11.7) a
j
(n)
j
(n) +
a
j
(n 1)
j
(n 1)
= b
j
(n).
Hence our approach gives us positive solutions
(11.8)
j
u
j
= 0, u
j
(n) =
n1
m=n0
j
(m) > 0
for free. Moreover, it shows that a given operator H
1
can only be factorized in this
manner if a positive solution u
1
of
1
u
1
= 0 exists. On the other hand, since we
have
o
(n) =
a
1
(n)
1
(n)
=
_
a
2
(n)
2
(n),
e
(n) =
_
a
1
(n)
1
(n) =
a
2
(n 1)
2
(n 1)
, (11.9)
a positive solution is the only necessary ingredient for such a factorization. Together
with Section 2.3, where we have seen that positive solutions exist if H
j
0 (see
also Remark 2.10) we obtain
Corollary 11.2. Suppose a(n) < 0. Then a Jacobi operator is nonnegative, H
0, if and only if there exists a positive solution of the corresponding Jacobi equation.
11.2. The single commutation method 189
11.2. The single commutation method
In this section we invert the process of the last section. This will result in a method
for inserting eigenvalues below the spectrum of a given Jacobi operator H. It is
usually referred to as single commutation method and a detailed study is the
topic of this and the next section.
To assure existence of the aforementioned necessary positive solutions, we will
need a < 0 throughout Sections 11.2 and 11.3.
Hypothesis H. 11.3. Suppose a, b
(
1
, n) constructed in Section 2.3
and set
(11.10) u
1
(
1
, n) =
1 +
1
2
u
+
(
1
, n) +
1
1
2
u
(
1
, n).
By the analysis of Section 2.3 (cf. (2.68)), any positive solution can be written in
this form up to normalization.
Motivated by (11.9) we consider the sequences
(11.11)
o,1
(n) =
a(n)u
1
(
1
, n)
u
1
(
1
, n + 1)
,
e,1
(n) =
a(n)u
1
(
1
, n + 1)
u
1
(
1
, n)
which are bounded because of
(11.12) [
a(n)u
1
(n + 1)
u
1
(n)
[ +[
a(n 1)u
1
(n 1)
u
1
(n)
[ = [
1
b(n)[.
Thus we can dene a corresponding operator A
1
on
2
(Z) together with its adjoint
A
1
,
A
1
f(n) =
o,1
(n)f(n + 1) +
e,1
(n)f(n),
A
1
f(n) =
o,1
(n 1)f(n 1) +
e,1
(n)f(n). (11.13)
For simplicity of notation we will not distinguish between the operator A
1
and its
dierence expression.
By Theorem 11.1, A
1
A
1
is a positive selfadjoint operator. A quick calcula
tion shows
(11.14) H = A
1
A
1
+
1
.
Commuting A
1
and A
1
yields a second positive selfadjoint operator A
1
A
1
and
further the commuted operator
(11.15) H
1
= A
1
A
1
+
1
.
The next theorem characterizes H
1
.
Theorem 11.4. Assume (H.11.3) and let
1
< inf (H),
1
[1, 1]. Then the
selfadjoint operator H
1
is associated with
(11.16) (
1
f)(n) = a
1
(n)f(n + 1) +a
1
(n 1)f(n 1) +b
1
(n)f(n),
190 11. Spectral deformations
where
a
1
(n) =
_
a(n)a(n + 1)u
1
(
1
, n)u
1
(
1
, n + 2)
u
1
(
1
, n + 1)
,
b
1
(n) =
1
a(n)
_
u
1
(
1
, n)
u
1
(
1
, n + 1)
+
u
1
(
1
, n + 1)
u
1
(
1
, n)
_
= b(n) +
a(n)u
1
(
1
, n)
u
1
(
1
, n + 1)
(11.17)
satisfy (H.11.3). The equation
1
v =
1
v has the positive solution
(11.18) v
1
(
1
, n) =
1
_
a(n)u
1
(
1
, n)u
1
(
1
, n + 1)
,
which is an eigenfunction of H
1
if and only if
1
(1, 1). H
1
and H
1
1
restricted to the orthogonal complements of their corresponding onedimensional
nullspaces are unitarily equivalent and hence
(11.19)
(H
1
) =
_
(H)
1
,
1
(1, 1)
(H),
1
1, 1
,
ac
(H
1
) =
ac
(H),
pp
(H
1
) =
_
pp
(H)
1
,
1
(1, 1)
pp
(H),
1
1, 1
,
sc
(H
1
) =
sc
(H).
In addition, the operator
(11.20) A
1
f(n) =
W
n
(u
1
(
1
), f)
_
a(n)u
1
(
1
, n)u
1
(
1
, n + 1)
acts as transformation operator
(11.21)
1
A
1
f = A
1
f.
Moreover, one obtains
(11.22) W
1
(A
1
u(z), A
1
v(z)) = (z
1
)W(u(z), v(z))
for solutions u(z), v(z) of u = zu, where W
1,n
(u, v) = a
1
(n)(u(n)v(n + 1)
u(n + 1)v(n)). The resolvent of H
1
for z C((H)
1
) is given by
(11.23) (H
1
z)
1
=
1
z
1
_
A
1
(H z)
1
A
1
1
_
,
or, in terms of Greens functions for n m, z C((H)
1
),
(11.24) G
1
(z, n, m) =
A
1
u
(z, m)A
1
u
+
(z, n)
(z
1
)W(u
(z), u
+
(z))
.
Hence, u
1,
(z, n) = A
1
u
(z, n) satisfy u
1,
(z, n)
nZ
v
1
(
1
, n)
2
=
4
1
2
1
W(u
(
1
), u
+
(
1
))
1
,
1
(1, 1),
and, if u() = u(), u(, .)
2
(Z),
(11.26)
nZ
(A
1
u(, n))
2
= (
1
)
nZ
u(, n)
2
.
11.2. The single commutation method 191
Proof. The unitary equivalence together with equation (11.23) follow from Theo
rem 11.1. Equations (11.20) (11.22) are straightforward calculations. Equation
(11.24) follows from the obvious fact that A
1
u
(z, n)
2
(
1
, n) =
1
1
2
W(u
(
1
), u
+
(
1
))v
1
(
1
, n)
shows that v
1
(
1
, n) is in
2
(Z) if and only if
1
(1, 1). In addition, we infer
A
1
u
(
1
, n) 0 since W(u
(
1
), u
+
(
1
)) < 0 (by G(
1
, n, n) < 0) and hence
a simple continuity argument implies A
1
u
1
A
1
u.
The rest is simple.
Remark 11.5. (i). Multiplying u
1
with a positive constant leaves all formulas
and, in particular, H
1
invariant.
(ii). We can also insert eigenvalues into the highest spectral gap, that is, above the
spectrum of H, upon considering H. Then > sup((H)) implies that we dont
have positive but rather alternating solutions and all our previous calculations carry
over with minor changes.
(iii). We can weaken (H.11.3) by requiring a(n) ,= 0 instead of a(n) < 0. Everything
stays the same with the only dierence that u
n=0
1
a(n)u
1
(
1
, n)u
1
(
1
, n + 1)
< if and only if
1
[1, 1)
(1, 1]
(cf. Lemma 2.9 (iv)).
(v). In the case
1
= 1 the second minimal solution for z =
1
is given by (cf.
(1.51))
(11.29) u
1,
(
1
, n) = v
1
(n)
n+1
n
j=
u
1
(
1
, n),
1
= 1.
(vi). The formula (11.22) can be generalized to
W
1
(A
1
u(z), A
1
v( z))
= (z
1
)W(u(z), v( z)) +
(z z)v( z, n + 1)
u(
1
, n + 1)
W
n
(u(
1
), u(z))
=
a(n)a(n + 1)
u(
1
, n + 1)
C
n
(u(
1
), u(z), v( z)), (11.30)
where C
n
(. . . ) denotes the Ndimensional Casoratian
(11.31) C
n
(u
1
, . . . , u
N
) = det (u
i
(n +j 1))
1i,jN
.
In the remainder of this section we will show some connections between the
single commutation method and some other theories. We start with the Weyl m
functions.
192 11. Spectral deformations
Lemma 11.6. Assume (H.11.3). The Weyl mfunctions m
1,
(z) of H
1
,
1
[1, 1] in terms of m
(z)
1
a(0)
(z)
_
, =
u
1
(
1
, 1)
u
1
(
1
, 0)
.
Proof. The above formulas follows after a straightforward calculation using (11.24)
and (2.2).
From Lemma 8.1 (iii) we obtain the useful corollary
Corollary 11.7. The operator H
1
is reectionless if and only if H is.
Moreover, we even have the following result.
Lemma 11.8. The operators H
n
1
and (H
1
)
1
n1
1
(resp. H
n
1
and
(H
1
)
n
1
) are unitarily equivalent when restricted to the orthogonal complements
of their corresponding onedimensional nullspaces, where
(11.33) =
u
1
(
1
, 1)
u
1
(
1
, 0)
,
1
=
v
1
(
1
, 1)
v
1
(
1
, 0)
.
Proof. We recall Remark 1.11 and we set
1
= 0 without loss of generality.
Choose
A
1
= P
n
A
1
P
n
+
e,1
(n)
n+1
, .
n
, then one computes
A
A
1
= P
n
HP
n
and
A
1
1
= P
1
n1
H
1
P
1
n1
. Similarly, choosing
A
1
= P
n
A
1
P
n
+
e,1
(n
1)
n
, .
n1
, one obtains
A
A
1
= P
n
HP
n
and
A
1
1
= P
n
H
1
P
n
.
Finally, we turn to scattering theory and assume (H.10.1) (cf. Remark 11.5).
We denote the set of eigenvalues by
p
(H) =
j
jJ
, where J N is a nite index
set, and use the notation introduced in Chapter 10.
Lemma 11.9. Suppose H satises (H.10.1) and let H
1
be constructed as in The
orem 11.4 with
(11.34) u
1
(
1
, n) =
1 +
1
2
f
+
(k
1
, n) +
1
1
2
f
(k
1
, n).
Then the transmission T
1
(k) and reection coecients R
,1
(k) of H
1
in terms
of the corresponding scattering data T(k), R
(k),
1
(1, 1),
T
1
(k) = T(k), R
,1
(k) =
k
1
1
k
1
k
1
1
k
1
R
(k),
1
1, 1, (11.35)
where k
1
=
1
+
_
2
1
1 (1, 0). Moreover, the norming constants
,1,j
associated with
j
pp
(H
1
) in terms of
,j
, corresponding to H, read
,1,j
= [k
j
[
1
1 k
j
k
1
(k
j
k
1
)
,j
, j J,
1
(1, 1),
,1,1
=
_
1
1
1 +
1
_
1
[1 k
2
1
[ T(k
1
),
1
(1, 1),
,1,j
= [k
1
1
k
1
j
[
,j
, j J,
1
1, 1. (11.36)
Proof. The claims follow easily after observing that up to normalization the Jost
solutions of H
1
are given by A
1
f
1,2
satisfy
(11.38) H
1
= A
1,2
A
1,2
2
and a further commutation then yields the operator
(11.39) H
1,2
= A
1,2
A
1,2
2
.
Clearly, if one works out the expression for a
1,2
and b
1,2
(by simply plugging
in) we get rather clumsy formulas. Moreover, these formulas will get even worse in
each step. The purpose of this section is to show that the resulting quantities after
N steps can be expressed in terms of determinants.
Theorem 11.10. Assume (H.11.3). Let H be as in Section 11.2 and choose
(11.40)
N
< <
2
<
1
< inf (H),
[1, 1], 1 N N.
Then we have
a
1,...,
N
(n) =
_
a(n)a(n +N)
_
C
n
(U
1,...,
N
)C
n+2
(U
1,...,
N
)
C
n+1
(U
1,...,
N
)
,
b
1,...,
N
(n) =
N
a(n)
C
n+2
(U
1,...,
N1
)C
n
(U
1,...,
N
)
C
n+1
(U
1,...,
N1
)C
n+1
(U
1,...,
N
)
a(n +N 1)
C
n
(U
1,...,
N1
)C
n+1
(U
1,...,
N
)
C
n+1
(U
1,...,
N1
)C
n
(U
1,...,
N
)
= b(n) +
a(n)
D
n
(U
1,...,
N
)
C
n+1
(U
1,...,
N
)
= b(n +N)
a(n +N)
D
n+1
(U
1,...,
N
)
C
n+1
(U
1,...,
N
)
, (11.41)
where (U
1,...,
N
) = (u
1
1
, . . . , u
N
N
) and
(11.42) u
(n) =
1 +
2
u
+
(
, n) + (1)
+1
1
2
u
, n).
Here C
n
(. . . ) denotes the Casoratian (11.31) and
D
n
(u
1
, . . . , u
N
) = det
_
u
i
(n), j = 1
u
i
(n +j), j > 1
_
1i,jN
,
D
n
(u
1
, . . . , u
N
) = det
_
u
i
(n +j 1), j < N
u
i
(n +N), j = N
_
1i,jN
. (11.43)
The corresponding Jacobi operator H
1,...,
N
is selfadjoint with spectrum
(11.44) (H
1,...,
N
) = (H)
(1, 1), 1 N.
194 11. Spectral deformations
In addition, we have
o,1,...,
N
(n) =
a(n)
C
n+2
(U
1,...,
N1
)C
n
(U
1,...,
N
)
C
n+1
(U
1,...,
N1
)C
n+1
(U
1,...,
N
)
,
e,1,...,
N
(n) =
a(n +N 1)
C
n
(U
1,...,
N1
)C
n+1
(U
1,...,
N
)
C
n+1
(U
1,...,
N1
)C
n
(U
1,...,
N
)
. (11.45)
and
(11.46) A
1,...,
N
A
1
f(n) =
_
1
j=0
_
a(n +j)
_
C
n
(U
1,...,
N
, f)
_
C
n
(U
1,...,
N
)C
n+1
(U
1,...,
N
)
.
The principal solutions of
1,...,
u = u for 1 N, <
are given by
(11.47) u
1,...,
,
(, n) = (1)
A
1,...,
A
1
u
(, n)
and
(11.48) u
1,...,
, n) =
1 +
2
u
1,...,
1
,+
(
, n) +
1
2
u
1,...,
1
,
(
, n)
is used to dene H
1,...,
.
To prove this theorem we need a special case of Sylvesters determinant identity
(cf. [91], Sect. II.3, [128]).
Lemma 11.11. Let M be an arbitrary nn matrix (n 2). Denote by M
(i,j)
the
matrix with the ith column and the jth row deleted and by M
(i,j,l,m)
the matrix
with the ith and lth column and the jth and mth row deleted. Then we have
det M
(n,n)
det M
(n1,n1)
det M
(n1,n)
det M
(n,n1)
= det M
(n,n,n1,n1)
det M. (11.49)
Proof. Obviously we can assume M to be nonsingular. First observe that inter
changing the ith row and the jth row for 1 i, j n 2 leaves the formula
invariant. Furthermore, adding an arbitrary multiple of the jth row to the ith
row with 1 j n 2, 1 i n, also leaves the formula unchanged. Thus we
may assume M to be of the form
(11.50)
_
_
_
_
_
_
_
_
M
11
M
12
. . . . . . M
1n
M
22
.
.
.
.
.
.
.
.
.
M
n1,n1
M
n1,n
M
n,n1
M
n,n
_
_
_
_
_
_
_
_
,
11.3. Iteration of the single commutation method 195
if M
(n,n,n1,n1)
is nonsingular, and of the form
(11.51)
_
_
_
_
_
_
_
_
0 M
12
. . . . . . M
1n
.
.
.
.
.
.
.
.
.
0 M
n2,2
.
.
.
M
n1,1
. . . . . . . . . M
n1,n
M
n,1
. . . . . . . . . M
n,n
_
_
_
_
_
_
_
_
,
if M
(n,n,n1,n1)
is singular. But for such matrices the claim can be veried directly.
Proof. (of Theorem 11.10) It is enough to prove (11.41) and (11.46), the remaining
assertions then follow from these two equations. We will use a proof by induction
on N. All formulas are valid for N = 1 and we have
f
1,...,
N+1
(n) =
_
a
1,...,
N
(n)C
n
(u
1,...,
N
(
N
), f)
_
u
1,...,
N
(
N
, n)u
1,...,
N
(
N
, n + 1)
,
a
1,...,
N+1
(n) =
_
a
1,...,
N
(n)a
1,...,
N
(n + 1)
_
u
1,...,
N
(
N
, n)u
1,...,
N
(
N
, n + 2)
u
1,...,
N
(
N
, n + 1)
, (11.52)
where f
1,...,
= A
1,...,
N
A
1
f. Equation (11.46) follows after a straightfor
ward calculation using (by Lemma 11.11)
C
n
(U
1,...,
N
, f)C
n+1
(U
1,...,
N+1
) C
n+1
(U
1,...,
N
, f)C
n
(U
1,...,
N+1
)
= C
n+1
(U
1,...,
N
)C
n
(U
1,...,
N+1
, f). (11.53)
The formula for a
1,...,
N
is then a simple calculation.
The rst formula for b
1,...,
N
follows after solving the corresponding Jacobi
equation (
1,...,
N
N
)f
1,...,
N
= 0 for it. The remaining two follow upon letting
in (11.47) and comparing with (6.27). The rest is straightforward.
For the sake of completeness we also extend Lemma 11.9. For brevity we assume
(1, 1).
Lemma 11.12. Suppose H satises (H.10.1) and let H
1,...,
N
,
(1, 1, ),
1 N, be constructed as in Theorem 11.10 with
u
(n) =
1 +
2
f
+
(k
, n) + (1)
+1
1
2
f
(k
, n). (11.54)
Then the transmission T
1,...,
N
(k) and reection coecients R
1,...,
N
,
(k) of the
operator H
1,...,
N
in terms of the corresponding scattering data T(k), R
(k) of H
are given by
T
1,...,
N
(k) =
_
N
=1
1 k k
k k
_
T(k),
R
1,...,
N
,
(k) = k
N
_
N
=1
k k
1 k k
_
R
(k), (11.55)
196 11. Spectral deformations
where k
+
_
,1,...,
N
,j
associated with
j
pp
(H
1,...,
N
) in terms of
,j
corresponding to
H read
,1,...,
N
,j
=
_
1j
1+j
_
1
[k
j
[
2(N1)
N
=1
[1 k
j
k
N
=1
=j
[k
j
k
l
[
T(k
j
), 1 j N,
,1,...,
N
,j
= [k
j
[
N
N
=1
1 k
j
k
[k
j
k
,j
, j J. (11.56)
Proof. Observe that
u
1,2
(
2
, n) =
1 +
2
2
A
1
f
+
(k
2
, n) +
1
2
2
A
1
f
(k
2
, n)
= c
_
1 +
2
2
f
+,1
(k
2
, n) +
1
2
2
f
,1
(k
2
, n)
_
, (11.57)
where c > 0 and
2
,
2
are related via
(11.58)
1 +
2
1
2
=
1
k
2
1 +
1
1
1
.
The claims now follow from Lemma 11.9 after extending this result by induction.
j=1
c
j
(g)
w
j
ln
_
(w +z(n))
(w +z(n 1))
_
w=0
,
(n) =
(z
0
(n 1))
(z
0
(n + 1))
(z(p
1
, n + 1))
(z(p
1
, n))
exp
_
_
p1
p0
+,
_
, (11.59)
(where all roots are chosen with positive sign) yields
o,1
(n) =
a(n)
(n)
=
a
(z(n + 1))
(z(n))
(z(p
1
, n))
(z(p
1
, n + 1))
exp
_
_
p1
p0
+,
2
_
,
e,1
(n) =
_
a(n)(n)
=
a
(z
0
(n 1))
(z
0
(n))
(z(p
1
, n + 1))
(z(p
1
, n))
exp
_
_
p1
p0
+,
2
_
. (11.60)
11.4. Application of the single commutation method 197
Next we calculate
1
(n) =
o,1
(n)
e,1
(n + 1)
=
(z(p
1
, n))
(z(p
1
, n + 2))
(z(n + 1))
(z(n))
exp
_
_
p1
p0
+,
_
. (11.61)
This suggests to dene
(11.62) z
1
(p, n) =
A
p0
(p)
p0
(T
1
) + 2n
A
p0
(
+
)
p0
,
associated with a new Dirichlet divisor T
1
given by
(11.63)
p0
(T
1
) =
p0
(T
) A
p0
(p
1
) A
p0
(
+
).
Lemma 11.8 says that we have p
1
=
g+1
(0) and
(11.64)
1,j
(0) =
j
(0), 1 j g,
where = (0). This can also be obtained directly by comparing (11.63) and
(9.11).
Upon using
(11.65) z(p
1
, n) = z
1
(
+
, n 1), z(
+
, n) = z
1
(p
1
, n),
we obtain (again all roots are chosen with positive sign)
a
1
(n) = a
(z
1
(n + 1))(z
1
(n 1))
(z
1
(n))
2
,
b
1
(n) =
b +
g
j=1
c
j
(g)
w
j
ln
_
(w +z
1
(n))
(w +z
1
(n 1))
_
w=0
,
1
(n) =
(z
1
(n 1))
(z
1
(n + 1))
(z
1
(p
1
, n + 1))
(z
1
(p
1
, n))
exp
_
_
p
1
p0
+,
_
. (11.66)
Remark 11.13. If we choose (p
1
) E
2g+2
instead of (p
1
) E
1
the operator
H
1
is welldened but the sequences
o,1
,
e,1
become purely imaginary.
Now let us choose (p
1
) [E
2j1
, E
2j
]. Then neither
o,1
,
e,1
nor H
1
are
welldened. But if we ignore this and make (formally) a second commutation at
a point p
2
= (
2
,
2
R
1/2
2g+2
(
2
)) with (p
2
) [E
2j1
, E
2j
], we get an operator H
2
associated with a Dirichlet divisor T
1
,
2
given by
p0
(T
1
,
2
) =
p0
(T
1
) A
p0
(p
1
) A
p0
(
+
)
=
p0
(T
) A
p0
(p
1
) A
p0
(p
2
) 2A
p0
(
+
). (11.67)
Since 2[i Im(A
p0
(p
1
))] = [0], H
2
is a welldened operator. Moreover, choosing
p
1
=
j
in the rst step implies that the Dirichlet eigenvalue
j
is formally replaced
by one at
2
= (, R
1/2
2g+2
())) and the factor 2A
p0
(
+
) shifts the sequences
198 11. Spectral deformations
by one. We will undo the shift and denote the corresponding sequences by a
(,)
,
b
(,)
. From Theorem 11.10 we know that they are equivalently given by
a
(,)
(n + 1) =
_
a(n)a(n + 2)
C
n
(
j
(
j
),
())C
n+2
(
j
(
j
),
())
C
n+1
(
j
(
j
),
())
2
,
b
(,)
(n + 1) = b(n)
j
(
j
, n)
(, n + 1)
C
n
(
j
(
j
),
())
, (11.68)
where the n + 1 on the lefthandside takes the aforementioned shift of reference
point into account. Thus, applying this procedure g times we can replace all Dirich
let eigenvalues (
j
,
j
) by (
j
,
j
) proving that the sequences associated with
j
,
1 j g, are given by
a
( 1, 1),...,( g, g)
(n) =
_
a(n g)a(n +g)
_
C
ng
(
( 1, 1),...,( g, g)
)C
ng+2
(
( 1, 1),...,( g, g)
)
C
ng+1
(
( 1, 1),...,( g, g)
)
,
b
( 1, 1),...,( g, g)
(n) = b(n)
a(n)
D
ng+1
(
( 1, 1),...,( g, g)
)
C
ng+1
(
( 1, 1),...,( g, g)
)
,
(11.69)
where (
( 1, 1),...,( g, g)
) = (
1
(
1
),
1
(
1
), . . . ,
g
(
g
),
g
(
g
)). We will
generalize this method to arbitrary operators in Section 11.8.
11.5. A formal second commutation
Our next aim is to remove the condition that H is bounded from below. To do this
we apply the single commutation method twice which will produce a method for
inserting eigenvalues into arbitrary spectral gaps of a given Jacobi operator (not
only into the lowest). This method is hence called the double commutation method.
Let H be a given operator (a < 0 for the moment) and let u
(
1
, .),
1
< (H),
be a positive solution. Dene
Af(n) =
o
(n)f(n + 1) +
e
(n)f(n),
A
f(n) =
o
(n 1)f(n 1) +
e
(n)f(n), (11.70)
where
(11.71)
o
(n) =
a(n)u
(
1
, n)
u
(
1
, n + 1)
,
e
(n) =
a(n)u
(
1
, n + 1)
u
(
1
, n)
.
Then we have H = A
A+
1
and we can dene a second operator
(11.72)
H = AA
+
1
.
Fix
(
1
, n)
_
a(n)u
(
1
, n)u
(
1
, n + 1)
11.5. A formal second commutation 199
solves
Hv =
1
v, where
(11.74) c
(
1
, n) =
1
j=
u
(
1
, j)
2
.
Now dene
o,
(n) =
e
(n + 1)
(
1
, n)
c
(
1
, n + 1)
,
e,
(n) =
o
(n)
(
1
, n + 1)
c
(
1
, n)
(11.75)
and introduce corresponding operators A
, A
on
2
(Z) by
(A
f)(n) =
o,
(n)f(n + 1) +
e,
(n)f(n),
(A
f)(n) =
o,
(n 1)f(n 1) +
e,
(n)f(n). (11.76)
A simple calculation shows that A
= A
1
A
1
and hence
(11.77) H
1
= A
+
1
.
Performing a second commutation yields the doubly commuted operator
(11.78) H
= A
+
1
.
Explicitly, one veries
(11.79) (H
f)(n) = a
(n)f(n + 1) +a
(n 1)f(n 1) b
(n)f(n),
with
a
(n) = a(n + 1)
_
c
(
1
, n)c
(
1
, n + 2)
c
(
1
, n + 1)
,
b
(n) = b(n + 1)
_
a(n)u
(
1
, n)u
(
1
, n + 1)
c
(
1
, n)
a(n + 1)u
(
1
, n + 1)u
(
1
, n + 2)
c
(
1
, n + 1)
_
. (11.80)
Note that
(11.81) u
(
1
, n) =
u
(
1
, n + 1)
_
c
(
1
, n)c
(
1
, n + 1)
is a solution of H
u =
1
u fullling u
(
1
)
2
(, n
0
) and because of
(11.82)
nZ
[u
(
1
, n)[
2
=
lim
n
1
c
(
1
, n)
,
we have u
(
1
)
2
(Z). Furthermore, if u is a solution of Hu = zu (for arbitrary
z C), then
u
(z, n) =
1
z
1
AA
u(z, n)
=
c
(
1
, n)u(z, n + 1)
1
z1
u
(z, n + 1)W
n
(u
(
1
), u(z))
_
c
(
1
, n)c
(
1
, n + 1)
(11.83)
200 11. Spectral deformations
is a solution of H
u = zu.
Now observe that H
is no longer positive.
This applies, in particular, in the case where u
(
1
) has zeros and hence all in
termediate operators A
1
, A
,
H, etc., become illdened. Thus, to dene H
,
it suces to assume the existence of a solution u
(
1
) which is square summable
near . This condition is much less restrictive than the existence of a positive
solution (e.g., the existence of a spectral gap for H around
1
is sucient in this
context).
One expects that formulas analogous to (11.19) will carry over to this more gen
eral setup. That this is actually the case will be shown in Theorem 11.16. Hence
the double commutation method (contrary to the single commutation method) en
ables one to insert eigenvalues not only below the spectrum of H but into arbitrary
spectral gaps of H.
11.6. The double commutation method
In this section we provide a complete characterization of the double commutation
method for Jacobi operators. Since Theorem 11.1 only applies when
1
< (H),
we need a dierent approach. We start with a linear transformation which turns
out to be unitary when restricted to proper subspaces of
2
(Z). We use this trans
formation to construct an operator H
1
from a given background operator H. This
operator H
1
will be the doubly commuted operator of H as discussed in the pre
vious section. The following abbreviation will simplify our notation considerably
(11.84) f, g
n
m
=
n
j=m
f(j)g(j).
Let u
2
(n) =
1
+u, u
n
, ,= 0
(setting
1
= 0). Consider the following linear transformation
(11.86)
U
:
2
(Z) (Z)
f(n)
_
c(n)
c(n1)
f(n) u
(n)u, f
n
,
(U
0
= 1l), where
(11.87) u
(n) =
u(n)
_
c
(n 1)c
(n)
,
(u
0
= 0). We note that U
can be dened on
2
(Z) and U
u =
1
u
, ,= 0.
Furthermore,
(11.88) u
(n)
2
=
1
c
(n)
,
and hence
(11.89) u

2
=
_
, u ,
2
(Z)
2
u
2
1+u
2
, u
2
(Z)
11.6. The double commutation method 201
implying u
2
(Z) if u
2
< < . If = u
2
, = we only have
u
(Z), u
2
+
(Z), respectively. In addition, we remark that for f
= U
f
we have
u
(n)f
(n) =
u, f
n
(n)
,
[f
(n)[
2
= [f(n)[
2
+
[u, f
n
[
2
c
(n)
. (11.90)
Summing over n and taking limits (if = use CauchySchwarz) shows
u
, f
=
_
c
(n)
1
u, f
n
, R
u,f)
u
2
c
(n)
1
u, f
n+1
, =
,
f
, f
= f, f
n
[u, f
n
[
2
c
(n)
. (11.91)
Clearly, the last equation implies U
:
2
(Z)
2
(Z). In addition, we remark that
this also shows U
:
2
(Z)
2
(Z).
Denote by P, P
(set P, P
= 0 if u, u
,
2
(Z)), respectively. Dene
(11.92)
U
1
:
2
(Z) (Z)
g(n)
_
_
_
_
c(n1)
c(n)
g(n) +u(n)u
, g
n
, R
_
c(n1)
c(n)
g(n) u(n)u
, g
n+1
, =
and note
(11.93) c
1
(n) =
_
u
, u
, R
u
2
+u
, u
n+1
, =
.
As before one can show U
1
: (1l P
)
2
(Z)
2
(Z) and one veries
(11.94)
U
U
1
= 1l, U
1
= 1l, R,
U
U
1
= 1l, U
1
= 1l P, = .
If P = 0, (u
2
, ), then U
U
1
)
2
(Z)
since U
1
,
2
(Z) by
(11.95) U
1
=
_
u, R
u
2
u, =
.
Summarizing,
Lemma 11.14. The operator U
)
2
(Z)
with inverse U
1
. If P, P
,= 0, then U
on
2
(Z) by
(11.96)
U
= U
(1l P) +
_
1 +u
2
U
P.
Proof. The rst identity in (11.91) shows that U
maps (1l P)
2
(Z) onto (1l
P
)
2
(Z). Unitarity follows from (11.91) and
(11.97) lim
n
[u, f
n
[
2
u, u
n
= 0
202 11. Spectral deformations
for any f
2
(Z) if u ,
2
(Z). In fact, suppose f = 1, pick m and n > m so large
that f, f
m
/2 and u, u
m1
/u, u
n
, g
= f, g
n
f, u
n
u, g
n
(n)
,
where f
= U
f, g
= U
g.
Next, we take a Jacobi operator associated with bounded sequences a, b as
usual. We assume (
1
,
1
) to be of the following kind.
Hypothesis H. 11.15. Suppose (, ) satises the following conditions.
(i). u
(, n) exists.
(ii). [u
()
2
, ) .
Note that by Lemma 2.2 a sucient condition for u
(
1
, n) to exist is
R
ess
(H
).
We now use Lemma 11.14 with u(n) = u
(
1
, n), =
1
, c
1
(
1
, n) = c
(n) to
prove
Theorem 11.16. Suppose (H.11.15) and let H
1
be the operator associated with
(11.99) (
1
f)(n) = a
1
(n)f(n + 1) +a
1
(n 1)f(n 1) +b
1
(n)f(n),
where
a
1
(n) = a(n)
_
c
1
(
1
, n 1)c
1
(
1
, n + 1)
c
1
(
1
, n)
,
b
1
(n) = b(n)
a(n)u
(
1
, n)u
(
1
, n + 1)
c
1
(
1
, n)
. (11.100)
Then
(11.101) H
1
(1l P
1
(
1
)) = U
1
HU
1
1
(1l P
1
(
1
))
and
1
u
1,
(
1
) =
1
u
1,
(
1
), where
(11.102) u
1,
(
1
, n) =
u
(
1
, n)
_
c
1
(
1
, n 1)c
1
(
1
, n)
.
Proof. Let f be a sequence which is square summable near implying that f
is also square summable near . Then a straightforward calculation shows
(11.103)
1
(U
1
f) = U
1
(f).
The rest follows easily.
Corollary 11.17. Suppose u
(
1
) ,
2
(Z).
(i). If
1
> 0, then H and (1l P
1
(
1
))H
1
are unitarily equivalent. Moreover,
H
1
has the additional eigenvalue
1
with eigenfunction u
1,
(
1
).
(ii). If
1
= , then H and H
1
are unitarily equivalent.
Suppose u
(
1
)
2
(Z) (i.e.,
1
is an eigenvalue of H).
(i). If
1
(u
(
1
)
2
, ), then H and H
1
are unitarily equivalent (using
U
1
from Lemma 11.14).
11.6. The double commutation method 203
(ii). If
1
= u
(
1
)
2
, , then (1l P(
1
))H, H
1
are unitarily equivalent,
that is, the eigenvalue
1
is removed.
Remark 11.18. (i). By choosing
1
ac
(H)
sc
(H) (provided the continuous
spectrum is not empty and a solution satisfying (H.11.15) exists) we can use the
double commutation method to construct operators with eigenvalues embedded in
the continuous spectrum.
(ii). Multiplying
1
by some constant c
2
> 0 amounts to the same as multiplying
u
(
1
, .) by c
1
.
The previous theorem tells us only how to transfer solutions of u = zu into
solutions of
1
v = zv if u is square summable near . The following lemma
treats the general case.
Lemma 11.19. The sequence
(11.104) u
1
(z, n) =
c
1
(
1
, n)u(z, n)
1
z1
u
(
1
, n)W
n
(u
(
1
), u(z))
_
c
1
(
1
, n 1)c
1
(
1
, n)
,
z C
1
solves
1
u = zu if u(z) solves u = zu. If u(z) is square summable
near , we have u
1
(z, n) = (U
1
u)(z, n), justifying our notation. Furthermore,
we note
(11.105) [u
1
(z, n)[
2
= [u(z, n)[
2
+
1
[z
1
[
2
[W
n
(u
(
1
), u(z))[
2
c
1
(
1
, n)
,
and
(11.106) W
1,n
(u
,1
(
1
), u
1
(z)) =
W
n
(u
(
1
), u(z))
c
1
(
1
, n)
,
where W
1,n
(u, v) = a
1
(n)(u(n)v(n + 1) u(n + 1)v(n)). Hence u
1
is square
summable near if u is. If u
1
( z) is constructed analogously, then
W
1,n
(u
1
(z), u
1
( z)) = W
n
(u(z), u( z)) +
1
c
1
(
1
, n)
z z
(z
1
)( z
1
)
W
n
(u
(
1
), u(z))W
n
(u
(
1
), u( z)). (11.107)
Proof. All facts are tedious but straightforward calculations.
As a consequence we get.
Corollary 11.20. The Green function G
1
(z, n, m) of H
1
is given by (n > m)
(11.108) G
1
(z, n, m) =
u
+,1
(z, n)u
,1
(z, m)
W(u
(z), u
+
(z))
,
where
(11.109) u
,1
(z, n) =
c
1
(
1
, n)u
(z, n)
1
z1
u
(
1
, n)W
n
(u
(
1
), u
(z))
_
c
1
(
1
, n 1)c
1
(
1
, n)
.
Next, we turn to Weyl mfunctions. Without loss of generality we assume
u
(
1
, n) = s
(
1
, n, 0), that is,
(11.110) u
(
1
, 0) = sin(), u
(
1
, 0) = cos(), [0, )
(change
1
if necessary see Remark 11.18 (ii)).
204 11. Spectral deformations
Theorem 11.21. Let m
(z), m
1,
(z) denote the Weyl mfunctions of H, H
1
respectively. Then we have
(11.111) m
1,
(z) =
c
1
(
1
, 0)
c
1
(
1
, 1)
1 +
2
1 +
2
_
m
(z)
c
1
(
1
, 0)
1
z
1
c
1
(
1
, 1)
1
a(0)(1 +
2
)
_
,
where = cot() and
(11.112)
=
c
1
(
1
, 1)
c
1
(
1
, 1)
.
Proof. Consider the sequences
_
c
1
(
1
, 1)c
1
(
1
, 1)
c
1
(
1
, 0)
s
,1
(z, n),
c
1
(
1
, 1)
c
1
(
1
, 0)
1 +
2
1 +
2
_
c
,1
(z, n) s
,1
(z, n)
_
, (11.113)
(where is dened in (11.116) below) constructed from the fundamental system
c
(z, n), s
1,+
(z) =
+
(
1
) +c
1
(
1
, 0)
1
z
1
+O(z
1
)
0
,
m
1,
(z) =
(c
1
(
1
, 0)(
1
c
1
(
1
, 0) 1))
1
z
1
+O(z
1
)
0
, (11.114)
where
+
(
1
) = (u
(
1
), u
(
1
)
1
)
1
0 is the norming constant of H
+
at
1
.
As a direct consequence of Theorem 11.21 we infer
Theorem 11.23. Given H and H
1
the respective Weyl Mmatrices M
(z) and
M
1
(z) are related by
M
1,0,0
(z) =
c
1
(
1
, 0)
c
1
(
1
, 1)
1 +
2
1 +
2
M
0,0
(z),
M
1,0,1
(z) =
c
1
(
1
, 0)
_
c
1
(
1
, 1)c
1
(
1
, 1)
_
M
0,1
(z) +M
0,0
(z)
_
,
M
1,1,1
(z) =
c
1
(
1
, 0)
c
1
(
1
, 1)
1 +
2
1 +
2
_
M
1,1
(z) + 2M
0,1
(z)
+
2
M
0,0
(z)
_
, (11.115)
where
(11.116) =
a(0)c
1
(
1
, 0)
1
z
1
c
1
(
1
, 1)
1
1 +
2
.
11.6. The double commutation method 205
The quantities M
1,0,0
(z) and M
1,0,1
(z) are both holomorphic near z =
1
. More
over, M
1,1,1
(z) has a simple pole with residue (
1
c
1
(
1
, 0))
1
at z =
1
.
Using the previous corollary plus weak convergence of
1
Im(F( + i))d to
the corresponding spectral measure d() as 0 (cf. Lemma B.3) implies
Corollary 11.24. The matrix measures d
, d
1
corresponding to M
(z), M
1
(z)
are related by
d
1,0,0
() =
c
1
(
1
, 0)
c
1
(
1
, 1)
1 +
2
1 +
2
d
0,0
(),
d
1,0,1
() =
c
1
(
1
, 0)
_
c
1
(
1
, 1)c
1
(
1
, 1)
_
d
0,1
() +d
0,0
()
_
,
d
1,1,1
() =
c
1
(
1
, 0)
c
1
(
1
, 1)
1 +
2
1 +
2
_
d
1,1
() + 2d
0,1
()
+
2
d
0,0
()
_
+ (
1
c
1
(
1
, 0))
1
d(
1
). (11.117)
Equivalently
(11.118) d
1
() = Ud
()U
T
+
1
1
c
1
(
1
, 0)
_
0 0
0 1
_
,
where
(11.119) U =
c
1
(
1
, 0)
c
1
(
1
, 1)
1 +
2
1 +
2
_
1+
2
1+
2
_
c
1
(1,1)
c
1
(1,1)
0
1
_
.
Remark 11.25. We can interchange the role of and in this section by
substituting , ., ..
n
., ..
n+1
, u
u
+
.
Our next aim is to show how the scattering data of the operators H, H
1
are
related, where H
1
is dened as in Theorem 11.16 using the Jost solution f
(k, n)
(see Theorem 10.2).
Lemma 11.26. Let H be a given Jacobi operator satisfying (H.10.1). Then the
doubly commuted operator H
1
,
1
> 0, dened via u
(
1
, n) = f
(k
1
, n),
1
=
(k
1
+k
1
1
)/2 as in Theorem 11.16, has the transmission and reection coecients
T
1
(k) = sgn(k
1
)
k k11
kk1
T(k),
R
,1
(k) = R
(k), R
+,1
(k) =
_
kk1
k k11
_
2
R
+
(k), (11.120)
where z and k are related via z = (k +k
1
)/2. Furthermore, the norming constants
,j
corresponding to
j
pp
(H), j J, (cf. (10.53)) remain unchanged except
for an additional eigenvalue
1
with norming constant
,1
=
1
if u
(
1
) ,
2
(Z)
respectively with norming constant
,1
=
,1
+
1
if u
(
1
)
2
(Z) and
,1
denotes the original norming constant of
1
p
(H).
Proof. By Lemma 11.19 the Jost solutions f
,1
(k, n) are up to a constant given
by
(11.121)
c
1
(
1
, n 1)f
(k, n)
1
z1
u
(
1
, n)W
n1
(u
(
1
), f
(k))
_
c
1
(
1
, n 1)c
1
(
1
, n)
.
206 11. Spectral deformations
This constant is easily seen to be 1 for f
,1
(k, n). Thus we can compute R
() us
ing (11.107) (the second unknown constant cancels). The rest follows by a straight
forward calculation.
11.7. Iteration of the double commutation method
Finally, we demonstrate how to iterate the double commutation method. We choose
a given background operator H and further (
1
,
1
) according to (H.11.15). Next
use u
(
1
) to dene the transformation U
1
and the operator H
1
. In the sec
ond step, we choose (
2
,
2
) and another function u
(
2
) to dene u
,1
(
2
) =
U
1
u
(
2
), a corresponding transformation U
1,2
, and an operator H
1,2
. Apply
ing this procedure Ntimes results in
Theorem 11.27. Let H be a given background operator and let (
), 1 N,
satisfy (H.11.15). Dene the following matrices (1 N)
C
(n) =
_
r,s
r
+u
(
r
), u
(
s
)
n
_
1r,s
,
C
(f, g)(n) =
_
_
_
_
C
1
(n)
r,s
, r,s1
f, u
(
s
)
n
, s1,r=
u
(
r
), g
n
, r1,s=
f, g
n
, r=s=
_
_
_
_
1r,s
,
D
j
(n) =
_
_
_
_
C
(n)
r,s
, r,s
u
(
s
, n j), s,r=+1
u
(
r
, n), r,s=+1
0,j
, r=s=+1
_
_
_
_
1r,s+1
,
U
(f(n)) =
_
_
_
_
C
(n)
r,s
, r,s
u
(
s
), f
n
, s,r=+1
u
(
r
, n), r,s=+1
f, r=s=+1
_
_
_
_
1r,s+1
. (11.122)
Then we have (set C
0
(n) = 1, U
0
(f) = f)
(11.123) U
1,...,
1
U
1
f, U
1,...,
1
U
1
g
n
=
det C
(f, g)(n)
det C
1
(n)
and
(11.124) U
1,...,
U
1
f(n) =
det U
(f)(n)
_
det C
(n 1) det C
(n)
.
In particular, we obtain
c
1,...,
, n) =
1
+U
1,...,
1
U
1
u
), U
1,...,
1
U
1
u
)
n
=
det C
(n)
det C
1
(n)
(11.125)
and hence
(11.126)
N
=1
c
1,...,
, n) = det C
N
(n).
11.7. Iteration of the double commutation method 207
The corresponding operator H
1,...,
N
is associated with
a
1,...,
N
(n) = a(n)
_
det C
N
(n 1) det C
N
(n + 1)
det C
N
(n)
,
b
1,...,
N
(n) = b(n)
a(n)
det D
N
1
(n + 1)
det C
N
(n)
, (11.127)
and we have
H
1,...,
N
(1l
N
j=1
P
1,...,
N
(
j
))
= (U
1,...,
N
U
1
)H(U
1
1
U
1
1,...,
N
)(1l
N
j=1
P
1,...,
N
(
j
)). (11.128)
Here P
1,...,
N
(
j
) denotes the projection onto the onedimensional subspace spanned
by
(11.129) u
1,...,
N
,
(
, n) =
det U
N
(n)
_
det C
N
(n 1) det C
N
(n)
,
where U
N
det C
(f, g) det C
(u
), g) det C
(f, u
))
= det C
1
det C
+1
(f, g), (11.130)
which proves (11.123) together with a look at (11.98) by induction on N. Next,
(11.125) easily follows from (11.123). Similarly,
det C
det U
1
(f) det U
1
(u
)) det C
(u
), f)
= det C
1
det U
(f), (11.131)
and (11.92) prove (11.124). To show the formula for b
1,...,
N
we use that (11.124)
with f = u
(z, 0) = 1 implying u
(), u
(z)
n
= (z )
1
W
n
(u
(), u
(z)) = O(z
n
)
as [z[ . Now we see that u
(), u
(z)
n
= u
(, n)u
(z, n) + u
(, n
1)u
(z, n 1) +O(z
n2
) and thus
(11.132) det U
N
(u
(
s
), u(z)) sN,r=N+1
u
(
r
, n) rN,s=N+1
u(z, n) r=s=N+1
_
_
_
_
1r,sN+1
satises
1,...,
N
u = zu if u(z, .) satises u = zu.
(ii). Equation (11.129) can be rephrased as
(u
1,...,
N
,
(
1
, n), . . . , u
1,...,
N
,
(
N
, n)) =
det C
N
(n)
det C
N
(n 1)
(C
N
(n))
1
(u
(
1
, n), . . . , u
(
N
, n)), (11.135)
where (C
(n))
1
is the inverse matrix of C
(n).
(iii). The ordering of the pairs (
j
,
j
), 1 j N, is clearly irrelevant (inter
changing row i, j and column i, j leaves all determinants unchanged). Moreover, if
i
=
j
, then (
i
,
i
), (
j
,
j
) can be replaced by (
i
,
i
+
j
) (by the rst assertion
it suces to verify this for N = 2).
Finally, we also extend Lemma 11.26. For simplicity we assume u
(
j
, n) ,
2
(Z),
j
> 0, 1 j N.
Lemma 11.29. Let H be a given Jacobi operator satisfying (H.10.1). Then the
operator H
1,...,
N
, dened via u
, n) = f
(k
, n) with
= (k
+ k
1
)/2 in
R(H
1,...,
1
), 1 N, has the transmission and reection coecients
T
1,...,
N
(k) =
_
N
=1
sgn(k
)
k k
1
k k
_
T(k),
R
,1,...,
N
(k) = R
(k),
R
+,1,...,
N
(k) =
_
N
=1
_
k k
k k
1
_
2
_
R
+
(k), (11.136)
where z = (k +k
1
)/2. Furthermore, the norming constants
,j
corresponding to
j
p
(H), j J, (cf. (10.53)) remain unchanged and the additional eigenvalues
.
Remark 11.30. Of special importance is the case a(n) = 1/2, b(n) = 0. Here
we have f
(k, n) = k
n
, T(k) = 1, and R
(k), [k[ = 1 together with the point spectrum and corresponding norming
constants uniquely determine a(n), b(n). Hence we infer from Lemma 11.12 that
H
1,...,
N
constructed from u
, n) = k
n
=
1+
2
k
n
+(1)
+1 1
2
k
n
N
=1
[1 k
j
k
N
=1
=j
[k
j
k
[
, 1 j N.
11.8. The Dirichlet deformation method 209
11.8. The Dirichlet deformation method
In this section we will see that the method of changing individual Dirichlet eigen
values found in Section 11.4 for quasiperiodic nitegap sequences works in a much
more general setting.
Let us formulate our basic hypothesis.
Hypothesis H. 11.31. (i). Let (E
0
, E
1
) be a spectral gap of H, that is, (E
0
, E
1
)
(H) = E
0
, E
1
.
(ii). Suppose
0
d
(H
0
) [E
0
, E
1
].
(iii). Let (, ) [E
0
, E
1
] and (E
0
, E
1
) or
d
(H).
Remark 11.32. If
0
is an eigenvalue of two of the operators H, H
, H
+
, then it
is also one of the third. Hence, if
0
d
(H
0
)
d
(H
0
), then
0
(E
0
, E
1
) and
if
0
d
(H
0
)
d
(H
0
), then
0
E
0
, E
1
. (The choice of
0
in the latter
case is irrelevant). Condition (ii) thus says that
0
= E
0,1
is only allowed if E
0,1
is
a discrete eigenvalue of H. Similarly in (iii) for .
Our next objective is to dene the operator H
(,)
as in Section 11.4 by perform
ing two single commutations using u
0
(
0
, .) and u
(, .). Since H
(0,0)
= H, we
will assume (, ) ,= (
0
,
0
) without restriction.
Due to our assumption (H.11.31), we can nd solutions u
0
(
0
, .), u
(, .) (cf.
Lemma 2.2) and dene
(11.138) W
(,)
(n) =
_
_
_
Wn(u
0
(0),u())
0
, ,=
0
n
m=0
u
0
(
0
, m)
2
, (, ) = (
0
,
0
)
,
where
n
m=+
=
m=n+1
. The motivation for the case (, ) = (
0
,
0
)
follows from (assuming u
m=0
u
0
(
0
, m)u
0
(, m)
=
n
m=0
u
0
(
0
, m)
2
. (11.139)
From the proof of Theorem 4.19 we infer
Lemma 11.33. Suppose (H.11.31), then
(11.140) W
(,)
(n + 1)W
(,)
(n) > 0, n Z.
Thus the sequences
a
(,)
(n) = a(n)
W
(,)
(n 1)W
(,)
(n + 1)
W
(,)
(n)
2
,
b
(,)
(n) = b(n)
a(n)u
0
(
0
, n)u
(, n + 1)
W
(,)
(n)
, (11.141)
are both welldened and we can consider the associated dierence expression
(11.142) (
(,)
u)(n) = a
(,)
(n)u(n + 1) +a
(,)
(n 1)u(n 1) +b
(,)
(n)u(n).
210 11. Spectral deformations
However, we do not know at this point whether a
(,)
(n), b
(,)
(n) are bounded
or not. This will be ignored for the moment. The next lemma collects some basic
properties which follow either from Section 11.3 (choosing N = 2 in Theorem 11.10)
or can be veried directly.
Lemma 11.34. Let
(11.143) (A
(,)
u)(z, n) =
W
(,)
(n)u(z, n)
1
z0
u
(, n)W
n
(u
0
(
0
), u(z))
_
W
(,)
(n 1)W
(,)
(n)
,
where u(z) solves u = zu for z C
0
. Then we have
(11.144)
(,)
(A
(,)
u)(z, n) = z(A
(,)
u)(z, n)
and
(11.145)
[(A
(,)
u)(z, n)[
2
= [u(z, n)[
2
+
1
[z
0
[
2
u
(, n)
u
0
(
0
, n)
[W
n
(u
0
(
0
), u(z))[
2
W
(,)
(n)
.
Moreover, the sequences
u
0
(n) =
u
(, n)
_
W
(,)
(n 1)W
(,)
(n)
,
u
(n) =
u
0
(
0
, n)
_
W
(,)
(n 1)W
(,)
(n)
, (11.146)
satisfy
(,)
u =
0
u,
(,)
u = u respectively. Note also u
0
(
0
, 0) = u
(0) = 0
and
(11.147) u
0
(n)u
(n) =
1
W
(,)
(n)
.
In addition, let u(z), u(z) satisfy u = zu, then
W
(,),n
(u
0
, A
(,)
u(z)) =
W
n
(u
(), u(z))
W
(,)
(n)
,
W
(,),n
(u
, A
(,)
u(z)) =
z
z
0
W
n
(u
0
(
0
), u(z))
W
(,)
(n)
,
W
(,),n
(A
(,)
u(z), A
(,)
u( z)) =
z
z
0
W
n
(u(z), u( z))
+
z z
(z
0
)( z
0
)
W
n
(u
(), u(z))W
n
(u
0
(
0
), u( z))
W
(,)
(n)
, (11.148)
where W
(,),n
(u, v) = a
(,)
(n)
_
u(n)v(n + 1) u(n + 1)v(n)
_
.
Since we do not know whether a
(,)
(n), b
(,)
(n) are bounded or not, we need
to (temporarily) introduce suitable boundary conditions for
(,)
(compare Sec
tion 2.6). They will soon turn out to be superuous.
Let and
(11.149) BC
(f) =
_
lim
n
W
n
(u
, f) = 0 if
(,)
is l.c. at
0 if
(,)
is l.p. at
,
where u
is u
= u
or u
= u
0
11.8. The Dirichlet deformation method 211
Using this boundary conditions we dene
(11.150)
H
(,)
: D(H
(,)
)
2
(Z)
f
(,)
f
,
where the domain of H
(,)
is explicitly given by
(11.151) D(H
(,)
) = f
2
(Z)[
(,)
f
2
(Z), BC
(f) = BC
+
(f) = 0.
Furthermore, H
(,),
denote the corresponding Dirichlet halfline operators with
respect to the base point n
0
= 0.
We rst give a complete spectral characterization of the halfline operators
H
(,),
. In addition, this will provide all necessary results for the investigation of
H
(,)
.
To begin with we compute the Weyl mfunctions of H
(,)
.
Theorem 11.35. Let m
(z) and m
(,),
(z) denote the Weyl mfunctions of H
and H
(,)
respectively. Then we have
(11.152) m
(,),
(z) =
1
1 +
(,)
_
z
0
z
m
(z)
(,)
z
_
,
where
(11.153)
(,)
=
_
_
_
(
0
) m
(), ,=
0
0u
0
(0,01)
2
n
0
N
u
0
(0,n)
2
, (, ) = (
0
,
0
)
.
In particular, H
(,),
and thus a
(,)
, b
(,)
are bounded.
Proof. We rst note that
c
(,)
(z, n) =
z
0
z
(A
(,)
c)(z, n)
(,)
z
a(0)(A
(,)
s)(z, n),
s
(,)
(z, n) =
_
1 +
(,)
(A
(,)
s)(z, n) (11.154)
constructed from the fundamental system c(z, n), s(z, n) for form a fundamental
system for
(,)
corresponding to the same initial conditions. Furthermore, note
(11.155)
W
(,)
(1)
W
(,)
(0)
= 1 +
(,)
,
W
(,)
(0)
W
(,)
(1)
= 1.
Now the result follows upon evaluating (cf. Section 2.6)
(11.156) m
(,),
(z) =
1
a
(,)
(0)
lim
n
W
(,),n
(c
(,)
(z), u
)
W
(,),n
(s
(,)
(z), u
)
.
Using (11.148) one obtains for u
= u
(n), u
0
(n) and v
= u
0
(
0
, n), u
0
(
0
, n),
respectively,
(11.157) m
(,),
(z) =
1
1 +
(,)
_
z
0
z
1
a(0)
lim
n
W
n
(c(z), v
)
W
n
(s(z), v
)
(,)
z
_
.
Hence the claim is evident.
In particular, since H
(,)
is bounded (away from ,
0
the spectrum of H
(,),
is unchanged), we can forget about the boundary conditions (11.149) and H
(,)
is
dened on
2
(Z). As a consequence we note
212 11. Spectral deformations
Corollary 11.36. The sequences
(11.158) u
(,),
(z, n) = (A
(,)
u
)(z, n), z C,
0
,
are square summable near . Moreover, the same is true for
u
(,),
(, n) = W
(,)
(n)u
0
(n) u
(n)
n
m=
u
(, m)
2
,
u
(,),0
(
0
, n) = W
(,)
(n)u
(n) u
0
(n)
n
m=0
u
0
(
0
, m)
2
(11.159)
and
u
(,),
(, n) = u
(n), ,
d
(H),
u
(,),0
(
0
, n) = u
0
(n),
0
,
d
(H). (11.160)
If or
0
d
(H) one has to replace the last formulas by
u
(,),
(, n) = W
(,)
(n)u
0
(n) u
(n)
n
m=
u
(, m)
2
,
u
(,),0
(
0
, n) = W
(,)
(n)u
(n) u
0
(n)
n
m=0
u
0
(
0
, m)
2
, (11.161)
respectively.
Proof. Follows immediately from
(11.162) u
(,),
(z, n) =
c
(z)
1 +
(,)
z
0
z
_
c
(,)
(z, n)
a
(,)
(0)
m
(,),
(z)s
(,)
(z, n)
_
if
(11.163) u
(z, n) = c
(z)
_
c(z, n)
a(0)
m
(z)s(z, n)
_
.
If z = ,
0
one can assume u
=
_
(
0
)( m
+
() + m
()), ,=
0
u(,1)
2
1
n=
u(,n)
2
+
u+(,1)
2
+
n=1
u+(,n)
2
, (, ) = (
0
,
0
)
0.
In addition,
= 0 if
d
(H)
0
and
> 0 otherwise. If ,=
0
, then
(11.166) m
(,),
(z) = O(z
0
)
0
,
and note
= a(0)
2
(
0
)g(, 0)
1
.
11.8. The Dirichlet deformation method 213
Using the previous corollary plus weak convergence of
1
Im( m
( + i))d
to the corresponding spectral measure d
() as 0 implies
Lemma 11.38. Let d
() and d
(,),
() be the respective spectral measures of
m
(z) and m
(,),
(z). Then we have
(11.167) d
(,),
() =
1
1 +
(,)
_
0
d
()+
_
d( ) =
0 =
_
_
,
where
(,)
,
(
0
), P
(,),
() denote the orthogonal projections onto the subspaces
spanned by u
0
(
0
, .), u
(.) in
2
(N), respectively. Then the above results clearly
imply
Theorem 11.39. The operators (1l P
(
0
))H
and (1l P
(,),
())H
(,),
are
unitarily equivalent. Moreover, , (H
(,),
) and
0
, (H
(,),
).
If ,
d
(H) or (, ) = (
0
,
0
), then (H
(,),
) and thus
(11.168) (H
(,),
) =
_
(H
)
0
_
, =
, =
.
Otherwise, that is, if
d
(H)
0
, then , (H
(,),
) and thus
(11.169) (H
(,),
) = (H
)
0
.
In essence, Theorem 11.39 says that, as long as ,
d
(H)
0
, the Dirichlet
datum (
0
,
0
) is rendered into (, ), whereas everything else remains unchanged.
If
d
(H)
0
, that is, if we are trying to move
0
to an eigenvalue, then
0
is
removed. This latter case reects the fact that we cannot move
0
to an eigenvalue
E without moving the Dirichlet eigenvalue on the other side of E to E at the same
time (compare Section 8.2).
Remark 11.40. (i). For f (N) set
(A
(,),+
f)(n) =
W
(,)
(n)
W
(,)
(n 1)
f(n)
u
0,(,)
(
0
, n)
n
j=1
u
0
(
0
, j)f(j),
(A
1
(,),+
f)(n) =
W
(,)
(n 1)
W
(,)
(n)
f(n)
u
(, n)
n
j=1
u
,(,)
(, j)f(j). (11.170)
Then we have A
(,),+
A
1
(,),+
= A
1
(,),+
A
(,),+
= 1l
(N)
and A
(,),+
acts as
transformation operator
(11.171)
(,),+
= A
(,),+
+
A
1
(,),+
.
214 11. Spectral deformations
Similarly, for f (N) set
(A
(,),
f)(n) =
W
(,)
(n)
W
(,)
(n 1)
f(n)
u
0,(,)
(
0
, n)
1
j=n+1
u
0
(
0
, j)f(j),
(A
1
(,),
f)(n) =
W
(,)
(n 1)
W
(,)
(n)
f(n)
u
(, n)
1
j=n+1
u
,(,)
(, j)f(j). (11.172)
Then we have A
(,),
A
1
(,),
= A
1
(,),
A
(,),
= 1l
(N)
and
(11.173)
(,),
= A
(,),
A
1
(,),
.
(ii). Note that the case (, ) = (
0
,
0
) corresponds to the double commutation
method with
1
= (cf. Section 11.6). Furthermore, the operators A
(,),
are
unitary when restricted to proper subspaces of
2
(N) in this case.
(iii). Due to the factor
z0
z
in front of m
(,),
(z), all norming constants (i.e., the
negative residues at each pole of m
(,),
(z)) are altered.
(iv). Clearly, by Theorem 11.35, the Dirichlet deformation method preserves reec
tionless properties (cf. Lemma 8.1).
Having these results at our disposal, we can now easily deduce all spectral
properties of the operator H
(,)
. First of all, Theorem 11.35 yields for the Weyl
matrix of H
(,)
.
Theorem 11.41. The respective Weyl Mmatrices M(z), M
(,)
(z) of H, H
(,)
are related by
M
(,),0,0
(z) =
1
(1 +
(,)
)
2
z
z
0
M
0,0
(z),
M
(,),0,1
(z) =
1
1 +
(,)
_
M
0,1
(z) +
(,)
z
0
a(0)M
0,0
(z)
_
,
M
(,),1,1
(z) =
z
0
z
M
1,1
(z) 2
(,)
z
0
a(0)M
0,1
(z)
+
2
(,)
(z
0
)(z )
a(0)
2
M
0,0
(z). (11.174)
Moreover, M
(,),j,k
(z, m, n), j, k 0, 1, are holomorphic near z = ,
0
.
Given the connection between M(z) and M
(,)
(z), we can compute the corre
sponding Herglotz matrix measure of M
(,)
(z) as in Lemma 11.38.
11.8. The Dirichlet deformation method 215
Lemma 11.42. The matrix measures d and d
(,)
corresponding to M(z) and
M
(,)
(z) are related by
d
(,),0,0
() =
1
(1 +
(,)
)
2
0
d
0,0
(),
d
(,),0,1
() =
1
1 +
(,)
_
d
0,1
() +
(,)
0
a(0)d
0,0
()
_
,
d
(,),1,1
() =
0
d
1,1
() 2
(,)
0
a(0)d
0,1
()
+
2
(,)
(
0
)( )
a(0)
2
d
0,0
(). (11.175)
Equivalently
d
(,)
() =
1
(z )(z
0
)
_
z
1+
(,)
0
a(0)
(,)
z
0
_
d()
_
z
1+
(,)
a(0)
(,)
0 z
0
_
. (11.176)
This nally leads to our main theorem in this section
Theorem 11.43. Let H, H
(,)
be dened as before. Denote by P(
0
) and P()
the orthogonal projections corresponding to the spaces spanned by u
0
(
0
, .) and
u
(, 0) in
2
(Z) respectively. Then (1l P(
0
) P())H and H
(,)
are unitarily
equivalent. In particular, H and H
(,)
are unitarily equivalent if ,
0
,
d
(H).
Remark 11.44. By inspection, Dirichlet deformations produce the commuting
diagram
(
1
,
1
)
(
0
,
0
) (
2
,
2
)
@
@
@
@R

for (
j
,
j
), 0 j 2, according to (H.11.31). In fact, this can be easily veried
using (the rst two) formulas for the spectral measures (11.175) which uniquely
determine the corresponding operator.
Remark 11.45. We have seen in Theorem 11.39 that the Dirichlet deformation
method cannot create a situation where a discrete eigenvalue E of H is rendered
into a Dirichlet eigenvalue (i.e., moving
0
to the eigenvalue E). However, one can
use the following threestep procedure to generate a prescribed degeneracy at an
eigenvalue E of H:
(i). Use the Dirichlet deformation method to move to a discrete eigenvalue
E of H. (This removes both the discrete eigenvalue E of H and the (Dirichlet)
eigenvalue of H
H
+
).
(ii) As a consequence of step (i), there is now another eigenvalue of H
H
+
in
the resulting larger spectral gap of H. Move to E using the Dirichlet deformation
method.
216 11. Spectral deformations
(iii) Use the double commutation method to insert an eigenvalue of H at E
and to change
0
into any allowed value.
Theorem 11.16 then shows that the resulting operator is unitarily equivalent
to the original operator H, and Theorem 11.21 proves that the remaining Dirichlet
eigenvalues remain invariant.
Finally, we briey comment on how to iterate the Dirichlet deformation method.
All ingredients have already been derived in Section 11.3. Suppose
(11.177) (E
0,j
, E
1,j
), (
0,j
,
0,j
), (
j
,
j
) [E
0,j
, E
1,j
]
satisfy (H.11.31) for each j = 1, . . . , N, N N. Then the Dirichlet deformation
result after N steps, denoted by H
(1,1),...,(
N
,
N
)
, is associated with the sequences
a
(1,1),...,(
N
,
N
)
(n) =
_
a(n N)a(n +N)
_
C
nN
(U
1,...,N
)C
nN+2
(U
1,...,N
)
C
nN+1
(U
1,...,N
)
,
b
(1,1),...,(
N
,
N
)
(n) = b(n)
a(n)
D
nN+1
(U
1,...,N
)
C
nN+1
(U
1,...,N
)
, (11.178)
where (U
1,...,N
) = (u
0,1
(
0,1
), u
1
(
1
), . . . , u
0,N
(
0,N
), u
N
(
N
)).
Notes on literature
Chapter 1. The literature on second order dierence equations is enormous and tracing
all results would require a historian. Hence I can only give some hints which serve as
starting points. Classical references are the books by Atkinson [19], Berezanski [27]
(Chapter 7), and Fort [87]. A more recent reference would be the book by Carmona and
Lacroix [40] which focuses on random operators. Clearly, there are far more books on this
topic available, but these are the ones I found most useful for my purpose. However, note
that not everything in the rst chapter can be found in these references. I have added
some results, needed in later chapters, which I could not nd in the literature (e.g., I dont
know a reference for Lemma 1.2).
In addition, recent and easy introductions to dierence equations can be found in,
for instance, Jordan [141], Kelley and Peterson [147], and Mickens [171]. A book which
contains numerous examples and special results is the one by Agarwal [4].
Chapter 2. The theory presented in this chapter is the discrete analog of what is known
as WeylTitchmarshKodaira theory for SturmLiouville operators. The continuous theory
originated in the celebrated work of Weyl [244] with subsequent extensions by Titchmarsh
([227], [228]), Kodaira and numerous others (see any book on SturmLiouville operators
for further details). The origins of the discrete theory go back to Hellinger [131] and
Nevanlinna [182]. Many results have been rediscovered several times and the literature
on this subject is far to extensive to be quoted in full.
Again classical references are Atkinson [19], Berezanski [27] (Chapter 7), Fort [87],
and the more recent monograph by Carmona and Lacroix [40]. My personal favorite
though is Berezanski [27]. For connections with spectral theory on the half line and
the moment problem I suggest the classical book by Akhiezer [9]. There is also a very
good and recent survey article on the moment problem by Simon [209]. For the view
point of orthogonal polynomials, the standard reference is the book of Szeg o [217]. You
might also have a look at [211], which has emphasis on applications, or [57], which uses
a RiemannHilbert approach.
Especially for relations between Weyl mfunctions and the theory of continued frac
tions see Atkinson [9] and also [24], [240]. A paper which closely follows the original
ideas of Titchmarsh is [18]. There is a large bulk of papers on the question of essential
selfadjointness of Jacobi operators (i.e., limit point, limit circle criteria). See for instance
[9] (addenda and problems to Chapter 1), [19], [27], [110] (Appendix E), [132], [135],
217
218 Notes on literature
[188]. The characterization of boundary conditions in the limit circle case can be found in
[242], [243] (however, a much more cumbersome approach is used there). A paper which
focuses on the contributions by Krein is [186].
The investigation of positive operators respectively positive solutions has attracted
quite some interest and many results have been discovered more than once. Our section
compiles results from [26], [111], and [188]. I have streamlined most arguments and
simplied the proofs. Further references are [6], [133], and [191].
The section on inverse problems is taken from [222]. Additional results can be found
in [88], [89], [92] and, in the case of nite Jacobi matrices, [104].
There are extensions of many results to block Jacobi matrices (i.e., Jacobi matrices
whose coecients are again matrices). The interested reader might look at (e.g.) [5], [7],
or [245].
Chapter 3. The rst section summarizes some wellknown results, which however are
often hard to nd. Some of them dont seem to be written up at all and I cant give a
good reference for this section. For the continuous case you should look at the beautiful
paper of Simon [204] and, again, at the book of Berezanski [27].
The second section extends results of [215] to the case a(n) = 1 and provides a few
basic facts for easy reference.
The absolutely (resp. singularly) continuous spectrum is one of the most investigated
object in spectral theory. The contents of this section clearly has its origin in the beautiful
theory of Gilbert and Pearson [120] which is known as the principle of subordinacy. This
theory has since then been simplied and extended by various people (see e.g., [119], [48],
[145]). I follow ideas of [68], [139], [163], and [207]. For some general results on singular
continuous spectra see [208].
Bounds on the number of eigenvalues are considered in [95] and [96] (see also [195],
Theorem XIII.10).
Additional topics like embedded eigenvalues, slowly decreasing potentials, or some
explicitly solvable problems can be found in [179], [180], [58], [149], [215], or [77]. A
dierent approach to nd classes of operators with absolutely continuous spectral measures
is presented in [71] (see also [72]).
Chapter 4. Oscillation theory originates in the work of Sturm [216]. Numerous papers
have been written on this subject in the case of dierential equations and several mono
graphs are devoted solely to this topic. In the discrete case there is a large amount of papers
investigating whether a given Jacobi equation is oscillatory or not (e.g., [128], [132], [133],
[134], [188], and [189]). The case of nite matrices is treated (e.g.) in the textbook by
Fort [87]. For the case of innite matrices only the connection between (non)oscillation
and the inmum of the essential spectrum was known ([121], Theorem 32). The precise
connection between the number of oscillations and the spectrum was only claried recently
in [223]. This article is also the main reference for this chapter. Additional references are
[19], [26], and [147].
Chapter 5. The rst main theorem, that the spectrum is a.s. the same for all operators
is due to Pastur [187]. The results concerning the Lyapunov exponent are mainly taken
from [49], the ones for the integrated density of states respectively the rotation number
are mainly collected from [69], [70], and [22]. It is sometimes claimed that the process
(5.48) is subadditve (which is true for dierential operators) so I tried to clarify this point.
The results on the absolutely continuous spectrum originate from work of Kotani and are
taken from [205]. The section on almost periodic operators summarizes some results from
[21] and [122]. There have recently been some eorts to prove uniform spectral results
(i.e, for all operators in the hull), see, for example, [51], [52] and the references therein. In
Notes on literature 219
this respect, it should be mentioned that the absolutely continuous spectrum is constant
for all operators in the hull ([163], Theorem 1.5). It is also known that the other parts
need not be constant (see [140]). If you want to nd out almost everything about the
almost Mathieu operator, I suggest [138] and [162].
The reader who wants to learn more about this subject should consult the monograph
by Carmona and Lacroix [40] which is devoted entirely to this subject. Another brief
introduction can be found in Cycon et al. [50] (Chapters 9 and 10). For some recent
generalizations to orthogonal polynomials on the unit circle see [100].
Chapter 6. The rst trace formulas for Schr odinger operators on a nite interval were
given by Gelfand and Levitan [93]. Since then numerous extensions have been made. In
particular, recently Gesztesy and Simon showed [105] that all main results can be viewed
as special cases of Kreins spectral shift theory [155]. Most parts of this section are taken
from [222].
Chapter 7. Floquet (resp. Bloch) theory is one of the classical results found in most
textbooks on dierential equations. In the case of periodic Jacobi operators the rst
spectral characterization was given by [173]. The rst two sections contain the analogs
of several results from periodic SturmLiouville operators. The third section is somewhat
inspired by [169] (see also Chapter 8). Further results on perturbed periodic operators
can be found in [98], [121] (Section 67), [150], and [197]. For extensions to complex or
operator valued coecients see [176] or [165], respectively.
Chapter 8. Reectionless operators have attracted several mathematicians since they
allow an explicit treatment. In the rst two sections I follow closely ideas of [113] and
[222]. However, I have added some additional material. Complementary results to this
sections, namely for the case of Cantor spectra, can be found in [12] [14], and [212],
[213].
The remaining sections are of importance for the second part. The approach is in
spired by ideas of McKean [169]. Some parts are taken from [35] (see also [173]), but
again the material has been extended to a larger class of operators and augmented.
Chapter 9. The use of Riemann surfaces usually appears in connection with the Toda
lattice. The rst results seem due to Krichever (see, e.g., [156] [160]). The main
reference for this chapter is again [35]. Additional references are [74], [169], [173], [230],
and [175]. This technique can also be extended to surfaces with innitely many branch
points [12] [14], [212] and [213].
For the explicit calculations in the genus one case, I have used the tables for elliptic
integrals in [38]. In addition, you might also nd [10], [123], and [127] useful.
Chapter 10. Discrete scattering theory has been rst developed on an informal level by
Case in a series of papers [42] [47]. The rst rigorous results were given by Guseinov
[124], but unfortunately without proof. Additional results are given in [125], [126], and
[224]. This chapter extends the results of [124] and gives (to the best of my knowledge)
the rst proof in print. Most of the other results are also new and have not previously
appeared elsewhere. Further references are [80] (Chap. 2.III), [94], [95], [97], [102], and
[230]. For generalizations to polynomials on the unit circle see [99].
Chapter 11. The single commutation method for Jacobi operators rst appeared in
Gesztesy et al. [117]. A special case of the double commutation method can be found in
[235]. The general case of the double commutation method seems due to [110] which is
220 Notes on literature
one of the the main sources for this chapter. It also contains the analogue of the FIT
formula (Section 11.4) derived in [83] for Schr odinger operators. The second main source
is [221] where the discrete Dirichlet deformation method is introduced.
Part 2
Completely Integrable
Nonlinear Lattices
Chapter 12
The Toda system
This chapter is devoted to the Toda hierarchy. The rst section gives an informal
introduction and is mainly for background and motivation. The following sections
contain a rigorous treatment based on the Lax pair formalism. The basic existence
and uniqueness theorem for solutions of the initial value problem is proven and the
connection with hyperelliptic Riemann surfaces is established.
12.1. The Toda lattice
The Toda lattice is a simple model for a nonlinear onedimensional crystal. It
describes the motion of a chain of particles with nearest neighbor interaction. The
equation of motion for such a system is given by
(12.1) m
d
2
dt
2
x(n, t) = V
t
(x(n + 1, t) x(n, t)) V
t
(x(n, t) x(n 1, t)),
where m denotes the mass of each particle, x(n, t) is the displacement of the nth
particle from its equilibrium position, and V (r) (V
t
(r) =
dV
dr
(r)) is the interaction
potential. As discovered by M. Toda, this system gets particularly interesting if
one chooses an exponential interaction,
(12.2) V (r) =
m
2
2
_
e
r/
+
r
1
_
=
m
2
2
2
_
(
r
)
2
+O(
r
)
3
_
, , R.
This model is of course only valid as long as the relative displacement is not too
large (i.e., at least smaller than the distance of the particles in the equilibrium
position). For small displacements it is equal to a harmonic crystal with force
constant
m
2
(cf. Section 1.5).
After a scaling transformation, t t/, x x/, we can assume m = = =
1. If we suppose x(n, t) x(n 1, t) 0, x(n, t) 0 suciently fast as [n[ ,
we can introduce the Hamiltonian (q = x, p = x)
(12.3) 1(p, q) =
nZ
_
p(n, t)
2
2
+ (e
(q(n + 1, t) q(n, t))
1)
_
223
224 12. The Toda system
and rewrite the equations of motion in the Hamiltonian form
d
dt
p(n, t) =
1(p, q)
q(n, t)
= e
(q(n, t) q(n 1, t))
e
(q(n + 1, t) q(n, t))
,
d
dt
q(n, t) =
1(p, q)
p(n, t)
= p(n, t). (12.4)
We remark that these equations are invariant under the transformation
(12.5) p(n, t) p(n, t) +p
0
and q(n, t) q(n, t) +q
0
+p
0
t, (p
0
, q
0
) R
2
,
which reects the fact that the dynamics remains unchanged by a uniform motion
of the entire crystal.
The fact which makes the Toda lattice particularly interesting is the existence
of soliton solutions. These are pulslike waves traveling through the crystal without
changing their shape. Such solutions are rather special since from a generic linear
equation one would expect spreading of wave packets (see Section 1.5) and from a
generic nonlinear wave equation one would expect that solutions only exist for a
nite time (breaking of waves).
The simplest example of such a solitary wave is the onesoliton solution
(12.6) q
1
(n, t) = q
0
ln
1 + exp(2n 2 sinh()t)
1 + exp(2(n 1) 2 sinh()t)
, , > 0.
It describes a single bump traveling trough the crystal with speed sinh()/ and
width proportional to 1/. That is, the smaller the soliton the faster it propagates.
It results in a total displacement
(12.7) lim
n
(q
1
(n, t) q
1
(n, t)) = 2
of the crystal, which can equivalently be interpreted as the total compression of the
crystal around the bump. The total moment and energy are given by
nZ
p
1
(n, t) = 2 sinh(),
1(p
1
, q
1
) = 2(sinh() cosh() ). (12.8)
Existence of such solutions is usually connected to complete integrability of the
system which is indeed the case here. To see this, we introduce Flaschkas variables
(12.9) a(n, t) =
1
2
e
(q(n + 1, t) q(n, t))/2
, b(n, t) =
1
2
p(n, t)
and obtain the form most convenient for us
a(n, t) = a(n, t)
_
b(n + 1, t) b(n, t)
_
,
b(n, t) = 2
_
a(n, t)
2
a(n 1, t)
2
_
. (12.10)
12.1. The Toda lattice 225
The inversion is given by
p(n, t) = 2b(n, t),
q(n, t) = q(n, 0) 2
_
t
0
b(n, s)ds
= q(0, 0) 2
_
t
0
b(0, s)ds 2
n1
j=0
ln(2a(j, t))
= q(0, 0) 2
_
t
0
b(n, s)ds 2
n1
j=0
ln(2a(j, 0)). (12.11)
To show complete integrability it suces to nd a socalled Lax pair, that is,
two operators H(t), P(t) such that the Lax equation
(12.12)
d
dt
H(t) = P(t)H(t) H(t)P(t)
is equivalent to (12.10). One can easily convince oneself that the choice
H(t) :
2
(Z)
2
(Z)
f(n) a(n, t)f(n + 1) +a(n 1, t)f(n 1) +b(n, t)f(n)
,
P(t) :
2
(Z)
2
(Z)
f(n) a(n, t)f(n + 1) a(n 1, t)f(n 1)
(12.13)
does the trick. Now the Lax equation implies that the operators H(t) for dierent
t R are all unitarily equivalent and that
(12.14) tr(H(t)
j
H
j
0
), j N,
are conserved quantities, where H
0
is the operator corresponding to the constant
solution a
0
(n, t) =
1
2
, b
0
(n, t) = 0 (it is needed to make the trace converge). For
example,
tr(H(t) H
0) =
nZ
b(n, t) =
1
2
nZ
p(n, t) and
tr(H(t)
2
H
2
0
) =
nZ
b(n, t)
2
+ 2(a(n, t)
2
1
4
) =
1
2
1(p, q) (12.15)
correspond to conservation of the total momentum and the total energy, respec
tively.
This reformulation of the Toda equations as a Lax pair is the key to methods
of solving the Toda equations based on spectral and inverse spectral theory for the
Jacobi operator H.
Using these methods one can nd the general Nsoliton solution
(12.16) q
N
(n, t) = q
0
ln
det(1l +C
N
(n, t))
det(1l +C
N
(n 1, t))
,
where
(12.17)
C
N
(n, t) =
_
j
1 e
(i+j)
e
(
i
+
j
)n (
i
sinh(
i
) +
j
sinh(
j
))t
_
1i,jN
226 12. The Toda system
with
j
,
j
> 0 and
j
1. One can also nd (quasi)periodic solutions us
ing techniques from Riemann surfaces (respectively algebraic curves). Each such
solution is associated with an hyperelliptic curve of the type
(12.18) w
2
=
2g+2
j=0
(z E
j
), E
j
R,
where E
j
, 0 j 2g + 1, are the band edges of the spectrum of H (which is
independent of t and hence determined by the initial conditions). One obtains
(12.19) q(n, t) = q
0
2(t
b +nln(2 a)) ln
(z
0
2nA
p0
(
+
) 2tc(g))
(z
0
2(n 1)A
p0
(
+
) 2tc(g))
,
where z
0
R
g
, : R
g
R is the Riemann theta function associated with the
hyperelliptic curve (12.18), and a,
b R, A
p0
(
+
), c(g) R
g
are constants de
pending only on the curve (i.e., on E
j
, 0 j 2g + 1). If q(n, 0), p(n, 0) are
(quasi) periodic with average 0, then a =
1
2
,
b = 0.
The rest of this monograph is devoted to a rigorous mathematical investigation
of these methods.
12.2. Lax pairs, the Toda hierarchy, and
hyperelliptic curves
In this section we introduce the Toda hierarchy using a recursive approach for the
standard Lax formalism and derive the BurchnallChaundy polynomials in connec
tion with the stationary Toda hierarchy.
We let the sequences a, b depend on an additional parameter t R and require
Hypothesis H. 12.1. Suppose a(t), b(t) satisfy
(12.20) a(t), b(t)
(Z)
(Z) be dierentiable.
We introduce the corresponding operator H(t) as usual, that is,
(12.21)
H(t) :
2
(Z)
2
(Z)
f(n) a(n, t)f(n + 1) +a(n 1, t)f(n 1) +b(n, t)f(n)
.
The idea of the Lax formalism is to nd a nite, skewsymmetric operator P
2r+2
(t)
such that the Lax equation
(12.22)
d
dt
H(t) [P
2r+2
(t), H(t)] = 0, t R,
(here [. , ..] denotes the commutator, i.e., [P, H] = PHHP) holds. More precisely,
we seek an operator P
2r+2
(t) such that the commutator with H(t) is a symmet
ric dierence operator of order at most two. Equation (12.22) will then give an
evolution equation for a(t) and b(t). Our rst theorem tells us what to choose for
P
2r+2
(t).
Theorem 12.2. Suppose P(t) is of order at most 2r +2 and the commutator with
H(t) is of order at most 2. Then P(t) is of the form
(12.23) P(t) =
r
j=0
_
c
rj
P
2j+2
(t) +d
rj
H(t)
j+1
_
+d
r+1
1l,
12.2. The Toda hierarchy 227
where c
j
, d
j
C, 0 j r, d
r+1
(Z), and
(12.24)
P
2j+2
(t) = [H(t)
j+1
]
+
[H(t)
j+1
]
(cf. the notation in (1.11)) is called homogeneous Lax operator. Moreover, denote
by P
2r+2
(t) the operator P(t) with c
0
= 1 and d
j
= 0, 0 j r +1. Then we have
(12.25) P
2r+2
(t) = H(t)
r+1
+
r
j=0
(2a(t)g
j
(t)S
+
h
j
(t))H(t)
rj
+g
r+1
(t),
where (g
j
(n, t))
0jr+1
and (h
j
(n, t))
0jr+1
are given by
g
j
(n, t) =
j
=0
c
j
n
, H(t)
n
,
h
j
(n, t) = 2a(n, t)
j
=0
c
j
n+1
, H(t)
n
+c
j+1
(12.26)
and satisfy the recursion relations
g
0
= 1, h
0
= c
1
,
2g
j+1
h
j
h
j
2bg
j
= 0, 0 j r,
h
j+1
h
j+1
2(a
2
g
+
j
(a
)
2
g
j
) b(h
j
h
j
) = 0, 0 j < r. (12.27)
For the commutator we obtain
[P
2r+2
(t), H(t)] = a(t)(g
+
r+1
(t) g
r+1
(t))S
+
+a
(t)(g
r+1
(t) g
r+1
(t))S
+ (h
r+1
(t) h
r+1
(t)). (12.28)
Proof. By Lemma 1.2 we can write
(12.29) P(t) = h
1
(t)H(t)
r+1
+
r
j=0
(2a(t)g
j
(t)S
+
h
j
(t))H(t)
rj
+g
r+1
(t),
where g
r+1
(t) is only added for convenience and hence can be chosen arbitrarily.
Now we insert this ansatz into [P, H]. Considering the term (S
)
r+2
we see that
h
1
(t) must be independent of n, say h
1
(t) = c
0
d
0
. Next, we obtain after a
long but straightforward calculation
[P, H] = 2a(g
+
0
g
0
)S
+
H
r+1
(h
0
h
0
)H
r+1
r1
j=0
a
_
(2g
j+1
h
j
h
j
2bg
j
)
_
S
+
H
rj
r1
j=0
_
h
j+1
h
j+1
2(a
2
g
+
j
(a
)
2
g
j
) b(h
j
h
j
)
_
H
rj
+a(g
+
r+1
g
r+1
)S
+
+S
a(g
+
r+1
g
r+1
) + (h
r+1
h
r+1
), (12.30)
where g
r+1
, h
r+1
have been chosen according to
(2g
r+1
h
r
h
r
2bg
r
) = 0,
h
r+1
h
r+1
2(a
2
g
+
r
(a
)
2
g
r
) b(h
r
h
r
) = 0. (12.31)
228 12. The Toda system
(Recall f = f
+
f.) But (12.30) is of order 2 if and only if (compare Lemma 1.2)
g
0
= c
0
, h
0
= c
1
d
1
,
2g
j+1
h
j
h
j
2bg
j
= 2d
j+1
, 0 j r 1,
h
j+1
h
j+1
2(a
2
g
+
j
(a
)
2
g
j
) b(h
j
h
j
) = 0, 0 j < r 2, (12.32)
where c
0
, c
1
, d
j
, 1 j r, are constants. By Lemma 6.4
(12.33) g
j
(n, t) =
n
, H(t)
j
n
,
h
j
(n, t) = 2a(n, t)
n+1
, H(t)
j
h
j
(n, t) are homogeneous of degree j, j + 1, respectively. The general solution of
the above system (12.32) is hence given by g
j
=
j
=0
c
g
j
, 1 j r, and
h
j
=
j
=0
c
h
j
+ c
j+1
d
j+1
, 1 j r 1. Introducing another arbitrary
sequence d
r+1
it is no restriction to assume that the formula for h
j
also holds for
j = r.
It remains to verify (12.24). We use induction on r. The case r = 0 is easy. By
(12.25) we need to show
(12.34)
P
2r+2
=
P
2r
H + (2a g
r
S
+
h
r) g
r
H + g
r+1
.
This can be done upon considering
m
,
P
2r+2
n
and making case distinctions
m < n 1, m = n 1, m = n, m = n + 1, m > n + 1. Explicitly, one veries, for
instance, in the case m = n,
n
,
P
2r+2
=
n
,
P
2r
(a
n1
+a
n+1
+b
n
)
h
r
(n) b(n) g
r
(n) + g
r+1
(n)
=
n
, ([H
r
]
+
[H
r
]
)(a
n1
+a
n+1
+b
n
)
h
r
(n) b(n) g
r
(n) + g
r+1
(n)
=
n
, [H
r
]
+
a
n+1
n
, [H
r
]
a
n1
h
r
(n) b(n) g
r
(n) + g
r+1
(n)
= a(n)
n
, H
r
n+1
a(n 1)
n
, H
r
n1
h
r
(n) b(n) g
r
(n) + g
r+1
(n)
=
h
r
(n) +
h
r
(n 1)
2
b(n) g
r
(n) + g
r+1
(n) = 0
(12.35)
using (12.27),
(12.36)
m
, [H
j
]
n
=
_
m
, H
j
n
, (n m) > 0
0, (n m) 0
,
and H
m
= a(m)
m+1
+a(m1)
m1
+b(m)
m
. This settles the case m = n. The
remaining cases are settled one by one in a similar fashion.
Remark 12.3. It is also easy to obtain (12.28) from (12.23) and (12.24). In fact,
simply evaluate
m
, [
P
2j+2
, H]
n
= H
m
, [H
j+1
]
+
n
+H
n
, [H
j+1
]
+
H
m
, [H
j+1
]
n
H
n
, [H
j+1
]
m
(12.37)
as in the proof of Theorem 12.2.
12.2. The Toda hierarchy 229
Since the selfadjoint part of P(t) does not produce anything interesting when
inserted into the Lax equation, we will set d
j
= 0, 0 j r + 1, and take
(12.38) P
2r+2
(t) =
r
j=0
c
rj
P
2j+2
(t)
as our Lax operator in (12.22). Explicitly we have
P
2
(t) = a(t)S
+
a
(t)S
P
4
(t) = a(t)a
+
(t)S
++
+a(t)(b
+
(t) +b(t))S
+
a
(t)(b(t) +b
(t))S
(t)a
(t)S
+c
1
(a(t)S
+
a
(t)S
)
etc. . (12.39)
Clearly, H(t) and iP
2r+2
(t) are bounded, selfadjoint operators.
Even though the expression (12.25) for P
2r+2
(t) looks much more complicated
and clumsy in comparison to (12.24), we will see that this ruse of expanding P
2r+2
(t)
in powers of H(t) will turn out most favorable for our endeavor. But before we can
see this, we need to make sure that the Lax equation is welldened, that is, that
H(t) is dierentiable.
First of all, please recall the following facts. Denote by B(
2
(Z)) the C
algebra
of bounded linear operators. Suppose A, B : R B(
2
(Z)) are dierentiable with
derivative
A,
B, respectively, then we have
A+B is dierentiable with derivative
A+
B,
AB is dierentiable with derivative
AB +A
B,
A
,
A
1
(provided A is invertible) is dierentiable with derivative A
1
AA
1
.
In addition, f : R
(Z) B(
2
(Z)) is
isometric).
For our original problem this implies that H(t) and P
2r+2
(t) are dierentiable
since they are composed of dierentiable operators. Hence the Lax equation (12.22)
is welldened and by (12.30) it is equivalent to
TL
r
(a(t), b(t))
1
= a(t) a(t)(g
+
r+1
(t) g
r+1
(t)) = 0,
TL
r
(a(t), b(t))
2
=
b(t) (h
r+1
(t) h
r+1
(t)) = 0, (12.40)
where the dot denotes a derivative with respect to t. Or, in integral form we have
a(t) = a(0) exp
_
_
t
0
(g
+
r+1
(t) g
r+1
(t))ds
_
,
b(t) = b(0) +
_
t
0
(h
r+1
(t) h
r+1
(t))ds. (12.41)
Varying r N
0
yields the Toda hierarchy (TL hierarchy)
(12.42) TL
r
(a, b) = (TL
r
(a, b)
1
, TL
r
(a, b)
2
) = 0, r N
0
.
Notice that multiplying P
2r+2
(t) with c
0
,= 0 gives only a rescaled version of the
Toda hierarchy which can be reduced to the original one by substituting t t/c
0
.
Hence our choice c
0
= 1.
230 12. The Toda system
Explicitly, one obtains from (12.27),
g
1
= b +c
1
,
h
1
= 2a
2
+c
2
,
g
2
= a
2
+ (a
)
2
+b
2
+c
1
b +c
2
,
h
2
= 2a
2
(b
+
+b) +c
1
2a
2
+c
3
,
g
3
= a
2
(b
+
+ 2b) + (a
)
2
(2b +b
) +b
3
+c
1
(a
2
+ (a
)
2
+b
2
) +c
2
b +c
3
,
h
3
= 2a
2
((a
+
)
2
+a
2
+ (a
)
2
+b
2
+b
+
b + (b
+
)
2
)
+c
1
2a
2
(b
+
+b) +c
2
2a
2
+c
4
,
etc. (12.43)
and hence
TL
0
(a, b) =
_
a a(b
+
b)
b 2(a
2
(a
)
2
)
_
,
TL
1
(a, b) =
_
a a((a
+
)
2
(a
)
2
+ (b
+
)
2
b
2
)
b 2a
2
(b
+
+b) + 2(a
)
2
(b +b
)
_
c
1
_
a(b
+
b)
2(a
2
(a
)
2
)
_
,
TL
2
(a, b) =
_
aa((a
+
)
2
(b
++
+2b
+
)+a
2
(2b
+
+b)+(b
+
)
3
a
2
(b
+
+2b)(a
)
2
(2b+b
)b
3
)
b+2((a
)
4
+(a
)
2
a
2
(a
2
+(a
+
)
2
+(b
+
)
2
+bb
+
+b
2
)+(a
)
2
(b
2
+b
b+(b
)
2
))
_
c
1
_
a((a
+
)
2
(a
)
2
+(b
+
)
2
b
2
)
2a
2
(b
+
+b)2(a
)
2
(b+b
)
_
c
2
_
a(b
+
b)
2(a
2
(a
)
2
)
_
,
etc. (12.44)
represent the rst few equations of the Toda hierarchy. We will require c
j
R even
though c
j
could depend on t. The corresponding homogeneous Toda equations
obtained by taking all summation constants equal to zero, c
0, 1 r, are
then denoted by
(12.45)
TL
r
(a, b) = TL
r
(a, b)
0, 1r
.
We are interested in investigating the initial value problem associated with
the Toda equations, that is,
(12.46) TL
r
(a, b) = 0, (a(t
0
), b(t
0
)) = (a
0
, b
0
),
where (a
0
, b
0
) are two given (bounded) sequences. Since the Toda equations are
autonomous, we will choose t
0
= 0 without restriction.
In order to draw a number of fundamental consequences from the Lax equation
(12.22), we need some preparations.
Let P(t), t R, be a family of bounded skewadjoint operators in
2
(Z). A two
parameter family of operators U(t, s), (t, s) R
2
, is called a unitary propagator
for P(t), if
(1) U(t, s), (s, t) R
2
, is unitary.
(2) U(t, t) = 1l for all t R.
(3) U(t, s)U(s, r) = U(t, r) for all (r, s, t) R
3
.
12.2. The Toda hierarchy 231
(4) The map t U(t, s) is dierentiable in the Banach space B(
2
(Z)) of
bounded linear operators and
(12.47)
d
dt
U(t, s) = P(t)U(t, s), (t, s) R
2
.
Note U(s, t) = U(t, s)
1
= U(t, s)
= (
d
dt
U(t, s))
= (P(t)U(t, s))
= U(t, s)
P(t)
and
(12.49)
d
dt
U(t, s)
1
= U(t, s)
1
(
d
dt
U(t, s))U(t, s)
1
= U(t, s)
1
P(t)
we infer U(t, s)
= U(t, s)
1
by unique solubility of the initial value problem and
U(s, s)
= U(s, s)
1
= 1l.
If P(t) = P is actually timeindependent (stationary solutions), then the uni
tary propagator is given by Stones theorem, that is, U(t, s) = exp((t s)P).
The situation for unbounded operators is somewhat more dicult and requires
the operators P(t), t R, to have a common dense domain (cf. [218], Corollary on
page 102, [193], Theorem X.69).
Now we can apply this fact to our situation.
Theorem 12.5. Let a(t), b(t) satisfy TL
r
(a, b) = 0 and (H.12.1). Then the Lax
equation (12.22) implies existence of a unitary propagator U
r
(t, s) for P
2r+2
(t) such
that
(12.50) H(t) = U
r
(t, s)H(s)U
r
(t, s)
1
, (t, s) R
2
.
Thus all operators H(t), t R, are unitarily equivalent and we might set
(12.51) (H) (H(t)) = (H(0)), (H) (H(t)) = (H(0)).
(Here (.) and (.) = C(.) denote the spectrum and resolvent set of an operator,
respectively.)
In addition, if (s)
2
(Z) solves H(s)(s) = z(s), then the function
(12.52) (t) = U
r
(t, s)(s)
fullls
(12.53) H(t)(t) = z(t),
d
dt
(t) = P
2r+2
(t)(t).
232 12. The Toda system
Proof. Let U
r
(t, s) be the unitary propagator for P
2r+2
(t). We need to show that
H(t) = U
r
(t, s)
1
H(t)U
r
(t, s) is equal to H(s). Since
H(s) = H(s) it suces to
show that
H(t) is independent of t, which follows from
(12.54)
d
dt
H(t) = U
r
(t, s)
1
(
d
dt
H(t) [P
2r+2
(t), H(t)])U
r
(t, s) = 0.
The rest is immediate from the properties of the unitary propagator.
To proceed with our investigation of the Toda equations, we ensure existence
and uniqueness of global solutions next. To do this, we consider the Toda equations
as a ow on the Banach space M =
(Z)
(Z).
Theorem 12.6. Suppose (a
0
, b
0
) M. Then there exists a unique integral curve
t (a(t), b(t)) in C
+ b(t)
Ker((t)z)
= 2a(t)G
r
(z, t)S
+
H
r+1
(z, t),
where G
r
(z, n, t) and H
r+1
(z, n, t) are monic (i.e. the highest coecient is one)
polynomials given by
G
r
(z, n, t) =
r
j=0
g
rj
(n, t)z
j
,
H
r+1
(z, n, t) = z
r+1
+
r
j=0
h
rj
(n, t)z
j
g
r+1
(n, t). (12.57)
One easily obtains
a = a(H
+
r+1
+H
r+1
2(z b
+
)G
+
r
),
b = 2(a
2
G
+
r
(a
)
2
G
r
) + (z b)
2
G
r
(z b)H
r+1
. (12.58)
12.3. Stationary solutions 233
As an illustration we record a few of the polynomials G
r
and H
r+1
,
G
0
= 1 =
G
0
,
H
1
= z b =
H
1
,
G
1
= z +b +c
1
=
G
1
+c
1
G
0
,
H
2
= z
2
+a
2
(a
)
2
b
2
+c
1
(z b) =
H
2
+c
1
H
1
,
G
2
= z
2
+bz +a
2
+ (a
)
2
+b
2
+c
1
(z +b) +c
2
=
G
2
+c
1
G
1
+c
2
G
0
,
H
3
= z
3
+ 2a
2
z 2(a
)
2
b b
3
+a
2
b
+
(a
)
2
b
+c
1
(z
2
+a
2
(a
)
2
b
2
) +c
2
(z b) =
H
3
+c
1
H
2
+c
2
H
1
,
etc. . (12.59)
Here
G
r
(z, n) and
H
r+1
(z, n) are the homogeneous quantities corresponding to
G
r
(z, n) and H
r+1
(z, n), respectively. By (12.38) we have
(12.60) G
r
(z, n) =
r
=0
c
r
G
(z, n), H
r+1
(z, n) =
r
=0
c
r
H
+1
(z, n).
Remark 12.7. (i). Since, by (12.27), a(t) enters quadratically in g
j
(t), h
j
(t),
respectively G
r
(z, ., t), H
r+1
(z, ., t), the Toda hierarchy (12.42) is invariant under
the substitution
(12.61) a(n, t) (n)a(n, t),
where (n) +1, 1. This result should be compared with Lemma 1.6.
(ii). If a(n
0
, 0) = 0 we have a(n
0
, t) = 0 for all t R (by (12.41)). This implies
H = H
,n0+1
H
+,n0
with respect to the decomposition
2
(Z) =
2
(, n
0
+1)
2
(n
0
, ). Hence P
2r+2
= P
,n0+1,2r+2
P
+,n0,2r+2
decomposes as well and we
see that the Toda lattice also splits up into two independent parts TL
,n0,r
. In
this way, the half line Toda lattice follows from our considerations as a special case.
Similarly, we can obtain Toda lattices on nite intervals.
12.3. Stationary solutions
In this section we specialize to the stationary Toda hierarchy characterized by
a =
b = 0 in (12.42) or, equivalently, by commuting dierence expressions
(12.62) [P
2r+2
, H] = 0
of order 2r + 2 and 2, respectively. Equations (12.58) then yield the equivalent
conditions
(z b)(H
r+1
H
r+1
) = 2a
2
G
+
r
2(a
)
2
G
r
,
H
+
r+1
+H
r+1
= 2(z b
+
)G
+
r
. (12.63)
Comparison with Section 8.3 suggests to dene
(12.64) R
2r+2
= (H
2
r+1
4a
2
G
r
G
+
r
).
A simple calculation using (12.63)
(z b)(R
2r+2
R
2r+2
)
= (z b)((H
r+1
+H
r+1
)(H
r+1
H
r+1
) 4G
r
(a
2
G
+
r
(a
)
2
G
r
))
= 2(H
r+1
+H
r+1
2(z b)G
r)(a
2
G
+
r
(a
)
2
G
r
) = 0 (12.65)
234 12. The Toda system
then proves that R
2r+2
is independent of n. Thus one infers
(12.66) R
2r+2
(z) =
2r+1
j=0
(z E
j
), E
j
0j2r+1
C.
The resulting hyperelliptic curve of (arithmetic) genus r obtained upon compacti
cation of the curve
(12.67) w
2
= R
2r+2
(z) =
2r+1
j=0
(z E
j
)
will be the basic ingredient in our algebrogeometric treatment of the Toda hierarchy
in Section 13.1.
Equations (12.63), (12.64) plus Theorem 2.31 imply
(12.68) g(z, n) =
G
r
(z, n)
R
1/2
2r+2
(z)
, h(z, n) =
H
r+1
(z, n)
R
1/2
2r+2
(z)
,
where g(z, n) =
n
, (Hz)
1
n
, h(z, n) = 2a(n)
n+1
, (Hz)
1
n
1 as usual.
Despite these similarities we need to emphasize that the numbers E
m
, 0
m 2r + 1, do not necessarily satisfy (8.54). This is no contradiction but merely
implies that there must be common factors in the denominators and numera
tors of (12.68) which cancel. That is, if the number of spectral gaps of H is
s + 2, then there is a monic polynomial Q
rs
(z) (independent of n) such that
G
r
(z, n) = Q
rs
(z)G
s
(z, n), H
r+1
(z, n) = Q
rs
(z)H
s+1
(z, n), and R
1/2
2r+2
(z) =
Q
rs
(z)
2
R
1/2
2s+2
(z).
We have avoided these factors in Section 8.3 (which essentially correspond
to closed gaps (cf. Remark 7.6)). For example, in case of the constant solution
a(n, t) = 1/2, b(n, t) = 0 of TL
r
(a, b) = 0 we have for r = 0, 1, . . .
G
0
(z) = 1, H
1
(z) = z, R
2
(z) = z
2
1,
G
1
(z) = z +c
1
, H
2
(z) = (z +c
1
)z, R
4
(z) = (z +c
1
)
2
(z
2
1),
etc. . (12.69)
Conversely, any given reectionless nite gap sequences (a, b) satisfy (12.63)
and hence give rise to a stationary solution of some equation in the Toda hierarchy
and we obtain
Theorem 12.8. The stationary solutions of the Toda hierarchy are precisely the
reectionless nitegap sequences investigated in Section 8.3.
In addition, with a little more work, we can even determine which equation
(i.e., the constants c
j
). This will be done by relating the polynomials G
r
(z, n),
H
r+1
(z, n) to the homogeneous quantities
G
r
(z, n),
H
r+1
(z, n). We introduce the
constants c
j
(E), E = (E
0
, . . . , E
2r+1
), by
(12.70) R
1/2
2r+2
(z) = z
r+1
j=0
c
j
(E)z
j
, [z[ > H,
12.3. Stationary solutions 235
implying
(12.71) c
0
(E) = 1, c
1
(E) =
1
2
2r+1
j=0
E
j
, etc. .
Lemma 12.9. Let a(n), b(n) be given reectionless nitegap sequences (see Sec
tion 8.3) and let G
r
(z, n), H
r+1
(z, n) be the associated polynomials (see (8.66) and
(8.70)). Then we have (compare (12.60))
(12.72) G
r
(z, n) =
r
=0
c
r
(E)
G
(z, n), H
r+1
(z, n) =
r
=0
c
r
(E)
H
+1
(z, n).
In addition, g
(n) =
1
_
b
()
(n) +
1
j=1
j
g
j
(n)b
(j)
(n)
_
, > 1,
where
(12.74) b
()
(n) =
1
2
2r+1
j=0
E
j
r
j=1
j
(n)
.
Proof. From (8.96) we infer for [z[ > H, using Neumanns expansion for the
resolvent of H and the explicit form of g,
=0
g
(n)z
,
H
r+1
(z, n) = R
1/2
2r+2
(z)
_
1
1
z
=0
(n)z
_
. (12.75)
This, together with (12.70), completes the rst part. The rest follows from (6.59)
and Theorem 6.10.
Corollary 12.10. Let a(n), b(n) be given reectionless nitegap sequences with
corresponding polynomial R
1/2
2s+2
(z). Then (a, b) is a stationary solution of TL
r
if and only if there is a constant polynomial Q
rs
(z) of degree r s such that
c
j
= c
j
(E), where E is the vector of zeros of R
1/2
2r+2
(z) = Q
rs
(z)
2
R
1/2
2s+2
(z).
It remains to show how all stationary solutions for a given equation of the Toda
hierarchy can be found. In fact, up to this point we dont even know whether the
necessary conditions (12.63) can be satised for arbitrary choice of the constants
(c
j
)
1jr
.
Suppose (c
j
)
1jr
is given and dene d
0
= c
0
= 1,
(12.76) d
j
= 2c
j
+
j1
=1
c
c
j
, 1 j r.
Choose d
j
R, r + 1 j 2r + 2, and dene (E
j
)
0j2r+1
by
(12.77) R
2r+2
(z) =
2r+2
j=0
d
2r+2j
z
j
=
2r+1
j=0
(z E
j
).
236 12. The Toda system
Note that our choice of d
j
implies c
j
(E) = c
j
, 1 j r, which is a necessary
condition by Lemma 12.9. Since g(z, n) cannot be meromorphic, R
2r+2
(z) must
not be a complete square. Hence those choices of d
j
R, r+1 j 2r+2, have to
be discarded. For any other choice we obtain a list of band edges satisfying (8.54)
after throwing out all closed gaps. In particular, setting
(12.78) (d) =
r
_
j=0
[E
2j
, E
2j+1
]
any operator in Iso
R
((d)) (see Section 8.3) produces a stationary solution. Thus,
the procedure of Section 8.3 can be used to compute all corresponding stationary
gap solutions.
Theorem 12.11. Fix TL
r
, that is, x (c
j
)
1jr
and r. Let (d
j
)
1jr
be given
by (12.76). Then all stationary solutions of TL
r
can be obtained by choosing
(d
j
)
r+1j2r+2
such that R
2r+2
(z) dened as in (12.77) is not a complete square
and then choosing any operator in the corresponding isospectral class Iso
R
((d)).
Remark 12.12. The case where R
2r+2
(z) is a complete square corresponds to
stationary solutions with a(n) = 0 for some n. For instance, G
0
(n) = 1, H
1
(n) =
z b(n), R
2r+2
(z) = (z b(n))
2
corresponds to a(n) = 0, n Z.
Finally, let us give a further interpretation of the polynomial R
2r+2
(z).
Theorem 12.13. The polynomial R
2r+2
(z) is the BurchnallChaundy polyno
mial relating P
2r+2
and H, that is,
(12.79) P
2
2r+2
= R
2r+2
(H) =
2r+1
j=0
(H E
j
).
Proof. Because of (12.62) one computes
_
P
2r+2
Ker(z)
_
2
=
_
(2aG
r
S
+
H
r+1
)
Ker(z)
_
2
=
_
2aG
r
(2(z b
+
)G
+
r
H
+
r+1
H
r+1
)S
+
+H
2
r+1
4a
2
G
r
G
+
r
_
Ker(z)
= R
2r+2
(z)
Ker(z)
(12.80)
and since z C is arbitrary, the rest follows from Corollary 1.3.
This result clearly shows again the close connection between the Toda hierarchy
and hyperelliptic curves of the type M = (z, w)[w
2
=
2r+2
j=0
(z E
j
).
12.4. Time evolution of associated quantities
For our further investigations in the next two chapters, it will be important to know
how several quantities associated with the Jacobi operator H(t) vary with respect
to t.
First we will try to calculate the time evolution of the fundamental solutions
c(z, ., t), s(z, ., t) if a(t), b(t) satisfy TL
r
(a, b) = 0. For simplicity of notation we
will not distinguish between H(t) and its dierential expression (t). To emphasize
12.4. Time evolution of associated quantities 237
that a solution of H(t)u = zu is not necessarily in
2
(Z) we will call such solutions
weak solutions. Similarly for P
2r+2
(t).
First, observe that (12.22) implies
(12.81) (H(t) z)(
d
dt
P
2r+2
(t))(z, ., t) = 0,
where (z, n, t) is the transfer matrix from (1.30). But this means
(12.82) (
d
dt
P
2r+2
(t))(z, ., t) = (z, ., t)C
r
(z, t)
for a certain matrix C
r
(z, t). If we evaluate the above expression at n = 0, using
(z, 0, t) = 1l, we obtain
C
r
(z, t) = (P
2r+2
(t)(z, ., t))(0)
=
_
H
r+1
(z, 0, t) 2a(0, t)G
r
(z, 0, t)
2a(0, t)G
r
(z, 1, t) 2(z b(1, t))G
r
(z, 1, t) H
r+1
(z, 1, t)
_
. (12.83)
The time evolutions of c(z, n, t) and s(z, n, t) now follow from
(12.84)
(z, ., t) = P
2r+2
(z, ., t) + (z, ., t)C
r
(z, t)
or more explicitly
c(z, n, t) = 2a(n, t)G
r
(z, n, t)c(z, n + 1, t) (H
r+1
(z, n, t)
+H
r+1
(z, 0, t))c(z, n, t) 2a(0, t)G
r
(z, 1, t)s(z, n, t),
s(z, n, t) = 2a(n, t)G
r
(z, n, t)s(z, n + 1, t) (H
r+1
(z, n, t) +H
r+1
(z, 1, t)
2(z b(1, t))G
r
(z, 1, t))s(z, n, t) + 2a(0, t)G
r
(z, 0, t)c(z, n, t). (12.85)
Remark 12.14. In case of periodic coecients, this implies for the time evolution
of the monodromy matrix M(z, t)
(12.86)
d
dt
M(z, t) = [M(z, t), C
r
(z, n, t)],
where
(12.87) C
r
(z, n, t) =
_
H
r+1
2aG
r
2aG
+
r
2(z b
+
)G
+
r
H
r+1
_
.
This shows (take the trace) that the discriminant is time independent and that
E
j
2N
j=1
, A, and B (cf. Section 7.1) are time independent.
Evaluating (12.86) explicitly and expressing everything in terms of our polyno
mials yields (omitting some dependencies)
d
dt
_
H/2 aG
aG
+
H/2
_
=
_
2a
2
(G
r
G
+
G
+
r
G) 2a(G
r
H (z b
+
)G
+
r
G)
2a(G
+
r
H (z b
+
)G
+
r
G
+
) 2a
2
(G
r
G
+
G
+
r
G)
_
. (12.88)
Equation (12.84) enables us to prove
Lemma 12.15. Assume (H.12.1) and suppose TL
r
(a, b) = 0. Let u
0
(z, n) be a
weak solution of H(0)u
0
= zu
0
. Then the system
(12.89) H(t)u(z, n, t) = zu(z, n, t),
d
dt
u(z, n, t) = P
2r+2
(t)u(z, n, t)
238 12. The Toda system
has a unique weak solution fullling the initial condition
(12.90) u(z, n, 0) = u
0
(z, n).
If u
0
(z, n) is continuous (resp. holomorphic) with respect to z, then so is u(z, n, t).
Furthermore, if u
1,2
(z, n, t) both solve (12.89), then
W(u
1
(z), u
2
(z)) = W
n
(u
1
(z, t), u
2
(z, t)) =
a(n, t)
_
u
1
(z, n, t)u
2
(z, n + 1, t) u
1
(z, n + 1, t)u
2
(z, n, t)
_
(12.91)
depends neither on n nor on t.
Proof. Any solution u(z, n, t) of the system (12.89) can be written as
(12.92) u(z, n, t) = u(z, 0, t)c(z, n, t) +u(z, 1, t)s(z, n, t)
and from (12.84) we infer that (12.89) is equivalent to the ordinary dierential
equation
(12.93)
_
u(z, 0, t)
u(z, 1, t)
_
= C
r
(z, t)
_
u(z, 0, t)
u(z, 1, t)
_
,
_
u(z, 0, 0)
u(z, 1, 0)
_
=
_
u
0
(z, 0)
u
0
(z, 1)
_
,
which proves the rst assertion. The second is a straightforward calculation using
(12.56) and (12.58).
The next lemma shows that solutions which are square summable near for
one t R remain square summable near for all t R, respectively.
Lemma 12.16. Let u
,0
(z, n) be a solution of H(0)u = zu which is square sum
mable near . Then the solution u
(z), u
+
(z))
_
u
+
(z, n, t)u
(z, m, t) for m n
u
+
(z, m, t)u
(z, n, t) for n m
.
Especially, if z < (H) and a(n, t) < 0 we can choose u
,0
(z, n) > 0, implying
u
(z, n, t) > 0.
Proof. We only prove the u
case (the u
+
case follows from reection) and drop
z for notational simplicity. By Lemma 12.15 we have a solution u(n, t) of (12.89)
with initial condition u(n, 0) = u
+,0
(n) and hence
(12.95) S(n, t) = S(n, 0) + 2
_
t
0
Re
0
j=n
u(j, s)P
2r+2
(s)u(j, s)ds,
where S(n, t) =
0
j=n
[u(j, t)[
2
. Next, by boundedness of a(t), b(t), we can nd
a constant C > 0 such that 4[H
r+1
(n, t)[ C and 8[a(n, t)G
r
(n, t)[ C. Using
(12.56) and the CauchySchwarz inequality yields
(12.96)
j=n
u(j, s)P
2r+2
(s)u(j, s)
C
2
_
[u(1, s)[
2
+S(n, s)
_
.
Invoking Gronwalls inequality shows
(12.97) S(n, t)
_
S(n, 0) +C
_
t
0
[u(1, s)[
2
e
Cs
ds
_
e
Ct
12.4. Time evolution of associated quantities 239
and letting n implies that u(z, ., t)
2
(Z).
Since u(z, n, t) = 0, z < (H), is not possible by Lemma 2.6, positivity follows
as well and we are done.
Using the Lax equation (12.22) one infers (z (H))
(12.98)
d
dt
(H(t) z)
1
= [P
2r+2
(t), (H(t) z)
1
].
Furthermore, using (12.94) we obtain for m n
d
dt
G(z, n, m, t) =
=
u
(z, m)(P
2r+2
u
+
(z, .))(n) +u
+
(z, n)(P
2r+2
u
(z, .))(m)
W(u
(z), u
+
(z))
. (12.99)
As a consequence (use (12.56) and (12.58)) we also have
d
dt
g(z, n, t) = 2(G
r
(z, n, t)h(z, n, t) g(z, n, t)H
r+1
(z, n, t)),
d
dt
h(z, n, t) = 4a(n, t)
2
(G
r
(z, n, t)g(z, n + 1, t)
g(z, n, t)G
r
(z, n + 1, t)) (12.100)
and
d
dt
g
j
(t) = 2 g
r+j+1
(t) + 2
r
=0
(g
r
(t)
h
+j
(t) g
+j
(t)h
r
(t))
+ 2g
r+1
(t) g
j
(t),
d
dt
h
j
(t) = 4a(t)
2
r
=0
(g
r
(t) g
+
+j
(t) g
+j
(t)g
+
r
(t)). (12.101)
Chapter 13
The initial value problem
for the Toda system
In the previous chapter we have seen that the initial value problem associated with
the Toda equations has a unique (global) solution. In this section we will consider
certain classes of initial conditions and derive explicit formulas for solutions in these
special cases.
13.1. Finitegap solutions of the Toda hierarchy
In this section we want to construct reectionless nitegap solutions (see Sec
tion 8.3) for the Toda hierarchy. Our starting point will be an rgap stationary
solution (a
0
, b
0
) of the type (8.83). This rgap stationary solution (a
0
, b
0
) repre
sents the initial condition for our Toda ow (cf. (12.42)),
(13.1)
TL
s
(a(t), b(t)) = 0, (a(0), b(0)) = (a
0
, b
0
)
for some s N
0
, whose explicit solution we seek. That is, we take a stationary
solution of the Toda hierarchy and consider the time evolution with respect to a (in
general) dierent equation of the Toda hierarchy. To stress this fact, we use a hat
for the Toda equation (and all associated quantities) which gives rise to the time
evolution.
From our treatment in Section 8.3 we know that (a
0
, b
0
) is determined by the
band edges (E
j
)
0j2r+1
and the Dirichlet eigenvalues (
0,j
(n
0
))
1jr
at a xed
point n
0
Z. Since we do not know whether the reectionless property of the initial
condition is preserved by the Toda ow, (a(t), b(t)) might not be reectionless.
However, if we suppose that (a(t), b(t)) is reectionless for all t, we have
(13.2) g(z, n
0
, t) =
r
j=1
z
j
(n
0
, t)
R
1/2
2r+2
(z)
.
241
242 13. The initial value problem
Plugging (13.2) into (12.100) and evaluating at
j
(n
0
, t) we arrive at the following
time evolution for the Dirichlet eigenvalues
j
(n
0
, t)
d
dt
j
(n
0
, t) = 2
G
s
(
j
(n
0
, t), n
0
, t)
j
(n
0
, t)R
j
(n
0
, t),
j
(n
0
, 0) =
0,j
(n
0
), 1 j r, t R, (13.3)
(recall (8.67)). Here
G
s
(z, n
0
, t) has to be expressed in terms of
j
(n
0
, t), 1 j r
(cf. Lemma 12.9).
Now our strategy consists of three steps.
(1) Show solvability of the system (13.3).
(2) Construct a(n, t), b(n, t) (and associated quantities) from (
j
(n
0
, t))
1jr
as in Section 8.3.
(3) Show that a(n, t), b(n, t) solve
TL
r
(a, b) = 0.
In order to show our rst step, that is, solubility of the system (13.3) we will rst
introduce a suitable manifold M
D
for the Dirichlet eigenvalues
j
. For the sake of
simplicity we will only consider the case r = 2 with E
0
= E
1
< E
2
= E
3
< E
4
< E
5
,
the general case being similar. We dene M
D
as the set of all points (
1
,
2
)
satisfying (H.8.12) together with the following charts (compare Section A.7).
(i). Set U
1, 2
= (
1
,
2
) M
D
[
1,2
,= E
2
,
2
,= E
4
,
1,2
=
1,2
, U
t
1, 2
=
(E
0
, E
2
) (E
2
, E
4
).
(13.4)
U
1, 2
U
t
1, 2
U
t
1, 2
U
1, 2
(
1
,
2
) (
1
,
2
) (
1
,
2
) ((
1
,
1
), (
2
,
2
))
.
(ii). Set U
E2
= (
1
,
2
) M
D
[
2
,= E
4
,
1
=
2
, U
t
E2
= (E
0
E
4
, E
4
E
0
)
(0, ).
U
E2
U
t
E2
(
1
,
2
) (
1
(
2
1
),
_
2E2
1E2
_
1
)
((E
2
, ), (E
2
, )) (0,
1
1+
)
,
U
t
E2
U
E2
(
1
> 0,
2
) ((E
2
1
1+2
, +), (E
2
+
12
1+2
, ))
(0,
2
) ((E
2
,
12
1+2
), (E
2
,
12
1+2
))
(
1
< 0,
2
) ((E
2
+
12
1+2
, ), (E
2
1
1+2
, +))
. (13.5)
(iii). Set U
E4,
= (
1
,
2
) M
D
[
1,2
,= E
2
,
1
= , U
t
E4,
= (E
0
, E
2
)
(
E
4
E
2
,
E
4
E
2
).
U
E4,
U
t
E4,
(
1
,
2
) (
1
,
2
E
4
2
)
,
U
t
E4,
U
E4,
(
1
,
2
) ((
1
, ), (E
4
2
2
, sgn(
2
)))
. (13.6)
13.1. Finitegap solutions of the Toda hierarchy 243
Next, we look at our system (13.3). In the chart (i) it reads (omitting the
dependence on t and n
0
)
1
= 2
G
s
(
1
)
1
R
1/2
6
(
1
)
2
,
2
= 2
G
s
(
2
)
2
R
1/2
6
(
2
)
1
. (13.7)
In the chart (ii) it reads (
1
,= 0)
1
=
1
1 +
2
_
2
F
s
(
1
2
1 +
2
) F
s
(
1
1 +
2
)
_
,
2
=
2
F
s
(
12
1+2
) F
s
(
1
1+2
)
1
, (13.8)
where
(13.9) F
s
(z) =
G
s
(z)R
1/2
6
(z)
z E
2
= 2
G
s
(z)(z E
0
)
_
z E
4
_
z E
5
.
Note that F
s
(z) remains invariant if
1
and
2
are exchanged. Moreover, the
singularity at
1
= 0 in the second equation is removable and hence this vector eld
extends to the whole of U
t
E2
.
Finally, in the chart (iii) it reads
1
= 2
G
s
(
1
)
1
R
1/2
6
(
1
)
1
E
4
+
2
2
,
2
= 2
G
s
(E
4
2
2
)
(E
4
E
0
+
2
2
)(E
4
E
2
+
2
2
)
_
E
5
E
4
2
2
1
E
4
+
2
2
. (13.10)
Hence our system (13.3) is a C
1,2
is close to E
2
and
1
=
2
. In this case we can make the coordinate change
1
=
1
+
2
,
2
= (E
2
1
)(
2
E
2
). Again we get
2
=
2
(. . . ) and hence
2
remains positive or equivalently neither
1
nor
2
can reach E
2
if
1
=
2
.
Concluding, for given initial conditions (
1
(0),
2
(0)) = (
0,1
,
0,2
) M
D
we
get a unique solution (
1
(t),
2
(t)) C
(R, M
D
). Due to the requirement
1
(E
0
, E
2
] and
2
[E
2
, E
4
], the functions
1,2
(t) are not C
2
cross E
2
. Hence if F(
1
,
2
) C
(R
2
, R) is symmetric with respect to
1
and
2
, then we have F(
1
(.),
2
(.)) C
(R, R).
In summary,
Theorem 13.1. The system (13.3) has a unique global solution
(13.11) (
j
(n
0
, .))
1jr
C
(R, M
D
)
for each initial condition satisfying (H.8.12).
244 13. The initial value problem
If s = r and c
= c
z=j(n0,t)
= 2
G
s
(
j
(n
0
, t), n
0
, t)H
r+1
(
j
(n
0
, t), n
0
, t),
1 j r. Since two polynomials of (at most) degree r 1 coinciding at r points
are equal, we infer
(13.13)
d
dt
G
r
(z, n
0
, t) = 2(
G
s
(z, n
0
, t)H
r+1
(z, n
0
, t) G
r
(z, n
0
, t)
H
s+1
(z, n
0
, t)),
provided we can show that the righthand side of (13.13) is a polynomial of degree
less or equal to r 1. By (12.60) it suces to consider the homogeneous case. We
divide (13.13) by R
1/2
2r+2
(z) and use (12.75) to express G
r
(z, n
0
, t) and H
r+1
(z, n
0
, t)
respectively (12.57) to express
G
s
(z, n
0
, t) and
H
s+1
(z, n
0
, t). Collecting powers of
z shows that the coecient of z
j
, 0 j r, is
(13.14) g
sj
s
=j+1
g
j1
h
s
g
sj
+
s
=j+1
h
s
g
j1
= 0,
which establishes the claim.
To obtain the time derivative of H
r+1
(z, n
0
, t) we use
(13.15) H
r+1
(z, n
0
, t)
2
4a(n
0
, t)
2
G
r
(z, n
0
, t)G
r
(z, n
0
+ 1, t) = R
1/2
2r+2
(z)
as in (8.73). Again, evaluating the time derivative rst at
j
(n
0
, t), one obtains
d
dt
H
r+1
(z, n
0
, t) = 4a(n
0
, t)
2
(
G
s
(z, n
0
, t)G
r
(z, n
0
+ 1, t)
G
r
(z, n
0
, t)
G
s
(z, n
0
+ 1, t)) (13.16)
for those t R such that
j
(n
0
, t) , E
2j1
, E
2j
, provided the righthand side
of (13.16) is of degree at most r 1. This can be shown as before. Since the
exceptional set is discrete, the identity follows for all t R by continuity.
Similarly, dierentiating (13.15) and evaluating (d/dt)G
r
(z, n
0
+ 1, t) at z =
j
(n
0
+ 1, t) (the zeros of G
r
(z, n
0
+ 1, t)), we see that (13.13) also holds with n
0
replaced by n
0
+ 1 and nally that (
j
(n
0
+ 1, t))
1jr
satises (13.3) with initial
condition
j
(n
0
+1, 0) =
0,j
(n
0
+1). Proceeding inductively, we obtain this result
for all n n
0
and with a similar calculation (cf. Section 8.3) for all n n
0
.
Summarizing, we have constructed the points (
j
(n, t))
1jr
for all (n, t)
Z R such that
d
dt
j
(n, t) = 2
G
s
(
j
(n, t), n, t)
j
(n, t)R
j
(n, t),
j
(n, 0) =
0,j
(n), 1 j r, (n, t) Z R, (13.17)
with (
j
(n, .))
1jr
C
(R, M
D
).
13.1. Finitegap solutions of the Toda hierarchy 245
Furthermore, we have corresponding polynomials G
r
(z, n, t) and H
r+1
(z, n, t)
satisfying
d
dt
G
r
(z, n, t) = 2
_
G
s
(z, n, t)H
r+1
(z, n, t)
G
r
(z, n, t)
H
s+1
(z, n, t)
_
(13.18)
and
d
dt
H
r+1
(z, n, t) = 4a(n, t)
2
_
G
s
(z, n, t)G
r
(z, n + 1, t)
G
r
(z, n, t)
G
s
(z, n + 1, t)
_
. (13.19)
It remains to verify that the sequences a(t), b(t) solve the Toda equation
TL
s
(a, b) = 0.
Theorem 13.2. The solution (a(t), b(t)) of the
TL
s
equations (13.1) with rgap
initial conditions (a
0
, b
0
) of the type (8.83) is given by
a(n, t)
2
=
1
2
r
j=1
R
j
(n, t) +
1
8
2r+1
j=0
E
2
j
1
4
r
j=1
j
(n, t)
2
1
4
b(n, t)
2
,
b(n, t) =
1
2
2r+1
j=0
E
j
r
j=1
j
(n, t), (13.20)
where (
j
(n, t))
1jr
is the unique solution of (13.17).
Proof. Dierentiating (13.15) involving (13.18), (13.19) and (8.82) yields the rst
equation of (12.58). Dierentiating (8.82) using (13.18), (13.19) and (8.101) yields
the second equation of (12.58).
Observe that this algorithm is constructive and can be implemented numerically
(it only involves nding roots of polynomials and integrating ordinary dierential
equations).
Next, let us deduce some additional information from the results obtained thus
far.
Clearly (13.18) and (13.19) imply
d
dt
K
r+1
(z, n, t) = 2a(n, t)
_
(
1
G
r
(z, n + 1, t) G
r
(z, n, t))
K
s+1
(z, n, t)
(
1
G
s
(z, n + 1, t)
G
s
(z, n, t))K
r+1
(z, n, t)
_
+
a(n, t)
a(n, t)
_
K
r+1
(z, n, t) +(
1
G
r
(z, n + 1, t) G
r
(z, n, t))
_
(13.21)
and as a consequence we note (use (8.113))
246 13. The initial value problem
Lemma 13.3. The time evolution of the zeros
j
(n, t), R0, of the polyno
mial K
r+1
(., n, t) is given by
d
dt
j
(n, t) =
_
a(n, t)
s+1
(
j
(n, t), n, t) +
a(n, t)
2a(n, t)
_
R
1/2
2r+2
(
j
(n, t))
r+1
,=j
j
(n, t)
(n, t)
. (13.22)
If
j
(n, t) =
j+1
(n, t) the righthand side has to be replaced by a limit as in (8.67).
Finally, let us compute the time dependent BakerAkhiezer function. Using
(13.23) (p, n, t) =
H
r+1
(p, n, t) +R
1/2
2r+2
(p)
2a(n, t)G
r
(p, n, t)
=
2a(n, t)G
r
(p, n + 1, t)
H
r+1
(p, n, t) R
1/2
2r+2
(p)
we dene for (p, n, n
0
, t)
(p, n, n
0
, t) = exp
_
_
t
0
(2a(n
0
, x)
G
s
(p, n
0
, x)(p, n
0
, x)
H
s+1
(p, n
0
, x))dx
_n1
m=n0
(p, m, t). (13.24)
Straightforward calculations then imply
(13.25) a(n, t)(p, n, t) +a(n 1, t)(p, n 1, t)
1
= (p) b(n, t),
d
dt
ln (p, n, t) = 2a(n, t)(
G
s
(p, n, t)(p, n, t) +
G
s
(p, n + 1, t)(p, n, t)
1
)
+ 2((p) b(n + 1, t))
G
s
(p, n + 1, t)
H
s+1
(p, n + 1, t) +
H
s+1
(p, n, t)
= 2a(n + 1, t)
G
s
(p, n + 1, t)(p, n + 1, t)
2a(n, t)
G
s
(p, n, t)(p, n, t)
H
s+1
(p, n + 1, t) +
H
s+1
(p, n, t). (13.26)
Similarly, (compare Lemma 12.15)
a(n, t)(p, n + 1, n
0
, t) +a(n 1, t)(p, n 1, n
0
, t)
= ((p) b(n, t))(p, n, n
0
, t), (13.27)
and
d
dt
(p, n, n
0
, t) = 2a(n, t)
G
s
(p, n, t)(p, n + 1, n
0
, t)
H
s+1
(p, n, t)(p, n, n
0
, t)
=
P
2s+2
(t)(p, ., n
0
, t)(n). (13.28)
The analogs of all relations found in Section 8.3 clearly extend to the present
time dependent situation.
13.2. Quasiperiodic nitegap solutions 247
13.2. Quasiperiodic nitegap solutions and the
timedependent BakerAkhiezer function
In this section we again consider the case with no eigenvalues, that is r = g, since we
do not want to deal with singular curves. The manifold for our Dirichlet eigenvalues
is in this case simply the submanifold
g
j=1
1
([E
2j1
, E
2j
]) M
g
introduced in
Lemma 9.1.
For simplicity we set n
0
= 0 and omit it in the sequel. In order to express
(p, n, t) and (p, n, t) in terms of theta functions, we need a bit more of notation.
Let
,j
be the normalized Abelian dierential of the second kind with a single
pole at
, j N
0
.
For
+,j
,j
we can make the following ansatz
(13.30)
+,j
,j
=
g+j+1
k=0
c
g+jk+1
k
d
R
1/2
2g+2
.
From the asymptotic requirement (13.29) we obtain c
k
= c
k
(E), 0 k j + 1,
where c
k
(E) is dened in (12.70). The remaining g constants c
k+j+1
, 1 k g,
have to be determined from the fact that the aperiods must vanish.
Given the summation constants c
1
, . . . , c
s
in
G
s
, see (12.38), we then dene
(13.31)
s
=
s
j=0
(j + 1) c
sj
(
+,j
,j
), c
0
= 1.
Since the dierentials
,j
were supposed to be normalized, we have
(13.32)
_
aj
s
= 0, 1 j g.
Next, let us write down the expansion for the abelian dierentials of the rst kind
j
,
(13.33)
j
=
_
=0
j,
(
)w
_
dw =
_
=0
j,
(
+
)w
_
dw near
.
Using (A.116) we infer
(13.34)
=
g
k=max1,g]
c(k)d
g+k
(E),
where d
k
(E) is dened by
(13.35)
1
R
1/2
2g+2
(z)
=
1
z
g+1
k=0
d
k
(E)
z
k
, d
0
(E) = 1, d
1
(E) =
1
2
2g+1
k=0
E
k
, etc. .
In addition, relation (A.19) yields
(13.36) U
s,j
=
1
2i
_
bj
s
= 2
s
=0
c
s
j,
(
+
), 1 j g.
As in Lemma 9.7 we will show uniqueness of the BakerAkhiezer function.
248 13. The initial value problem
Lemma 13.4. Let (., n, t), (n, t) ZR, be meromorphic on M
+
,
with
essential singularities at
such that
(13.37)
(p, n, t) = (p, n, t) exp
_
t
_
p
p0
s
_
is multivalued meromorphic on M and its divisor satises
(13.38) (
(., n, t)) T
(0,0)
+n(T
+
T
).
Dene a divisor T
zer
(n, t) by
(13.39) (
(., n, t)) = T
zer
(n, t) T
(0,0)
+n(T
+
T
).
If T
(0,0)
is nonspecial, then so is T
zer
(n, t) for all (n, t) Z R, that is,
(13.40) i(T
zer
(n, t)) = 0, (n, t) Z R,
and (., n, t) is unique up to a constant multiple (which may depend on n and t).
Proof. By the RiemannRoch theorem for multivalued functions (cf. [81], III.9.12)
there exists at least one BakerAkhiezer function. The rest follows as in Lemma 9.7.
p0
,
z(n, t) = z(
+
, n, t). (13.41)
Then we have
(p, n, t) = C(n, t)
(z(p, n + 1, t))
(z(p, n, t))
exp
_
_
p
p0
+,
_
,
(p, n, t) = C(n, 0, t)
(z(p, n, t))
(z(p, 0, 0))
exp
_
n
_
p
p0
+,
+t
_
p
p0
s
_
, (13.42)
where C(n, t), C(n, 0, t) are realvalued,
(13.43) C(n, t)
2
=
(z(n 1, t))
(z(n + 1, t))
, C(n, 0, t)
2
=
(z(0, 0))(z(1, 0))
(z(n, t))(z(n 1, t))
,
and the sign of C(n, t) is opposite to that of a(n, t). Moreover,
(13.44)
p0
(T
(n,t)
) =
p0
(T
(0,0)
) 2nA
p0
(
+
) [tU
s
],
where [.] denotes the equivalence class in J(M) (cf. (A.53)).
Hence the ows (13.11) are linearized by the Abel map
(13.45)
d
dt
p0
(T
(n,t)
) = U
s
.
Proof. First of all note that both functions in (13.42) are welldened due to (9.34),
(13.32), (13.36), and (A.69).
Denoting the righthand side of (13.42) by (p, n, t), our goal is to prove = .
By Theorem 9.2 it suces to identify
(13.46) (p, 0, t) = (p, 0, t).
13.2. Quasiperiodic nitegap solutions 249
We start by noting that (13.23), (13.24), and (13.18) imply
(p, 0, t) = exp
__
t
0
_
2a(0, x)
G
s
(z, 0, x)(p, 0, x)
H
s+1
(z, 0, x)
_
dx
_
= exp
_
_
t
0
_
G
s
(z, 0, x)(R
1/2
2g+2
(p) +H
g+1
(z, 0, x))
G
g
(z, 0, x)
H
s+1
(z, 0, x)
_
dx
_
. (13.47)
In order to spot the zeros and poles of on M
+
,
j
(0, x)
j
(0, x) (p)
+O(1)
_
dx
_
=
_
_
_
(
j
(0, t) (p))O(1) for p near
j
(0, t) ,=
j
(0, 0)
O(1) for p near
j
(0, t) =
j
(0, 0)
(
j
(0, 0) (p))
1
O(1) for p near
j
(0, 0) ,=
j
(0, t)
, (13.48)
with O(1) ,= 0. Hence all zeros and all poles of (p, 0, t) on M
+
,
are
simple and the poles coincide with those of (p, 0, t). Next, we need to identify
the essential singularities of (p, 0, t) at
G
s
(z, 0, x)
G
g
(z, 0, x)
_
dx
_
=
_
G
g
(z, 0, t)
G
g
(z, 0, 0)
_
1/2
exp
_
R
1/2
2g+2
(p)
_
t
0
G
s
(z, 0, x)
G
g
(z, 0, x)
dx
_
. (13.49)
We claim that
(13.50) R
1/2
2g+2
(p)
G
s
(z, n, t)/G
g
(z, n, t) =
s
=0
c
s
z
1+
+O(1) for p near
.
By (12.38), in order to prove (13.50), it suces to prove the homogeneous case
c
0
= 1, c
= 0, 1 s, that is
G
s
=
G
s
. Using (8.96), we may rewrite (13.50) in
the form
G
s
(z, n, t)z
s1
= z
1
s
=0
g
s
(n, t)z
s
= g(z, n, t) +O(z
s1
) as z . (13.51)
But this follows from (6.7) and hence we conclude = const(t). The remaining
constant can be determined as in Theorem 9.2.
It is instructive to give a direct proof that (13.17) is linearized by Abels map.
It suces to assume that all
j
(n, t) are away from the band edges E
m
(the general
case follows from continuity) and that c
j
= 0, 1 j s. Therefore we calculate
250 13. The initial value problem
(in the chart induced by the projection) using (9.17), (13.17)
d
dt
p0
(T
(n,t)
) =
g
j=1
j
(n, t)
g
k=1
c(k)
j
(n, t)
k1
R
1/2
2g+2
(
j
(n, t))
= 2
g
j,k=1
c(k)
G
s
(
j
(n, t), n, t)
,=j
(
j
(n, t)
(n, t))
j
(n, t)
k1
=
2
2i
g
k=1
c(k)
_
G
s
(z, n, t)
g
=1
(z
(n, t))
z
k1
dz, (13.52)
where is a closed path in C encircling all Dirichlet eigenvalues
j
(n, t). The last
integral can be evaluated by calculating the residue at . Using (see (13.51))
(13.53)
G
s
(z, n, t)
g
j=1
(z
k
(n, t))
=
G
s
(z, n, t)
g(z, n, t)
1
R
1/2
2g+2
(z)
=
z
s+1
(1 +O(z
s
))
R
1/2
2g+2
(z)
we infer (cf. (13.34))
(13.54)
d
dt
p0
(T
(n,t)
) = 2
g
k=max1,gs]
c(k)d
sg+k
(E) = 2
s
= U
s
.
Equation (9.13) shows that the ows (13.22) are linearized by the Abel map as well
(13.55)
d
dt
p0
(T
(n,t)
) = U
s
.
Finally, the function representation for the time dependent quasiperiodic
ggap solutions of the Toda hierarchy follows as in Theorem 9.4.
Theorem 13.6. The solution (a(t), b(t)) of the
TL
s
equations (13.1) with ggap
initial conditions (a
0
, b
0
) is given by
a(n, t)
2
= a
2
(z(n 1, t))(z(n + 1, t))
(z(n, t))
2
,
b(n, t) =
b +
g
j=1
c
j
(g)
w
j
ln
_
(w +z(n, t))
(w +z(n 1, t))
_
w=0
= E
0
+ a
(z(n 1, t))(z(p
0
, n + 1, t))
(z(n, t))(z(p
0
, n, t))
+ a
(z(n, t))(z(p
0
, n 1, t))
(z(n 1, t))(z(p
0
, n, t))
, (13.56)
with a,
b introduced in (9.42).
13.3. A simple example continued
We now continue our simple example of Section 1.3. The sequences
(13.57) a(n, t) = a
0
=
1
2
, b(n, t) = b
0
= 0,
13.4. Inverse scattering transform 251
solve the Toda equation TL
r
(a
0
, b
0
) = 0 since we have
G
0,r
(z, n, t) = G
0,r
(z) =
r
j=1
(z E
2j
),
H
0,r+1
(z, n, t) = H
0,r+1
(z) = z
r
j=1
(z E
2j
),
R
2r+2
(z) =
2r1
j=0
(z E
j
), (13.58)
where E
0
= 1, E
2r+1
= 1, E
2j1
= E
2j
, 1 j r. In addition, we have
(13.59) g
0,2j
=
_
2j
j
_
, g
0,2j+1
= 0,
h
0,j
= g
0,j1
.
Furthermore, the sequences
(13.60)
(z, n, t) = k
n
exp
_
r
(k)t
2
_
, z =
k +k
1
2
,
where
(13.61)
r
(k) = 2(kG
0,r
(z) H
0,r+1
(z)) = (k k
1
)G
0,r
(z)
satisfy (compare Section 1.3)
H
0
(t)
(z, n, t) = z
(z, n, t),
d
dt
(z, n, t) = P
0,2r+2
(t)
(z, n, t)
= 2a
0
G
0,r
(z)
(z, n + 1, t) H
0,r+1
(z)
G
0,2r
(z) =
r
j=0
(2j)!
4
j
(j!)
2
z
2(rj)
,
G
0,2r+1
(z) = z
G
0,2r
(z),
H
0,r+1
(z) = z
G
0,r
(z)
and hence
0
(k) = k k
1
,
1
(k) = k
2
k
2
+c
1
(k k
1
),
etc. . (13.64)
13.4. Inverse scattering transform
This section is an application of Chapter 10 to the Toda case. Our strategy is
similar to that of the previous section. Again we will look for the time evolution of
suitable spectral data, namely the scattering data associated with H(t). We rst
show that it suces to check (10.1) for one t
0
R.
Lemma 13.7. Suppose a(n, t), b(n, t) is a solution of the Toda system satisfying
(10.1) for one t
0
R, then (10.1) holds for all t R, that is,
(13.65)
nZ
[n[([1 2a(n, t)[ +[b(n, t)[) < .
252 13. The initial value problem
Proof. Without loss of generality we choose t
0
= 0. Shifting a a
1
2
we can
consider the norm
(13.66) (a, b)
nZ
(1 +[n[)([1 2a(n)[ +[b(n)[).
Proceeding as in Theorem 12.6 we conclude local existence of solutions with respect
to this norm. Next, we note that by (12.26) we have the estimate
nZ
(1 +[n[)[g
r
(n, t) g
0,r
[ C
r
(H(0))(a(t), b(t))
nZ
(1 +[n[)[h
r
(n, t) h
0,r
[ C
r
(H(0))(a(t), b(t))
, (13.67)
where C
r
(H(0)) is some positive constant. It suces to consider the homo
geneous case c
j
= 0, 1 j r. In this case the claim follows by induction
from equations (6.9) and (6.13) (note that we have g
i
(n, t)g
j
(m, t) g
0,i
g
0,j
=
(g
i
(n, t) g
0,i
)g
j
(m, t) g
0,i
(g
j
(m, t) g
0,j
)). Hence we infer from (12.40)
[a(n, t)
1
2
[ [a(n, 0)
1
2
[ +H(0)
_
t
0
[g
r+1
(n, s) g
0,r+1
[
+[g
r+1
(n + 1, s) g
0,r+1
[ds,
[b(n, t)[ [b(n, 0)[ +
_
t
0
[h
r+1
(n, s) h
0,r+1
[
+[h
r+1
(n 1, s) h
0,r+1
[ds (13.68)
and thus
(13.69) (a(t), b(t))
(a(0), b(0))
+
C
_
t
0
(a(s), b(s))
ds,
where
C = 2(1 +H(0))C
r+1
(H(0)). The rest follows from Gronwalls inequal
ity.
Thus we can dene Jost solutions, transmission and reection coecients as
in Chapter 10 with the only dierence that they now depend on an additional
parameter t R. For example, we set
(13.70) S
(H(t)) = R
,
,
(t), 1 N.
Clearly we are interested how the scattering data vary with respect to t.
Theorem 13.8. Suppose a(n, t), b(n, t) is a solution of the Toda system satisfying
(10.1) for one t
0
R. Then the functions
(13.71) exp(
r
(k)t
2
)f
(k, n, t)
fulll (12.89) with z = (k + k
1
)/2, where f
r
(k) is dened in (13.61). In addition, we have
T(k, t) = T(k, 0),
R
(k, t) = R
(k, 0) exp(
r
(k)t),
,
(t) =
,
(0) exp(
r
(k
)t), 1 N. (13.72)
13.4. Inverse scattering transform 253
Proof. As in Lemma 7.10 one shows that f
(t)f
(k, n, t).
Inserting this into (12.89), multiplying with k
n
, and evaluating as n yields
(13.74) C
(t) = exp(
r
(k)t/2).
The general result for all [k[ 1 now follows from continuity. This immediately
implies the formulas for T(k, t), R
. Then we have
(13.75) exp(
r
(k
)t/2)f
(k
, n, t) = U
r
(t, 0)f
(k
, n, 0)
(compare (12.52)), which implies
(13.76)
d
dt
exp(
r
(k
)t)
,
(t)
=
d
dt
U
r
(t, 0)f
(k
, ., 0)
2
= 0
and concludes the proof.
Thus S
(H(0)) as follows
(13.77) S
(H(t)) = R
(k, 0) exp(
r
(k)t), [k[ = 1;
k
,
,
(0) exp(
r
(k
)t), 1 N
and a(n, t), b(n, t) can be computed from a(n, 0), b(n, 0) using the GelfandLevitan
Marchenko equations. This procedure is known as inverse scattering transform and
can be summarized as follows:
(a(0), b(0)) (a(t), b(t))
6
scattering
theory
?
GelfandLevitan
Marchenko equations
S
(H(0)) S
(H(t))

time evolution
Hence, in the short range situation, the initial value problem for the Toda
lattice equations can be reduced to solving three linear problems. The mathematical
pillars of this method are formed by Theorem 12.6, Lemma 13.7, Theorem 13.8,
and Theorem 10.10.
Since the transmission coecient is time independent, so is the perturbation
determinant (10.36)
(13.78) (k(z), t) = (k(z), 0) = exp
_
_
R
()d
z
_
.
254 13. The initial value problem
Moreover, the traces
(13.79) tr
_
H(t)
(H
0
)
_
= tr
_
H(0)
(H
0
)
_
=
_
R
()d.
produce an innite series of conservation laws for the Toda lattice. Or, one could
also take the expansion coecients K
m
(see (10.17)) as conserved quantities.
13.5. Some additions in case of the Toda lattice
The rest of this chapter is devoted to the Toda lattice, that is, to the case TL
0
(a, b) =
0.
Combining
(13.80)
d
dt
p0
(T
(n,t)
) = 2c(g).
with (9.16) we obtain
(13.81)
p0
(T
(n,t)
) =
p0
(T
) 2nA
p0
(
+
) 2[tc(g)].
Hence we can rewrite (13.56) as
a(n, t)
2
= a
2
(z(n + 1, t))(z(n 1, t))
(z(n, t))
2
,
b(n, t) =
b +
1
2
d
dt
ln
_
(z(n, t))
(z(n 1, t))
_
. (13.82)
Now let us verify TL
0
(a, b) = 0 directly. We start with the rst equation
d
dt
a(n, t)
2
= a
2
d
dt
(z(n + 1, t))(z(n 1, t))
(z(n, t))
2
= a
2
(z(n + 1, t))(z(n 1, t))
(z(n, t))
2
d
dt
ln
_
(z(n + 1, t))(z(n 1, t))
(z(n, t))
2
_
= 2a(n, t)
2
_
b(n + 1, t) b(n, t)
_
. (13.83)
The second equation is more complicated. We compare the expansions (6.26) and
(9.45), (9.42) to obtain
a
2
(n 1, t) =
b
2
c +
b
g
j=1
B
j
(n, t) +
1
2
_
_
g
j=1
B
j
(n, t)
_
2
j=1
_
2g+2
k=1
E
k
2
+
c
j
(g 1)
c
j
(g)
_
B
j
(n, t) +
g
j,k=1
BB
j,k
(n, t)
_
,
a
2
(n, t) +b
2
(n, t) = c +
b
g
j=1
B
j
(n, t) +
1
2
_
_
g
j=1
B
j
(n, t)
_
2
+
g
j=1
_
2g+2
k=1
E
k
2
+
c
j
(g 1)
c
j
(g)
_
B
j
(n, t) +
g
j,k=1
BB
j,k
(n, t)
_
. (13.84)
13.5. Some additions in case of the Toda lattice 255
Adding these equations and eliminating b(n, t)
2
yields
a
2
(n, t) a
2
(n 1, t) =
g
j,k=1
BB
j,k
(n, t) =
1
4
d
2
dt
2
ln
_
(z(n, t))
(z(n 1, t))
_
=
1
2
d
dt
b(n, t). (13.85)
The BakerAkhiezer function for the Toda lattice is given by
(13.86) (p, n, t) = C(n, 0, t)
(z(p, n, t))
(z(p, 0, 0))
exp
_
n
_
p
p0
+,
+t
_
p
p0
0
_
,
where
(13.87) C(n, 0, t)
2
=
(z(0, 0))(z(1, 0))
(z(n, t))(z(n 1, t))
and
0
=
+,0
,0
. Since we have the following expansion near
(13.88)
_
p
p0
0
= (z +
d +O(
1
z
)),
d R,
we can make the following ansatz for
0
(13.89)
0
=
g
j=0
(
j
)
R
1/2
2g+2
d.
The constants
j
have to be determined from the normalization
(13.90)
_
aj
0
= 2
E2j
_
E2j1
g
=0
(z
)
R
1/2
2g+2
(z)
dz = 0,
and the requirement
(13.91)
g
j=0
j
=
1
2
2g+1
j=0
E
j
.
It follows that the bperiods are given by
(13.92)
_
bj
0
= 2
E2j1
_
E2j2
g
=0
(z
)
R
1/2
2g+2
(z)
dz = 4i c
j
(g).
Remark 13.9. Expanding equation (13.86) around
shows that
(13.93)
_
p
p0
0
=
_
z +
b +O(z
1
)
_
near
,
where
b is dened in (9.42). Conversely, proving (13.28) as in (9.61) (by expanding
both sides in (13.28) around
0
,
1
from
(13.94)
0
+
1
=
E
0
+E
1
+E
2
+E
3
2
and ([38] 254.00, 254.10, 340.01, 340.02, 336.02)
_
a1
0
= 2
_
E2
E1
x
2
(
0
+
1
)x +
1
R
1/2
4
(x)
dx
=
C
2
_
(E
0
E
3
+E
1
E
2
2
1
)K(k)
+ (E
2
E
0
)(E
3
E
1
)E(k)
_
= 0, (13.95)
which implies
(13.96) 2
1
= (E
2
E
0
)(E
3
E
1
)
E(k)
K(k)
+ (E
0
E
3
+E
1
E
2
).
The bperiod is given by (cf. (13.92))
(13.97)
_
b1
0
= 2
_
E1
E0
x
2
(
0
+
1
)x +
1
R
1/2
4
(x)
dx = 4i c(1) =
2i
CK(k)
.
And for p = (z, R
1/2
4
(z)) near p
0
we have ([38] 251.00, 251.03, 340.01, 340.02,
336.02)
_
p
p0
0
=
_
z
E0
x
2
(
0
+
1
)x +
1
R
1/2
4
(x)
dx
=
C(E
2
E
0
)
2
_
(E
3
E
1
)
E(k)F(u(z), k) K(k)E(u(z), k)
K(k)
+ (E
3
E
0
)
u(z)
_
1 u(z)
2
_
1 k
2
u(z)
2
1
E3E0
E3E1
u(z)
2
_
. (13.98)
Finally we remark that (9.85) now reads
(13.99) (n, t) = E
1
1
E2E1
E2E0
E0
E1
sn(2K(k) 2nF(
_
E3E1
E3E0
, k)
2t
C
)
2
1
E2E1
E2E0
sn(2K(k) 2nF(
_
E3E1
E3E0
, k)
2t
C
)
2
,
since we have
(13.100) A
(E0,0)
( (n, t)) = A
(E0,0)
( (0, 0)) 2nA
(E0,0)
(
+
) 2[tc(1)].
Chapter 14
The Kacvan Moerbeke
system
In this chapter we will introduce the Kacvan Moerbeke (KM) hierarchy. Using
a commutation (sometimes also called a supersymmetric) approach we will show
that the Kacvan Moerbeke hierarchy is a modied Toda hierarchy precisely in
the way that the modied Kortewegde Vries (mKdV) hierarchy is related to the
Kortewegde Vries (KdV) hierarchy. The connection between the TL hierarchy and
its modied counterpart, the KM hierarchy, is given by a Backlund transformation
(the Miura transformation in case of the KdV and mKdV equation) based on the
factorization of dierence expressions found in Section 11.1. This will allow us to
compute new solutions from known ones. In particular, we will be able to compute
the Nsoliton solution starting from the trivial solution (see Section 13.3) of the
TL equation.
14.1. The Kacvan Moerbeke hierarchy and its
relation to the Toda hierarchy
This section is devoted to the Kacvan Moerbeke hierarchy and its connection with
the Toda hierarchy.
We will suppose that (t) satises the following hypothesis throughout this
chapter.
Hypothesis H. 14.1. Let
(14.1) (t)
(Z) be dierentiable.
Dene the even and odd parts of (t) by
(14.2)
e
(n, t) = (2n, t),
o
(n, t) = (2n + 1, t), (n, t) Z R,
and consider the bounded operators (in
2
(Z))
(14.3) A(t) =
o
(t)S
+
+
e
(t), A(t)
o
(t)S
+
e
(t).
257
258 14. The Kacvan Moerbeke system
As in Section 11.1 we have the factorization
(14.4) H
1
(t) = A(t)
A(t), H
2
(t) = A(t)A(t)
,
with
(14.5) H
k
(t) = a
k
(t)S
+
+a
k
(t)S
+b
k
(t), k = 1, 2,
and
a
1
(t) =
e
(t)
o
(t), b
1
(t) =
e
(t)
2
+
o
(t)
2
,
a
2
(t) =
+
e
(t)
o
(t), b
2
(t) =
e
(t)
2
+
o
(t)
2
. (14.6)
In the sequel it will be important to observe that the transformation (n, t) =
(n + 1, t) implies
(14.7)
e
(n, t) =
o
(n, t),
o
(n, t) =
e
(n + 1, t)
and hence a
1
(t) = a
2
(t),
b
1
(t) = b
2
(t), a
2
(t) = a
+
1
(t),
b
2
(t) = b
+
1
(t). In addition, we
have g
1,j
(t) = g
2,j
(t),
h
1,j
(t) = h
2,j
(t) and g
2,j
(t) = g
+
1,j
(t),
h
2,j
(t) = h
+
1,j
(t).
Using
m
, H
j+1
2
n
=
m
, AH
j
1
A
n
we can express g
2,j
, h
2,j
, j N, in terms
of g
1,j
, h
1,j
, j N. Explicit evaluation with m = n gives (provided we pick the
same summation constant c
1,
= c
2,
c
, 1 r)
(14.8) g
2,j+1
(t) =
o
(t)
2
g
+
1,j
(t) +
e
(t)
2
g
1,j
(t) +h
1,j
(t).
Similarly, for m = n + 1 we obtain
h
2,j+1
(t) h
1,j+1
=
o
(t)
2
_
2(
+
e
(t)
2
g
+
1,j
(t)
e
(t)
2
g
1,j
(t))
+ (h
+
1,j
(t) h
1,j
(t))
_
. (14.9)
By virtue of (14.7) we can even get two more equations
g
1,j+1
(t) =
e
(t)
2
g
2,j
(t) +
o
(t)
2
g
2,j
(t) +h
2,j
(t),
h
2,j+1
(t) h
1,j+1
=
e
(t)
2
_
2(
o
(t)
2
g
2,j
(t)
o
(t)
2
g
2,j
(t))
(h
2,j
(t) h
2,j
(t))
_
. (14.10)
These relations will turn out useful below.
Now we dene matrixvalued operators D(t), Q
2r+2
(t) in
2
(Z, C
2
) as follows,
D(t) =
_
0 A(t)
A(t) 0
_
,
Q
2r+2
(t) =
_
P
1,2r+2
(t) 0
0 P
2,2r+2
(t)
_
, (14.11)
r N
0
. Here P
k,2r+2
(t), k = 1, 2, are dened as in (12.38), that is,
P
k,2r+2
(t) = H
k
(t)
r+1
+
r
j=0
(2a
k
(t)g
k,j
(t)S
+
h
k,j
(t))H
k
(t)
j
+g
k,r+1
,
P
k,2r+2
(t)
Ker(
k
(t)z)
= 2a
k
(t)G
k,r
(z, t)S
+
H
k,r+1
(z, t). (14.12)
The sequences (g
k,j
(n, t))
0jr
, (h
k,j
(n, t))
0jr+1
are dened as in (12.27), and
the polynomials G
k,r
(z, n, t), H
k,r+1
(z, n, t) are dened as in (12.57). Moreover,
14.1. The Kacvan Moerbeke hierarchy 259
we choose the same summation constants in P
1,2r+2
(t) and P
2,2r+2
(t) (i.e., c
1,
=
c
2,
c
e
(t)
o
(t)S
+
e
(t)
o
(t)S
0
0
+
e
(t)
o
(t)S
+
e
(t)
o
(t)S
_
etc. . (14.13)
Analogous to equation (12.22) we now look at
(14.14)
d
dt
D = [Q
2r+2
, D] =
_
0 P
1,2r+2
A
P
2,2r+2
P
2,2r+2
AAP
1,2r+2
0
_
.
Using H
j
1
A
= A
H
j
2
and S
=
1
a
2
(H
2
a
2
S
+
b
2
) we obtain after a little
calculation
A
P
2,2r+2
P
1,2r+2
A
=
r
j=1
+
e
e
_
2
2
e
(g
2,j
g
1,j
) + (h
2,j
h
1,j
)
_
S
+
H
rj
2
+
r
j=1
1
e
_
(h
2,j+1
h
1,j+1
) + 2((a
1
)
2
g
1,j
(a
2
)
2
g
2,j
)
2
o
(h
2,j
h
1,j
) +
2
e
(h
2,j
h
2,j
)
_
H
rj
2
+ (h
2,r+1
h
1,r+1
)
o
(g
2,r+1
g
1,r+1
)S
+
e
(g
2,r+1
g
1,r+1
). (14.15)
Evaluating
n
, (H
1
A
2
)
n
= 0 explicitly gives
(14.16) h
2,j
h
1,j
+ 2
2
e
(g
2,j
g
1,j
) = 0
and the second identity in (14.10) together with (14.16) implies
(14.17)
(h
2,j+1
h
1,j+1
) + 2((a
1
)
2
g
1,j
(a
2
)
2
g
2,j
)
2
o
(h
2,j
h
1,j
) +
2
e
(h
2,j
h
2,j
) = 0.
Hence we obtain
(14.18) P
1,2r+2
A
P
2,2r+2
=
o
(g
2,r+1
g
1,r+1
)S
+
e
(g
2,r+1
g
1,r+1
).
Or, in other words,
(14.19)
d
dt
D(t) [Q
2r+2
(t), D(t)] = 0
is equivalent to
KM
r
() = (KM
r
()
e
, KM
r
()
o
)
=
_
e
e
(g
2,r+1
g
1,r+1
)
o
+
o
(g
2,r+1
g
+
1,r+1
)
_
= 0. (14.20)
Here the dot denotes a derivative with respect to t. One look at the transformation
(14.7) veries that the equations for
o
,
e
are in fact one equation for . More
explicitly, combining g
k,j
, respectively h
k,j
, into one sequence
(14.21)
G
j
(2n) = g
1,j
(n)
G
j
(2n + 1) = g
2,j
(n)
, respectively
H
j
(2n) = h
1,j
(n)
H
j
(2n + 1) = h
2,j
(n)
,
260 14. The Kacvan Moerbeke system
we can rewrite (14.20) as
(14.22) KM
r
() = (G
+
r+1
G
r+1
).
From (12.27) we see that G
j
, H
j
satisfy the recursions
G
0
= 1, H
0
= c
1
,
2G
j+1
H
j
H
j
2(
2
+ (
)
2
)G
j
= 0, 0 j r,
H
j+1
H
j+1
2((
+
)
2
G
++
j
(
)
2
G
j
)
(
2
+ (
)
2
)(H
j
H
j
) = 0, 0 j < r. (14.23)
As in the Toda context (12.42), varying r N
0
yields the Kacvan Moerbeke
hierarchy (KM hierarchy) which we denote by
(14.24) KM
r
() = 0, r N
0
.
Using (14.23) we compute
G
1
=
2
+ (
)
2
+c
1
,
H
1
= 2(
+
)
2
+c
2
,
G
2
= (
+
)
2
+ (
2
+ (
)
2
)
2
+ (
)
2
+c
1
(
2
+ (
)
2
) +c
2
,
H
2
= 2(
+
)
2
((
++
)
2
+ (
+
)
2
+
2
(
)
2
) +c
1
2(
+
)
2
+c
3
,
etc. . (14.25)
and hence
KM
0
() = ((
+
)
2
(
)
2
) = 0,
KM
1
() = ((
+
)
4
(
)
4
+ (
++
)
2
(
+
)
2
+ (
+
)
2
2
(
)
2
(
)
2
(
)
2
) +c
1
()((
+
)
2
(
)
2
) = 0,
etc. . (14.26)
Again the Lax equation (14.19) implies
Theorem 14.2. Let satisfy (H.14.1) and KM() = 0. Then the Lax equation
(14.19) implies the existence of a unitary propagator V
r
(t, s) such that we have
(14.27) D(t) = V
r
(t, s)D(s)V
r
(t, s)
1
, (t, s) R
2
,
and thus all operators D(t), t R, are unitarily equivalent. Clearly, we have
(14.28) V
r
(t, s) =
_
U
1,r
(t, s) 0
0 U
2,r
(t, s)
_
.
And as in Theorem 12.6 we infer
Theorem 14.3. Suppose
0
(R,
(t) = (n)(n, t)
nZ
, (n) +1, 1, n Z.
This result should again be compared with Lemma 1.6 and Lemma 14.9 below.
14.1. The Kacvan Moerbeke hierarchy 261
As in Section 12.2 (cf. (12.45)) we use a tilde to distinguish between inhomo
geneous and homogeneous Kacvan Moerbeke equations, that is,
(14.30)
KM
r
() = KM
r
()
0, 1r
,
with c
e
= 2
e
2
o
(G
+
1,r
(z) G
2,r
(z)) +
e
(H
1,r+1
(z) H
2,r+1
(z)),
d
dt
o
= 2
o
(
+
e
)
2
(G
+
1,r
(z) G
+
2,r
(z))
o
(H
+
1,r+1
(z) H
2,r+1
(z)), (14.31)
r N
0
. One only has to use (14.16) and
(14.32) h
2,j
h
1,j
+ 2
2
o
(g
2,j
g
+
1,j
) = 0,
which follows from (14.16) invoking the transformation (14.7).
The connection between P
k,2r+2
(t), k = 1, 2, and Q
2r+2
(t) is clear from (14.11),
the corresponding connection between H
k
(t), k = 1, 2, and D(t) is provided by the
elementary observation
(14.33) D(t)
2
=
_
H
1
(t) 0
0 H
2
(t)
_
.
Moreover, since we have
(14.34)
d
dt
D(t)
2
= [Q
2r+2
(t), D(t)
2
]
we obtain the implication
(14.35) KM
r
() = 0 TL
r
(a
k
, b
k
) = 0, k = 1, 2.
Using (14.16) and (14.32) this can also be written as
(14.36) TL
r
(a
k
, b
k
) = W
k
KM
r
(), k = 1, 2,
where W
k
(t) denote the matrixvalued dierence expressions
(14.37) W
1
(t) =
_
o
(t)
e
(t)
2
e
(t) 2
o
(t)S
_
, W
2
(t) =
_
o
(t)S
+
+
e
(t)
2
e
(t) 2
o
(t)
_
.
That is, given a solution of the KM
r
equation (14.24) (respectively (14.31)), one
obtains two solutions, (a
1
, b
1
) and (a
2
, b
2
), of the TL
r
equations (12.42) related
to each other by the Miuratype transformations (14.6). Hence we have found a
Backlund transformation relating the Toda and Kacvan Moerbeke systems.
Note that due to (H.14.1), (a
1
, b
1
) and (a
2
, b
2
) both fulll (H.12.1).
In addition, we can dene
(14.38)
1
(n, t) =
e
(n, t)
o
(n, t)
,
2
(n, t) =
o
(n, t)
e
(n + 1, t)
.
This implies
(14.39) a
k
(n, t)
k
(n, t) +
a
k
(n 1, t)
k
(n 1, t)
+b
k
(n, t) = 0
262 14. The Kacvan Moerbeke system
and using (14.8), the rst identity in (14.10), and (12.27) one veries (cf. (13.26))
d
dt
ln
k
(n, t) = 2a
k
(n, t)(g
k,r
(n, t)
k
(n, t) +g
k,r
(n + 1, t)
k
(n, t)
1
)
2b
k
(n + 1, t)g
k,r
(n + 1, t) + (g
k,r+1
(n + 1, t) g
k,r+1
(n, t))
(h
k,r
(n + 1, t) h
k,r
(n, t)). (14.40)
Hence we infer
(14.41) H
k
(t)u
k
(n, t) = 0,
d
dt
u
k
(n, t) = P
k,2r+2
(t)u
k
(n, t)
(in the weak sense, i.e., u
k
is not necessarily square summable), where
u
k
(n, t) = exp
_
_
t
0
(2a
k
(n
0
, x)g
k,r
(n
0
, x)
k
(n
0
, x)
h
k,r
(n
0
, x) +g
k,r+1
(n
0
, x))dx
_n1
m=n0
k
(m, t). (14.42)
Furthermore, explicitly writing out (14.19) shows that if
(14.43) H
k
(t)u
k
(z, n, t) = zu
k
(z, n, t),
d
dt
u
k
(z, n, t) = P
k,2r+2
(t)u
k
(z, n, t)
holds for the solution u
1
(z, n, t) (resp. u
2
(z, n, t)), then it also holds for the solution
u
2
(z, n, t) = A(t)u
1
(z, n, t) (resp. u
1
(z, n, t) = A(t)
u
2
(z, n, t)).
Summarizing,
Theorem 14.5. Suppose satises (H.14.1) and KM
r
() = 0. Then (a
k
, b
k
),
k = 1, 2, satises (H.12.1) and TL
r
(a
k
, b
k
) = 0, k = 1, 2. In addition, if (14.43)
holds for the solution u
1
(z, n, t) (resp. u
2
(z, n, t)), then it also holds for the solution
u
2
(z, n, t) = A(t)u
1
(z, n, t) (resp. u
1
(z, n, t) = A(t)
u
2
(z, n, t)).
Finally, let us extend Lemma 12.16 to the case (H).
Lemma 14.6. Suppose (H) and a(n, t) < 0. Then u
0
(, n) > 0 implies that
the solution u(, n, t) of (12.89) with initial condition u
0
(, n) is positive.
Proof. Shifting H(t) H(t) we can assume = 0. Now use u
0
(0, n) > 0 to
dene
0
(n) by
0,o
(n) =
a(n, 0)u
0
(0, n)
u
0
(0, n + 1)
,
0,e
(n) =
a(n, 0)u
0
(0, n + 1)
u
0
(0, n)
. (14.44)
By Theorem 14.3 we have a corresponding solution (n, t) of the KM hierar
chy and hence (by (14.35)) two solutions a
k
(n, t), b
k
(n, t) of the TL hierarchy.
Since a
1
(n, 0) = a(n, 0) and b
1
(n, 0) = b(n, 0) we infer a
1
(n, t) = a(n, t) and
b
1
(n, t) = b(n, t) by uniqueness (Theorem 12.6). Finally, we conclude u(0, n, t) =
u
0
(0, n
0
)u
1
(n, t) > 0 (with u
1
(n, t) as in (14.42)) again by uniqueness (Theo
rem 12.15).
Another important consequence is
14.2. Kac and van Moerbekes original equations 263
Lemma 14.7. Let (H(0)) and a(n, t) < 0. Suppose u(, n, t) solves (12.89)
and is a minimal positive solution near for one t = t
0
, then this holds for all
t R. In particular, H(t) is critical (resp. subcritical) for all t R if and only
if it is critical (resp. subcritical) for one t = t
0
.
Proof. Since linear independence and positivity is preserved by the system (12.89)
(by (12.91) and Lemma 14.6) H(t) is critical (resp. subcritical) for all t R if
and only if it is critical (resp. subcritical) for one t = t
0
. If H(t) is subcritical,
we note that the characterization (2.70) of minimal solutions is independent of t.
Hence it could only happen that u
+
(, n, t) and u
j=0
(D
_
E
j
)(D +
_
E
j
) =
2r+1
j=0
(D
2
E
j
)
and
(14.47) y
2
=
2r+1
j=0
(w
_
E
j
)(w +
_
E
j
) =
2r+1
j=0
(w
2
E
j
).
Here
_
E
j
are the band edges of the spectrum of D, which appear in pairs since
D and D are unitarily equivalent (see Lemma 14.9 below).
We note that the curve (14.47) becomes singular if and only if E
j
= 0 for some
0 j 2r +1. Since we have 0 E
0
< E
1
< < E
2r+1
this happens if and only
if E
0
= 0.
We omit further details at this point since we will show in the following sections
how solutions of the KM hierarchy can be computed from the corresponding ones
of the TL hierarchy.
14.2. Kac and van Moerbekes original equations
The Kacvan Moerbeke equation is originally given by
(14.48)
d
dt
R(n, t) =
1
2
_
e
R(n1,t)
e
R(n+1,t)
_
.
The transformation
(14.49) (n, t) =
1
2
e
R(n,t)/2
,
yields the form most convenient for our purpose
(14.50) (n, t) = (n, t)
_
(n + 1, t)
2
(n 1, t)
2
_
,
264 14. The Kacvan Moerbeke system
which is precisely KM
0
() = 0. The form obtained via the transformation c(n, t) =
2(n, t)
2
is called Langmuir lattice,
(14.51) c(n, t) = c(n, t)
_
c(n + 1, t) c(n 1, t)
_
,
and is used for modeling Langmuir oscillations in plasmas.
14.3. Spectral theory for supersymmetric
Diractype dierence operators
In this section we briey study spectral properties of selfadjoint
2
(Z, C
2
) realiza
tions associated with supersymmetric Diractype dierence expressions.
Given
o
S
+
e
,
with
e
(n) = (2n),
o
(n) = (2n + 1). We denote by D the bounded selfadjoint
supersymmetric Dirac operator
(14.53)
D :
2
(Z, C
2
)
2
(Z, C
2
)
f
_
0 A
A 0
_
f
and by U the corresponding unitary involution
(14.54)
U :
2
(Z, C
2
)
2
(Z, C
2
)
f
_
0 1l
1l 0
_
f
.
Theorem 14.8. Suppose
(H
2
w
2
)
1
A(H
1
w
2
)
1
w(H
2
w
2
)
1
_
, w
2
C(D),
where H
1
= A
A, H
2
= AA
.
Proof. The spectrum of D follows from the spectral mapping theorem since D
2
=
H
1
H
2
and since D and D are unitarily equivalent (see the lemma below). The
formula for the resolvent is easily vindicated using A(H
1
w
2
)
1
= (H
2
w
2
)
1
A
and (H
1
w
2
)
1
A
= A
(H
2
w
2
)
1
(since AH
1
= H
2
A and H
1
A
= A
H
2
) (see
also Theorem 11.1).
Hence one can reduce the spectral analysis of supersymmetric Dirac operators
D to that of H
1
and H
2
. Moreover, since H
1
[
Ker(H1)
and H
2
[
Ker(H2)
are unitarily
equivalent by Theorem 11.1, a complete spectral analysis of D in terms of that of
H
1
and Ker(H
2
) can be given.
The analog of Lemma 1.6 reads
Lemma 14.9. Suppose satises (H.14.1) and introduce
R
(Z) by
(14.57)
in
2
(Z, C
2
) as in (14.53) with replaced by
. Then D and D
are
unitarily equivalent, that is, there exists a unitary operator U
in
2
(Z, C
2
) such
that
(14.58) D = U
U
1
.
Especially taking (n) = 1 shows that D and D are unitarily equivalent.
Proof. U
is explicitly represented by
(14.59) U
=
_
U
1,
0
0 U
2,
_
, U
k,
=
k
(n)
m,n
m,nZ
, k = 1, 2,
1
(n + 1)
2
(n) = (2n + 1),
1
(n)
2
(n) = (2n), n Z.
14.4. Associated solutions
In Theorem 14.5 we saw, that from one solution of KM
r
() = 0 we can get two
solutions (a
1
, b
1
), (a
2
, b
2
) of TL
r
(a, b). In this section we want to invert this process
(cf. Section 11.2).
Suppose (a
1
, b
1
) satises (H.12.1), a
1
(n, t) < 0 and TL
r
(a
1
, b
1
) = 0. Suppose
H
1
> 0 and let u
1
(n, t) > 0 solve
(14.60) H
1
(t)u
1
(n, t) = 0,
d
dt
u
1
(n, t) = P
1,2r+2
(t)u
1
(n, t)
(cf. Lemma 12.16). This implies that
1
(n, t) = u
1
(n + 1, t)/u
1
(n, t) fullls
(14.61) a
1
(n, t)
1
(n, t) +
a
1
(n 1, t)
1
(n 1, t)
= b
1
(n, t),
and
d
dt
ln
1
(n, t) = 2a
1
(n, t)(g
1,r
(n, t)
1
(n, t) +g
1,r
(n + 1, t)
1
(n, t)
1
)
+ 2b
1
(n + 1, t)g
1,r
(n + 1, t) + (g
1,r+1
(n + 1, t) g
1,r+1
(n, t))
(h
1,r
(n + 1, t) h
1,r
(n, t)). (14.62)
Now dene
(14.63)
o
(n) =
a
1
(n, t)
1
(n, t)
,
e
(n, t) =
_
a
1
(n, t)
1
(n, t).
Then, using (12.40), (12.27), and (14.8) a straightforward calculation shows that
the sequence
(14.64) (n, t) =
_
e
(m, t) for n = 2m
o
(m, t) for n = 2m+ 1
= (1)
n+1
_
a
1
(n, t)
1
(n, t)
(1)
n
,
fullls (H.14.1) and KM
r
() = 0. Hence by (14.35)
(14.65) a
2
(n, t) =
e
(n + 1, t)
o
(n, t), b
2
(n, t) =
e
(n, t)
2
+
o
(n, t)
2
satisfy TL
r
(a
2
, b
2
) = 0.
266 14. The Kacvan Moerbeke system
Since we already know some solutions of TL
r
(a
1
, b
1
) = 0, we can illustrate this
process. Taking a nitegap solution
a
1
(n, t) = a
(z
1
(n + 1, t))(z
1
(n 1, t))
(z
1
(n, t))
2
,
b
1
(n, t) =
b +
g
j=1
c
j
(g)
w
j
ln
_
(w +z
1
(n, t))
(w +z
1
(n 1, t))
_
w=0
,
1
(n, t) =
(z
1
(n 1, t))
(z
1
(n + 1, t))
(z
1
(p
1
, n + 1, t))
(z
1
(p
1
, n, t))
exp
_
_
p1
p0
+,
_
, (14.66)
with (p
1
) E
0
yields as in Section 11.4
o
(n, t) =
a
1
(n, t)
1
(n, t)
=
a
(z
1
(n + 1, t))
(z
1
(n, t))
(z
1
(p
1
, n, t))
(z
1
(p
1
, n + 1, t))
exp
_
1
2
_
p1
p0
+,
_
,
e
(n, t) =
_
a
1
(n, t)
1
(n, t)
=
a
(z
1
(n 1, t))
(z
1
(n, t))
(z
1
(p
1
, n + 1, t))
(z
1
(p
1
, n, t))
exp
_
1
2
_
p1
p0
+,
_
, (14.67)
and
a
2
(n, t) = a
(z
2
(n + 1, t))(z
2
(n 1, t))
(z
2
(n, t))
2
,
b
2
(n, t) =
b +
g
j=1
c
j
(g)
w
j
ln
_
(w +z
2
(n, t))
(w +z
2
(n 1, t))
_
w=0
,
2
(n, t) =
(z
2
(n 1, t))
(z
2
(n + 1, t))
(z
2
(p
1
, n + 1, t))
(z
2
(p
1
, n, t))
exp
_
_
p
1
p0
+,
_
, (14.68)
where z
2
(n) = z
1
(p
1
, n + 1) respectively z
2
(p
1
, n) = z
1
(n). We remark that if we
interchange the role of a
1
and a
2
and use p
2
_
[ b[
_
[ b[
2
4a
2
and hence
(14.70) (n, t) =
(1)
n+1
2
_
[ b[
_
[ b[
2
4a
2
.
14.5. Nsoliton solutions on arbitrary background 267
14.5. Nsoliton solutions on arbitrary background
In this section we want to extend the calculations of Section 14.4 and of Chapter 11.
Suppose (a, b) satises (H.12.1), a(n, t) < 0, and TL
r
(a, b) = 0. Suppose
1
(H), let u
(
1
, n, t) > 0 be the minimal positive solutions of (12.89) found
in Lemma 14.7, and set
(14.71) u
1
(
1
, n, t) =
1 +
1
2
u
+
(
1
, n, t) +
1
1
2
u
(
1
, n, t).
Note that the dependence on
1
will drop out in what follows if u
+
(
1
, n, t) and
u
(
1
, n, t) are linearly dependent (for one and hence for all t). Now dene
(14.72)
1,o
(n, t) =
a(n, t)
1
(
1
, n, t)
,
1,e
(n, t) =
_
a(n, t)
1
(
1
, n, t),
where
1
(
1
, n, t) = u
1
(
1
, n + 1, t)/u
1
(
1
, n, t).
Then, proceeding as in the proof of Lemma 14.6 shows that the sequence
(14.73)
1
(n, t) =
_
1,e
(m, t) for n = 2m
1,o
(m, t) for n = 2m+ 1
,
fullls (H.14.1) and KM
r
() = 0. Hence by (14.35)
(14.74)
a
1
(n, t) =
1,e
(n + 1, t)
1,o
(n, t), b
1
(n, t) =
1,e
(n, t)
2
+
1,o
(n, t)
2
satisfy TL
r
(a
1
, b
1
) = 0.
We summarize this result in our rst main theorem.
Theorem 14.10. Suppose (a, b) satises (H.12.1) and TL
r
(a, b) = 0. Pick
1
(H),
1
[1, 1] and let u
(
1
, n, t) be the minimal positive solutions of (12.89).
Then the sequences
a
1
(n, t) =
a(n, t)u
1
(
1
, n, t)
u
1
(
1
, n + 1, t)
(14.75)
with
(14.76) u
1
(
1
, n + 1, t) =
1 +
1
2
u
+
(
1
, n, t) +
1
1
2
u
(
1
, n, t),
satisfy TL
r
(a
1
, b
1
) = 0. Here
1
to express the quantities g
1,j
, h
1,j
in terms
of g
j
, h
j
.
(ii). We have omitted the requirement a(n, t) < 0 since the formulas for a
1
, b
1
are actually independent of the sign of a(n, t). In addition, we could even allow
1
(H). However,
1,e
(n, t) and
1,o
(n, t) would be purely imaginary in this
case.
268 14. The Kacvan Moerbeke system
Iterating this procedure (cf. Theorem 11.10) gives
Theorem 14.12. Let a(t), b(t) satisfy (H.12.1) and TL
r
(a, b) = 0. Let H(t) be the
corresponding Jacobi operators and choose
(14.78)
N
< <
2
<
1
(H),
[1, 1], 1 N, N N.
Suppose u
_
C
n
(U
1,...,
N
(t))C
n+2
(U
1,...,
N
(t))
C
n+1
(U
1,...,
N
(t))
,
b
1,...,
N
(n, t) = b(n, t) +
a(n, t)
D
n
(U
1,...,
N
(t))
C
n+1
(U
1,...,
N
(t))
(14.79)
satises TL
r
(a
1,...,
N
, b
1,...,
N
) = 0. Here C
n
denotes the ndimensional Casora
tian
C
n
(u
1
, . . . , u
N
) = det (u
i
(n +j 1))
1i,jN
,
D
n
(u
1
, . . . , u
N
) = det
_
u
i
(n), j = 1
u
i
(n +j), j > 1
_
1i,jN
, (14.80)
and (U
1,...,
N
(t)) = (u
1
1
(t), . . . , u
N
N
(t)) with
(14.81) u
(n, t) =
1 +
2
u
+
(
, n, t) + (1)
+1
1
2
u
, n, t).
Moreover, for 1 N, <
,
(14.82) u
,1,...,
(, n, t) =
j=0
_
a(n +j, t)C
n
(U
1,...,
(t), u
(, t))
_
C
n
(U
1,...,
(t))C
n+1
(U
1,...,
(t))
,
are the minimal solutions of
1,...,
(, n, t) = P
1,...,
,2r+2
(t)u
,1,...,
(, n, t).
Dening
1,...,
N
,o
(n, t) =
a(n, t)
C
n+2
(U
1,...,
N1
(t))C
n
(U
1,...,
N
(t))
C
n+1
(U
1,...,
N1
(t))C
n+1
(U
1,...,
N
(t))
,
1,...,
N
,e
(n, t) =
a(n +N 1, t)
C
n
(U
1,...,
N1
(t))C
n+1
(U
1,...,
N
(t))
C
n+1
(U
1,...,
N1
(t))C
n
(U
1,...,
N
(t))
, (14.84)
the corresponding sequence
1,...,
N
(n) solves KM
r
(
1,...,
N
) = 0.
Clearly, if we drop the requirement (H) the solution u
1
(
1
, n, t) used to
perform the factorization will no longer be positive. Hence the sequences a
1
(n, t),
b
1
(n, t) can be complex valued and singular. Nevertheless there are two situations
where a second factorization step produces again realvalued nonsingular solutions.
14.5. Nsoliton solutions on arbitrary background 269
Firstly we perform two steps with
1,2
in the same spectral gap of H(0) (com
pare Section 11.8).
Theorem 14.13. Suppose (a, b) satises (H.12.1) and TL
r
(a, b) = 0. Pick
1,2
,
1,2
1 and let
1,2
lie in the same spectral gap of H(0) ( (
1
,
1
) ,= (
2
,
2
)
to make sure we get something new). Then the sequences
a
1,2
(n, t) = a(n, t)
W
1,2
(n 1, t)W
1,2
(n + 1, t)
W
1,2
(n, t)
2
,
b
1,2
(n, t) = b(n, t)
a(n, t)u
1
(
1
, n, t)u
2
(
2
, n + 1, t)
W
1,2
(n, t)
, (14.85)
where
(14.86) W
1,2
(n, t) =
_
_
_
Wn(u
1
(1,t),u
2
(2,t))
21
,
1
,=
2
n
m=1
u
1
(
1
, m, t)
2
, (
1
,
1
) = (
2
,
2
)
,
are realvalued nonsingular solutions of TL
r
(a
1,2
, b
1,2
) = 0. In addition, the
sequence
(14.87)
W
1,2
(n, t)u(z, n, t)
1
z1
u
2
(
2
, n, t)W
n
(u
1
(
1
, t), u(z, t))
_
W
1,2
(n 1, t)W
1,2
(n, t)
,
satises H
1,2
(t)u = zu, d/dt u = P
1,2,2r+2
(t)u (weakly).
Proof. Theorem 4.19 implies W
1,2
(n, t)W
1,2
(n + 1, t) > 0 and hence the se
quences a
1,2
(t), b
1,2
(t) satisfy (H.12.1). The rest follows from the previous the
orem (with N = 2) as follows. Replace
1
by z (
1
,
1
+) and observe that
a
1,2
(n, t), b
1,2
(n, t) and a
1,2
(n, t),
b
1,2
(n, t) are meromorphic with respect
to z. From the algebraic structure we have simply performed two single commu
tation steps. Hence, provided Theorem 14.10 applies to this more general setting
of meromorphic solutions, we can conclude that our claims hold except for a dis
crete set with respect to z where the intermediate operators are illdened due to
singularities of the coecients. However, the proof of Theorem 14.10 uses these
intermediate operators and in order to see that Theorem 14.10 still holds, one has
to resort to the direct algebraic proof outlined in Remark 14.11(i). Continuity with
respect to z takes care of the remaining points.
Secondly, we consider again two commutation steps but now with
1
=
2
(compare Section 11.6).
Theorem 14.14. Suppose (a, b) satises (H.12.1) and TL
r
(a, b) = 0. Pick
1
in
a spectral gap of H(0) and
1
[u
(
1
)
2
, ) . Then the sequences
a
1
(n, t) = a(n, t)
_
c
1
(
1
, n 1, t)c
1
(
1
, n + 1, t)
c
1
(
1
, n, t)
,
b
1
(n, t) = b(n, t)
a(n, t)u
(
1
, n, t)u
(
1
, n + 1, t)
c
1
(
1
, n, t)
. (14.88)
270 14. The Kacvan Moerbeke system
satisfy TL
r
(a
1
, b
1
) = 0, where
(14.89) c
1
(
1
, n, t) =
1
1
+
n
m=
u
(
1
, m, t)
2
.
In addition, the sequence
(14.90)
c
1
(
1
, n, t)u(z, n, t)
1
z1
u
(
1
, n, t)W
n
(u
(
1
, t), u(z, t))
_
c
1
(
1
, n 1, t)c
1
(
1
, n, t)
,
satises H
1
(t)u = zu, d/dt u = P
1,2r+2
(t)u (weakly).
Proof. Following Section 11.5 we can obtain the double commutation method from
two single commutation steps. We pick
1
= 1 for the rst factorization. Consid
ering A
1
u
(z, n + 1, t)/(z
1
) and performing the limit z
1
shows that
(14.91) v(
1
, n, t) =
c
1
(
1
, n, t)
_
a(n, t)u
(
1
, n, t)u
(
1
, n + 1, t)
is a solution of the new (singular) operator which can be used to perform a second
factorization. The resulting operator is associated with a
1
, b
1
. Now argue as
before.
Again we point out that one can also prove this theorem directly as follows.
Without restriction we choose
1
= 0. Then one computes
d
dt
c
1
(0, n, t)=2a(n, t)
2
_
g
r1
(n + 1, t)u
(0, n, t)
2
+g
r1
(n, t)u
(0, n + 1, t)
2
_
+ 2h
r1
(n, t)a(n, t)u
(0, n, t)u
(0, n + 1, t) (14.92)
and it remains to relate g
1,j
, h
1,j
and g
j
, h
j
. Since these quantities arise as coef
cients of the Neumann expansion of the respective Green functions it suces to
relate the Green functions of H
1
and H. This can be done using (11.108).
Iterating this procedure (cf. Theorem 11.27) gives
Theorem 14.15. Let a(n, t), b(n, t) satisfy (H.12.1). Suppose TL
r
(a, b) = 0 and
let H(t) be the corresponding Jacobi operators. Choose
j
(H) and
j
[u
(
j
)
2
, ) , 1 j N, and assume
(14.93)
d
dt
u
(
j
, n, t) = P
2r+2
(t)u
(
j
, n, t).
We dene the following matrices
C
N
(n, t) =
_
ij
i
+
n
m=
u
(
i
, m, t)u
(
j
, m, t)
_
1i,jN
,
D
N
(n, t) =
_
_
_
_
C
N
(n, t)
i,j
, i,jN
u
(
j
, n 1, t), jN,i=N+1
u
(
i
, n, t), iN,j=N+1
0, i=j=N+1
_
_
_
_
1i,jN+1
. (14.94)
14.5. Nsoliton solutions on arbitrary background 271
Then the sequences
a
1,...,
N
(n, t) = a(n, t)
_
det C
N
(n 1, t) det C
N
(n + 1, t)
det C
N
(n, t)
,
b
1,...,
N
(n, t) = b(n, t)
a(n, t)
det D
N
(n + 1, t)
det C
N
(n, t)
(14.95)
satisfy TL
r
(a
1,...,
N
, b
1,...,
N
) = 0. Moreover, (1 N)
(14.96) u
1,...,
(z, n, t) =
det U
N
(u(z, n, t))
_
det C
(n 1, t) det C
(n, t)
satises
(14.97) H
1,...,
N
u =
j
u,
d
dt
u = P
1,...,
N
,2r+2
(t)u
(in the weak sense) if u(z, n, t) satises (12.89).
Remark 14.16. In the case r = 0 we even obtain
(14.98)
d
dt
c
1
(
1
, n, t) = 2a(n, t)u
(
1
, n, t)u
(
1
, n + 1, t)
and hence
(14.99) b
1
(n, t) = b(n, t) +
1
2
d
dt
ln
c
1
(
1
, n, t)
c
1
(
1
, n 1, t)
,
where c
1
(
1
, n, t) =
1
1
+
n
m=
u
(
1
, m, t). Or, by induction,
(14.100) b
1,...,
N
(n, t) = b(n, t) +
1
2
d
dt
ln
det C
N
(n, t)
det C
N
(n 1, t)
.
We conclude this section with an example; the Nsoliton solution of the TL
and KM hierarchies. We take the constant solution of the Toda hierarchy
(14.101) a
0
(n, t) =
1
2
, b
0
(n, t) = 0,
as our background. Let H
0
, P
0,2r+2
denote the associated Lax pair and recall
(14.102)
H
0
(t)u
0,
(z, n, t) = zu
0,
(z, n, t),
d
dt
u
0,
(z, n, t) = P
0,2r+2
(t)u
0,
(z, n, t),
where
(14.103) u
0,
(z, n, t) = k
n
exp
_
r
(k)t
2
_
, k = z
_
z
2
1, [k[ 1,
and
(14.104)
r
(k) = 2(kG
0,r
(z) H
r+1,0
(z)) = (k k
1
)G
0,r
(z).
Then the Nsoliton solution of the Toda hierarchy is given by
a
0,1,...,
N
(n, t) =
_
C
n
(U
1,...,
N
(t))C
n+2
(U
1,...,
N
(t))
2C
n+1
(U
1,...,
N
(t))
,
b
0,1,...,
N
(n, t) =
D
n
(U
1,...,
N
(t))
2C
n+1
(U
1,...,
N
(t))
, (14.105)
272 14. The Kacvan Moerbeke system
where (U
0,1,...,
N
(t)) = (u
1
1
(t), . . . , u
N
N
(t)) with
(14.106) u
j
0,j
(n, t) = k
n
j
+(1)
j+1
1
j
1 +
j
exp(
r
(k
j
)t)k
n
j
, k
j
=
j
2
j
1.
The corresponding Nsoliton solution
1,...,
N
(n) of the Kacvan Moerbeke hier
archy reads
0,1,...,
N
,o
(n, t) =
C
n+2
(U
0,1,...,
N1
(t))C
n
(U
0,1,...,
N
(t))
2C
n+1
(U
0,1,...,
N1
(t))C
n+1
(U
0,1,...,
N
(t))
,
0,1,...,
N
,e
(n, t) =
C
n
(U
0,1,...,
N1
(t))C
n+1
(U
0,1,...,
N
(t))
2C
n+1
(U
0,1,...,
N1
(t))C
n
(U
0,1,...,
N
(t))
. (14.107)
Introducing the time dependent norming constants
(14.108)
j
(t) =
j
exp(
r
(k
j
)t)
we obtain the following alternate expression for the Nsoliton solution of the Toda
hierarchy
a
0,1,...,
N
(n, t) =
_
det C
N
0
(n 1, t) det C
N
0
(n + 1, t)
2 det C
N
0
(n, t)
,
b
0,1,...,
N
(n, t) =
det D
N
0
(n + 1, t)
2 det C
N
0
(n, t)
, (14.109)
where
C
N
0
(n, t) =
_
rs
r
(t)
+
(k
r
k
s
)
n
1 k
r
k
s
_
1r,sN
,
D
N
0
(n, t) =
_
_
_
_
C
N
0
(n, t)
r,s
, r,sN
k
1n
s
, sN,r=N+1
k
n
r
, rN,s=N+1
0, r=s=N+1
_
_
_
_
1r,sN+1
. (14.110)
The sequences a
0,1,...,
N
, b
0,1,...,
N
coincide with a
0,1,...,
N
, b
0,1,...,
N
provided
(cf. Remark 11.30)
(14.111)
j
=
_
1
j
1 +
j
_
1
[k
j
[
1N
N
=1
[1 k
j
k
N
=1
=j
[k
j
k
[
, 1 j N.
Notes on literature
Chapter 12. The Toda lattice was rst introduced by Toda in [233]. For further material
see his selected papers [232] or his monograph [230].
The Toda hierarchy was rst introduced in [237] and [238] (see also [236]). Our intro
duction uses a recursive approach for the standard Lax formalism ([164]) which was rst
advocated by Alber [11], Jacobi [137], McKean [169], and Mumford [178] (Sect. III a). It
follows essentially the monograph [35] (with some extensions). An existence and unique
ness theorem for the semi innite case seems to appear rst in [65], the form stated
here is taken from [220]. The BurchnallChaundy polynomial in the case of commuting
dierential expressions was rst obtained in [36], [37] (cf. also [236]).
Further references I refer to a few standard monographs such as [78], [79] [184], and
[190].
Chapter 13. Section 1 and 2 closely follow [35] with several additions made. The Baker
Akhiezer function, the fundamental object of our approach, goes back to the work of
Baker [23], Burchnall and Chaundy [36], [37], and Akhiezer [8]. The modern approach
was initiated by Its and Matveev [136] in connection with the Kortewegde Vries equation
and further developed into a powerful machinery by Krichever (see, e.g., [156][158])
and others. Gesztesy and Holden are also working on a monograph on hierarchies of
soliton equations and their algebrogeometric solutions [103]. We refer, in particular, to
the extensive treatments in [25], [74], [75], [76], [168], [177], and [184]. In the special
context of the Toda equations we refer to [3], [75], [76], [156], [159], [168], [170], and
[175].
The periodic case was rst investigated by Kac and van Moerbeke [143], [144]. Fur
ther references for the periodic case are [3], [33], [53], [56], [146], [168], and [173][175].
In the case of the Toda lattice (r = 0) the formulas for the quasi periodic nite gap
case in terms of Riemann theta functions seem due to [210]. However, some norming
constants and the fact
b =
d (see Remark 13.9) are not established there.
In case of the Toda equation on the half line a dierent approach using the spectral
measure of H+ is introduced and extended in [28] [32] (see also [15]). It turns out
that the time evolution of the spectral measure can be computed explicitly as the Freud
transform of the initial measure.
The inverse scattering transform for the case r = 0 has been formally developed by
Flaschka [85] (see also [230] and [80] for the case of rapidly decaying sequences) for the
273
274 Notes on literature
Toda lattice. In addition, Flaschka also worked out the inverse procedure in the reection
less case (i.e., R(k, t) = 0). His formulas clearly apply to the entire Toda hierarchy upon
using the t dependence of the norming constants given in (13.72) and coincide with those
obtained using the double commutation method (i.e., the last example in Section 14.5).
All results from the fourth section are taken from [219] and [224].
For additional results where the Toda equation is driven from one end by a particle
see [63], [64] and also [67]. For results on a continuum limit of the Toda lattice see [59].
For connections of the RiemannHilbert problem and the inverse scattering method and
applications to the Toda lattice see [61] and [62]. In addition, there are also very nice
lecture notes by Deift [57] which give an introduction to the RiemannHilbert problem
and its applications.
Chapter 14. The Kacvan Moerbeke equation has been rst introduced in [142]. A con
nection between the Kacvan Moerbeke and Toda systems was already known to Henon
in 1973. The B acklund transformation connecting the Toda and the Kacvan Moerbeke
equations has rst been considered in [234] and [239] (see also [56], [151], [152]). The re
sults presented here are mainly taken from [35], [117] and [220]. An alternative approach
to the modied Toda hierarchy, using the discrete analog of the formal pseudo dierential
calculus, can be found in [161], Ch. 4.
For applications of the Kacvan Moerbeke equation I refer to(e.g.) [25] (Ch. 8), [55],
[66], [130], [159], [183], or [246]
Appendix A
Compact Riemann surfaces
a review
The facts presented in this section can be found in almost any book on Riemann
surfaces and their theta functions (cf., e.g. [74], [81], [82], [86], [154], [178], [181]).
However, no book contains all results needed. Hence, in order to save you the
tedious work of looking everything up from dierent sources and then translate
results from one notation to another, I have included this appendix. In addition, I
have worked out some facts on hyperelliptic surfaces in more detail than you will
nd them in the references. Most of the time we will closely follow [81] in the rst
sections.
A.1. Basic notation
Let M be a compact Riemann surface (i.e., a compact, second countable, con
nected Hausdor space together with a holomorphic structure) of genus g N.
We restrict ourselves to g 1 since most of our denitions only make sense in this
case. We will think of our Riemann surface M often as its fundamental polygon
(e.g., for g = 2):
275
276 A. Compact Riemann surfaces a review
@
@@R
?
@
@ @I
6
M
b
1
1
a
1
1
b
1
a
1
b
1
2
a
1
2
b
2
a
2
We denote the closure of the polygon by
M and its boundary by
M =
g
=1
(a
a
1
b
1
), where a
, b
1g
are (xed) representatives of a canonical
homology basis for M. We get M from
M by identifying corresponding points,
that is, M =
M/ .
We denote the set of meromorphic functions on M by /(M), the set of
meromorphic dierentials by /
1
(M), and the set of holomorphic dieren
tials by 1
1
(M). If we expand f /(M) respectively /
1
(M) in a local chart
(U, z) centered at p M we get (z(p) = 0)
(A.1) f =
=m
c
, respectively =
_
=m
c
_
dz, c
m
,= 0, m Z.
The number m is independent of the chart chosen and so we can dene ord
p
f = m
respectively ord
p
= m. The number m is called the order of f, respectively , at
p. We will also need the residue res
p
= c
1
, which is welldened and satises
(A.2)
pM
res
p
= 0.
The above sum is nite since can only have nitely many poles (M is compact!).
Next, we dene the branch number b
f
(p) = ord
p
(f f(p)) 1 if f(p) ,=
and b
f
(p) = ord
p
(f) 1 if f(p) = . There is a connection between the total
branching number B =
pM
b
f
(p) (the sum is again nite), the degree n of f
(A.3) n = degree(f) =
pf
1
(c)
(1 +b
f
(p)), c C
(we will show that n is independent of c), and the genus g, the RiemannHurwitz
relation
(A.4) 2g 2 = B 2n.
We omit the proof which is not dicult but involves a lot of denitions which are
of no further use for our purpose. (Triangulate C such that each pole and
each zero of f is a vertex. Lift the triangulation to M via f and calculate the Euler
characteristic . . . )
A.2. Abelian dierentials 277
A.2. Abelian dierentials
Now consider two closed dierentials , on M and let us try to compute
(A.5)
__
M
.
We dene
(A.6)
A(p) =
_
p
p0
p
0
, p
M,
where the hat indicates that we regard as a dierential on to
M.
A is well
dened since
M is simply connected and d = 0 (of course we require the path of
integration to lie in
M). Using Stokes theorem we obtain
__
M
=
__
M
d
A =
__
M
d(
A) =
_
A
=
g
=1
_
_
a
A+
_
b
A+
_
a
1
A+
_
b
1
A
_
. (A.7)
Now, what can we say about
A(p) and
A( p) if p, p are corresponding points on a
,
a
1
@
@ @I
6
a
1
p
0
p
p
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ...................................
...................................................................................................................................................
Thus we have
A( p) =
A(p) +
_
b
(the parts on a
and a
1
cancel). A similar
argument for corresponding points p, p on b
, b
1
yields
A( p) =
A(p)
_
a
and
we nally get
(A.8)
__
M
=
g
=1
__
a
_
b
_
a
_
b
_
.
We will mainly be interested in meromorphic dierentials /
1
(M), also
called abelian dierentials. We call of the rst kind if is holomorphic, of
the second kind if all residues vanish, and of the third kind else.
Unfortunately, formula (A.8) makes only sense in the special case of holomor
phic dierentials. Nevertheless there is one thing we can do. Let 1
1
(M),
/
1
(M) and assume that has no poles on the boundary which can al
ways be achieved by slightly (continuously) deforming the (representatives of the)
homology basis. The residue theorem now yields
(A.9) 2i
pM
res
p
A =
_
A =
g
=1
__
a
_
b
_
a
_
b
_
.
278 A. Compact Riemann surfaces a review
For 1
1
(M) we dene
(A.10) 
2
= i
__
M
= i
g
=1
__
a
_
b
_
a
_
b
_
,
(the bar denotes complex conjugation) which is indeed a norm (in local coordinates
we get: i = 2[[
2
dx dy with = dz, z = x + iy).
Therefore we conclude that 0 if and only if (e.g.) all the aperiods are
zero and further that the map from the space of holomorphic dierentials 1
1
(M)
(which is gdimensional) to the aperiods (
= C
g
) is an isomorphism. Thus we may
choose a basis
j
g
j=1
for 1
1
(M) fullling
(A.11)
_
a
k
j
=
jk
, j, k = 1, . . . , g.
For the bperiods we get
(A.12)
_
b
k
j
=
j,k
, j, k = 1, . . . , g,
with
j,k
C. Now (A.9) implies
(A.13)
j,k
=
k,j
and (for =
g
=1
c
1
1
(M)0 with c
C)
0 < 
2
= i
j,k
c
j
c
k
__
a
j
_
b
_
b
j
_
a
k
_
= 2
j,k
c
j
c
k
Im(
j,k
). (A.14)
Especially for c
=
j
we have Im(
j,j
) > 0. Thus is symmetric with positive
denite imaginary part.
We will now ensure the existence of meromorphic dierentials (without proof,
cf. [81], II.5.3):
Theorem A.1. For every set (p
j
, (c
j,
)
1nj
)
1jm
with distinct p
j
M, c
j,
C, and m, n
j
N we can nd a meromorphic dierential with poles of the form
(A.15) = (
nj
=1
c
j,
z
j
+O(1))dz
j
,
near each p
j
in a given chart (U
j
, z
j
) centered at p
j
, 1 j m, and is holomorphic
elsewhere, if and only if
(A.16)
pM
res
p
=
m
j=1
c
j,1
= 0.
Hence, for a meromorphic dierential we can prescribe the poles and the Lau
rent principal parts at each pole with the only restriction, that the sum of the
residues has to be zero. Two dierentials constructed in that way may only dier
by a holomorphic dierential. From now on we will assume that no poles lie on
A.3. Divisors and the RiemannRoch theorem 279
the boundary otherwise just deform the homology basis so that we can make it
unique by the following normalization
(A.17)
_
a
= 0, = 1, . . . , g.
We are interested in some special cases. First let
p,n
be holomorphic in Mp
and let (U, z) be a chart centered at p
(A.18)
p,n
= (
1
z
2+n
+O(1))dz,
_
a
p,n
= 0, = 1, . . . , g, n N
0
.
This denition is obviously depending on the chart around p. The bperiods follow
easily from (A.9) with =
and =
p,n
(A.19)
_
b
p,n
= 2i
pM
res
p
A
p,n
=
2i
(n + 1)!
_
d
dz
_
n
(z)
z=0
,
where
pq
= 1 ord
q
pq
= 1
res
p
pq
= +1 res
q
pq
= 1
,
_
a
pq
= 0, = 1, . . . , g.
This denition depends on the specic homology basis chosen. Because if we de
form, for example, a
pq
. The
bperiods again follow from (A.9) with =
and =
pq
(A.21)
_
b
pq
= 2i
pM
res
p
A
pq
= 2i(
A(p)
A(q)) = 2i
_
p
q
.
Recall that the path of integration in the last integral lies in
M.
A.3. Divisors and the RiemannRoch theorem
This section is also valid for g = 0. A divisor T is a map
(A.22)
T : M Z
p T(p)
being nonzero only for a nite number of points in M. The set of all divisors Div(M)
becomes a module over the ring Z by dening each operation in the natural way
(i.e., pointwise). The point divisors T
p
, p M,
(A.23) T
p
(p) = 1, T
p
(q) = 0, q Mp
form a basis for Div(M). We can also establish a group homomorphism by assigning
each divisor a degree
(A.24)
deg : Div(M) Z
T deg(T) =
pM
T(p)
(the sum is nite and thus welldened). We associate a divisor with every nonzero
meromorphic function and with every nonzero meromorphic dierential on M
(A.25)
( ) : /(M)0 Div(M)
f (f)
,
( ) : /
1
(M)0 Div(M)
()
,
280 A. Compact Riemann surfaces a review
where (f)(p) = ord
p
f and ()(p) = ord
p
.
It is easy to verify the following formulas
(A.26) (f
1
f
2
) = (f
1
) + (f
2
), (
1
f
1
) = (f
1
), f
1
, f
2
/(M)0,
implying (f
1
) = (f
2
) (f
1
/f
2
) = 0 f
1
= c f
2
, c C since the only holomorphic
functions on a compact Riemann surface are constants. Similarly,
(A.27) (f
1
) = (f) + (
1
), (
2
) = (
1
) (
2
),
1
,
2
/
1
(M)0
implying (
1
) = (
2
) (
1
/
2
) = 0
1
= c
2
, c C.
We claim
deg((f)) = 0, f /(M)0,
deg(()) = 2g 2, /
1
(M)0. (A.28)
To obtain the rst equation observe that deg((f)) =
pM
res
p
df/f = 0. This
implies that f attains every value exactly n times for a certain integer n called the
degree of f (counting multiplicities). In fact, if 0 is attained n times the same is
true for (otherwise deg((f)) would not be zero). Now consider the function f c
with c C. It attains as often as f and thus zero as often as f.
For the second equation observe deg((
1
)) = deg((
2
)) which follows from
1
/
2
/(M)0. Now let f be a meromorphic function of degree n and let
us calculate (df/f). At points p with f(p) , 0, , ord
p
df/f is just the branch
number b
f
(p) and at each other point it is 1. The desired result now follows from
the RiemannHurwitz relation (A.4) and
deg((
df
f
)) =
pM
b
f
(p)
pf
1
(0)
(1 +b
f
(p))
pf
1
()
(1 +b
f
(p))
= B 2n. (A.29)
Divisors corresponding to meromorphic functions T = (f) are called principal,
divisors corresponding to meromorphic dierentials T = () canonical. It follows
from the above considerations that the principal divisors form a subgroup. It is
clearly normal (since all groups in sight are abelian) and we may consider its factor
group induced by the following equivalence relation
(A.30) T
1
T
2
T
1
T
2
= (f).
The factor group is called divisor class group and the equivalence classes [T]
divisor classes. Observe, that all canonical divisors lie in the same divisor class (by
(A.27)). The subset formed by the divisors of degree zero Div
0
(M) is a subgroup
of Div(M) and its factor group is called Picard group Pic(M).
The set Div(M) can be partially ordered by dening
(A.31) T
1
T
2
T
1
(p) T
2
(p), p M.
We write T
1
> T
2
if in addition T
1
,= T
2
. Note: T
1
T
2
deg(T
1
) deg(T
2
)
and T
1
> T
2
deg(T
1
) > deg(T
2
).
A divisor is called positive if T 0 and strictly positive if T > 0. Two
divisors T
1
, T
2
are called relatively prime if T
1
(p)T
2
(p) = 0, p M. A divisor
A.3. Divisors and the RiemannRoch theorem 281
can always be uniquely written as the dierence of two relatively prime positive
divisors
(A.32) T = T
+
T
, T
+
0, T
0.
For a meromorphic function (f)
+
and (f)
= /
1
(T + (f)), i(T) = i(T + (f)),
/
1
(T)
= C ).
Proof. Let p M. Since r(T
p
) = 2, there exists a (non constant) meromorphic
function with a single pole. It is injective (its degree is 1) and surjective (since every
non constant holomorphic map between compact Riemann surfaces is). Hence it
provides the desired isomorphism.
Corollary A.4. (g 1) There are no meromorphi
Viel mehr als nur Dokumente.
Entdecken, was Scribd alles zu bieten hat, inklusive Bücher und Hörbücher von großen Verlagen.
Jederzeit kündbar.