Beruflich Dokumente
Kultur Dokumente
TODD KEMP
C ONTENTS 1. Wigner Matrices 2. Wigners Semicircle Law 3. Markovs Inequality and Convergence of Expectation 4. First Proof of Wigners Semicircle Law 4.1. Convergence of matrix moments in Expectation 4.2. Convergence of Matrix Moments in Probability 4.3. Weak Convergence 5. Removing Moment Assumptions 6. Largest Eigenvalue 7. The Stiletjes Transform 8. The Stieltjes Transform and Convergence of Measures 8.1. Vague Convergence 8.2. Robustness of the Stieltjes transform 8.3. The Stieltjes Transform of an Empirical Eigenvalue Measure 9. Gaussian Random Matrices 10. Logarithmic Sobolev Inequalities 10.1. Heat kernels 10.2. Proof of the Gaussian log-Sobolev inequality, Theorem 10.3 11. The Herbst Concentration Inequality 12. Concentration for Random Matrices 12.1. Continuity of Eigenvalues 12.2. Concentration of the Empirical Eigenvalue Distribution 12.3. Return to Wigners Theorem 13. Gaussian Wigner Matrices, and the Genus Expansion 13.1. GOEn and GU En 13.2. Covariances of GU En 13.3. Wicks Theorem 13.4. The Genus Expansion 14. Joint Distribution of Eigenvalues of GOEn and GU En 14.1. Change of Variables on Lie Groups 14.2. Change of Variables for the Diagonalization Map 14.3. Joint Law of Eigenvalues and Eigenvectors for GOEn and GU En 15. The Vandermonde Determinant, and Hermite Polynomials 15.1. Hermite polynomials 15.2. The Vandermonde determinant and the Hermite kernel
Date: November 4, 2013.
1
2 4 6 7 7 12 15 17 21 31 34 34 37 39 40 44 47 51 54 57 57 59 61 61 61 64 65 67 70 71 72 75 77 78 81
TODD KEMP
15.3. Determinants of reproducing kernels 15.4. The averaged empirical eigenvalue distribution 16. Fluctuations of the Largest Eigenvalue of GU En , and Hypercontractivity 16.1. Lp -norms of Hermite polynomials 16.2. Hypercontractivity 16.3. (Almost) optimal non-asymptotic bounds on the largest eigenvalue 16.4. The Harer-Zagier recursion, and optimal tail bounds 17. The Tracy-Widom Law and Level Spacing of Zeroes
83 85 86 86 88 92 94 99
For the next ten weeks, we will be studying the eigenvalues of random matrices. A random matrix is simply a matrix (for now square) all of whose entries are random variables. That is: X is an n n matrix with {Xij }1i,j n random variables on some probability space (, F , P). 2 2 Alternatively, one can think of X as a random vector in Rn (or Cn ), although it is better to think of it as taking values in Mn (R) or Mn (C) (to keep the matrix structure in mind). For any xed instance , then, X ( ) is an n n matrix and (maybe) has eigenvalues 1 ( ), . . . , n ( ). So the eigenvalues are also random variables. We seek to understand aspects of the distributions of the i from knowledge of the distribution of the Xij . To begin, in order to guarantee that the matrix actually has eigenvalues (and they accurately capture the behavior of the linear transformation X ), we will make the assumption that X is a symmetric / Hermitian matrix. 1. W IGNER M ATRICES We begin by xing an innite family of real-valued random variables {Yij }j i1 . Then we can dene a sequence of symmetric random matrices Yn by [Yn ]ij = Yij , Yji , ij . i>j
The matrix Yn is symmetric and so has n real eigenvalues, which we write in increasing order 1 (Yn ) n (Yn ). In this very general setup, little can be said about these eigenvalues. The class of matrices we are going to begin studying, Wigner matrices, are given by the following conditions. We assume that the {Yij }1ij are independent. We assume that the diagonal entries {Yii }i1 are identically-distributed, and the off-diagonal entries {Yij }1i<j are identically-distributed. 2 2 2 We assume that E(Yij ) < for all i, j . (I.e. r2 = max{E(Y11 ), E(Y12 )} < .) It is not just for convenience that we separate out the diagonal terms; as we will see, they really do not contribute to the eigenvalues in the limit as n . It will also be convenient, at least at the start, to strengthen the nal assumption to moments of all orders: for k 1, set
k k rk = max{E(Y11 ), E(Y12 )}.
To begin, we will assume that rk < for each k ; we will weaken this assumption later. Note: in the presence of higher moments, much of the following will not actually require identicallydistributed entries. Rather, uniform bounds on all moments sufce. Herein, we will satisfy ourselves with the i.i.d. case.
One variation we will allow from i.i.d. is a uniform scaling of the matrices as n increases. That is: let n be a sequence of positive scalars, and set X n = n Y n . In fact, there is a natural choice for the scaling behavior, in order for there to be limiting behavior of eigenvalues. That is to say: we would like to arrange (if possible) that both 1 (Xn ) and n (Xn ) converge to nite numbers (different from each other). An easy way to test this is suggested by the following simple calculation:
2 2 2 2 2 1 + + n n max{j } = n max{1 , n }. j =1 n
Hence, in order for 1 and n to both converge to distinct constants, it is necessary for the sequence 1 2 (2 1 + + n ) to be bounded. Fortunately, this sequence can be calculated without explicit n reference to eigenvalues: it is the (normalized) Hilbert-Schmidt norm (or Fr obenius norm) of a symmetric matrix X: n 1 1 2 2 i (Xn )2 = Xij . X 2= n i=1 n 1i,j n Exercise 1.0.1. Verify the second equality above, by showing (using the spectral theorem) that 1 both expressions are equal to the quantity n Tr (X2 ). For our random matrix Xn above, then, we can calculate the expected value of this norm: E Xn 2 2
2 1 2 n 2 E(Yij ) = = n n 1i,j n n n 2 2n E(Yii ) + n 2 2 E(Yij ) 1i<j n
i=1
2 n
2 E(Y11 )
2 2 + (n 1)n E(Y12 ).
2 We now have two cases. If E(Y12 ) = 0 (meaning the off-diagonal terms are all 0 a.s.) then we see the correct scaling for n is n 1. This is a boring case: the matrices Xn are diagonal, with all diagonal entries identically distributed. Thus, these entries are also the eigenvalues, and so the distribution of eigenvalues is given by the common distribution of the diagonal entries. We ignore 2 ) > 0. Hence, in order for E Xn 2 this case, and therefore assume that E(Y12 2 to be a bounded sequence (that does not converge to 0), we must have n n1/2 . 2 ) > 0. Then the Denition 1.1. Let {Yij }1ij and Yn be as above, with r2 < and E(Y12 matrices Xn = n1/2 Yn are Wigner matrices.
It is standard to abuse notation and refer to the sequence Xn as a a Wigner matrix. The preceding calculation shows that, if Xn is a Wigner matrix, then the expected Hilbert-Schmidt norm E Xn 2 2 converges (as n ) to the second moment of the (off-diagonal) entries. As explained above, this is prerequisite to the bulk convergence of the eigenvalues. As we will shortly see, it is also sufcient. Consider the following example. Take the entires Yij to be N (0, 1) normal random variables. These are easy to simulate with MATLAB. Figure 1 shows the histogram of all n = 4000 eigenvalues of one instance of the corresponding Gaussian Wigner matrix X4000 . The plot suggests that 1 (Xn ) 2 while n (Xn ) 2 in this case. Moreover, although random uctuations remain, it appears that the histogram of eigenvalues (sometimes called the density of eigenvalues) converges to a deterministic shape. In fact, this is universally true. There is a universal probability distribution
TODD KEMP
t such that the density of eigenvalues of any Wigner matrix (with second moment t) converges to t . The limiting distribution is known as Wigners semicircle law: 1 t (dx) = (4t x2 )+ dx. 2t
(%
$%
'%
&%
!%
"%
% !