Random Matrix
Random Matrix
Random Matrix
Valkó
Our aim in this course to study the asymptotic behavior of the spectrum of certain random matrices.
Wigner Matrices
Definition 1 (real Wigner matrices). For 1 ≤ i < j < ∞ let Xi,j be i.i.d. (real) random variables
with mean 0 and variance 1 and set Xj,i = Xi,j . Let Xi,i be i.i.d. (real) random variables (with
possibly a different distribution) with mean 0 and variance 1. Then Mn = [Xi,j ]ni,j=1 will be a
random n × n symmetric matrix.
Definition 2 (complex Wigner matrices). For 1 ≤ i < j < ∞ let Xi,j be i.i.d. (complex) random
variables with mean 0, E|Xi,j |2 = 1 and set Xj,i = Xi,j . Let Xi,i be i.i.d. (real) random variables
with mean 0 and variance 1. Then Mn = [Xi,j ]ni,j=1 will be a random n × n hermitian matrix.
λ1 ≤ λ2 ≤ . . . λn .
(We will denote the dependence on n). Fact (which we will prove later): these are continuous
functions of Mn hence they are random variables themselves.
We would like to study the scaling limit of the empirical spectral measure
n
1X
νn∗ = δ λi .
n
i=1
This is a random discrete probability measure which puts n−1 mass to each (random) eigenvalue.
The following picture shows the histogram of eigenvalues for a certain 200 × 200 Wigner matrix.
14
12
10
-20 -10 0 10 20
1
The picture suggests that there is a nice deterministic limiting behavior. In order to figure out
the right scaling, we first compute the order of the empirical mean and second moment of the
eigenvalues.
n n
1X 1 1X
λi = TrMn = Xi,i
n n n
i=1 i=1
n n
1X 1 1 X
λi = TrMn2 = 2
Xi,j
n n n
i=1 i,j=1
The first moment converges to 0 by the strong law of large numbers. However the second moment
is of O(n) as we have about n2 /2 independent terms in the sum with a normalization of n1 instead
of n12 . This suggests that in order to see a meaningful limit, we need to scale the eigenvalues (or
the matrix) by √1n .
The following theorem states that in case we indeed have a deterministic limit.
Theorem 3 (Wigner’s semicircle law). Let
n
1X
νn = δ √λi .
n n
i=1
Gaussian Ensembles
It is easy to check the following useful fact: if C ∈ Rn×n is orthogonal (i.e. CC T = I) the C T Mn C
has the same distribution as Mn . (The GOE is invariant to orthogonal conjugation.) It is a bit
harder (we will prove it later) that one can actually compute the joint eigenvalue density which is
given by
1 Y 1 Pn 2
f (λ1 , . . . , λn ) = |λj − λi |e− 4 i=1 λi .
Z1
i<j
2
Definition 5 (GUE). Consider a complex Wigner matrix where Xi,j is standard complex Gaussian
(i.e. Xi,j ∼ N (0, 12 ) + iN (0, 12 )) and Xi,i ∼ N (0, 1) (real). The resulting random hermitian matrix
model is called Gaussian Unitary Ensemble (or GUE).
Another construction: let ai,j , i, j ∈ Z be i.i.d. standard complex Gaussians and An = [ai,j ]ni,j=1 .
An +A∗n
(Note that this is not a symmetric matrix!). Then the distribution of Mn = √
2
is GUE.
As the name suggests, GUE is invariant under unitary conjugation. If C ∈ Cn×n is unitary
(i.e. CC ∗ = I) the C T Mn C has the same distribution as Mn . (We will later show that the joint
eigenvalue density is given by
1 Y 1 Pn 2
f (λ1 , . . . , λn ) = |λj − λi |2 e− 2 i=1 λi .
Z2
i<j
One can see the similarity between the two densities: they are contained in the following one-
parameter family of densities:
1 Y β Pn 2
f (λ1 , . . . , λn ) = |λj − λi |β e− 4 i=1 λi . (1)
Zβ
i<j
For a given β > 0 the resulting distribution (on ordered n-tuples in R) is called Dyson’s β-ensemble.
For β = 1 one gets the eigenvalue density of GOE, for β = 2 we get the GUE. The β = 4 case is also
special: it is related another classical random matrix model, the Gaussian Symplectic Ensemble
(GSE), which can be defined using quaternions.
For other values of β there are no ’nice’ random matrices in the background. (We will see that one
can still build random matrices from which we get the general β-ensemble, but they won’t have
such nice symmetry properties.)
Later in the semester we will show that if one scales the β ensembles properly (’zooming in’ to see
the individual eigenvalues near a point) then one gets a point process limit. The limiting point
process is especially nice in the β = 2 case (GUE). It is conjectured that its distribution appears
among the critical line zeros of the Riemann-ζ function.
Another way of constructing a symmetric random matrix is the following. Let ai,j be i.i.d random
real random variables with mean 0 and variance 1. Let A = [ai,j ]n,m
i,j=1 ∈ R
n×m be a random matrix
with n rows and m columns (with n ≤ m). Then Mn = AAT is a (positive) symmetric random
matrix. We will show that the appropriately normalized empirical spectral measure will converge
to a deterministic limit. (This is the Marchenko-Pastur law). A similar statement holds if we
construct our matrix from i.i.d. complex random variables.