random unitary matrices and friends
TRANSCRIPT
![Page 1: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/1.jpg)
Random Unitary Matrices and Friends
Elizabeth Meckes
Case Western Reserve University
LDHD Summer SchoolSAMSI
August, 2013
![Page 2: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/2.jpg)
What is a random unitary matrix?
I A unitary matrix is an n× n matrix U with entries in C, suchthat
UU∗ = I,
where U∗ is the conjugate transpose of U.
That is, a unitary matrix is an n × n matrix over C whosecolumns (or rows) are orthonormal in Cn.
I The set of all n × n unitary matrices is denoted U (n); thisset is a group and a manifold.
![Page 3: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/3.jpg)
What is a random unitary matrix?
I A unitary matrix is an n× n matrix U with entries in C, suchthat
UU∗ = I,
where U∗ is the conjugate transpose of U.
That is, a unitary matrix is an n × n matrix over C whosecolumns (or rows) are orthonormal in Cn.
I The set of all n × n unitary matrices is denoted U (n); thisset is a group and a manifold.
![Page 4: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/4.jpg)
What is a random unitary matrix?
I A unitary matrix is an n× n matrix U with entries in C, suchthat
UU∗ = I,
where U∗ is the conjugate transpose of U.
That is, a unitary matrix is an n × n matrix over C whosecolumns (or rows) are orthonormal in Cn.
I The set of all n × n unitary matrices is denoted U (n); thisset is a group and a manifold.
![Page 5: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/5.jpg)
What is a random unitary matrix?
I A unitary matrix is an n× n matrix U with entries in C, suchthat
UU∗ = I,
where U∗ is the conjugate transpose of U.
That is, a unitary matrix is an n × n matrix over C whosecolumns (or rows) are orthonormal in Cn.
I The set of all n × n unitary matrices is denoted U (n); thisset is a group and a manifold.
![Page 6: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/6.jpg)
What is a random unitary matrix?
I Metric Structure:I U (n) sits inside Cn2
and inherits a geodesic metric dg(·, ·)from the Euclidean metric on Cn2
.
I U (n) also has its own Euclidean (Hilbert-Schmidt) metricfrom the inner product 〈U,V 〉 = Tr(UV ∗).
I The two metrics are equivalent:
dHS(U,V ) ≤ dg(U,V ) ≤ π
2dHS(U,V ).
I Randomness:There is a unique translation-invariant probability measurecalled Haar measure on U (n): if U is a Haar-distributedrandom unitary matrix, so are AU and UA, for A a fixedunitary matrix.
![Page 7: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/7.jpg)
What is a random unitary matrix?
I Metric Structure:I U (n) sits inside Cn2
and inherits a geodesic metric dg(·, ·)from the Euclidean metric on Cn2
.
I U (n) also has its own Euclidean (Hilbert-Schmidt) metricfrom the inner product 〈U,V 〉 = Tr(UV ∗).
I The two metrics are equivalent:
dHS(U,V ) ≤ dg(U,V ) ≤ π
2dHS(U,V ).
I Randomness:There is a unique translation-invariant probability measurecalled Haar measure on U (n): if U is a Haar-distributedrandom unitary matrix, so are AU and UA, for A a fixedunitary matrix.
![Page 8: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/8.jpg)
What is a random unitary matrix?
I Metric Structure:I U (n) sits inside Cn2
and inherits a geodesic metric dg(·, ·)from the Euclidean metric on Cn2
.
I U (n) also has its own Euclidean (Hilbert-Schmidt) metricfrom the inner product 〈U,V 〉 = Tr(UV ∗).
I The two metrics are equivalent:
dHS(U,V ) ≤ dg(U,V ) ≤ π
2dHS(U,V ).
I Randomness:There is a unique translation-invariant probability measurecalled Haar measure on U (n): if U is a Haar-distributedrandom unitary matrix, so are AU and UA, for A a fixedunitary matrix.
![Page 9: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/9.jpg)
What is a random unitary matrix?
I Metric Structure:I U (n) sits inside Cn2
and inherits a geodesic metric dg(·, ·)from the Euclidean metric on Cn2
.
I U (n) also has its own Euclidean (Hilbert-Schmidt) metricfrom the inner product 〈U,V 〉 = Tr(UV ∗).
I The two metrics are equivalent:
dHS(U,V ) ≤ dg(U,V ) ≤ π
2dHS(U,V ).
I Randomness:There is a unique translation-invariant probability measurecalled Haar measure on U (n): if U is a Haar-distributedrandom unitary matrix, so are AU and UA, for A a fixedunitary matrix.
![Page 10: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/10.jpg)
A couple ways to build a random unitary matrix
1. I Pick the first column U1 uniformly from S1C ⊆ Cn.
I Pick the second column U2 uniformly from U⊥1 ⊆ S1C.
...I Pick the last column Un uniformly from
(span{U1, . . . ,Un−1})⊥ ⊆ S1C.
2. I Fill an n × n array with i.i.d. standard complex Gaussianrandom variables.
I Stick the result into the QR algorithm; the resulting Q isHaar-distributed on U (n).
![Page 11: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/11.jpg)
A couple ways to build a random unitary matrix
1. I Pick the first column U1 uniformly from S1C ⊆ Cn.
I Pick the second column U2 uniformly from U⊥1 ⊆ S1C.
...I Pick the last column Un uniformly from
(span{U1, . . . ,Un−1})⊥ ⊆ S1C.
2. I Fill an n × n array with i.i.d. standard complex Gaussianrandom variables.
I Stick the result into the QR algorithm; the resulting Q isHaar-distributed on U (n).
![Page 12: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/12.jpg)
A couple ways to build a random unitary matrix
1. I Pick the first column U1 uniformly from S1C ⊆ Cn.
I Pick the second column U2 uniformly from U⊥1 ⊆ S1C.
...I Pick the last column Un uniformly from
(span{U1, . . . ,Un−1})⊥ ⊆ S1C.
2. I Fill an n × n array with i.i.d. standard complex Gaussianrandom variables.
I Stick the result into the QR algorithm; the resulting Q isHaar-distributed on U (n).
![Page 13: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/13.jpg)
A couple ways to build a random unitary matrix
1. I Pick the first column U1 uniformly from S1C ⊆ Cn.
I Pick the second column U2 uniformly from U⊥1 ⊆ S1C.
...I Pick the last column Un uniformly from
(span{U1, . . . ,Un−1})⊥ ⊆ S1C.
2. I Fill an n × n array with i.i.d. standard complex Gaussianrandom variables.
I Stick the result into the QR algorithm; the resulting Q isHaar-distributed on U (n).
![Page 14: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/14.jpg)
A couple ways to build a random unitary matrix
1. I Pick the first column U1 uniformly from S1C ⊆ Cn.
I Pick the second column U2 uniformly from U⊥1 ⊆ S1C.
...I Pick the last column Un uniformly from
(span{U1, . . . ,Un−1})⊥ ⊆ S1C.
2. I Fill an n × n array with i.i.d. standard complex Gaussianrandom variables.
I Stick the result into the QR algorithm; the resulting Q isHaar-distributed on U (n).
![Page 15: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/15.jpg)
A couple ways to build a random unitary matrix
1. I Pick the first column U1 uniformly from S1C ⊆ Cn.
I Pick the second column U2 uniformly from U⊥1 ⊆ S1C.
...I Pick the last column Un uniformly from
(span{U1, . . . ,Un−1})⊥ ⊆ S1C.
2. I Fill an n × n array with i.i.d. standard complex Gaussianrandom variables.
I Stick the result into the QR algorithm; the resulting Q isHaar-distributed on U (n).
![Page 16: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/16.jpg)
Meet U (n)’s kid sister: The orthogonal group
I An orthogonal matrix is an n × n matrix U with entries in R,such that
UUT = I,
where UT is the transpose of U. That is, a unitary matrix isan n × n matrix over R whose columns (or rows) areorthonormal in Rn.
I The set of all n × n unitary matrices is denoted O (n); thisset is a subgroup and a submanifold of U (n).
I O (n) has two connected components: SO (n) (det(U) = 1)and SO− (n) (det(U) = −1).
I There is a unique translation-invariant (Haar) probabilitymeasure on each of O (n), SO (n) and SO− (n).
![Page 17: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/17.jpg)
Meet U (n)’s kid sister: The orthogonal groupI An orthogonal matrix is an n × n matrix U with entries in R,
such thatUUT = I,
where UT is the transpose of U.
That is, a unitary matrix isan n × n matrix over R whose columns (or rows) areorthonormal in Rn.
I The set of all n × n unitary matrices is denoted O (n); thisset is a subgroup and a submanifold of U (n).
I O (n) has two connected components: SO (n) (det(U) = 1)and SO− (n) (det(U) = −1).
I There is a unique translation-invariant (Haar) probabilitymeasure on each of O (n), SO (n) and SO− (n).
![Page 18: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/18.jpg)
Meet U (n)’s kid sister: The orthogonal groupI An orthogonal matrix is an n × n matrix U with entries in R,
such thatUUT = I,
where UT is the transpose of U. That is, a unitary matrix isan n × n matrix over R whose columns (or rows) areorthonormal in Rn.
I The set of all n × n unitary matrices is denoted O (n); thisset is a subgroup and a submanifold of U (n).
I O (n) has two connected components: SO (n) (det(U) = 1)and SO− (n) (det(U) = −1).
I There is a unique translation-invariant (Haar) probabilitymeasure on each of O (n), SO (n) and SO− (n).
![Page 19: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/19.jpg)
Meet U (n)’s kid sister: The orthogonal groupI An orthogonal matrix is an n × n matrix U with entries in R,
such thatUUT = I,
where UT is the transpose of U. That is, a unitary matrix isan n × n matrix over R whose columns (or rows) areorthonormal in Rn.
I The set of all n × n unitary matrices is denoted O (n); thisset is a subgroup and a submanifold of U (n).
I O (n) has two connected components: SO (n) (det(U) = 1)and SO− (n) (det(U) = −1).
I There is a unique translation-invariant (Haar) probabilitymeasure on each of O (n), SO (n) and SO− (n).
![Page 20: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/20.jpg)
Meet U (n)’s kid sister: The orthogonal groupI An orthogonal matrix is an n × n matrix U with entries in R,
such thatUUT = I,
where UT is the transpose of U. That is, a unitary matrix isan n × n matrix over R whose columns (or rows) areorthonormal in Rn.
I The set of all n × n unitary matrices is denoted O (n); thisset is a subgroup and a submanifold of U (n).
I O (n) has two connected components: SO (n) (det(U) = 1)and SO− (n) (det(U) = −1).
I There is a unique translation-invariant (Haar) probabilitymeasure on each of O (n), SO (n) and SO− (n).
![Page 21: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/21.jpg)
The symplectic group:the weird uncle no one talks about
I A symplectic matrix is an 2n × 2n matrix with entries in C,such that
UJU∗ = J,
where U∗ is the conjugate transpose of U and
J =
[0 I−I 0
].
(It is really the quaternionic unitary group.)
I The group of 2n × 2n symplectic matrices is denotedSp (2n).
![Page 22: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/22.jpg)
The symplectic group:the weird uncle no one talks about
I A symplectic matrix is an 2n × 2n matrix with entries in C,such that
UJU∗ = J,
where U∗ is the conjugate transpose of U and
J =
[0 I−I 0
].
(It is really the quaternionic unitary group.)
I The group of 2n × 2n symplectic matrices is denotedSp (2n).
![Page 23: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/23.jpg)
The symplectic group:the weird uncle no one talks about
I A symplectic matrix is an 2n × 2n matrix with entries in C,such that
UJU∗ = J,
where U∗ is the conjugate transpose of U and
J =
[0 I−I 0
].
(It is really the quaternionic unitary group.)
I The group of 2n × 2n symplectic matrices is denotedSp (2n).
![Page 24: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/24.jpg)
The symplectic group:the weird uncle no one talks about
I A symplectic matrix is an 2n × 2n matrix with entries in C,such that
UJU∗ = J,
where U∗ is the conjugate transpose of U and
J =
[0 I−I 0
].
(It is really the quaternionic unitary group.)
I The group of 2n × 2n symplectic matrices is denotedSp (2n).
![Page 25: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/25.jpg)
Concentration of measure
Theorem (G/M;B/E;L;M/M)Let G be one of SO (n), SO− (n), SU (n), U (n), Sp (2n), and letF : G→ R be L-Lipschitz (w.r.t. the geodesic metric or theHS-metric). Let U be distributed according to Haar measure onG. Then there are universal constants C, c such that
P[∣∣F (U)− EF (U)
∣∣ > Lt]≤ Ce−cnt2
,
for every t > 0.
![Page 26: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/26.jpg)
The entries of a random orthogonal matrix
Note: permuting the rows or columns of a random orthogonalmatrix U corresponds to left- or right-multiplication by apermutation matrix (which is itself orthogonal).
=⇒ The entries {uij} of U all have the same distribution.
Classical fact: A coordinate of a random point on the sphere inRn is approximately Gaussian, for large n.
=⇒ The entries {uij} of U are
individually approximately Gaussian
if U is large.
![Page 27: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/27.jpg)
The entries of a random orthogonal matrix
Note: permuting the rows or columns of a random orthogonalmatrix U corresponds to left- or right-multiplication by apermutation matrix (which is itself orthogonal).
=⇒ The entries {uij} of U all have the same distribution.
Classical fact: A coordinate of a random point on the sphere inRn is approximately Gaussian, for large n.
=⇒ The entries {uij} of U are
individually approximately Gaussian
if U is large.
![Page 28: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/28.jpg)
The entries of a random orthogonal matrix
Note: permuting the rows or columns of a random orthogonalmatrix U corresponds to left- or right-multiplication by apermutation matrix (which is itself orthogonal).
=⇒ The entries {uij} of U all have the same distribution.
Classical fact: A coordinate of a random point on the sphere inRn is approximately Gaussian, for large n.
=⇒ The entries {uij} of U are
individually approximately Gaussian
if U is large.
![Page 29: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/29.jpg)
The entries of a random orthogonal matrix
Note: permuting the rows or columns of a random orthogonalmatrix U corresponds to left- or right-multiplication by apermutation matrix (which is itself orthogonal).
=⇒ The entries {uij} of U all have the same distribution.
Classical fact: A coordinate of a random point on the sphere inRn is approximately Gaussian, for large n.
=⇒ The entries {uij} of U are
individually approximately Gaussian
if U is large.
![Page 30: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/30.jpg)
The entries of a random orthogonal matrix
A more modern fact (Diaconis–Freedman): If X is a randomlydistributed point on the sphere of radius
√n in Rn, and Z is a
standard Gaussian random vector in Rn, then
dTV
((X1, . . . ,Xk ), (Z1, . . . ,Zk )
)≤ 2(k + 3)
n − k − 3.
=⇒ Any k entries within one row (or column) of U ∈ U (n) areapproximately independent Gaussians, if k = o(n).
Diaconis’question: How many entries of U can besimultaneously approximated by independent Gaussians?
![Page 31: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/31.jpg)
The entries of a random orthogonal matrix
A more modern fact (Diaconis–Freedman): If X is a randomlydistributed point on the sphere of radius
√n in Rn, and Z is a
standard Gaussian random vector in Rn, then
dTV
((X1, . . . ,Xk ), (Z1, . . . ,Zk )
)≤ 2(k + 3)
n − k − 3.
=⇒ Any k entries within one row (or column) of U ∈ U (n) areapproximately independent Gaussians, if k = o(n).
Diaconis’question: How many entries of U can besimultaneously approximated by independent Gaussians?
![Page 32: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/32.jpg)
The entries of a random orthogonal matrix
A more modern fact (Diaconis–Freedman): If X is a randomlydistributed point on the sphere of radius
√n in Rn, and Z is a
standard Gaussian random vector in Rn, then
dTV
((X1, . . . ,Xk ), (Z1, . . . ,Zk )
)≤ 2(k + 3)
n − k − 3.
=⇒ Any k entries within one row (or column) of U ∈ U (n) areapproximately independent Gaussians, if k = o(n).
Diaconis’question: How many entries of U can besimultaneously approximated by independent Gaussians?
![Page 33: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/33.jpg)
Jiang’s answer(s)
It depends on what you mean by approximated.
Theorem (Jiang)Let {Un} be a sequence of random orthogonal matrices withUn ∈ O (n) for each n, and suppose that pn,qn = o(
√n).
Let L(√
nU(pn,qn)) denote the joint distribution of the pnqnentries of the top-left pn × qn block of
√nUn, and let Z (pn,qn)
denote a collection of pnqn i.i.d. standard normal randomvariables. Then
limn→∞
dTV (L(√
nU(pn,qn)),Z (pn,qn)) = 0.
That is, a pn × qn principle submatrix can be approximated intotal variation by a Gaussian random matrix, as long aspn,qn �
√n.
![Page 34: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/34.jpg)
Jiang’s answer(s)It depends on what you mean by approximated.
Theorem (Jiang)Let {Un} be a sequence of random orthogonal matrices withUn ∈ O (n) for each n, and suppose that pn,qn = o(
√n).
Let L(√
nU(pn,qn)) denote the joint distribution of the pnqnentries of the top-left pn × qn block of
√nUn, and let Z (pn,qn)
denote a collection of pnqn i.i.d. standard normal randomvariables. Then
limn→∞
dTV (L(√
nU(pn,qn)),Z (pn,qn)) = 0.
That is, a pn × qn principle submatrix can be approximated intotal variation by a Gaussian random matrix, as long aspn,qn �
√n.
![Page 35: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/35.jpg)
Jiang’s answer(s)It depends on what you mean by approximated.
Theorem (Jiang)Let {Un} be a sequence of random orthogonal matrices withUn ∈ O (n) for each n, and suppose that pn,qn = o(
√n).
Let L(√
nU(pn,qn)) denote the joint distribution of the pnqnentries of the top-left pn × qn block of
√nUn, and let Z (pn,qn)
denote a collection of pnqn i.i.d. standard normal randomvariables. Then
limn→∞
dTV (L(√
nU(pn,qn)),Z (pn,qn)) = 0.
That is, a pn × qn principle submatrix can be approximated intotal variation by a Gaussian random matrix, as long aspn,qn �
√n.
![Page 36: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/36.jpg)
Jiang’s answer(s)It depends on what you mean by approximated.
Theorem (Jiang)Let {Un} be a sequence of random orthogonal matrices withUn ∈ O (n) for each n, and suppose that pn,qn = o(
√n).
Let L(√
nU(pn,qn)) denote the joint distribution of the pnqnentries of the top-left pn × qn block of
√nUn, and let Z (pn,qn)
denote a collection of pnqn i.i.d. standard normal randomvariables. Then
limn→∞
dTV (L(√
nU(pn,qn)),Z (pn,qn)) = 0.
That is, a pn × qn principle submatrix can be approximated intotal variation by a Gaussian random matrix, as long aspn,qn �
√n.
![Page 37: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/37.jpg)
Jiang’s answer(s)
Theorem (Jiang)For each n, let Yn =
[yij]n
i,j=1 be an n × n matrix of independent
standard Gaussian random variables and let Γn =[γij]n
i,j=1 bethe matrix obtained from Yn by performing the Gram-Schmidtprocess; i.e., Γn is a random orthogonal matrix. Let
εn(m) = max1≤i≤n,1≤j≤m
∣∣√nγij − yij∣∣.
Thenεn(mn)
P−−−→n→∞
0
if and only if mn = o(
nlog(n)
).
That is, in an “in probability” sense, n2
log(n) entries of U can besimultaneously approximated by independent Gaussians.
![Page 38: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/38.jpg)
Jiang’s answer(s)
Theorem (Jiang)For each n, let Yn =
[yij]n
i,j=1 be an n × n matrix of independent
standard Gaussian random variables and let Γn =[γij]n
i,j=1 bethe matrix obtained from Yn by performing the Gram-Schmidtprocess; i.e., Γn is a random orthogonal matrix. Let
εn(m) = max1≤i≤n,1≤j≤m
∣∣√nγij − yij∣∣.
Thenεn(mn)
P−−−→n→∞
0
if and only if mn = o(
nlog(n)
).
That is, in an “in probability” sense, n2
log(n) entries of U can besimultaneously approximated by independent Gaussians.
![Page 39: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/39.jpg)
A more geometric viewpoint
Choosing a principle submatrix of an n × n orthogonal matrix Ucorresponds to a particular type of orthogonal projection from alarge matrix space to a smaller one.
(Note that the result is no longer orthogonal.)
In general, a rank k orthogonal projection of O (n) looks like
U 7→(
Tr(A1U), . . . ,Tr(AkU)),
where A1, . . . ,Ak are orthonormal matrices in O (n); i.e.,
Tr(AiATj ) = δij .
![Page 40: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/40.jpg)
A more geometric viewpoint
Choosing a principle submatrix of an n × n orthogonal matrix Ucorresponds to a particular type of orthogonal projection from alarge matrix space to a smaller one.
(Note that the result is no longer orthogonal.)
In general, a rank k orthogonal projection of O (n) looks like
U 7→(
Tr(A1U), . . . ,Tr(AkU)),
where A1, . . . ,Ak are orthonormal matrices in O (n); i.e.,
Tr(AiATj ) = δij .
![Page 41: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/41.jpg)
A more geometric viewpoint
Choosing a principle submatrix of an n × n orthogonal matrix Ucorresponds to a particular type of orthogonal projection from alarge matrix space to a smaller one.
(Note that the result is no longer orthogonal.)
In general, a rank k orthogonal projection of O (n) looks like
U 7→(
Tr(A1U), . . . ,Tr(AkU)),
where A1, . . . ,Ak are orthonormal matrices in O (n); i.e.,
Tr(AiATj ) = δij .
![Page 42: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/42.jpg)
A more geometric viewpoint
Choosing a principle submatrix of an n × n orthogonal matrix Ucorresponds to a particular type of orthogonal projection from alarge matrix space to a smaller one.
(Note that the result is no longer orthogonal.)
In general, a rank k orthogonal projection of O (n) looks like
U 7→(
Tr(A1U), . . . ,Tr(AkU)),
where A1, . . . ,Ak are orthonormal matrices in O (n); i.e.,
Tr(AiATj ) = δij .
![Page 43: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/43.jpg)
A more geometric viewpoint
Theorem (Chatterjee–M.)Let A1, . . . ,Ak be orthonormal (w.r.t. the Hilbert-Schmidt innerproduct) in O (n), and let U ∈ O (n) be a random orthogonalmatrix. Consider the random vector
X := (Tr(A1U), . . . ,Tr(AkU)) ,
and let Z := (Z1, . . . ,Zk ) be a standard Gaussian randomvector in Rk . Then for all n ≥ 2,
dW (X ,Z ) ≤√
2kn − 1
.
Here, dW (·, ·) denotes the L1-Wasserstein distance.
![Page 44: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/44.jpg)
Eigenvalues – The empirical spectral measure
Let U be a Haar-distributed matrix in U (N).
Then U has (random) eigenvalues {eiθj}Nj=1.
Note: The distribution of the set of eigenvalues isrotation-invariant.
To understand the behavior of the ensemble of randomeigenvalues, we consider the empirical spectral measure of U:
µN :=1N
N∑j=1
δeiθj .
![Page 45: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/45.jpg)
Eigenvalues – The empirical spectral measure
Let U be a Haar-distributed matrix in U (N).
Then U has (random) eigenvalues {eiθj}Nj=1.
Note: The distribution of the set of eigenvalues isrotation-invariant.
To understand the behavior of the ensemble of randomeigenvalues, we consider the empirical spectral measure of U:
µN :=1N
N∑j=1
δeiθj .
![Page 46: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/46.jpg)
Eigenvalues – The empirical spectral measure
Let U be a Haar-distributed matrix in U (N).
Then U has (random) eigenvalues {eiθj}Nj=1.
Note: The distribution of the set of eigenvalues isrotation-invariant.
To understand the behavior of the ensemble of randomeigenvalues, we consider the empirical spectral measure of U:
µN :=1N
N∑j=1
δeiθj .
![Page 47: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/47.jpg)
Eigenvalues – The empirical spectral measure
Let U be a Haar-distributed matrix in U (N).
Then U has (random) eigenvalues {eiθj}Nj=1.
Note: The distribution of the set of eigenvalues isrotation-invariant.
To understand the behavior of the ensemble of randomeigenvalues, we consider the empirical spectral measure of U:
µN :=1N
N∑j=1
δeiθj .
![Page 48: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/48.jpg)
Eigenvalues – The empirical spectral measure
Let U be a Haar-distributed matrix in U (N).
Then U has (random) eigenvalues {eiθj}Nj=1.
Note: The distribution of the set of eigenvalues isrotation-invariant.
To understand the behavior of the ensemble of randomeigenvalues, we consider the empirical spectral measure of U:
µN :=1N
N∑j=1
δeiθj .
![Page 49: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/49.jpg)
E. Rains
100 i.i.d. uniform randompoints
The eigenvalues of a100× 100 random unitarymatrix
![Page 50: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/50.jpg)
Diaconis/Shahshahani
Theorem (D–S)Let Un ∈ U (n) be a random unitary matrix, and let µUn denotethe empirical spectral measure of Un. Let ν denote the uniformprobability measure on S1. Then
µUnn→∞−−−→ ν,
weak-* in probability.
I The theorem follows from explicit formulae for the mixedmoments of the random vector
(Tr(Un), . . . ,Tr(Uk
n ))
forfixed k , which have been useful in many other contexts.
I They showed in particular that(
Tr(Un), . . . ,Tr(Ukn ))
isasymptotically distributed as a standard complex Gaussianrandom vector.
![Page 51: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/51.jpg)
Diaconis/Shahshahani
Theorem (D–S)Let Un ∈ U (n) be a random unitary matrix, and let µUn denotethe empirical spectral measure of Un. Let ν denote the uniformprobability measure on S1. Then
µUnn→∞−−−→ ν,
weak-* in probability.
I The theorem follows from explicit formulae for the mixedmoments of the random vector
(Tr(Un), . . . ,Tr(Uk
n ))
forfixed k , which have been useful in many other contexts.
I They showed in particular that(
Tr(Un), . . . ,Tr(Ukn ))
isasymptotically distributed as a standard complex Gaussianrandom vector.
![Page 52: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/52.jpg)
Diaconis/Shahshahani
Theorem (D–S)Let Un ∈ U (n) be a random unitary matrix, and let µUn denotethe empirical spectral measure of Un. Let ν denote the uniformprobability measure on S1. Then
µUnn→∞−−−→ ν,
weak-* in probability.
I The theorem follows from explicit formulae for the mixedmoments of the random vector
(Tr(Un), . . . ,Tr(Uk
n ))
forfixed k , which have been useful in many other contexts.
I They showed in particular that(
Tr(Un), . . . ,Tr(Ukn ))
isasymptotically distributed as a standard complex Gaussianrandom vector.
![Page 53: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/53.jpg)
The number of eigenvalues in an arcTheorem (Wieand)Let Ij := (eiαj ,eiβj ) be intervals on S1 and for Un ∈ U (n) arandom unitary matrix, let
Yn,k :=µUn (Ik )− EµUn (Ik )
1π
√log(n)
.
Then as n tends to infinity, the random vector(Yn,1, . . . ,Yn,k
)converges in distribution to a jointly Gaussian random vector(Z1, . . . ,Zk ) with covariance
Cov(Zj ,Zk ) =
0, αj , αk , βj , βk all distict ;12 αj = αk or βj = βk (but not both);
−12 αj = βk or βj = αk (but not both);
1 αj = αk and βj = βk ;
−1 αj = βk and βj = αk .
![Page 54: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/54.jpg)
About that weird covariance structure...
Another Gaussian process that has it: Again suppose thatIj := (eiαj ,eiβj ) are intervals on S1, and suppose that{Gθ}θ∈[0,2π) are i.i.d. standard Gaussians. Define
Xn,k = Gβk −Gαk ;
then
Cov(Xj ,Xk ) =
0, αj , αk , βj , βk all distict ;12 αj = αk or βj = βk (but not both);
−12 αj = βk or βj = αk (but not both);
1 αj = αk and βj = βk ;
−1 αj = βk and βj = αk .
![Page 55: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/55.jpg)
About that weird covariance structure...
Another Gaussian process that has it:
Again suppose thatIj := (eiαj ,eiβj ) are intervals on S1, and suppose that{Gθ}θ∈[0,2π) are i.i.d. standard Gaussians. Define
Xn,k = Gβk −Gαk ;
then
Cov(Xj ,Xk ) =
0, αj , αk , βj , βk all distict ;12 αj = αk or βj = βk (but not both);
−12 αj = βk or βj = αk (but not both);
1 αj = αk and βj = βk ;
−1 αj = βk and βj = αk .
![Page 56: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/56.jpg)
About that weird covariance structure...
Another Gaussian process that has it: Again suppose thatIj := (eiαj ,eiβj ) are intervals on S1, and suppose that{Gθ}θ∈[0,2π) are i.i.d. standard Gaussians. Define
Xn,k = Gβk −Gαk ;
then
Cov(Xj ,Xk ) =
0, αj , αk , βj , βk all distict ;12 αj = αk or βj = βk (but not both);
−12 αj = βk or βj = αk (but not both);
1 αj = αk and βj = βk ;
−1 αj = βk and βj = αk .
![Page 57: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/57.jpg)
Where’s the white noise in U?
Theorem (Hughes–Keating–O’Connel)Let Z (θ) be the characteristic polynomial of U and fix θ1 . . . , θk .Then
1√12 log(n)
(log(Z (θ1)), . . . , log(Z (θk ))
)converges in distribution to a standard Gaussian random vectorin Ck , as n→∞.
HKO in particular showed that Wieand’s result follows fromtheirs by the argument principle.
![Page 58: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/58.jpg)
Where’s the white noise in U?
Theorem (Hughes–Keating–O’Connel)Let Z (θ) be the characteristic polynomial of U and fix θ1 . . . , θk .Then
1√12 log(n)
(log(Z (θ1)), . . . , log(Z (θk ))
)converges in distribution to a standard Gaussian random vectorin Ck , as n→∞.
HKO in particular showed that Wieand’s result follows fromtheirs by the argument principle.
![Page 59: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/59.jpg)
Where’s the white noise in U?
Theorem (Hughes–Keating–O’Connel)Let Z (θ) be the characteristic polynomial of U and fix θ1 . . . , θk .Then
1√12 log(n)
(log(Z (θ1)), . . . , log(Z (θk ))
)converges in distribution to a standard Gaussian random vectorin Ck , as n→∞.
HKO in particular showed that Wieand’s result follows fromtheirs by the argument principle.
![Page 60: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/60.jpg)
Powers of U
The eigenvalues of Um for m = 1,5,20,45,80, for U arealization of a random 80× 80 unitary matrix.
![Page 61: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/61.jpg)
Rains’ Theorems
Theorem (Rains 1997)Let U ∈ U (n) be a random unitary matrix, and let m ≥ n. Thenthe eigenvalues of Um are distributed exactly as n i.i.d. uniformpoints on S1.
Theorem (Rains 2003)Let m ≤ N be fixed. Then
[U (N)]me.v .d .
=⊕
0≤j<m
U(⌈
N − jm
⌉),
where e.v .d .= denotes equality of eigenvalue distributions.
![Page 62: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/62.jpg)
Rains’ Theorems
Theorem (Rains 1997)Let U ∈ U (n) be a random unitary matrix, and let m ≥ n. Thenthe eigenvalues of Um are distributed exactly as n i.i.d. uniformpoints on S1.
Theorem (Rains 2003)Let m ≤ N be fixed. Then
[U (N)]me.v .d .
=⊕
0≤j<m
U(⌈
N − jm
⌉),
where e.v .d .= denotes equality of eigenvalue distributions.
![Page 63: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/63.jpg)
Rains’ Theorems
Theorem (Rains 1997)Let U ∈ U (n) be a random unitary matrix, and let m ≥ n. Thenthe eigenvalues of Um are distributed exactly as n i.i.d. uniformpoints on S1.
Theorem (Rains 2003)Let m ≤ N be fixed. Then
[U (N)]me.v .d .
=⊕
0≤j<m
U(⌈
N − jm
⌉),
where e.v .d .= denotes equality of eigenvalue distributions.
![Page 64: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/64.jpg)
The eigenvalues of Um for m = 1,5,20,45,80, for U arealization of a random 80× 80 unitary matrix.
![Page 65: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/65.jpg)
Theorem (E.M./M. Meckes)Let ν denote the uniform probability measure on the circle and
Wp(µ, ν) := inf{(∫
|x − y |p dπ(x , y)) 1
p
∣∣∣∣ π(A× C) = µ(A)π(C× A) = ν(A)
}.
Then
I E[Wp(µm,N , ν)
]≤
Cp√
m[log( Nm )+1]
N .
I For 1 ≤ p ≤ 2,
P
[Wp(µm,N , ν) ≥
C√
m[log( Nm )+1]
N + t
]≤ exp
[−N2t2
24m
].
I For p > 2,
P
[Wp(µm,N , ν) ≥
Cp√
m[log( Nm )+1]
N + t
]≤ exp
[−N1+ 2
p t2
24m
].
![Page 66: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/66.jpg)
Theorem (E.M./M. Meckes)Let ν denote the uniform probability measure on the circle and
Wp(µ, ν) := inf{(∫
|x − y |p dπ(x , y)) 1
p
∣∣∣∣ π(A× C) = µ(A)π(C× A) = ν(A)
}.
Then
I E[Wp(µm,N , ν)
]≤
Cp√
m[log( Nm )+1]
N .
I For 1 ≤ p ≤ 2,
P
[Wp(µm,N , ν) ≥
C√
m[log( Nm )+1]
N + t
]≤ exp
[−N2t2
24m
].
I For p > 2,
P
[Wp(µm,N , ν) ≥
Cp√
m[log( Nm )+1]
N + t
]≤ exp
[−N1+ 2
p t2
24m
].
![Page 67: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/67.jpg)
Theorem (E.M./M. Meckes)Let ν denote the uniform probability measure on the circle and
Wp(µ, ν) := inf{(∫
|x − y |p dπ(x , y)) 1
p
∣∣∣∣ π(A× C) = µ(A)π(C× A) = ν(A)
}.
Then
I E[Wp(µm,N , ν)
]≤
Cp√
m[log( Nm )+1]
N .
I For 1 ≤ p ≤ 2,
P
[Wp(µm,N , ν) ≥
C√
m[log( Nm )+1]
N + t
]≤ exp
[−N2t2
24m
].
I For p > 2,
P
[Wp(µm,N , ν) ≥
Cp√
m[log( Nm )+1]
N + t
]≤ exp
[−N1+ 2
p t2
24m
].
![Page 68: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/68.jpg)
Theorem (E.M./M. Meckes)Let ν denote the uniform probability measure on the circle and
Wp(µ, ν) := inf{(∫
|x − y |p dπ(x , y)) 1
p
∣∣∣∣ π(A× C) = µ(A)π(C× A) = ν(A)
}.
Then
I E[Wp(µm,N , ν)
]≤
Cp√
m[log( Nm )+1]
N .
I For 1 ≤ p ≤ 2,
P
[Wp(µm,N , ν) ≥
C√
m[log( Nm )+1]
N + t
]≤ exp
[−N2t2
24m
].
I For p > 2,
P
[Wp(µm,N , ν) ≥
Cp√
m[log( Nm )+1]
N + t
]≤ exp
[−N1+ 2
p t2
24m
].
![Page 69: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/69.jpg)
Almost sure convergence
CorollaryFor each N, let UN be distributed according to uniform measureon U (N) and let mN ∈ {1, . . . ,N}. There is a C such that, withprobability 1,
Wp(µmN ,N , ν) ≤Cp√
mN log(N)
N12+
1max(2,p)
eventually.
![Page 70: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/70.jpg)
A miraculous representation of the eigenvaluecounting function
Fact: The set {eiθj}Nj=1 of eigenvalues of U (uniform in U (N)) isa determinantal point process.
Theorem (Hough/Krishnapur/Peres/Virag 2006)Let X be a determinantal point process in Λ satisfying someniceness conditions. For D ⊆ Λ, let ND be the number of pointsof X in D. Then
NDd=∑
k
ξk ,
where {ξk} are independent Bernoulli random variables withmeans given explicitly in terms of the kernel of X .
![Page 71: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/71.jpg)
A miraculous representation of the eigenvaluecounting function
Fact: The set {eiθj}Nj=1 of eigenvalues of U (uniform in U (N)) isa determinantal point process.
Theorem (Hough/Krishnapur/Peres/Virag 2006)Let X be a determinantal point process in Λ satisfying someniceness conditions. For D ⊆ Λ, let ND be the number of pointsof X in D. Then
NDd=∑
k
ξk ,
where {ξk} are independent Bernoulli random variables withmeans given explicitly in terms of the kernel of X .
![Page 72: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/72.jpg)
A miraculous representation of the eigenvaluecounting function
Fact: The set {eiθj}Nj=1 of eigenvalues of U (uniform in U (N)) isa determinantal point process.
Theorem (Hough/Krishnapur/Peres/Virag 2006)Let X be a determinantal point process in Λ satisfying someniceness conditions. For D ⊆ Λ, let ND be the number of pointsof X in D. Then
NDd=∑
k
ξk ,
where {ξk} are independent Bernoulli random variables withmeans given explicitly in terms of the kernel of X .
![Page 73: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/73.jpg)
A miraculous representation of the eigenvaluecounting function
That is, if Nθ is the number of eigenangles of U between 0 andθ, then
Nθd=
N∑j=1
ξj
for a collection {ξj}Nj=1 of independent Bernoulli randomvariables.
![Page 74: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/74.jpg)
A miraculous representation of the eigenvaluecounting function
Recall Rains’ second theorem:
[U (N)]me.v .d .
=⊕
0≤j<m
U(⌈
N − jm
⌉),
So: if Nm,N(θ) denotes the number of eigenangles of Um in[0, θ), then
Nm,N(θ)d=
N∑j=1
ξj ,
for {ξj}Nj=1 independent Bernoulli random variables.
![Page 75: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/75.jpg)
A miraculous representation of the eigenvaluecounting function
Recall Rains’ second theorem:
[U (N)]me.v .d .
=⊕
0≤j<m
U(⌈
N − jm
⌉),
So: if Nm,N(θ) denotes the number of eigenangles of Um in[0, θ), then
Nm,N(θ)d=
N∑j=1
ξj ,
for {ξj}Nj=1 independent Bernoulli random variables.
![Page 76: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/76.jpg)
Consequences of the miracle
I From Bernstein’s inequality and the representation of Nm,N(θ) as∑Nj=1 ξj ,
P[∣∣Nm,N(θ)− ENm,N(θ)
∣∣ > t]≤ 2 exp
[−min
{t2
4σ2 ,t2
}],
where σ2 = VarNm,N(θ).
I ENm,N(θ) = Nθ2π (by rotation invariance).
I Var[N1,N(θ)
]≤ log(N) + 1 (e.g., via explicit computation with the
kernel of the determinantal point process), and so
Var(Nm,N(θ)
)=∑
0≤j<m
Var(N
1,⌈
N−jm
⌉(θ)
)≤ m
(log(
Nm
)+ 1).
![Page 77: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/77.jpg)
Consequences of the miracle
I From Bernstein’s inequality and the representation of Nm,N(θ) as∑Nj=1 ξj ,
P[∣∣Nm,N(θ)− ENm,N(θ)
∣∣ > t]≤ 2 exp
[−min
{t2
4σ2 ,t2
}],
where σ2 = VarNm,N(θ).
I ENm,N(θ) = Nθ2π (by rotation invariance).
I Var[N1,N(θ)
]≤ log(N) + 1 (e.g., via explicit computation with the
kernel of the determinantal point process), and so
Var(Nm,N(θ)
)=∑
0≤j<m
Var(N
1,⌈
N−jm
⌉(θ)
)≤ m
(log(
Nm
)+ 1).
![Page 78: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/78.jpg)
Consequences of the miracle
I From Bernstein’s inequality and the representation of Nm,N(θ) as∑Nj=1 ξj ,
P[∣∣Nm,N(θ)− ENm,N(θ)
∣∣ > t]≤ 2 exp
[−min
{t2
4σ2 ,t2
}],
where σ2 = VarNm,N(θ).
I ENm,N(θ) = Nθ2π (by rotation invariance).
I Var[N1,N(θ)
]≤ log(N) + 1 (e.g., via explicit computation with the
kernel of the determinantal point process), and so
Var(Nm,N(θ)
)=∑
0≤j<m
Var(N
1,⌈
N−jm
⌉(θ)
)≤ m
(log(
Nm
)+ 1).
![Page 79: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/79.jpg)
Consequences of the miracle
I From Bernstein’s inequality and the representation of Nm,N(θ) as∑Nj=1 ξj ,
P[∣∣Nm,N(θ)− ENm,N(θ)
∣∣ > t]≤ 2 exp
[−min
{t2
4σ2 ,t2
}],
where σ2 = VarNm,N(θ).
I ENm,N(θ) = Nθ2π (by rotation invariance).
I Var[N1,N(θ)
]≤ log(N) + 1 (e.g., via explicit computation with the
kernel of the determinantal point process), and so
Var(Nm,N(θ)
)=∑
0≤j<m
Var(N
1,⌈
N−jm
⌉(θ)
)≤ m
(log(
Nm
)+ 1).
![Page 80: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/80.jpg)
The concentration of Nm,N leads to concentration of individualeigenvalues about their predicted values:
P[∣∣∣∣θj −
2πjN
∣∣∣∣ > 4πtN
]≤ 4 exp
−min
t2
m(
log(
Nm
)+ 1) , t ,
for each j ∈ {1, . . . ,N}:
P[θj >
2πjN
+4πN
u]
= P[N (m)
2π(j+2u)N
< j]
= P[j + 2u −N (m)
2π(j+2u)N
> 2u]
≤ P[∣∣∣∣N (m)
2π(j+2u)N
− EN (m)2π(j+2u)
N
∣∣∣∣ > 2u].
![Page 81: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/81.jpg)
The concentration of Nm,N leads to concentration of individualeigenvalues about their predicted values:
P[∣∣∣∣θj −
2πjN
∣∣∣∣ > 4πtN
]≤ 4 exp
−min
t2
m(
log(
Nm
)+ 1) , t ,
for each j ∈ {1, . . . ,N}:
P[θj >
2πjN
+4πN
u]
= P[N (m)
2π(j+2u)N
< j]
= P[j + 2u −N (m)
2π(j+2u)N
> 2u]
≤ P[∣∣∣∣N (m)
2π(j+2u)N
− EN (m)2π(j+2u)
N
∣∣∣∣ > 2u].
![Page 82: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/82.jpg)
The concentration of Nm,N leads to concentration of individualeigenvalues about their predicted values:
P[∣∣∣∣θj −
2πjN
∣∣∣∣ > 4πtN
]≤ 4 exp
−min
t2
m(
log(
Nm
)+ 1) , t ,
for each j ∈ {1, . . . ,N}:
P[θj >
2πjN
+4πN
u]
= P[N (m)
2π(j+2u)N
< j]
= P[j + 2u −N (m)
2π(j+2u)N
> 2u]
≤ P[∣∣∣∣N (m)
2π(j+2u)N
− EN (m)2π(j+2u)
N
∣∣∣∣ > 2u].
![Page 83: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/83.jpg)
The concentration of Nm,N leads to concentration of individualeigenvalues about their predicted values:
P[∣∣∣∣θj −
2πjN
∣∣∣∣ > 4πtN
]≤ 4 exp
−min
t2
m(
log(
Nm
)+ 1) , t ,
for each j ∈ {1, . . . ,N}:
P[θj >
2πjN
+4πN
u]
= P[N (m)
2π(j+2u)N
< j]
= P[j + 2u −N (m)
2π(j+2u)N
> 2u]
≤ P[∣∣∣∣N (m)
2π(j+2u)N
− EN (m)2π(j+2u)
N
∣∣∣∣ > 2u].
![Page 84: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/84.jpg)
Bounding EWp(µm,N , ν)
If νN := 1N∑N
j=1 δexp(
i 2πjN
), then Wp(νN , ν) ≤ πN and
EW pp (µm,N , νN) ≤ 1
N
N∑j=1
E∣∣∣∣θj −
2πjN
∣∣∣∣p
≤ 8Γ(p + 1)
4π√
m[log(
Nm
)+ 1]
N
p
,
using the concentration result and Fubini’s theorem.
![Page 85: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/85.jpg)
Bounding EWp(µm,N , ν)
If νN := 1N∑N
j=1 δexp(
i 2πjN
), then Wp(νN , ν) ≤ πN and
EW pp (µm,N , νN) ≤ 1
N
N∑j=1
E∣∣∣∣θj −
2πjN
∣∣∣∣p
≤ 8Γ(p + 1)
4π√
m[log(
Nm
)+ 1]
N
p
,
using the concentration result and Fubini’s theorem.
![Page 86: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/86.jpg)
Bounding EWp(µm,N , ν)
If νN := 1N∑N
j=1 δexp(
i 2πjN
), then Wp(νN , ν) ≤ πN and
EW pp (µm,N , νN) ≤ 1
N
N∑j=1
E∣∣∣∣θj −
2πjN
∣∣∣∣p
≤ 8Γ(p + 1)
4π√
m[log(
Nm
)+ 1]
N
p
,
using the concentration result and Fubini’s theorem.
![Page 87: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/87.jpg)
Concentration of Wp(µm,N , ν)
The Idea: Consider the function Fp(U) = Wp (µUm , ν), whereµUm is the empirical spectral measure of Um.
I By Rains’ theorem, it is distributionally the same asFp(U1, . . . ,Um) =
(1m∑m
j=1 µUj , ν)
.
I Fp(U1, . . . ,Um) is Lipschitz (w.r.t. the L2 sum of the
Euclidean metrics) with Lipschitz constant N−1
max(p,2) .
I If we had a general concentration phenomenon on⊕0≤j<m U
(⌈N−jm
⌉), concentration of Wp (µUm , ν) would
follow.
![Page 88: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/88.jpg)
Concentration of Wp(µm,N , ν)
The Idea: Consider the function Fp(U) = Wp (µUm , ν), whereµUm is the empirical spectral measure of Um.
I By Rains’ theorem, it is distributionally the same asFp(U1, . . . ,Um) =
(1m∑m
j=1 µUj , ν)
.
I Fp(U1, . . . ,Um) is Lipschitz (w.r.t. the L2 sum of the
Euclidean metrics) with Lipschitz constant N−1
max(p,2) .
I If we had a general concentration phenomenon on⊕0≤j<m U
(⌈N−jm
⌉), concentration of Wp (µUm , ν) would
follow.
![Page 89: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/89.jpg)
Concentration of Wp(µm,N , ν)
The Idea: Consider the function Fp(U) = Wp (µUm , ν), whereµUm is the empirical spectral measure of Um.
I By Rains’ theorem, it is distributionally the same asFp(U1, . . . ,Um) =
(1m∑m
j=1 µUj , ν)
.
I Fp(U1, . . . ,Um) is Lipschitz (w.r.t. the L2 sum of the
Euclidean metrics) with Lipschitz constant N−1
max(p,2) .
I If we had a general concentration phenomenon on⊕0≤j<m U
(⌈N−jm
⌉), concentration of Wp (µUm , ν) would
follow.
![Page 90: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/90.jpg)
Concentration of Wp(µm,N , ν)
The Idea: Consider the function Fp(U) = Wp (µUm , ν), whereµUm is the empirical spectral measure of Um.
I By Rains’ theorem, it is distributionally the same asFp(U1, . . . ,Um) =
(1m∑m
j=1 µUj , ν)
.
I Fp(U1, . . . ,Um) is Lipschitz (w.r.t. the L2 sum of the
Euclidean metrics) with Lipschitz constant N−1
max(p,2) .
I If we had a general concentration phenomenon on⊕0≤j<m U
(⌈N−jm
⌉), concentration of Wp (µUm , ν) would
follow.
![Page 91: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/91.jpg)
Concentration of Wp(µm,N , ν)
The Idea: Consider the function Fp(U) = Wp (µUm , ν), whereµUm is the empirical spectral measure of Um.
I By Rains’ theorem, it is distributionally the same asFp(U1, . . . ,Um) =
(1m∑m
j=1 µUj , ν)
.
I Fp(U1, . . . ,Um) is Lipschitz (w.r.t. the L2 sum of the
Euclidean metrics) with Lipschitz constant N−1
max(p,2) .
I If we had a general concentration phenomenon on⊕0≤j<m U
(⌈N−jm
⌉), concentration of Wp (µUm , ν) would
follow.
![Page 92: Random Unitary Matrices and Friends](https://reader031.vdocuments.us/reader031/viewer/2022020701/61f97ab1c91baf5d4b098bc3/html5/thumbnails/92.jpg)
Concentration on U (N1)⊕ · · · ⊕ U (Nk)
Theorem (E. M./M. Meckes)Given N1, . . . ,Nk ∈ N, denote by M = U (N1)× · · ·U (Nk )equipped with the L2-sum of Hilbert–Schmidt metrics.
Suppose that F : M → R is L-Lipschitz, and that Uj ∈ U(Nj)
are independent, uniform random unitary matrices, for1 ≤ j ≤ k. Then for each t > 0,
P[F (U1, . . . ,Uk ) ≥ EF (U1, . . . ,Uk ) + t
]≤ e−Nt2/12L2
,
where N = min{N1, . . . ,Nk}.