Gaussian ensemble
In-game article clicks load inline without leaving the challenge.
In random matrix theory, the Gaussian ensembles are specific probability distributions over self-adjoint matrices whose entries are independently sampled from the gaussian distribution. They are among the most-commonly studied matrix ensembles, fundamental to both mathematics and physics. The three main examples are the Gaussian orthogonal (GOE), unitary (GUE), and symplectic (GSE) ensembles. These are classified by the Dyson index β, which takes values 1, 2, and 4 respectively, counting the number of real components per matrix element (1 for real elements, 2 for complex elements, 4 for quaternions). The index can be extended to take any real positive value.
The gaussian ensembles are also called the Wigner ensembles, or the Hermite ensembles.
Definitions
Conventions
There are many conventions for defining the Gaussian ensembles. In this article, we specify exactly one of them.
In all definitions, the Gaussian ensemble have zero expectation.
- β {\displaystyle \beta }: a positive real number. Called the Dyson index. The cases of β = 1 , 2 , 4 {\displaystyle \beta =1,2,4} are special.
- N {\displaystyle N}: the side-length of a matrix. Always a positive integer.
- W N {\displaystyle W_{N}}: a matrix sampled from a Gaussian ensemble with size N × N {\displaystyle N\times N}. The letter W {\displaystyle W} stands for "Wigner".
- M ∗ {\displaystyle M^{*}}: the adjoint of a matrix. We assume W N = W N ∗ {\displaystyle W_{N}=W_{N}^{*}} (self-adjoint) when W N {\displaystyle W_{N}} is sampled from a gaussian ensemble. If M {\displaystyle M} is real, then M ∗ {\displaystyle M^{*}} is its transpose. If M {\displaystyle M} is complex or quaternionic, then M ∗ {\displaystyle M^{*}} is its conjugate transpose.
- λ 1 , … , λ N {\displaystyle \lambda _{1},\dots ,\lambda _{N}}: the eigenvalues of the matrix, which are all real, since the matrices are always assumed to be self-adjoint.
- σ d 2 {\displaystyle \sigma _{d}^{2}}: the variance of on-diagonal matrix entries. We assume that for each N {\displaystyle N}, all on-diagonal matrix entries have the same variance. It is always defined as E [ | W N , | 2 ] {\displaystyle \mathbb {E} [|W_{N,}|^{2}]}.
- σ o d 2 {\displaystyle \sigma _{od}^{2}}: the variance of off-diagonal matrix entries. We assume that for each N {\displaystyle N}, all off-diagonal matrix entries have the same variance. It is always defined as E [ | W N , i j | 2 ] {\displaystyle \mathbb {E} [|W_{N,ij}|^{2}]} where i ≠ j {\displaystyle i\neq j}. For a complex number, | a + b i | 2 = a 2 + b 2 {\displaystyle |a+bi|^{2}=a^{2}+b^{2}}. For a quaternion, | a + b i + c j + d k | 2 = a 2 + b 2 + c 2 + d 2 {\displaystyle |a+bi+cj+dk|^{2}=a^{2}+b^{2}+c^{2}+d^{2}}.
- Z {\displaystyle Z}: the partition function.
| Name | GOE(N) | GUE(N) | GSE(N) | GβE(N) |
|---|---|---|---|---|
| Full name | Gaussian orthogonal ensemble | Gaussian unitary ensemble | Gaussian symplectic ensemble | Gaussian beta ensemble |
| β {\displaystyle \beta } | 1 | 2 | 4 | β |
| σ d 2 {\displaystyle \sigma _{d}^{2}} | 2 | 1 | 1/2 | 2/β |
| σ o d 2 {\displaystyle \sigma _{od}^{2}} | 1 | 1 | 1 | 1 |
| matrix density | 1 Z e − 1 4 T r W N 2 {\displaystyle {\frac {1}{Z}}e^{-{\frac {1}{4}}\mathrm {Tr} W_{N}^{2}}} | 1 Z e − 1 2 T r W N 2 {\displaystyle {\frac {1}{Z}}e^{-{\frac {1}{2}}\mathrm {Tr} W_{N}^{2}}} | 1 Z e − T r W N 2 {\displaystyle {\frac {1}{Z}}e^{-\mathrm {Tr} W_{N}^{2}}} | 1 Z e − 1 4 β T r W N 2 {\displaystyle {\frac {1}{Z}}e^{-{\frac {1}{4}}\beta \mathrm {Tr} W_{N}^{2}}} |
| Z {\displaystyle Z} | 2 1 4 N ( N + 3 ) π 1 4 N ( N + 1 ) {\displaystyle 2^{{\frac {1}{4}}N(N+3)}\pi ^{{\frac {1}{4}}N(N+1)}} | 2 1 2 N π 1 2 N 2 {\displaystyle 2^{{\frac {1}{2}}N}\pi ^{{\frac {1}{2}}N^{2}}} | 2 − N ( N − 1 ) π 1 2 N ( 2 N − 1 ) {\displaystyle 2^{-N(N-1)}\pi ^{{\frac {1}{2}}N(2N-1)}} | 2 1 2 N ( 2 π β ) 1 2 N + 1 4 β N ( N − 1 ) {\displaystyle 2^{{\frac {1}{2}}N}\left({\frac {2\pi }{\beta }}\right)^{{\frac {1}{2}}N+{\frac {1}{4}}\beta N(N-1)}} |
When referring to the main reference works, it is necessary to translate the formulas from them, since each convention leads to different constant scaling factors for the formulas.
| Name | σ d 2 {\displaystyle \sigma _{d}^{2}} | σ o d 2 {\displaystyle \sigma _{od}^{2}} |
|---|---|---|
| Wikipedia (this page) | 2/β | 1 |
| (Deift 2000) (β = 2 only) | 1/2 | 1/2 |
| (Mehta 2004) | 1/β | 1/2 |
| (Anderson, Guionnet & Zeitouni 2010) | 2/β | 1 |
| (Forrester 2010) for β = 1, 2, 4 | 1/β | 1/2 |
| (Forrester 2010) for β ≠ 1, 2, 4 | 1 | β/2 |
| (Tao 2012) (β = 2 only) | 1 | 1 |
| (Mingo & Speicher 2017) (β = 2 only) | 1/N | 1/N |
| (Livan, Novaes & Vivo 2018) | 1 | β/2 |
| (Potters & Bouchaud 2020) | 2 σ 2 β N {\displaystyle {\frac {2\sigma ^{2}}{\beta N}}} | σ 2 N {\displaystyle {\frac {\sigma ^{2}}{N}}} |
There are equivalent definitions for the GβE(N) ensembles, given below.
By sampling
For all β = 1 , 2 , 4 {\displaystyle \beta =1,2,4} cases, the GβE(N) ensemble is defined by how it is sampled:
- Sample a gaussian matrix X N {\displaystyle X_{N}}, such that all its entries are IID sampled from the corresponding standard normal distribution. If β = 1 {\displaystyle \beta =1}, then X N , i j ∼ N ( 0 , 1 ) {\displaystyle X_{N,ij}\sim {\mathcal {N}}(0,1)}. If β = 2 {\displaystyle \beta =2}, then X N , i j ∼ N ( 0 , 1 / 2 ) + i N ( 0 , 1 / 2 ) {\displaystyle X_{N,ij}\sim {\mathcal {N}}(0,1/2)+i{\mathcal {N}}(0,1/2)}. If β = 4 {\displaystyle \beta =4}, then X N , i j ∼ N ( 0 , 1 / 4 ) + i N ( 0 , 1 / 4 ) + j N ( 0 , 1 / 4 ) + k N ( 0 , 1 / 4 ) {\displaystyle X_{N,ij}\sim {\mathcal {N}}(0,1/4)+i{\mathcal {N}}(0,1/4)+j{\mathcal {N}}(0,1/4)+k{\mathcal {N}}(0,1/4)}.
- Let W N = 1 2 ( X + X ∗ ) {\displaystyle W_{N}={\frac {1}{\sqrt {2}}}(X+X^{*})}.
By density
For all β = 1 , 2 , 4 {\displaystyle \beta =1,2,4} cases, the GβE(N) ensemble is defined with density function ρ ( W N ) = 1 Z e − β 4 ∑ i = 1 N W N , i i 2 − β 2 ∑ 1 ≤ i < j ≤ N | W N , i j | 2 = 1 Z e − β 4 T r W N 2 {\displaystyle \rho (W_{N})={\frac {1}{Z}}e^{-{\frac {\beta }{4}}\sum _{i=1}^{N}W_{N,ii}^{2}-{\frac {\beta }{2}}\sum _{1\leq i<j\leq N}|W_{N,ij}|^{2}}={\frac {1}{Z}}e^{-{\frac {\beta }{4}}\mathrm {Tr} W_{N}^{2}}}where the partition function is Z = 2 1 2 N ( 2 π β ) 1 2 N + 1 4 β N ( N − 1 ) {\displaystyle Z=2^{{\frac {1}{2}}N}\left({\frac {2\pi }{\beta }}\right)^{{\frac {1}{2}}N+{\frac {1}{4}}\beta N(N-1)}}.
The Gaussian orthogonal ensemble GOE(N) is defined as the probability distribution over N × N {\displaystyle N\times N} symmetric matrices with density functionρ ( W N ) = 1 Z e − 1 4 ∑ i = 1 N W N , i i 2 − 1 2 ∑ 1 ≤ i < j ≤ N W N , i j 2 = 1 Z e − 1 4 T r W N 2 {\displaystyle \rho (W_{N})={\frac {1}{Z}}e^{-{\frac {1}{4}}\sum _{i=1}^{N}W_{N,ii}^{2}-{\frac {1}{2}}\sum _{1\leq i<j\leq N}W_{N,ij}^{2}}={\frac {1}{Z}}e^{-{\frac {1}{4}}\mathrm {Tr} W_{N}^{2}}}where the partition function is Z = 2 1 4 N ( N + 3 ) π 1 4 N ( N + 1 ) {\displaystyle Z=2^{{\frac {1}{4}}N(N+3)}\pi ^{{\frac {1}{4}}N(N+1)}}.
Explicitly, since there are only 1 2 N ( N + 1 ) {\displaystyle {\frac {1}{2}}N(N+1)} degrees of freedom, the parameterization is as follows:ρ ( W N ) ∏ 1 ≤ i ≤ j ≤ N d W N , i j {\displaystyle \rho (W_{N})\prod _{1\leq i\leq j\leq N}dW_{N,ij}}where we pick the upper diagonal entries { W i j } 1 ≤ i ≤ j ≤ N {\displaystyle \{W_{ij}\}_{1\leq i\leq j\leq N}} as the degrees of freedom.
The Gaussian unitary ensemble GUE(N) is defined as the probability distribution over N × N {\displaystyle N\times N} Hermitian matrices with density functionρ ( W N ) = 1 Z e − 1 2 ∑ i = 1 N W N , i i 2 − ∑ 1 ≤ i < j ≤ N | W N , i j | 2 = 1 Z e − 1 2 T r W N 2 . {\displaystyle \rho (W_{N})={\frac {1}{Z}}e^{-{\frac {1}{2}}\sum _{i=1}^{N}W_{N,ii}^{2}-\sum _{1\leq i<j\leq N}|W_{N,ij}|^{2}}={\frac {1}{Z}}e^{-{\frac {1}{2}}\mathrm {Tr} \,W_{N}^{2}}.}where the partition function is Z = 2 1 2 N π 1 2 N 2 {\displaystyle Z=2^{{\frac {1}{2}}N}\pi ^{{\frac {1}{2}}N^{2}}}.
Explicitly, since there are only N 2 {\displaystyle N^{2}} degrees of freedom, the parameterization is as follows: ρ ( W N ) ∏ i = 1 N d W N , i i ∏ 1 ≤ i < j ≤ N d ( R e W N , i j ) d ( I m W N , i j ) {\displaystyle \rho (W_{N})\,\prod _{i=1}^{N}dW_{N,ii}\;\prod _{1\leq i<j\leq N}d(\mathrm {Re} \,W_{N,ij})\,d(\mathrm {Im} \,W_{N,ij})}where we pick the upper diagonal entries { W N , i i } 1 ≤ i ≤ N ∪ { R e W N , i j , I m W N , i j } 1 ≤ i < j ≤ N {\displaystyle \{W_{N,ii}\}_{1\leq i\leq N}\cup \{\mathrm {Re} \,W_{N,ij},\,\mathrm {Im} \,W_{N,ij}\}_{1\leq i<j\leq N}} as the degrees of freedom.
The Gaussian symplectic ensemble GSE(N) is defined as the probability distribution over N × N {\displaystyle N\times N} self‑adjoint quaternionic matrices with density functionρ ( W N ) = 1 Z e − ∑ i = 1 N W N , i i 2 − 2 ∑ 1 ≤ i < j ≤ N | W N , i j | 2 = 1 Z e − T r W N 2 . {\displaystyle \rho (W_{N})={\frac {1}{Z}}e^{-\sum _{i=1}^{N}W_{N,ii}^{2}-2\sum _{1\leq i<j\leq N}|W_{N,ij}|^{2}}={\frac {1}{Z}}e^{-\mathrm {Tr} \,W_{N}^{2}}.}where the partition function is Z = 2 − N ( N − 1 ) π 1 2 N ( 2 N − 1 ) {\displaystyle Z=2^{-N(N-1)}\pi ^{{\frac {1}{2}}N(2N-1)}}.
Explicitly, since there are only N ( 2 N − 1 ) {\displaystyle N(2N-1)} degrees of freedom, the parameterization is as follows:ρ ( W N ) ∏ i = 1 N d W N , i i ∏ 1 ≤ i < j ≤ N ∏ a = 0 3 d W N , i j ( a ) {\displaystyle \rho (W_{N})\,\prod _{i=1}^{N}dW_{N,ii}\;\prod _{1\leq i<j\leq N}\prod _{a=0}^{3}dW_{N,ij}^{(a)}}where we write W N , i j = W N , i j ( 0 ) + i W N , i j ( 1 ) + j W N , i j ( 2 ) + k W N , i j ( 3 ) {\displaystyle W_{N,ij}=W_{N,ij}^{(0)}+i\,W_{N,ij}^{(1)}+j\,W_{N,ij}^{(2)}+k\,W_{N,ij}^{(3)}} and pick the upper diagonal entries { W N , i i } 1 ≤ i ≤ N ∪ { W N , i j ( a ) } 1 ≤ i < j ≤ N , 0 ≤ a ≤ 3 {\displaystyle \{W_{N,ii}\}_{1\leq i\leq N}\cup \{W_{N,ij}^{(a)}\}_{1\leq i<j\leq N,\;0\leq a\leq 3}} as the degrees of freedom.
By invariance
For all β = 1 , 2 , 4 {\displaystyle \beta =1,2,4} cases, the GβE(N) ensemble is uniquely characterized (up to affine transform) by its symmetries, or invariance under appropriate transformations.
For GOE, consider a probability distribution over N × N {\displaystyle N\times N} symmetric matrices satisfying the following properties:
- Invariance under orthogonal transformation: For any fixed (not random) N × N {\displaystyle N\times N} orthogonal matrix O {\displaystyle O}, let M {\displaystyle M} be a random sample from the distribution. Then O M O T {\displaystyle OMO^{T}} has the same distribution as M {\displaystyle M}.
- Independence: The entries { M i j } 1 ≤ i ≤ j ≤ N {\displaystyle \{M_{ij}\}_{1\leq i\leq j\leq N}} are independently distributed.
For GUE, consider a probability distribution over N × N {\displaystyle N\times N} Hermitian matrices satisfying the following properties:
- Invariance under unitary transformation: For any fixed (not random) N × N {\displaystyle N\times N} unitary matrix U {\displaystyle U}, let M {\displaystyle M} be a random sample from the distribution. Then U M U ∗ {\displaystyle UMU^{*}} has the same distribution as M {\displaystyle M}.
- Independence: The entries { M i j } 1 ≤ i ≤ j ≤ N {\displaystyle \{M_{ij}\}_{1\leq i\leq j\leq N}} are independently distributed.
For GSE, consider a probability distribution over N × N {\displaystyle N\times N} self-adjoint quaternionic matrices satisfying the following properties:
- Invariance under symplectic transformation: For any fixed (not random) N × N {\displaystyle N\times N} symplectic matrix S {\displaystyle S}, let M {\displaystyle M} be a random sample from the distribution. Then S M S ∗ {\displaystyle SMS^{*}} has the same distribution as M {\displaystyle M}.
- Independence: The entries { M i j } 1 ≤ i ≤ j ≤ N {\displaystyle \{M_{ij}\}_{1\leq i\leq j\leq N}} are independently distributed.
In all 3 cases, these conditions force the distribution to have the form ρ ( M ) = 1 Z e − a Tr ( M 2 ) + b Tr ( M ) {\displaystyle \rho (M)={\frac {1}{Z}}e^{-a\operatorname {Tr} (M^{2})+b\operatorname {Tr} (M)}}, where a > 0 {\displaystyle a>0} and b , Z ∈ R {\displaystyle b,Z\in \mathbb {R} }. Thus, with the further specification of 1 N E [ Tr ( M ) ] = 0 , 1 N 2 E [ Tr ( M 2 ) ] = 1 + 2 / β − 1 N {\displaystyle {\frac {1}{N}}\mathbb {E} [\operatorname {Tr} (M)]=0,{\frac {1}{N^{2}}}\mathbb {E} [\operatorname {Tr} (M^{2})]=1+{\frac {2/\beta -1}{N}}}, we recover the GOE, GUE, GSE. Notably, if mere invariance is demanded, then any spectral distribution can be produced by multiplying with a function of form f ( Tr ( X ) , Tr ( X 2 ) , Tr ( X 3 ) , … ) {\displaystyle f(\operatorname {Tr} (X),\operatorname {Tr} (X^{2}),\operatorname {Tr} (X^{3}),\dots )}.
More succinctly stated, each of GOE, GUE, GSE is uniquely specified by invariance, independence, the mean, and the variance.
By spectral distribution
For all β = 1 , 2 , 4 {\displaystyle \beta =1,2,4} cases, the GβE(N) ensemble is defined as the ensemble obtained by A D A ∗ {\displaystyle ADA^{*}}, where
- D = diag ( λ 1 , … , λ N ) {\displaystyle D=\operatorname {diag} (\lambda _{1},\dots ,\lambda _{N})} is a diagonal real matrix with its entries sampled according to the spectral density, defined below;
- A {\displaystyle A} is an orthogonal/unitary/symplectic matrix sampled uniformly, that is, from the normalized Haar measure of the orthogonal/unitary/symplectic group.
In this way, the GβE(N) ensemble may be defined after the spectral density is defined first, so that any method to motivate the spectral density then motivates the GβE(N) ensemble, and vice versa.
By maximal entropy
For all β = 1 , 2 , 4 {\displaystyle \beta =1,2,4} cases, the GβE(N) ensemble is uniquely characterized as the absolutely continuous probability distribution ρ {\displaystyle \rho } over N × N {\displaystyle N\times N} real/complex/quaternionic symmetric/orthogonal/symplectic matrices that maximizes entropy E M ∼ ρ [ − ln ρ ( M ) ] {\displaystyle \mathbb {E} _{M\sim \rho }[-\ln \rho (M)]}, under the constraint of 1 N 2 E M ∼ ρ [ Tr ( M 2 ) ] = 1 + 2 / β − 1 N {\displaystyle {\frac {1}{N^{2}}}\mathbb {E} _{M\sim \rho }[\operatorname {Tr} (M^{2})]=1+{\frac {2/\beta -1}{N}}}.
Spectral density
For eigenvalues λ 1 , … , λ N {\displaystyle \lambda _{1},\dots ,\lambda _{N}} the joint density of GβE(N) isρ β , N ( λ 1 , … , λ N ) = 1 Z β , N e − β 4 ∑ i = 1 N λ i 2 ∏ 1 ≤ i < j ≤ N | λ i − λ j | β = 1 Z β , N e − β 4 ‖ λ ‖ 2 2 | Δ N ( λ ) | β {\displaystyle \rho _{\beta ,N}(\lambda _{1},\dots ,\lambda _{N})={\frac {1}{Z_{\beta ,N}}}e^{-{\frac {\beta }{4}}\sum _{i=1}^{N}\lambda _{i}^{2}}\prod _{1\leq i<j\leq N}|\lambda _{i}-\lambda _{j}|^{\beta }={\frac {1}{Z_{\beta ,N}}}e^{-{\frac {\beta }{4}}\|\lambda \|_{2}^{2}}|\Delta _{N}(\lambda )|^{\beta }}where Δ N {\displaystyle \Delta _{N}} is the Vandermonde determinant, and the partition function Z β , N {\displaystyle Z_{\beta ,N}} is explicitly evaluated as a Selberg integral:Z β , N = ∫ R N e − β 4 ∑ i = 1 N λ i 2 ∏ 1 ≤ i < j ≤ N | λ i − λ j | β d λ = ( 2 π ) N 2 ( 2 β ) 1 2 N + 1 4 β N ( N − 1 ) ∏ j = 1 N Γ ( 1 + j β 2 ) Γ ( 1 + β 2 ) {\displaystyle {\begin{aligned}Z_{\beta ,N}&=\int _{\mathbb {R} ^{N}}e^{-{\frac {\beta }{4}}\sum _{i=1}^{N}\lambda _{i}^{2}}\prod _{1\leq i<j\leq N}|\lambda _{i}-\lambda _{j}|^{\beta }d\lambda \\&=(2\pi )^{\frac {N}{2}}\left({\frac {2}{\beta }}\right)^{{\frac {1}{2}}N+{\frac {1}{4}}\beta N(N-1)}\prod _{j=1}^{N}{\frac {\Gamma \left(1+j{\frac {\beta }{2}}\right)}{\Gamma \left(1+{\frac {\beta }{2}}\right)}}\end{aligned}}}where Γ {\displaystyle \Gamma } is the Euler Gamma function. The expression is particularly simple when β = 2 {\displaystyle \beta =2}, where we have a superfactorial:Z 2 , N = ( 2 π ) N 2 ∏ j = 1 N j ! {\displaystyle Z_{2,N}=(2\pi )^{\frac {N}{2}}\prod _{j=1}^{N}j!}
Determinantal point process

Define functions ψ n ( x ) := e − 1 4 x 2 n ! 2 π He n ( x ) {\displaystyle \psi _{n}(x):={\frac {e^{-{\frac {1}{4}}x^{2}}}{\sqrt {n!{\sqrt {2\pi }}}}}\operatorname {He} _{n}(x)}, where He {\displaystyle \operatorname {He} } is the probabilist's Hermite polynomial. These are the wavefunction states of the quantum harmonic oscillator.
The spectrum of GUE(N) is a determinantal point process with kernel K N ( x , x ′ ) := ∑ n = 0 N − 1 ψ n ( x ) ψ n ( x ′ ) {\displaystyle K_{N}(x,x'):=\sum _{n=0}^{N-1}\psi _{n}(x)\psi _{n}(x')}, and by the Christoffel–Darboux formula,K N ( x , x ′ ) = e − 1 4 ( x 2 + x ′ 2 ) ( N − 1 ) ! 2 π He N ( x ) He N − 1 ( x ′ ) − He N − 1 ( x ) He N ( x ′ ) x − x ′ {\displaystyle K_{N}(x,x')={\frac {e^{-{\frac {1}{4}}\left(x^{2}+x^{\prime 2}\right)}}{(N-1)!{\sqrt {2\pi }}}}{\frac {\operatorname {He} _{N}(x)\operatorname {He} _{N-1}\left(x^{\prime }\right)-\operatorname {He} _{N-1}(x)\operatorname {He} _{N}\left(x^{\prime }\right)}{x-x^{\prime }}}}Using the confluent form of Christoffel–Darboux and the three-term recurrence of Hermite polynomials, the spectral density of GUE(N) for finite values of N {\displaystyle N}:ρ ( x ) = 1 N K N ( x , x ) = 1 N 2 π e − 1 2 x 2 ∑ n = 0 N − 1 1 n ! He n ( x ) 2 = e − x 2 / 2 2 π N ! ( He N ( x ) 2 − He N + 1 ( x ) He N − 1 ( x ) ) {\displaystyle {\begin{aligned}\rho (x)&={\frac {1}{N}}K_{N}(x,x)\\&={\frac {1}{N{\sqrt {2\pi }}}}e^{-{\frac {1}{2}}x^{2}}\sum _{n=0}^{N-1}{\frac {1}{n!}}\operatorname {He} _{n}(x)^{2}\\&={\frac {e^{-x^{2}/2}}{{\sqrt {2\pi }}N!}}\left(\operatorname {He} _{N}(x)^{2}-\operatorname {He} _{N+1}(x)\operatorname {He} _{N-1}(x)\right)\end{aligned}}}The spectral distribution of β = 1 , 4 {\displaystyle \beta =1,4} can also be written as a quaternionic determinantal point process involving skew-orthogonal polynomials.
Tridiagonalization
For all β = 1 , 2 , 4 {\displaystyle \beta =1,2,4} cases, given a sampled matrix W N {\displaystyle W_{N}} from the GβE(N) ensemble, we can perform a Householder transformation tridiagonalization on it to obtain a tridiagonal matrix T β , N {\displaystyle T_{\beta ,N}}, which has the same distribution as1 β [ a N 2 b N − 1 0 ⋯ 0 b N − 1 a N − 1 2 b N − 2 ⋱ ⋮ 0 b N − 2 ⋱ ⋱ 0 ⋮ ⋱ ⋱ a 2 2 b 1 0 ⋯ 0 b 1 a 1 2 ] {\displaystyle {\sqrt {\frac {1}{\beta }}}{\begin{bmatrix}a_{N}{\sqrt {2}}&b_{N-1}&0&\cdots &0\\b_{N-1}&a_{N-1}{\sqrt {2}}&b_{N-2}&\ddots &\vdots \\0&b_{N-2}&\ddots &\ddots &0\\\vdots &\ddots &\ddots &a_{2}{\sqrt {2}}&b_{1}\\0&\cdots &0&b_{1}&a_{1}{\sqrt {2}}\end{bmatrix}}}where each a 1 , … , a N ∼ N ( 0 , 1 ) {\displaystyle a_{1},\dots ,a_{N}\sim {\mathcal {N}}(0,1)} is gaussian-distributed, and each b i ∼ χ i β {\displaystyle b_{i}\sim \chi _{i\beta }} is chi-distributed, and all a 1 , … , a N , b 1 , … , b N − 1 {\displaystyle a_{1},\dots ,a_{N},b_{1},\dots ,b_{N-1}} are independent. The β = 1 {\displaystyle \beta =1} case was first noted in 1984, and the general case was noted in 2002. Like how the Laplace differential operator can be discretized to the Laplacian matrix, this tridiagonal form of the gaussian ensemble allows a reinterpretation of the gaussian ensembles as an ensemble over not matrices, but over differential operators, specifically, a "stochastic Airy operator". This leads more generally to the study of random matrices as stochastic operators.
Computationally, this allows efficient sampling of eigenvalues, from O ( N 3 ) {\displaystyle O(N^{3})} on the full matrix, to just O ( N 2 ) {\displaystyle O(N^{2})} on the tridiagonal matrix. If one only requires a histogram of the eigenvalues with m {\displaystyle m} bins, the time can be further decreased to O ( N m ) {\displaystyle O(Nm)}, by using the Sturm sequences. Theoretically, this definition allows extension to all β > 0 {\displaystyle \beta >0} cases, leading to the gaussian beta ensembles, and "anti-symmetric" gaussian beta ensembles.
Relatedly, let X N {\displaystyle X_{N}} be a N × N {\displaystyle N\times N} matrix, with all entries IID sampled from the corresponding standard normal distribution – for example, if β = 2 {\displaystyle \beta =2}, then X N , i j ∼ N ( 0 , 1 / 2 ) + i N ( 0 , 1 / 2 ) {\displaystyle X_{N,ij}\sim {\mathcal {N}}(0,1/2)+i{\mathcal {N}}(0,1/2)}. Then applying repeated Housholder transform on only the left side of a results in R N = H 1 … H N X N {\displaystyle R_{N}=H_{1}\dots H_{N}X_{N}}, where each H i {\displaystyle H_{i}} is a Householder matrix, and R N {\displaystyle R_{N}} is an upper triangular matrix with independent entries, such that each β R N , i i ∼ χ N + 1 − i {\displaystyle {\sqrt {\beta }}R_{N,ii}\sim \chi _{N+1-i}} for 1 ≤ i ≤ N {\displaystyle 1\leq i\leq N}, and each R N , i j ∼ N ( 0 , 1 / β ) ⊗ β {\displaystyle R_{N,ij}\sim {\mathcal {N}}(0,1/\beta )^{\otimes \beta }} for 1 ≤ i < j ≤ N {\displaystyle 1\leq i<j\leq N}.
Global law

The Wigner semicircle law states that the empirical eigenvalue distribution of 1 N W N {\displaystyle {\frac {1}{\sqrt {N}}}W_{N}} converges in distribution to the Wigner semicircle distribution with radius 2. That is, the distribution on [ − 2 , + 2 ] {\displaystyle [-2,+2]} with probability density function ρ s c ( x ) = 4 − x 2 2 π {\displaystyle \rho _{sc}(x)={\frac {\sqrt {4-x^{2}}}{2\pi }}}
The requirement that the matrix ensemble to be a gaussian ensemble is too strong for the Wigner semicircle law. Indeed, the theorem applies generally for much more generic matrix ensembles.
As Coulomb gas
The joint density ρ β , N {\displaystyle \rho _{\beta ,N}} can be written as a Gibbs measure:ρ β , N = 1 Z β , N CG e − β E N {\displaystyle \rho _{\beta ,N}={\frac {1}{Z_{\beta ,N}^{\text{CG}}}}e^{-\beta E_{N}}}with the energy function (also called the Hamiltonian) E N = 1 4 ∑ i = 1 N λ i 2 − ∑ 1 ≤ i < j ≤ N ln | λ i − λ j | {\displaystyle E_{N}={\frac {1}{4}}\sum _{i=1}^{N}\lambda _{i}^{2}-\sum _{1\leq i<j\leq N}\ln |\lambda _{i}-\lambda _{j}|}. This can be interpreted physically as a Boltzmann distribution of a physical system consisting of N {\displaystyle N} identical unit electric charges constrained to move on the real line, repelling each other via the two-dimensional Coulomb potential − ln | x − y | {\displaystyle -\ln |x-y|}, while being attracted to the origin via a quadratic potential 1 4 x 2 {\displaystyle {\frac {1}{4}}x^{2}}. This is the Coulomb gas model for the eigenvalues.
In the macroscopic limit, one rescales λ i = N 1 / 2 x i {\displaystyle \lambda _{i}=N^{1/2}x_{i}} and defines the empirical measure μ N = N − 1 ∑ i = 1 N δ x i {\displaystyle \mu _{N}=N^{-1}\sum _{i=1}^{N}\delta _{x_{i}}}, obtaining E N ≈ 1 2 N 2 ( E [ μ ] + 1 2 ln N ) {\displaystyle E_{N}\approx {\frac {1}{2}}N^{2}\left({\mathcal {E}}[\mu ]+{\frac {1}{2}}\ln N\right)}, where the mean-field functional E [ μ ] = 1 2 ∫ R x 2 μ ( d x ) − ∬ R 2 ln | x − y | μ ( d x ) μ ( d y ) {\displaystyle {\mathcal {E}}[\mu ]={\frac {1}{2}}\int _{\mathbb {R} }x^{2}\mu (dx)-\iint _{\mathbb {R} ^{2}}\ln |x-y|\mu (dx)\mu (dy)}yields the leading order N 2 {\displaystyle N^{2}} term in ln Z β , N {\displaystyle \ln Z_{\beta ,N}}, termed the Coulomb gas free energy. The Coulomb gas free energy is minimized by the Wigner semicircle law d μ s c ( x ) = ( 2 π ) − 1 4 − x 2 1 { | x | ≤ 2 } d x {\displaystyle d\mu _{sc}(x)=(2\pi )^{-1}{\sqrt {4-x^{2}}}1_{\{|x|\leq 2\}}dx}, which gives the limiting eigenvalue density.
Alternatively, suppose that there exists a ρ b {\displaystyle \rho _{b}}, such that the quadratic electric potential can be recreated (up to an additive constant) via∫ − 2 N 2 N − ln | x − y | ρ b ( y ) d y = 1 4 x 2 + C , x ∈ [ − 2 N , 2 N ] . {\displaystyle \int _{-2{\sqrt {N}}}^{2{\sqrt {N}}}-\ln |x-y|\rho _{b}(y)dy={\frac {1}{4}}x^{2}+C,\quad x\in [-2{\sqrt {N}},2{\sqrt {N}}].}Then, imposing a fixed background negative electric charge of density | ρ b ( y ) | {\displaystyle |\rho _{b}(y)|} exactly cancels out the electric repulsion between the freely moving positive charges. Such a function does exist: ρ b ( y ) = − 4 N − y 2 2 N π {\displaystyle \rho _{b}(y)=-{\frac {\sqrt {4N-y^{2}}}{2N\pi }}}, which can be found by solving an integral equation. This indicates that the Wigner semicircle distribution is the equilibrium distribution.
Gaussian fluctuations about μ s c {\displaystyle \mu _{sc}} obtained by expanding E N {\displaystyle E_{N}} to second order produce the sine kernel in the bulk and the Airy kernel at the soft edge after proper rescaling.
Extreme value
The largest eigenvalue for GβE(N) follows the Tracy–Widom distribution after proper translation and scaling. It can be efficiently sampled by the shift-invert Lanczos algorithm on the 10 n 1 / 3 × 10 n 1 / 3 {\displaystyle 10n^{1/3}\times 10n^{1/3}} upper left corner of the tridiagonal matrix form.
Level spacings
From ordered eigenvalues λ 1 < ⋯ < λ n < λ n + 1 < ⋯ < λ N {\displaystyle \lambda _{1}<\dots <\lambda _{n}<\lambda _{n+1}<\dots <\lambda _{N}}, define normalized spacings s n = λ n + 1 − λ n ⟨ s ⟩ {\displaystyle s_{n}={\frac {\lambda _{n+1}-\lambda _{n}}{\langle s\rangle }}} with mean spacing ⟨ s ⟩ {\displaystyle \langle s\rangle }. This normalizes the spacings by:∫ 0 ∞ p β ( s ) d s = 1 , ∫ 0 ∞ s p β ( s ) d s = 1 , β = 1 , 2 , 4. {\displaystyle \int _{0}^{\infty }p_{\beta }(s)\,ds=1,\qquad \int _{0}^{\infty }s\,p_{\beta }(s)\,ds=1,\qquad \beta =1,2,4.}With this, the approximate spacing distributions are p β ( s ) = { π 2 s exp ( − π 4 s 2 ) β = 1 32 π 2 s 2 exp ( − 4 π s 2 ) β = 2 2 18 3 6 π 3 s 4 exp ( − 64 9 π s 2 ) β = 4 {\displaystyle p_{\beta }(s)={\begin{cases}{\frac {\pi }{2}}s\exp \left(-{\frac {\pi }{4}}s^{2}\right)&\beta =1\\{\frac {32}{\pi ^{2}}}s^{2}\exp \left(-{\frac {4}{\pi }}s^{2}\right)&\beta =2\\{\frac {2^{18}}{3^{6}\pi ^{3}}}s^{4}\exp \left(-{\frac {64}{9\pi }}s^{2}\right)&\beta =4\\\end{cases}}}
Moments
For GOE(N), its moment generating function is E [ e Tr ( V W N ) ] = e 1 4 ‖ V + V T ‖ F 2 {\textstyle \mathbb {E} \left[e^{\operatorname {Tr} (VW_{N})}\right]=e^{{\frac {1}{4}}\|V+V^{\text{T}}\|_{F}^{2}}}, where ‖ ⋅ ‖ F {\displaystyle \|\cdot \|_{F}} is the Frobenius norm.
Rotationally symmetric ensemble
The GUE(N) ensemble can be understood as a particularly representative member of the general family of rotationally symmetric ensembles over Hermitian matrices. The general theory of rotationally symmetric ensembles has allowed researchers to prove some universality results.
Physics
The Wigner surmise
The Gaussian ensemble was first motivated in theoretical physics. In the 1940s, Eugene Wigner studied the irregular spacings of slow-neutron resonances in heavy nuclei. Working with the few dozen levels then available, he noticed a pronounced repulsion between neighbouring lines.
In 1951, he modelled the Hamiltonian of a compound-nucleus in a minimal way. He noted that by symmetry considerations, it must be a real symmetric operator, so he modelled it as a random sample from the GOE(N). He solved the 2×2 case and found the two-level spacing law P ( s ) = π 2 s e − π s 2 / 4 {\displaystyle P(s)={\frac {\pi }{2}}se^{-\pi s^{2}/4}}, which matched well with the data. He disseminated his guess ("the Wigner surmise") during a conference on Neutron Physics by Time-of-Flight in 1956:
Perhaps I am now too courageous when I try to guess the distribution of the distances between successive levels (of energies of heavy nuclei). Theoretically, the situation is quite simple if one attacks the problem in a simpleminded fashion. The question is simply what are the distances of the characteristic values of a symmetric matrix with random coefficients.
— Eugene Wigner, Results and theory of resonance absorption
Freeman Dyson stated the project as a statistical theory of nuclear energy levels, to be contrasted with precise calculations based on an analytic model of the nucleus. He argued that a statistical theory is necessary, because the energy levels then measured were on the order of millions, and for such a high order, precise calculations was simply impossible. The idea was different from the then-understood form of statistical mechanics, for instead of having a system with precisely stated dynamical laws, with too many particles interacting under it, thus the particles need to be treated statistically, he would model the dynamical laws themselves as unknown, and thus treated statistically.
Threefold Way
In 1962, Dyson proposed the "Threefold Way" to motivate the three ensembles, by showing that in 3 fields (group representation, quantum mechanics, random matrix theory), there is a 3-fold disjunction, which he traced back to the Frobenius theorem stating that there are only 3 real division algebras: the real, the complex, and the quaternionic. A random matrix representing a Hamiltonian H {\displaystyle H} can be classified by an anti-unitary operator T {\displaystyle T} that describes time-reversal symmetry. The classification depends on whether T {\displaystyle T} exists present and, if so, the value of T 2 {\displaystyle T^{2}}. Each symmetry class produces a constraint on the possible form of H {\displaystyle H}, and the corresponding gaussian ensemble can then be motivated as a maximal entropy distribution, as described previously.
| Symmetry | Matrix basis where H {\displaystyle H} is... | Group representation | Ensemble |
|---|---|---|---|
| T 2 = + 1 {\displaystyle T^{2}=+1} (e.g., integer spin, no strong spin–orbit interaction) | real symmetric | real | GOE |
| No T {\displaystyle T} (e.g., presence of a magnetic field, magnetic impurities, chiral gauge potential) | complex Hermitian | complex | GUE |
| T 2 = − 1 {\displaystyle T^{2}=-1} (e.g., half-integer spin with spin-orbit interaction) | quaternionic self-adjoint (symplectic) | pseudoreal | GSE |
If T 2 = + 1 {\displaystyle T^{2}=+1}, the Hamiltonian H {\displaystyle H} must be real symmetric. This typically occurs in systems with no magnetic field and either spinless particles or integer spin particles with negligible spin–orbit interaction. This occurs in level spacing distribution in nuclear compound states, the original motivation for Wigner.
If T {\displaystyle T} does not exist, then H {\displaystyle H} is only required to be Hermitian. Time-reversal symmetry can be broken by a homogeneous magnetic field, random magnetic fluxes, or spin-selective lasers. In these cases, the off-diagonal matrix elements acquire independent complex phases.
- Chaotic microwave cavities with a ferrite: Adding a strong axial magnetic field causes the level statistics to transition continuously from GOE to GUE, which was a confirmation of the BGS conjecture.
- Quantum Hall effect: The physics of quantum Hall edge states and Landau levels is modelled by the GUE due to the strong perpendicular magnetic field breaking time-reversal symmetry.
- Anderson localization in 3-D: Applying an Aharonov–Bohm flux can drive a system's statistics from GOE to GUE at a disorder-induced metal-insulator transition.
If T 2 = − 1 {\displaystyle T^{2}=-1}, then this is a consequence of Kramers' theorem for systems with half-integer spin and significant spin–orbit interaction. The resulting Hamiltonians are naturally described by quaternion-Hermitian matrices. It has been observed in Kramers doublet and many quantum chaotic systems. It is also possible to construct such a system without spin.
- Deift, Percy (2000). Orthogonal polynomials and random matrices: a Riemann-Hilbert approach. Courant lecture notes in mathematics. Providence, R.I: American Mathematical Society. ISBN 978-0-8218-2695-9.
- Mehta, M.L. (2004). Random Matrices. Amsterdam: Elsevier/Academic Press. ISBN 0-12-088409-7.
- Deift, Percy; Gioev, Dimitri (2009). Random matrix theory: invariant ensembles and universality. Courant lecture notes in mathematics. New York : Providence, R.I: Courant Institute of Mathematical Sciences; American Mathematical Society. ISBN 978-0-8218-4737-4.
- Forrester, Peter (2010). Log-gases and random matrices. London Mathematical Society monographs. Princeton: Princeton University Press. ISBN 978-0-691-12829-0.
- Anderson, G.W.; Guionnet, A.; Zeitouni, O. (2010). An introduction to random matrices. Cambridge: Cambridge University Press. ISBN 978-0-521-19452-5.
- Bai, Zhidong; Silverstein, Jack W. (2010). Spectral analysis of large dimensional random matrices. Springer series in statistics (2 ed.). New York; London: Springer. doi:. ISBN 978-1-4419-0660-1. ISSN .
- Akemann, Gernot; Baik, J.; Di Francesco, Philippe, eds. (2011). The Oxford Handbook of Random Matrix Theory. Oxford: Oxford University Press. ISBN 978-0-19-957400-1.
- Tao, Terence (2012). Topics in random matrix theory. Graduate studies in mathematics. Providence, R.I: American Mathematical Society. ISBN 978-0-8218-7430-1.
- Mingo, James A.; Speicher, Roland (2017). Free Probability and Random Matrices. Fields Institute Monographs. New York, NY: Springer. ISBN 978-1-4939-6942-5.
- Livan, Giacomo; Novaes, Marcel; Vivo, Pierpaolo (2018). . SpringerBriefs in Mathematical Physics. Vol. 26. Cham: Springer International Publishing. arXiv:. doi:. ISBN 978-3-319-70883-6.
- Potters, Marc; Bouchaud, Jean-Philippe (2020-11-30). A First Course in Random Matrix Theory: for Physicists, Engineers and Data Scientists. Cambridge University Press. doi:. ISBN 978-1-108-76890-0.