This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Random sorting networks: edge limit

Vadim Gorin    Jiaming Xu Departments of Statistics and Mathematics, University of California, Berkeley. [email protected] Department of Mathematics, University of Wisconsin - Madison. [email protected]
Abstract

A sorting network is a shortest path from 1 2n1\ 2\ \dots\ n to n 2 1n\ \dots\ 2\ 1 in the Cayley graph of the symmetric group 𝔖n\mathfrak{S}_{n} spanned by adjacent transpositions. The paper computes the edge local limit of the uniformly random sorting networks as nn\to\infty. We find the asymptotic distribution of the first occurrence of a given swap (k,k+1)(k,k+1) and identify it with the law of the smallest positive eigenvalue of a 2k×2k2k\times 2k aGUE (an aGUE matrix has purely imaginary Gaussian entries that are independently distributed subject to skew-symmetry). Next, we give two different formal definitions of a spacing — the time distance between the occurrence of a given swap (k,k+1)(k,k+1) in a uniformly random sorting network. Two definitions lead to two different expressions for the asymptotic laws expressed in terms of derivatives of Fredholm determinants.

Abstract

Un réseau de tri est un chemin le plus court de 1 2n1\ 2\ \dots\ n à n 2 1n\ \dots\ 2\ 1 dans le graphe de Cayley du groupe symétrique 𝔖n\mathfrak{S}_{n}, engendré par des transpositions des éléments adjacents. Dans cet article nous calculons la limite locale au bord des réseaux de tri choisi uniformément quand nn\to\infty. Nous trouvons la distribution asymptotique de la première occurrence d’une transposition donnée (k,k+1)(k,k+1) et l’identifions avec la loi de la plus petite valeur propre positive d’un 2k×2k2k\times 2k aGUE (une matrice aGUE a des entrées gaussiennes purement imaginaires qui sont distribuées indépendamment sous condition d’antisymétrie). Ensuite, nous considerons des espacements entre deux occurrences consecutives d’un échange donné (k, k + 1) pour un réseau de tri aléatoire choisi uniformément. Nous prenons deux formalisations pour un choix aléatoire d’un tel espacement. En passant à limite, ces deux définitions conduisent à deux expressions différentes pour des lois asymptotiques exprimées en termes de dérivées des déterminants de Fredholm.

60B20,
60B15,
Sorting network,
spacing,
antisymmetric Gaussian Unitary Ensemble,
keywords:
[class=MSC]
keywords:

1 Introduction

1.1 Motivation

The main object in this article is the uniformly random sorting network which we abbreviate as RSN. Let us start by giving basic definitions. Consider the permutation group of nn elements 𝔖n\mathfrak{S}_{n}. We use the one-row notation for the elements of 𝔖n\mathfrak{S}_{n} representing them as sequences (a1a2an)(a_{1}\ a_{2}\ ...\ a_{n}). We let τk\tau_{k} be the adjacent swap (k,k+1)(k,k+1), 1kn11\leq k\leq n-1, so that

(a1a2an)τk=(a1a2ak1ak+1akak+2an).(a_{1}\ a_{2}\ \dots\ a_{n})\cdot\tau_{k}=(a_{1}\ a_{2}\ a_{k-1}\ a_{k+1}\ a_{k}\ a_{k+2}\ \dots\ a_{n}).

A sorting network of size nn is a shortest sequence of swaps, whose product equals the reverse permutation (nn1 2 1)(n\ n-1\ ...\ 2\ 1). By counting inversions in permutations, one shows that the length NN of such shortest sequence is N=(n2)N=\binom{n}{2}. Let Ωn\Omega_{n} denote the set of all sorting networks of size nn. Thus, elements of Ωn\Omega_{n} are sequences (s1,,sN)(s_{1},\dots,s_{N}) such that

τs1τs2τsN=(nn1 2 1).\tau_{s_{1}}\tau_{s_{2}}\cdots\tau_{s_{N}}=(n\ n-1\ ...\ 2\ 1).

Sorting networks can be drawn as wiring diagrams, see Figure 1. Whenever st=ks_{t}=k, we say that swap (k,k+1)(k,k+1) occurs at time tt; in the wiring diagram this corresponds to two wires intersecting at the point (t,k+12)(t,k+\tfrac{1}{2}).

Refer to caption
Figure 1: A sorting network (s1sN)(s_{1}\dots s_{N}), N=(n2)N={n\choose 2} can be represented as a diagram of nn wires, with wires at heights kk and k+1k+1 being swapped at time ii whenever si=ks_{i}=k. The figure shows the wiring diagram of the sorting network (2,1,3,2,4,3,4,1,2,1)(2,1,3,2,4,3,4,1,2,1) with n=5n=5. The blue double arrow shows a spacing in row 3, which is a time interval between two adjacent swaps τ3\tau_{3} and it has length 3 in our example.

The study of sorting networks is a fruitful topic, with the first mathematical results going back to [S] where Stanley computed the number of elements in Ωn\Omega_{n}:

|Ωn|=(n2)!j=1n1(2n12j)j.|\Omega_{n}|=\frac{{n\choose 2}!}{\prod_{j=1}^{n-1}(2n-1-2j)^{j}}.

A recent point of interest is the study of uniformly random sorting networks initiated by Angel, Holroyd, Romik, and Virag in [AHRV]. The probabilistic results can be split into two groups: global and local. On the global side, [AHRV] proved that the density of swaps approximates the semi-circle law111There is no known direct connection to the Wigner semi-circle law of the random matrix theory and the match seems coincidental. as nn\to\infty; [AHRV] predicted and [DVi, D] proved (among other results) that individual trajectories in the wiring diagram become (random) sine curves; see also [K, RVV] for related results.

Our paper belongs to the second group of results, [AGH, R, ADHV, GR], which studies local limits. The most basic local characteristic of the sorting network is spacing in row kk, which is a distance between two occurrences of the same swap si=ks_{i}=k in the sorting network, cf. Figure 1. [R] and [GR] discovered a link between the asymptotic laws of spacings as nn\to\infty and random matrix theory. Namely, [R] matched the asymptotic law of the spacing in row 11 with the distance between two eigenvalues of 2×22\times 2 matrix of Gaussian Orthogonal Ensemble (GOE) of real symmetric random matrices. [GR] dealt with spacings in row αn\alpha n with 0<α<10<\alpha<1 and showed that after proper rescaling they converge to the Gaudin–Mehta law, which is the universal asymptotic law for spacings between eigenvalues of real symmetric matrices of very large sizes.

Comparing the results of [R] and [GR], we see that the former links the spacings in the extreme row (i.e. at the end-point of the edge) to 2×22\times 2 real symmetric matrices, while the latter links the spacings in the bulk rows to the real symmetric matrices of infinite sizes. The observed gap in the asymptotics between 2×22\times 2 and ×\infty\times\infty matrices motivated the central question of our paper: we would like to understand, whether distributions related to k×kk\times k random matrices can be also found in the asymptotic laws for random sorting networks.

The answer presented in this text is both Yes and No. On one side, we are so far unable to find any connections of sorting networks to eigenvalues of real symmetric k×kk\times k matrices (which would be the most direct interpolation between [R] and [GR]). On the other hand, by slightly adjusting our point view, we find that the law of the asymptotic spacing in row kk can be expressed in terms of the eigenvalues of 2k×2k2k\times 2k random anti-symmetric Gaussian matrices — they are known in the literature as aGUE, since their analysis reveals connections to the tools used for the Gaussian Unitary Ensemble222However, up to multiplication by 𝐢\mathbf{i}, matrix elements of aGUE are real, rather than complex., see [M, Chapter 13] and [FN].

1.2 Main Results

Here is the precise definition of the random matrix object appearing in our asymptotic results.

Definition 1.1.

Let YY be an ×\ell\times\ell matrix whose entries are i.i.d. standard Gaussian random variables N(0,1)N(0,1). Then

M=𝐢2(YYT)M_{\ell}=\frac{\mathbf{i}}{2}\,(Y-Y^{T})

is called ×\ell\times\ell anti-symmetric GUE or aGUE for being short.

Note that the eigenvalues of MM_{\ell} are real and come in pairs: λ\lambda is an eigenvalue if and only if so is λ-\lambda. When \ell is odd, MM_{\ell} is necessary degenerate, i.e. it has a zero eigenvalue; for even \ell, zero is not an eigenvalue almost surely.

We also would like to deal with eigenvalues of all MM_{\ell} together.

Definition 1.2.

Let MM be an infinite aGUE, i.e, an ×\infty\times\infty matrix whose lthl^{\text{th}} corner is a l×ll\times l aGUE. The aGUE–corners process is a point process X\pazocal{X} on 0×0\mathbb{Z}_{\geq 0}\times\mathbb{R}_{\geq 0}: we put a particle at (l,u)(l,u) whenever uu is one of the l2\lfloor\frac{l}{2}\rfloor positive eigenvalues of the top-left l×ll\times l corner of MM.

Remark 1.3.

aGUE is studied in [FN], and it turns out that its corners process X\pazocal{X} has a determinantal structure. We give the correlation kernel of X\pazocal{X} in Theorem 2.9.

Definition 1.4.

𝐓(k)\mathbf{T}(k) is the smallest positive eigenvalue of a 2k×2k2k\times 2k aGUE.

The determinantal structure for the distribution of eigenvalues of M2kM_{2k} (see [M, Chapter 13] and [FN]) leads to an expression for the distribution of 𝐓(k)\mathbf{T}(k) as a series, which can be identified with a Fredholm determinant:

[𝐓(k)>t]=1+l=1(1)ll!0t0tdet[K(k)(ui,uj)]du1dul,\mathbb{P}\left[\mathbf{T}(k)>t\right]=1+\sum_{l=1}^{\infty}\frac{(-1)^{l}}{l!}\int_{0}^{t}\dots\int_{0}^{t}\det[K^{(k)}(u_{i},u_{j})]\,\mathrm{d}u_{1}\cdots\mathrm{d}u_{l},

where K(k)K^{(k)} is a correlation kernel, expressed through the Hermite polynomials Hi(x)H_{i}(x) as

K(k)(u1,u2)=e12u1212u22=0k1H2(u1)H2(u2)π(2)!221.K^{(k)}(u_{1},u_{2})=e^{-\tfrac{1}{2}u_{1}^{2}-\tfrac{1}{2}u_{2}^{2}}\sum_{\ell=0}^{k-1}\frac{H_{2\ell}(u_{1})H_{2\ell}(u_{2})}{\sqrt{\pi}(2\ell)!2^{2\ell-1}}. (1.1)

Here we use the “physicist’s normalization”, so that Hi(x)H_{i}(x), i=0,1,i=0,1,\dots, is a polynomial of degree ii with leading coefficient 2i2^{i} and such that

Hi(x)Hj(x)ex2𝑑x=δijj!2jπ.\int_{\mathbb{R}}H_{i}(x)H_{j}(x)e^{-x^{2}}dx=\delta_{ij}j!2^{j}\sqrt{\pi}. (1.2)

We can now state our first theorem.

Theorem 1.5 (First swapping law).

Fix k1k\in\mathbb{Z}_{\geq 1}, and let 𝐓FS,n(k)\mathbf{T}_{\rm{FS},n}(k) denote the first occurrence time of the swap (k,k+1)(k,k+1) in a uniformly random sorting network (s1,,sN)Ωn(s_{1},\dots,s_{N})\in\Omega_{n}, N=(n2)N={n\choose 2}:

𝐓FS,n(k)=min{t1:st=k}.\mathbf{T}_{\rm{FS},n}(k)=\min\{t\geq 1:\,s_{t}=k\}.

Then we have the following convergence in law:

limn𝐓FS,n(k)n32=12𝐓(k).\lim_{n\rightarrow\infty}\frac{\mathbf{T}_{\rm{FS},n}(k)}{n^{\frac{3}{2}}}\,=\frac{1}{2}\mathbf{T}(k). (1.3)

In order to connect the first swaps to the spacings, we are going to use translation invariance of the random sorting networks (cf. [AHRV, Theorem 1, (i)]), which is based on the following combinatorial fact:

Proposition 1.6.

Let N=(n2)N={n\choose 2}. Then (s1,,sN)Ωn(s_{1},...,s_{N})\in\Omega_{n} if and only if (s2,,sN,ns1)Ωn(s_{2},...,s_{N},n-s_{1})\in\Omega_{n}.

Applying Proposition 1.6 one readily proves that for all 1rr+tN1\leq r\leq r+t\leq N,

[𝐓FS,n(k)>t]=[there are no swaps (k,k+1) at times r+1,r+2,,r+t],\mathbb{P}\left[\mathbf{T}_{\rm{FS},n}(k)>t\right]=\mathbb{P}\left[\text{there are no swaps }(k,k+1)\text{ at times }r+1,r+2,\dots,r+t\right],

which can be used to identify the distribution of a spacing in row kk. Before doing so, we need to specify the definition of a spacing. In fact, there are two natural definitions, which lead to distinct random variables. For the definition it is convenient to extend a sorting network to a 2N2N–periodic \mathbb{Z}–indexed sequence:

Definition 1.7.

Given (s1sN)Ωn(s_{1}\dots s_{N})\in\Omega_{n}, we extend it to a sequence (st)t(s_{t})_{t\in\mathbb{Z}} by requiring st+N=nsts_{t+N}=n-s_{t} for all tt\in\mathbb{Z}. We call (st)t(s_{t})_{t\in\mathbb{Z}} the periodic extension of (s1sN)Ωn(s_{1}\dots s_{N})\in\Omega_{n}.

For instance (1 2 1)Ω3(1\,2\,1)\in\Omega_{3} is extended to the infinite \mathbb{Z}–indexed sequence (2 1 2 1 2)(\dots 2\,1\,2\,1\,2\,\dots) with 11s at odd positions and 22s at even positions. The reason for this particular definition is contained in the following straightforward corollary of Proposition 1.6:

Corollary 1.8.

If (s1,,sN)(s_{1},\dots,s_{N}) is a uniformly random element of Ωn\Omega_{n}, then its periodic extension (st)t(s_{t})_{t\in\mathbb{Z}} is translation-invariant: for each aa\in\mathbb{Z} we have a distributional identity

(st+a)t=d(st)t.(s_{t+a})_{t\in\mathbb{Z}}\stackrel{{\scriptstyle d}}{{=}}(s_{t})_{t\in\mathbb{Z}}.
Definition 1.9 (First definition of the spacing).

Fix n=1,2,n=1,2,\dots, k{1,2,,n1}k\in\{1,2,...,n-1\} and aa\in\mathbb{Z}. Let (st)t(s_{t})_{t\in\mathbb{Z}} be the periodic extension of a uniformly random sorting network in Ωn\Omega_{n}. Set

X=max{ta:st=k},Y=min{t>a:st=k}.X=\max\{t\leq a:s_{t}=k\},\qquad Y=\min\{t>a:s_{t}=k\}.

We define the spacing on the kthk^{\text{th}} row of a sorting network of size nn as Spk,n:=YX\mathrm{Sp}_{k,n}:=Y-X.

Remark 1.10.

While the definition depends on the choice of aa, Corollary 1.8 implies that the distribution of Spk,n\mathrm{Sp}_{k,n} is the same for all aa\in\mathbb{Z}. Hence, we omit aa from the notation.

The nn\to\infty limit of Spk,n\mathrm{Sp}_{k,n} turns out to be connected to 𝐓(k)\mathbf{T}(k) of Definition 1.4.

Theorem 1.11.

Fix k1k\in\mathbb{Z}_{\geq 1}. We have the following convergence in distribution:

limnSpk,nn32=12Zk,\lim_{n\to\infty}\frac{\mathrm{Sp}_{k,n}}{n^{\frac{3}{2}}}=\frac{1}{2}Z_{k},

where ZkZ_{k} is a positive random variable of density gk(x)g_{k}(x), x>0x>0, given by

gk(x)=x2x2([𝐓(k)>x]).g_{k}(x)=x\,\frac{\partial^{2}}{\partial x^{2}}\bigl{(}\mathbb{P}\left[\mathbf{T}(k)>x\right]\bigr{)}. (1.4)

Here is an alternative definition of the spacing.

Definition 1.12 (Second definition of the spacing).

Fix n=1,2,n=1,2,\dots, k{1,2,,n1}k\in\{1,2,...,n-1\} and aa\in\mathbb{Z}. Let (st)t(s_{t})_{t\in\mathbb{Z}} be the periodic extension of a uniformly random sorting network in Ωn\Omega_{n}. Set

Y=min{t>a:st=k}.Y=\min\{t>a:s_{t}=k\}.

We define the spacing Sp^k,n\widehat{\mathrm{Sp}}_{k,n} on the kthk^{\text{th}} row of a sorting network of size nn as a random variable whose law is the distribution of YaY-a conditional on the event {sa=k}\{s_{a}=k\}.

Remark 1.13.

As in Definition 1.9, the choice of aa does not affect the distribution of Sp^k,n\widehat{\mathrm{Sp}}_{k,n}.

Both definitions of the spacing have their own merits. Definition 1.12 is the one preferred in theoretical physics and random matrix literature, and it has been used implicitly in [R], which studies the scaling limit of the spacing of RSN on the 1st1^{st} row. However, a disadvantage of Definition 1.12 is that it is hard to sample or observe Sp^k,n\widehat{\mathrm{Sp}}_{k,n}, as it fails to be a random variable on the state space of all sorting networks; on the other hand, Spk,n\mathrm{Sp}_{k,n} is a function of a sorting network, and this is the definition used in [GR] which studies the limiting behavior of RSN in the bulk. The nn\to\infty limit of Sp^k,n\widehat{\mathrm{Sp}}_{k,n} is still connected to 𝐓(k)\mathbf{T}(k) of Definition 1.4, but in a slightly different way.

Theorem 1.14.

Fix k1k\in\mathbb{Z}_{\geq 1}. We have the following convergence in distribution:

limnSp^k,nn32=12Z^k,\lim_{n\to\infty}\frac{\widehat{\mathrm{Sp}}_{k,n}}{n^{\frac{3}{2}}}=\frac{1}{2}\widehat{Z}_{k},

where Z^k\widehat{Z}_{k} is a positive random variable of density g^k(x)\widehat{g}_{k}(x), x>0x>0, given by

g^k(x)=π2(2k2)!!(2k1)!!2x2([𝐓(k)>x]).\widehat{g}_{k}(x)=\frac{\sqrt{\pi}}{2}\frac{(2k-2)!!}{(2k-1)!!}\,\frac{\partial^{2}}{\partial x^{2}}\bigl{(}\mathbb{P}\left[\mathbf{T}(k)>x\right]\bigr{)}. (1.5)
Remark 1.15.

In k=1k=1 case 𝐓(k)\mathbf{T}(k) becomes the absolute value of a Gaussian random variable, |N(0,12)||N(0,\tfrac{1}{2})|. Hence, Z^1\widehat{Z}_{1} has density 2xex22xe^{-x^{2}}, x>0x>0, which matches the result of [R].

In addition, we present an alternative expression for Z^k\widehat{Z}_{k} by identifying it with a marginal of a certain two-dimensional determinantal point process; we refer to [B] and references therein for general discussion of the determinantal point processes.

Definition 1.16.

Let Xk\pazocal{X}_{k}^{{}^{\prime}} be a determinantal point process on 2×0\mathbb{Z}_{\geq 2}\times\mathbb{R}_{\geq 0}, such that with respect to the product of the counting measure on the first coordinate and the Lebesgue measure on the second coordinate, the correlation kernel is

Kk(x1,u1;x2,u2)\displaystyle K^{{}^{\prime}}_{k}(x_{1},u_{1};\,x_{2},u_{2}) =𝟏{u2<u1,x2<x1}(u2u1)x1x21(x1x21)!\displaystyle=\mathbf{1}_{\{u_{2}<u_{1},\,x_{2}<x_{1}\}}\,\frac{(u_{2}-u_{1})^{x_{1}-x_{2}-1}}{(x_{1}-x_{2}-1)!}
+1(2π𝐢)2Cz[0,)𝑑zCw[0,x1)𝑑wΓ(w)Γ(z+1)Γ(z+x2+1)Γ(x1w)Γ(z+x22)Γ(x1+w+12)\displaystyle+\frac{1}{(2\pi\mathbf{i})^{2}}\oint\limits_{C_{z}[0,\infty)}dz\oint\limits_{C_{w}[0,x_{1})}dw\,\frac{\Gamma(-w)}{\Gamma(z+1)}\frac{\Gamma(z+x_{2}+1)}{\Gamma(x_{1}-w)}\frac{\Gamma(-\frac{z+x_{2}}{2})}{\Gamma(\frac{-x_{1}+w+1}{2})}
×z+x22kz+x22k+1x1w2kx1w2k1u1wu2zw+z+x2x1+1.\displaystyle\quad\times\frac{z+x_{2}-2k}{z+x_{2}-2k+1}\frac{x_{1}-w-2k}{x_{1}-w-2k-1}\frac{u_{1}^{w}u_{2}^{z}}{w+z+x_{2}-x_{1}+1}.

Here u1,u2>0u_{1},u_{2}\in\mathbb{R}_{>0} and x1,x22x_{1},x_{2}\in\mathbb{Z}_{\geq 2}; when u1u_{1} or u2u_{2} is equal to 0, the kernel is defined as the limit as u1u_{1} or u2u_{2} tends to 0. Cz[0,),Cw[0,x1)C_{z}[0,\infty),C_{w}[0,x_{1}) are counterclockwise–oriented contours which enclose the integers in [0,)[0,\infty) and [0,x1)[0,x_{1}), respectively, and are arranged so that CzC_{z} and x1x21Cwx_{1}-x_{2}-1-C_{w} are disjoint, as in Figure 5.

Remark 1.17.

For m1m\geq 1, there are mm particles in the 2mth2m^{\text{th}} and 2m+1th2m+1^{\text{th}} levels of Xk\pazocal{X}^{{}^{\prime}}_{k} (i.e. with first coordinates 2m2m and 2m+12m+1, respectively), except that on the 2kth2k^{\text{th}} level there are only k1k-1 particles.

If we denote the jthj^{\text{th}} particle(from top to bottom) on the lthl^{\text{th}} level by xjlx^{l}_{j} and set xjl=0x^{l}_{j}=0 for j>#j>\# particles in the lthl^{\text{th}} level, then the particles of Xk\pazocal{X}^{{}^{\prime}}_{k} interlace, i.e,

xj+1l+1xjlxjl+1.x^{l+1}_{j+1}\leq x^{l}_{j}\leq x^{l+1}_{j}.

While we do not prove this, we expect that compared to X\pazocal{X}, Xk\pazocal{X}^{{}^{\prime}}_{k} can be thought as the corner process of aGUE conditioned on the event that the smallest nonnegative eigenvalue of the 2kth2k^{\text{th}} corner is 0, see Remark 3.2.

Theorem 1.18.

Let Z^k\widehat{Z}_{k} be as in Theorem 1.14 and let ZlZ^{\prime}_{l} denote the coordinate of the closest to the origin particle on the lthl^{\text{th}} level in the point process Xk\pazocal{X}^{\prime}_{k}.

  • If k=1k=1, then the law of Z^1\widehat{Z}_{1} coincides with that of Z3Z^{\prime}_{3}.

  • If k2k\geq 2, then the law of Z^k\widehat{Z}_{k} coincides with that of min{Z2k1,Z2k+1}\min\{Z^{\prime}_{2k-1},Z^{\prime}_{2k+1}\}.

1.3 Methods

Here is a sketch of our proof of Theorem 1.5:

  • The Edelman-Greene bijection [EG] maps the problem to the study of the asymptotic distribution of the entries of uniformly random standard Young tableau of staircase shape, see Section 4.2.

  • We replace standard Young tableau by a continuous version, which we call Poissonized Young tableau. In Section 3.2 we show that the error of this replacement is negligible in our limit regime.

  • We use the formula of [GR] for the correlation kernel of the determinantal point process, describing the entries of a Poissonized Young tableau.

  • Asymptotic analysis of this formula leads in Theorems 3.5 and 3.6 to the limiting process described in terms of double contour integral formulas for its correlation kernel.

  • Expanding the integrals in residues, performing resummations, and using the Gibbs (conditional uniformity) properties of the point processes under consideration, we reveal in Section 3.4 a match to the distribution of eigenvalues of aGUE matrices.

We remark that (in contrast to the results in the bulk, i.e., for kk of order αn\alpha n, 0<α<10<\alpha<1, as in [GR, ADHV]) we do not claim any results for the joint asymptotic distribution of several swaps; Theorem 1.5 only deals with one-dimensional marginal distribution. A technical reason is that the Edelman-Green bijection does not have any straightforward continuous version acting on the eigenvalues of aGUE of finite sizes: it seems that one needs to deal simultaneously with aGUE of all sizes, which is not a particularly transparent operation. In future, it would be interesting to find a way to overcome this difficulty.

Theorems 1.11 and 1.14 are proven in Section 4 as corollaries of Theorem 1.5. The central idea is to develop discrete versions of (1.4), (1.5) valid for random sorting networks of finite sizes. In fact, these discrete statements are valid for any periodic point processes and, thus, can be of independent interest, see Proposition 4.2.

Finally, Theorem 1.18 is proven in Section 4 by repeating the arguments of Theorem 1.5 for the sorting networks conditional on the event {s1=k}\{s_{1}=k\}. This requires asymptotic analysis of entries of standard Young tableau of a slightly different shape, which is still possible by our methods. Thus, we arrive at the identities of Theorem 1.18 in an indirect way, by showing that the two sides of the identity are nn\to\infty limits of the same discrete distributions.

Acknowledgements

We thank the anonymous referee for helpful comments. The work of V.G. was partially supported by NSF Grants DMS-1664619, DMS-1949820, DMS-2246449, and by the Office of the Vice Chancellor for Research and Graduate Education at the University of Wisconsin–Madison with funding from the Wisconsin Alumni Research Foundation.

2 Preliminaries

One of the key tools in studying sorting networks is the Edelman-Green bijection (see Section 4.2 for the details), which maps them to Standard Young Tableaux (SYT). In this and the next section we develop asymptotic results for SYT, which will be ultimately used in Section 4 to prove the theorems about the random sorting networks announced in the introduction.

In addition, in the last subsection of this section we recall the properties of the eigenvalues of aGUE, which will be useful later on.

2.1 Young diagrams and Young tableaux

A partition is a sequence of non-negative integers λ1λ20\lambda_{1}\geq\lambda_{2}\geq...\geq 0 such that |λ|:=i=1λi<|\lambda|:=\sum_{i=1}^{\infty}\lambda_{i}<\infty. The length of λ\lambda, denoted by l(λ)l(\lambda), is the number of strictly positive λis\lambda_{i}s.

We identify a partition λ\lambda with a collection of N=|λ|N=|\lambda| boxes, such that there are λi\lambda_{i} boxes in row ii. We use coordinate (i,j)(i,j) to denote the j-th box at row i. In other words, the coordinates of boxes are {(i,j)1jλi,i=1,2,}\{(i,j)\mid 1\leq j\leq\lambda_{i},\,i=1,2,\dots\}. Such a combinatorial object is named a Young diagram, still denoted as λ\lambda. In particular, we say the Young diagram is of staircase shape and write λ=Δn\lambda=\Delta_{n} for some n2n\in\mathbb{Z}_{\geq 2}, if λi=ni\lambda_{i}=n-i for i=1,2,,ni=1,2,\dots,n. We also use the diagram λ=Δn(nk,k)\lambda=\Delta_{n}\setminus(n-k,k) for 1kn11\leq k\leq n-1, which has λi=ni\lambda_{i}=n-i for inki\neq n-k and λnk=k1\lambda_{n-k}=k-1.

A standard Young tableau (SYT) of shape λ\lambda is an insertion of numbers 1,2,…,|λ||\lambda| into boxes of λ\lambda without repetitions, such that the numbers in each row and column strictly increase from left to right and from top to bottom. The numbers within a SYT are its entries. Fixing the Young diagram λ\lambda, we can get a uniformly random SYT of shape λ\lambda, denoted as 𝐓λ\mathbf{T}_{\lambda} by considering a uniform measure on the space of all SYT of shape λ\lambda. See Figure 2 for an example with λ=Δ6\lambda=\Delta_{6}.

A Poissonized Young tableau (PYT) of shape λ\lambda is an insertion of |λ||\lambda| real numbers in [0,1][0,1] into boxes of λ\lambda, such that the numbers strictly increase along each row and column. We use PYT(λ)\mathrm{PYT}(\lambda) to denote the space of all PYT of shape λ\lambda. Note that a given PYT of shape λ\lambda can be transformed canonically to a SYT of the same shape, by replacing the k-th smallest entry with k. In the opposite direction, we can get a uniformly random PYT of shape λ\lambda, denoted as Tλ\pazocal{T}_{\lambda} by the following steps: first generate a uniformly random SYT of shape λ\lambda, then independently sample |λ||\lambda| i.i.d. random variables uniform in [0,1], and replace the entry kk with the k-th smallest random variable.

Refer to caption
Figure 2: An example of staircase shaped standard Young tableau, with n=6n=6 in two different coordinate systems: the right one is defined in Section 2.1 and left one is defined in Section 2.2.

2.2 Rotation of Young Tableaux

Let TΔnT_{\Delta_{n}} denote a standard Young tableau of shape Δn\Delta_{n}. We make a change of coordinates (i,j)(l,m)(i,j)\to(l,m) for the entries in TΔnT_{\Delta_{n}} by letting

l=ni+j,m={nij2+1,nijis even,nij+12,nijis odd.\displaystyle l=n-i+j,\qquad m=\begin{cases}\frac{n-i-j}{2}+1,&n-i-j\;\;\text{is even},\\ \frac{n-i-j+1}{2},&n-i-j\;\;\text{is odd}.\end{cases}

We let TΔn(l,m)T_{\Delta_{n}}(l,m) denote the entry of TΔnT_{\Delta_{n}} on the ithi^{\text{th}} row and the jthj^{\text{th}} column, where ii and jj are reconstructed from ll and mm by inverting the above formulas. In more detail, we have

i={nl2m+1,l is even,nl2m+12,l is odd;j={l2m+1,l is even,l2m+12,l is odd.i=\begin{cases}n-\frac{l}{2}-m+1,&l\text{ is even},\\ n-\frac{l}{2}-m+\frac{1}{2},&l\text{ is odd};\end{cases}\qquad j=\begin{cases}\frac{l}{2}-m+1,&l\text{ is even},\\ \frac{l}{2}-m+\frac{1}{2},&l\text{ is odd}.\end{cases}

The allowed values for (i,j)(i,j) are: i1i\geq 1, j1j\geq 1, i+jni+j\leq n. The allowed values for (l,m)(l,m) are l=2,3,4,,2n2l=2,3,4,\dots,2n-2 and m=1,2,,min(l,2nl)2m=1,2,\dots,\lfloor\frac{\min(l,2n-l)}{2}\rfloor. The transformation (i,j)(l,m)(i,j)\to(l,m) is essentially a clockwise rotation by 45 degrees, as can be seen in Figure 2.


Similarly, for standard Young tableau TΔn(nk,k)T_{\Delta_{n}\setminus(n-k,k)} we make the same change of coordinates as above and deal with TΔn(nk,k)(l,m)T_{\Delta_{n}\setminus(n-k,k)}(l,m). Formally, the entry at (l,m)=(2k,1)(l,m)=(2k,1) does not exist, because it corresponds to the removed box at (nk,k)(n-k,k), but we can also think about this entry as being N=(n2)N={n\choose 2}; in this way TΔn(nk,k)T_{\Delta_{n}\setminus(n-k,k)} becomes TΔnT_{\Delta_{n}} with the additional restriction TΔn(2k,1)=NT_{\Delta_{n}}(2k,1)=N.

For Poissonized Young tableau (PYT) Tλ\pazocal{T}_{\lambda} of shapes λ=Δn\lambda=\Delta_{n} or λ=Δn(nk,k)\lambda=\Delta_{n}\setminus(n-k,k), we define the change of coordinate in exactly the same way and use (m,l)(m,l) coordinate system.

2.3 Point processes associated with PYT

Let Tλ\pazocal{T}_{\lambda} be a PYT\mathrm{PYT} of shape λ\lambda. As in Section 2.2, we will focus on the case λ=Δn\lambda=\Delta_{n} or λ=Δn(nk,k)\lambda=\Delta_{n}\setminus(n-k,k), for which we induce a point process on 0×0\mathbb{Z}_{\geq 0}\times\mathbb{R}_{\geq 0} from its entries.

Definition 2.1.

The projection of a PYT\mathrm{PYT} T\pazocal{T} with shape Δn{\Delta_{n}} or Δn(nk,k)\Delta_{n}\setminus(n-k,k) is a point configuration on 0×[0,1]\mathbb{Z}_{\geq 0}\times[0,1], such that there’s a particle at (l,u)(l,u), if for some l2l\in\mathbb{Z}_{\geq 2} and m1m\in\mathbb{Z}_{\geq 1},

u=1T(l,m).u=1-\pazocal{T}(l,m).

By definition, u=1T(l,m)u=1-\pazocal{T}(l,m) is the mthm^{\text{th}} lowest particle on level {l}×0\{l\}\times\mathbb{R}_{\geq 0} (except for level l=2kl=2k in TΔn(nk,k)\pazocal{T}_{\Delta_{n}\setminus(n-k,k)}, where T(2k,1)\pazocal{T}(2k,1) is missing), and the particles on neighboring levels interlace by the setting of PYT, see Figure 3.

Refer to caption
Refer to caption
Figure 3: Producing a point configuration from a PYT\mathrm{PYT} TT of shape Δ6\Delta_{6} with entries given in the figure as in Section 2.2 and 2.3: first rotate Δ6\Delta_{6} clockwise by 45 degrees, then project its entries into a configuration of interlacing particles.

If we take Tλ\pazocal{T}_{\lambda} to be uniformly random, then the projection of Tλ\pazocal{T}_{\lambda} becomes a point process on 0×[0,1]\mathbb{Z}_{\geq 0}\times[0,1]. Note that this random point configuration is simple almost surely. We also would like to rescale the uu–coordinate.

Definition 2.2.

The point process Xn\pazocal{X}^{{}^{\prime}}_{n} on 0×0\mathbb{Z}_{\geq 0}\times\mathbb{R}_{\geq 0} is obtained by taking a uniformly random PYT\mathrm{PYT} TΔn\pazocal{T}_{\Delta_{n}} , projecting its entries on 0×[0,1]\mathbb{Z}_{\geq 0}\times[0,1], and then rescaling the second coordinate of each particle by the map (l,u)(l,n12u)(l,u)\mapsto(l,n^{\frac{1}{2}}\cdot u). Similarly, Xk,n\pazocal{X}^{{}^{\prime}}_{k,n} is obtained from uniformly random PYT TΔn(nk,k)\pazocal{T}_{\Delta_{n}\setminus(n-k,k)} by the same procedure.

We study the asymptotic behavior of Xn\pazocal{X}_{n}^{{}^{\prime}} and Xk,n\pazocal{X}^{{}^{\prime}}_{k,n} in Section 3.3. Our analysis uses the technique of determinantal point processes with correlation kernels given by double contour integrals, which we present next.

2.4 Determinantal Point Process

Here’s a brief review of some standard definitions, for a thorough introduction see [B, DVe]. Let SS be a locally compact Polish space, say \mathbb{Z} or n\mathbb{R}^{n}. We consider the space of discrete subsets in SS, which consists of countable subsets XX of SS without accumulation points, and identify each such point configuration with the measure xXδx\sum_{x\in X}\delta_{x}. The topology of this space is given as the weak topology of Radon measures on SS. We further use Borel σ\sigma-algebra corresponding to this topology.

A point process is a random discrete subset of SS, and we assume that it is simple a.s, i.e, all the points have distinct positions. A determinantal point process X\pazocal{X} on SS is a point process with correlation kernel K:S×SK:S\times S\rightarrow\mathbb{R}, and a reference measure μ\mu on SS, such that for any f:Skf:S^{k}\rightarrow\mathbb{R} with compact support,

𝔼[x1,,xkXx1,,xkare distinctf(x1,,xk)]=Skdet[K(xi,xj)]f(x1,,xk)μk(dx1,,dxk).\mathbb{E}\,\Bigg{[}\sum_{\begin{subarray}{c}x_{1},\ldots,x_{k}\in\pazocal{X}\\ \,x_{1},\ldots,x_{k}\;\text{are distinct}\end{subarray}}f(x_{1},\ldots,x_{k})\Bigg{]}=\int\limits_{S^{k}}\det\left[K(x_{i},x_{j})\right]f(x_{1},\ldots,x_{k})\,\mu^{\otimes k}(dx_{1},\ldots,dx_{k}). (2.1)

Expectations of the form given by the l.h.s. of (2.1) determine the law of X\pazocal{X} under mild conditions on KK, see [Le]. This will always be the case in this text as the correlation kernels we consider will be continuous.

The determinantal point processes appearing in this paper live in space S=×[0,1]S=\mathbb{Z}\times[0,1] or S=×0S=\mathbb{Z}\times\mathbb{R}_{\geq 0}, with reference measure being the product of counting measure and Lebesgue measure, denoted by #×L([0,1])\#_{\mathbb{Z}}\times\pazocal{L}([0,1]) or #×L(0)\#_{\mathbb{Z}}\times\pazocal{L}(\mathbb{R}_{\geq 0}) resp. We use the following lemma, whose proof we omit, see [DVe], [Le].

Lemma 2.3.

Let YnY_{n} be a determinantal point process on ×>0\mathbb{Z}\times\mathbb{R}_{>0} with reference measure #×L(>0)\#_{\mathbb{Z}}\times\pazocal{L}(\mathbb{R}_{>0}) and correlation kernel KnK_{n}. Then the sequence YnY_{n} converges weakly as nn\to\infty to a determinantal point process X\pazocal{X} with correlation kernel KK on ×0\mathbb{Z}\times\mathbb{R}_{\geq 0} (with reference measure #×0\#_{\mathbb{Z}}\times\mathbb{R}_{\geq 0}) and with almost surely no particles at (i,0)(i,0), ii\in\mathbb{Z}, if

KnKK_{n}\rightarrow K uniformly on compact subsets of ×0\mathbb{Z}\times\mathbb{R}_{\geq 0}.

2.5 Uniformly Random Projection

When PYT\mathrm{PYT} TΔn\pazocal{T}_{\Delta_{n}} and TΔn(nk,k)\pazocal{T}_{\Delta_{n}\setminus(n-k,k)} are uniformly random, their projections are point processes on 0×[0,1]\mathbb{Z}_{\geq 0}\times[0,1], whose correlation functions were computed in [GR]. We restate the result there for our case in the following theorem, which is a stepping stone of the proofs of our main results.

Theorem 2.4.

The processes Xn\pazocal{X}^{{}^{\prime}}_{n} and Xk,n\pazocal{X}^{{}^{\prime}}_{k,n} of Definition 2.2 are determinantal point process on {2,3,,2n2}×>0\{2,3,\dots,2n-2\}\times\mathbb{R}_{>0} with correlation kernel Kλ(x1,u1;x2,u2)K_{\lambda}(x_{1},u_{1};x_{2},u_{2}) given for x1,x2{2,3,,2n2}x_{1},x_{2}\in{\{2,3,\dots,2n-2\}} and u1,u2>0u_{1},u_{2}\in\mathbb{R}_{>0}, by

Kλ(x1,u1;x2;u2)=𝟏{u2<u1,x2<x1}(u2u1)x1x21n12(x1x2)(x1x21)!\displaystyle K_{\lambda}(x_{1},u_{1};x_{2};u_{2})=\mathbf{1}_{\{u_{2}<u_{1},x_{2}<x_{1}\}}\frac{(u_{2}-u_{1})^{x_{1}-x_{2}-1}n^{-\frac{1}{2}(x_{1}-x_{2})}}{(x_{1}-x_{2}-1)!} (2.2)
+1(2πi)2Cz[0,λ1+nx2)𝑑zCw[0,x1)𝑑wΓ(w)Γ(z+1)Gλ(z+x2n)Gλ(x1n1w)u2zu1wn12(z+w+1)w+z+x2x1+1,\displaystyle+\frac{1}{(2\pi i)^{2}}\oint_{C_{z}[0,\lambda_{1}+n-x_{2})}dz\oint_{C_{w}[0,x_{1})}dw\frac{\Gamma(-w)}{\Gamma(z+1)}\frac{G_{\lambda}(z+x_{2}-n)}{G_{\lambda}(x_{1}-n-1-w)}\cdot\frac{u_{2}^{z}\,u_{1}^{w}\,n^{-\frac{1}{2}(z+w+1)}}{w+z+x_{2}-x_{1}+1},

where

Gλ(u)=Γ(u+1)i=1u+iuλi+i=Γ(u+1+n)i=1n(uλi+i),G_{\lambda}(u)=\Gamma\left(u+1\right)\prod_{i=1}^{\infty}\frac{u+i}{u-\lambda_{i}+i}=\frac{\Gamma\left(u+1+n\right)}{\prod\limits_{i=1}^{n}(u-\lambda_{i}+i)},

with λ=Δn\lambda=\Delta_{n} for Xn\pazocal{X}^{{}^{\prime}}_{n} and λ=Δn(nk,k)\lambda=\Delta_{n}\setminus(n-k,k) for Xk,n\pazocal{X}^{{}^{\prime}}_{k,n}. The contours Cz[0,λ1+nx2)C_{z}[0,\lambda_{1}+n-x_{2}) and Cw[0,x1)C_{w}[0,x_{1}) are as shown in Figure 4. Both are counter-clockwise, encloses only the integers in the respective half open intervals [0,λ1+nx2)[0,\lambda_{1}+n-x_{2}) and [0,x1)[0,x_{1}), and are arranged so that CzC_{z} and x1x21Cwx_{1}-x_{2}-1-C_{w} are disjoint, as in Figure 4.

zz-contour CzC_{z}ww-contour CwC_{w}01-1x11x_{1}-1λ1+n1x2\lambda_{1}+n-1-x_{2}
Refer to caption
Figure 4: The contours in the statement of Theorem 2.4, correspond to the case 2x1x22\leq x_{1}\leq x_{2} (top) and 2x2<x12\leq x_{2}<x_{1} (bottom). In the bottom picture, ww-contour splits into two components, in order to guarantee that CzC_{z} and x1x21Cwx_{1}-x_{2}-1-C_{w} are disjoint
Remark 2.5.

The Γ\Gamma–functions in double-contour integrals are used for the convenience of notations, but the integral is, in fact, a rational function of zz and ww multiplied by u2zu1wn12(z+w+1)u_{2}^{z}\,u_{1}^{w}\,n^{-\frac{1}{2}(z+w+1)}. Indeed, we have

Γ(w)Γ(z+1)Gλ(z+x2n)Gλ(x1n1w)=(z+1)(z+2)(z+x2)i=1n(z+x2nλi+i)i=1n(w+x1n1λi+i)(w)(1w)(x11w).\frac{\Gamma(-w)}{\Gamma(z+1)}\frac{G_{\lambda}(z+x_{2}-n)}{G_{\lambda}(x_{1}-n-1-w)}=\frac{(z+1)(z+2)\cdots(z+x_{2})}{\prod\limits_{i=1}^{n}(z+x_{2}-n-\lambda_{i}+i)}\cdot\frac{\prod\limits_{i=1}^{n}(-w+x_{1}-n-1-\lambda_{i}+i)}{(-w)(1-w)\cdots(x_{1}-1-w)}. (2.3)
Proof of Theorem 2.4.

This is a corollary of [GR, Theorem 1.5]. In the notation of [GR], we choose:

t1=1u1n12;t2=1u2n12;x1[GR]=x1n;x2[GR]=x2n;λ=ΔnorΔn(nk,k).t_{1}=1-\frac{u_{1}}{n^{\frac{1}{2}}};\ t_{2}=1-\frac{u_{2}}{n^{\frac{1}{2}}};\ x_{1}^{[GR]}=x_{1}-n;\ x_{2}^{[GR]}=x_{2}-n;\qquad\lambda=\Delta_{n}\;\text{or}\;\Delta_{n}\setminus(n-k,k).

Since our match of parameters involves rescaling of the real spatial coordinate, we also need to use the following Lemma 2.6, whose proof we omit. ∎

Lemma 2.6.

Let YnY_{n} be a determinantal point process on ×(0,1)\mathbb{Z}\times(0,1) with reference measure #×L((0,1))\#_{\mathbb{Z}}\times\pazocal{L}((0,1)) and correlation kernel KnK_{n}. For cnc_{n}\in\mathbb{Z}, α\alpha\in\mathbb{R}, we define the rescaled and shifted point process

Yn={(xcn,nαu)(x,u)Yn}.Y^{{}^{\prime}}_{n}=\{(x-c_{n},n^{\alpha}u)\mid(x,u)\in Y_{n}\}.

Then, the correlation kernel of YnY^{{}^{\prime}}_{n} with reference measure #×L(>0)\#_{\mathbb{Z}}\times\pazocal{L}(\mathbb{R}_{>0}) is

Kn(x1,u1;x2,u2)=nαKn(x1+cn,u1nα;x2+cn,u2nα).K^{{}^{\prime}}_{n}(x_{1},u_{1};\,x_{2},u_{2})=n^{-\alpha}K_{n}\Bigg{(}x_{1}+c_{n},\frac{u_{1}}{n^{\alpha}};\,x_{2}+c_{n},\frac{u_{2}}{n^{\alpha}}\Bigg{)}.
Remark 2.7.

We can extend Xn\pazocal{X}_{n}^{{}^{\prime}} and Xk,n\pazocal{X}_{k,n}^{{}^{\prime}} to point processes on {2,3,,2n2}×0{\{2,3,\dots,2n-2\}}\times\mathbb{R}_{\geq 0}, in such a way that almost surely there is no particle at (x,0)(x,0) for any x{2,3,,2n2}x\in{\{2,3,\dots,2n-2\}}. Moreover, the kernel KλK_{\lambda} can be extended to the same space based on its regular behavior at u1u_{1} or u2=0u_{2}=0.

In order to see that the double contour integral in the definition of the kernel is well-behaved as u10u_{1}\to 0 or u20u_{2}\to 0, we expand the integral as a sum of residues at the poles inside the integration contours. We start from the case 2x1x22\leq x_{1}\leq x_{2} with contours of the top panel of Figure 4. In this case the denominator w+z+x2x1+1w+z+x_{2}-x_{1}+1 is never singular inside the integration contours, the ww–residues are at simple poles at finitely many non-negative integers and zz–residues are also at simple poles at finitely many non-negative integers. The residue at w=m1w=m_{1}, z=m2z=m_{2} for m1,m20m_{1},m_{2}\in\mathbb{Z}_{\geq 0} has the form c(m1,m2)u1m1u2m2c(m_{1},m_{2})\cdot u_{1}^{m_{1}}u_{2}^{m_{2}}, where c(m1,m2)c(m_{1},m_{2}) does not depend on u1,u2u_{1},u_{2}. Hence, the residue is continuous as u1,u20u_{1},u_{2}\to 0 and so is the double integral.

Proceeding to the case 2x2<x12\leq x_{2}<x_{1} with contours of the bottom panel of Figure 4, the additional feature is the potential singularity of the denominator w+z+x2x1+1w+z+x_{2}-x_{1}+1. Let us first compute the ww–integral as a sum of the residues, getting a sum of the terms of the form u1m1×u_{1}^{m_{1}}\times (one-dimensional zz-integral) with non-negative values for m1m_{1}. Up to the factors not depending on u1u_{1} or u2u_{2}, the corresponding zz–integral has the form

Cz[0,λ1+nx2)Γ(z+x2+1)Γ(z+1)1i=1n(z+x2λi+in)u2zn12zz+m1+x2x1+1𝑑z.\oint_{C_{z}[0,\lambda_{1}+n-x_{2})}\frac{\Gamma(z+x_{2}+1)}{\Gamma(z+1)}\cdot\frac{1}{\prod\limits_{i=1}^{n}(z+x_{2}-\lambda_{i}+i-n)}\cdot\frac{u_{2}^{z}\,n^{-\frac{1}{2}z}}{z+m_{1}+x_{2}-x_{1}+1}dz.

The simple poles of the last integral lead to u2u_{2}–dependence of the form u2m2u_{2}^{m_{2}}, m20m_{2}\geq 0, which is again continuous at u2=0u_{2}=0. It would be more complicated, if the integral had a double pole: the residue at such a pole would have involved the zz–derivative of u2zu_{2}^{z}, which leads to the appearance of ln(u2)\ln(u_{2}) factor that is singular at u2=0u_{2}=0. However, we claim that there is no double pole. Indeed, using (2.3), m1{x1n1λi+i}i=1nm_{1}\notin\{x_{1}-n-1-\lambda_{i}+i\}_{i=1}^{n}. Therefore, for the pole of 1z+m1+x2x1+1\frac{1}{z+m_{1}+x_{2}-x_{1}+1}, we have (m1x2+x11){x2+λii+n}i=1n(-m_{1}-x_{2}+x_{1}-1)\notin\{-x_{2}+\lambda_{i}-i+n\}_{i=1}^{n}. Hence, this pole is outside the set of poles of 1i=1n(z+x2λi+in)\frac{1}{\prod_{i=1}^{n}(z+x_{2}-\lambda_{i}+i-n)}.

2.6 Anti-symmetric GUE

Let us recall a statement from [FN, Theorems 2.9 and 3.3]:

Proposition 2.8.

For MM in Definition 1.2 and l=1,2,l=1,2,\dots we have:

  1. (a)

    The lthl^{\text{th}} corner of MM has ll real eigenvalues.

  2. (b)

    The eigenvalues come in pairs, i.e, λ\lambda is an eigenvalue of the lthl^{\text{th}} corner if and only if λ-\lambda is an eigenvalue of the lthl^{\text{th}} corner. When ll is odd, 0 is necessarily an eigenvalue of the lthl^{\text{th}} corner.

  3. (c)

    The eigenvalues of the lthl^{\text{th}} corner interlace with the eigenvalues of the (l+1)st(l+1)^{\text{st}} corner a.s, which means that

    λj+1l+1λjlλjl+1,\lambda^{l+1}_{j+1}\leq\lambda^{l}_{j}\leq\lambda^{l+1}_{j}, (2.4)

    where j=1,2,,lj=1,2,\dots,l and λjl\lambda^{l}_{j} denotes the jthj^{\text{th}} largest eigenvalue of the lthl^{\text{th}} corner.

  4. (d)

    Conditionally on the positive eigenvalues of the lthl^{\text{th}} corner, the positive eigenvalues of the ithi^{\text{th}} corners, i=2,,l1i=2,...,l-1, are jointly distributed uniformly on the polytope determined by the interlacing inequalities (2.4).

  5. (e)

    The joint density of positive eigenvalues (λ1lλl2l)\bigl{(}\lambda^{l}_{1}\geq\dots\geq\lambda^{l}_{\lfloor\frac{l}{2}\rfloor}\bigr{)} of the lthl^{\text{th}} corner is

    p(λ1l,,λl2l)={Cl1i<jl2((λil)2(λjl)2)j=1l2e(λjl)2,liseven;Cl1i<jl12((λil)2(λjl)2)j=1l12[λjle(λjl)2],lisodd,\displaystyle p(\lambda^{l}_{1},...,\lambda^{l}_{\lfloor\frac{l}{2}\rfloor})=\begin{dcases}C_{l}\prod_{1\leq i<j\leq\frac{l}{2}}\left((\lambda^{l}_{i})^{2}-(\lambda^{l}_{j})^{2}\right)\,\prod_{j=1}^{\frac{l}{2}}e^{-(\lambda^{l}_{j})^{2}},&\text{l}\;\;\text{is}\;\;\text{even};\\ \quad\\ C_{l}\prod_{1\leq i<j\leq\frac{l-1}{2}}\left((\lambda^{l}_{i})^{2}-(\lambda^{l}_{j})^{2}\right)\,\prod_{j=1}^{\frac{l-1}{2}}\left[\lambda^{l}_{j}\,e^{-(\lambda^{l}_{j})^{2}}\right],&\text{l}\;\;\text{is}\;\;\text{odd},\end{dcases}

    where ClC_{l} is an explicitly known normalization constant.

Note that the above five properties uniquely characterize the joint distribution of {λjk}1jk<\{\lambda^{k}_{j}\}_{1\leq j\leq k<\infty}. An alternative characterization is through the correlation functions of the corresponding point process X\pazocal{X}:

Theorem 2.9 ([FN, Proposition 4.3]).

The aGUE corners process X\pazocal{X} of Definition 1.2 is a determinantal point process with correlation kernel K(x,u;y,v)K(x,u;y,v) (with respect to the reference measure #L(0)\#_{\mathbb{Z}}\otimes\pazocal{L}(\mathbb{R}_{\geq 0})), such that for xyx\geq y,

K(x,u;y,v)=eu2l=1[y/2]Hx2l(u)Hy2l(v)Ny2l,K(x,u;y,v)=e^{-u^{2}}\sum_{l=1}^{[y/2]}\frac{H_{x-2l}(u)H_{y-2l}(v)}{N_{y-2l}}, (2.5)

and for x<yx<y,

K(x,u;y,v)=eu2l=0Hx2l(u)Hy2l(v)Ny2l.K(x,u;y,v)=-e^{-u^{2}}\sum_{l=-\infty}^{0}\frac{H_{x-2l}(u)H_{y-2l}(v)}{N_{y-2l}}. (2.6)

Here Hj(u)H_{j}(u) is the jthj^{\text{th}} “physicist’s Hermite polynomial” with leading coefficient 2j2^{j}, so that

Hi(u)Hj(u)eu2𝑑u=2δijNj,Nj=j!2j1π.\int_{\mathbb{R}}H_{i}(u)H_{j}(u)e^{-u^{2}}du=2\delta_{ij}N_{j},\qquad N_{j}=j!2^{j-1}\sqrt{\pi}.
Remark 2.10.

The x=yx=y case of (2.5) differs from the kernel of (1.1) by the factor e12u2+12v2e^{-\tfrac{1}{2}u^{2}+\tfrac{1}{2}v^{2}}, which cancels out when we compute determinants.

3 Local Limit of Standard Staircase Shaped Tableaux

The aim of this section is to investigate the scaling limit near the bottom-left corner for the uniformly random standard Young tableaux of shapes Δn\Delta_{n} and Δn(nk,k)\Delta_{n}\setminus(n-k,k) as nn\to\infty.

3.1 Statement of the result

We use the (l,m)(l,m) coordinate system for SYT, as in Section 2.2. We recall that X\pazocal{X} is the point process corresponding to the corners of aGUE as in Definition 1.2. Let X(l,m)\pazocal{X}(l,m) denote the mthm^{\text{th}} particle of X\pazocal{X} in the level ll, sorted in the increasing order; in other words, X(l,m)\pazocal{X}(l,m) is the mthm^{\text{th}} positive eigenvalue of l×ll\times l corner of the infinite aGUE matrix MM (with m=1m=1 corresponding to the smallest positive and m=l2m=\lfloor\tfrac{l}{2}\rfloor to the largest positive eigenvalue).

Further, we set Xk\pazocal{X}^{{}^{\prime}}_{k} to be the point process of Definition 1.16 with an extra particle inserted at (2k,0)(2k,0). We let Xk(l,m)\pazocal{X}^{{}^{\prime}}_{k}(l,m) denote the mthm^{\text{th}} particle of Xk\pazocal{X}^{{}^{\prime}}_{k} in the level x=lx=l, sorted in the increasing order.

Theorem 3.1.

For each n2n\in\mathbb{Z}_{\geq 2}, let TΔn(l,m)T_{\Delta_{n}}(l,m) and TΔn(nk,k)(l,m)T_{\Delta_{n}\setminus(n-k,k)}(l,m) be uniformly random standard Young tableaux of shapes Δn\Delta_{n} and Δn(nk,k){\Delta_{n}\setminus(n-k,k)}, respectively, in the coordinate system of Section 2.2. Recall N=(n2)N={n\choose 2}. Then for each fixed kk, as nn\to\infty we have

{n12(1TΔn(l,m)N)}l2,  1ml2{X(l,m)}l2,  1ml2 and \left\{n^{\frac{1}{2}}\left(1-\frac{T_{\Delta_{n}}(l,m)}{N}\right)\right\}_{l\geq 2,\,\,1\leq m\leq\lfloor\frac{l}{2}\rfloor}\longrightarrow\bigl{\{}\pazocal{X}(l,m)\bigr{\}}_{l\geq 2,\,\,1\leq m\leq\lfloor\frac{l}{2}\rfloor}\;\qquad\text{ and } (3.1)
{n12(1TΔn(nk,k)(l,m)N)}l2,  1ml2{Xk(l,m)}l2,  1ml2,\left\{n^{\frac{1}{2}}\left(1-\frac{T_{\Delta_{n}\setminus(n-k,k)}(l,m)}{N}\right)\right\}_{l\geq 2,\,\,1\leq m\leq\lfloor\frac{l}{2}\rfloor}\longrightarrow\bigl{\{}\pazocal{X}^{{}^{\prime}}_{k}(l,m)\bigr{\}}_{l\geq 2,\,\,1\leq m\leq\lfloor\frac{l}{2}\rfloor}, (3.2)

in the sense of convergence of finite-dimensional distributions.

Remark 3.2.

The SYT TΔn(nk,k)T_{\Delta_{n}\setminus(n-k,k)} can be thought as TΔnT_{\Delta_{n}} conditioned on its largest entry being at (nk,k)(n-k,k), and in (3.2), TΔn(nk,k)(2k,0)T_{\Delta_{n}\setminus(n-k,k)}(2k,0) is set to be NN, corresponding to X(2k,1)=0\pazocal{X}(2k,1)=0 on the right. Hence, Theorem 3.1 suggests a conjecture: Xk\pazocal{X}^{{}^{\prime}}_{k} should have the same law as X\pazocal{X} conditioned on having a particle at (2k,0)(2k,0), i.e. on X(2k,1)=0\pazocal{X}(2k,1)=0. Note that additional efforts are needed to prove this, because the topology of the convergence in Theorem 3.1 is not strong enough to make conclusions about the conditional distributions.

The rest of this section is devoted to the proof of Theorem 3.1. We first couple the two types of SYT with the uniformly random PYT of the same shape, and induce two point processes Xn,Xk,n\pazocal{X}_{n}^{{}^{\prime}},\pazocal{X}_{k,n}^{{}^{\prime}} on 0×0\mathbb{Z}_{\geq 0}\times\mathbb{R}_{\geq 0} respectively from these two PYT, as in Definition 2.2. Then we prove that XnX\pazocal{X}_{n}^{{}^{\prime}}\rightarrow\pazocal{X}^{{}^{\prime}} and Xk,nXk\pazocal{X}_{k,n}^{{}^{\prime}}\rightarrow\pazocal{X}^{{}^{\prime}}_{k} as nn\to\infty, where X\pazocal{X}^{{}^{\prime}} and Xk\pazocal{X}^{{}^{\prime}}_{k} are determinantal point processes on 0×0\mathbb{Z}_{\geq 0}\times\mathbb{R}_{\geq 0} with explicit correlation kernels given by double contour integrals. Next, we identify X\pazocal{X}^{{}^{\prime}} with X\pazocal{X}, the corners process of aGUE of Definition 1.2.

3.2 Coupling with PYT

The first step of the proof is to introduce a coupling between SYT and PYT of the same shape and to show that under this coupling the difference between random SYT and PYT is negligible in the asymptotic regime of our interest.

Let TΔnT_{\Delta_{n}} and TΔn\pazocal{T}_{\Delta_{n}} be uniformly random SYT and PYT, respectively, of shape Δn\Delta_{n}. We couple these two tableaux in the following way: for N=(n2)N={n\choose 2}, let (P1<P2<<PN)(P_{1}\ <\ P_{2}\ <\cdots<\ P_{N}) be a uniformly random point on the simplex {(x1,,xN)0<x1<x2<<xN<1}\{(x_{1},\cdots,x_{N})\mid 0<x_{1}<x_{2}<\dots<x_{N}<1\}. Given a uniformly random SYT TΔnT_{\Delta_{n}}, we replace entry ll in TΔnT_{\Delta_{n}} by PlP_{l} for l=1,2,,Nl=1,2,\cdots,N; it is straightforward to check that the result is a uniformly random PYT TΔn\pazocal{T}_{\Delta_{n}}. In the opposite direction, TΔnT_{\Delta_{n}} is reconstructed by TΔn\pazocal{T}_{\Delta_{n}} by replacing the llth largest entry in TΔn\pazocal{T}_{\Delta_{n}} by ll for each l=1,2,,Nl=1,2,\dots,N.

Lemma 3.3.

Fix any pair of integers (i,j)(i,j) with i2i\geq 2 and 1ji21\leq j\leq\lfloor\tfrac{i}{2}\rfloor. Using the coupling described above, we have for each δ>0\delta>0

n1δ|TΔn(i,j)TΔn(i,j)N|0n^{1-\delta}\left|\pazocal{T}_{\Delta_{n}}(i,j)-\frac{T_{\Delta_{n}}(i,j)}{N}\right|\rightarrow 0

in probability, as nn\rightarrow\infty.

Proof.

Let TΔn(i,j)=ξ{1,2,,N}T_{\Delta_{n}}(i,j)=\xi\in\{1,2,\dots,N\}, where ξ\xi is a random variable. Then, by the construction of the coupling, TΔn(i,j)=Pξ\pazocal{T}_{\Delta_{n}}(i,j)=P_{\xi}. By Chebyshev’s inequality,

[n1δ|PξξN|>ϵ](n1δϵ)2𝔼[(PξξN)2]=n22δϵ2j=1N𝔼[(PξξN)2|ξ=l][ξ=l]\begin{split}&\quad\mathbb{P}\left[n^{1-\delta}\left|P_{\xi}-\frac{\xi}{N}\right|>\epsilon\right]\\ &\leq\left(\frac{n^{1-\delta}}{\epsilon}\right)^{2}\mathbb{E}\left[\left(P_{\xi}-\frac{\xi}{N}\right)^{2}\right]\\ &=\frac{n^{2-2\delta}}{\epsilon^{2}}\sum_{j=1}^{N}\mathbb{E}\left[\left(P_{\xi}-\frac{\xi}{N}\right)^{2}\,\Big{|}\,\xi=l\right]\cdot\mathbb{P}\left[\xi=l\right]\end{split} (3.3)

For a fixed ll, the random variable PlP_{l} is distributed according to Beta distribution with parameters ll and N+1lN+1-l, which has mean lN+1\frac{l}{N+1} and variance (see, e.g., [JKB, Chapter 25]).

𝔼[(PllN+1)2]=l(N+1l)(N+1)2(N+2)(N+12)2(N+1)2(N+2)=14(N+2).\mathbb{E}\left[\left(P_{l}-\frac{l}{N+1}\right)^{2}\right]=\frac{l(N+1-l)}{(N+1)^{2}(N+2)}\leq\frac{\left(\frac{N+1}{2}\right)^{2}}{(N+1)^{2}(N+2)}=\frac{1}{4(N+2)}. (3.4)

Thus,

𝔼[(PllN)2]=𝔼[(PllN+1)2]+(lN(N+1))214(N+2)+1(N+1)21N.\mathbb{E}\left[\left(P_{l}-\frac{l}{N}\right)^{2}\right]=\mathbb{E}\left[\left(P_{l}-\frac{l}{N+1}\right)^{2}\right]+\left(\frac{l}{N(N+1)}\right)^{2}\leq\frac{1}{4(N+2)}+\frac{1}{(N+1)^{2}}\leq\frac{1}{N}. (3.5)

Since N=n(n1)2N=\frac{n(n-1)}{2}, (3.3) and (3.5) imply that as nn\rightarrow\infty

[n1δ|PξξN|>ϵ]n22δϵ2N0.\mathbb{P}\left[n^{1-\delta}\left|P_{\xi}-\frac{\xi}{N}\right|>\epsilon\right]\leq\frac{n^{2-2\delta}}{\epsilon^{2}N}\rightarrow 0.\qed

In a similar way, we couple uniformly random SYT TΔn(nk,k)T_{\Delta_{n}\setminus(n-k,k)} and PYT TΔn(nk,k)\pazocal{T}_{\Delta_{n}\setminus(n-k,k)} of shape Δn(nk,k)\Delta_{n}\setminus(n-k,k): we let (P1<P2<<PN1)(P_{1}\ <\ P_{2}\ <\cdots<\ P_{N-1}) be a uniformly random point on the simplex {(x1,,xN1)0<x1<x2<<xN1<1}\{(x_{1},\cdots,x_{N-1})\mid 0<x_{1}<x_{2}<\dots<x_{N-1}<1\}. Given a uniformly random SYT TΔn(nk,k)T_{\Delta_{n}\setminus(n-k,k)}, we replace entry kk in TΔn(nk,k)T_{\Delta_{n}\setminus(n-k,k)} by PkP_{k} for k=1,2,,N1k=1,2,\cdots,N-1; it is straightforward to check that the result is a uniformly random PYT TΔn(nk,k)\pazocal{T}_{\Delta_{n}\setminus(n-k,k)}. Repeating the proof of Lemma 3.3 we arrive at:

Lemma 3.4.

Fix any pair of integers (i,j)(i,j) with i2i\geq 2 and 1ji21\leq j\leq\lfloor\tfrac{i}{2}\rfloor. Using the coupling described above, we have for each δ>0\delta>0

n1δ|TΔn(nk,k)(i,j)TΔn(nk,k)(i,j)N|0n^{1-\delta}\left|\pazocal{T}_{\Delta_{n}\setminus(n-k,k)}(i,j)-\frac{T_{\Delta_{n}\setminus(n-k,k)}(i,j)}{N}\right|\rightarrow 0

in probability, as nn\rightarrow\infty.

3.3 Limiting Processes

In this section we analyze asymptotic behavior of the point processes associated to random PYT TΔn\pazocal{T}_{\Delta_{n}} and TΔn(nk,k)\pazocal{T}_{\Delta_{n}\setminus(n-k,k)} by the procedure of Section 2.3.

Theorem 3.5.

The point process Xn\pazocal{X}^{{}^{\prime}}_{n} (corresponding to random PYT TΔn\pazocal{T}_{\Delta_{n}} via Definition 2.2) converges weakly as nn\to\infty to a point process X\pazocal{X}^{{}^{\prime}} on 2×0\mathbb{Z}_{\geq 2}\times\mathbb{R}_{\geq 0}. X\pazocal{X}^{{}^{\prime}} is a determinantal point process (with respect to reference measure #2L(0)\#_{\mathbb{Z}_{\geq 2}}\otimes\pazocal{L}(\mathbb{R}_{\geq 0})) with correlation kernel

K\displaystyle K^{{}^{\prime}} (x1,u1;x2,u2)=I{u2<u1,x2<x1}(u2u1)x1x21(x1x21)!\displaystyle(x_{1},u_{1};x_{2},u_{2})=I_{\{u_{2}<u_{1},x_{2}<x_{1}\}}\frac{(u_{2}-u_{1})^{x_{1}-x_{2}-1}}{(x_{1}-x_{2}-1)!}
+1(2πi)2Cz[0,)𝑑zCw[0,x1)𝑑wΓ(w)Γ(z+1)Γ(z+x2+1)Γ(x1w)Γ(z+x22)Γ(x1+w+12)u1wu2zw+z+x2x1+1.\displaystyle+\frac{1}{(2\pi i)^{2}}\oint_{C_{z}[0,\infty)}dz\oint_{C_{w}[0,x_{1})}dw\frac{\Gamma(-w)}{\Gamma(z+1)}\frac{\Gamma(z+x_{2}+1)}{\Gamma(x_{1}-w)}\frac{\Gamma(-\frac{z+x_{2}}{2})}{\Gamma(\frac{-x_{1}+w+1}{2})}\frac{u_{1}^{w}u_{2}^{z}}{w+z+x_{2}-x_{1}+1}.

Here u1,u2>0u_{1},u_{2}\in\mathbb{R}_{>0} and x1,x22x_{1},x_{2}\in\mathbb{Z}_{\geq 2}; when u1u_{1} or u2u_{2} is equal to 0, the kernel is defined as the limit as u1u_{1} or u2u_{2} tends to 0. Cz[0,),Cw[0,x1)C_{z}[0,\infty),C_{w}[0,x_{1}) are counterclockwise–oriented contours which enclose the integers in [0,)[0,\infty) and [0,x1)[0,x_{1}), respectively, and are arranged so that CzC_{z} and x1x21Cwx_{1}-x_{2}-1-C_{w} are disjoint, as in Figure 5.

Refer to caption
Refer to caption
Figure 5: The contours in the statements of Theorem 3.5 and 3.6, for the case 1x1x21\leq x_{1}\leq x_{2} and 1x2<x11\leq x_{2}<x_{1} respectively.
Theorem 3.6.

The point process Xk,n\pazocal{X}^{{}^{\prime}}_{k,n} (corresponding to random PYT TΔn(nk,k)\pazocal{T}_{\Delta_{n}\setminus(n-k,k)} via Definition 2.2) converges weakly to a point process Xk\pazocal{X}^{{}^{\prime}}_{k} on 2×0\mathbb{Z}_{\geq 2}\times\mathbb{R}_{\geq 0}. Xk\pazocal{X}^{{}^{\prime}}_{k} is a determinantal point process (with respect to reference measure #2L(0)\#_{\mathbb{Z}_{\geq 2}}\otimes\pazocal{L}(\mathbb{R}_{\geq 0})) with correlation kernel given by

Kk\displaystyle K^{{}^{\prime}}_{k} (x1,u1;x2,u2)=I{u2<u1,x2<x1}(u2u1)x1x21(x1x21)!\displaystyle(x_{1},u_{1};x_{2},u_{2})=I_{\{u_{2}<u_{1},x_{2}<x_{1}\}}\frac{(u_{2}-u_{1})^{x_{1}-x_{2}-1}}{(x_{1}-x_{2}-1)!}
+1(2πi)2Cz[0,)𝑑zCw[0,x1)𝑑wΓ(w)Γ(z+1)Γ(z+x2+1)Γ(x1w)Γ(z+x22)Γ(x1+w+12)\displaystyle+\frac{1}{(2\pi i)^{2}}\oint_{C_{z}[0,\infty)}dz\oint_{C_{w}[0,x_{1})}dw\frac{\Gamma(-w)}{\Gamma(z+1)}\frac{\Gamma(z+x_{2}+1)}{\Gamma(x_{1}-w)}\frac{\Gamma(-\frac{z+x_{2}}{2})}{\Gamma(\frac{-x_{1}+w+1}{2})}
×z+x22kz+x22k+1x1w2kx1w2k1u1wu2zw+z+x2x1+1.\displaystyle\quad\times\frac{z+x_{2}-2k}{z+x_{2}-2k+1}\cdot\frac{x_{1}-w-2k}{x_{1}-w-2k-1}\cdot\frac{u_{1}^{w}u_{2}^{z}}{w+z+x_{2}-x_{1}+1}.

Here u1,u2>0u_{1},u_{2}\in\mathbb{R}_{>0} and x1,x20x_{1},x_{2}\in\mathbb{Z}_{\geq 0}; the value of KkK^{{}^{\prime}}_{k} when u1u_{1} or u2u_{2} equals 0 is to be understood as the limit as u1u_{1} or u2u_{2} tends to 0. The contours Cz[0,)C_{z}[0,\infty) and Cw[0,x1)C_{w}[0,x_{1}) are the same as in Theorem 3.5.

In the proof we use a known asymptotic property of the Gamma-function:

Γ(y+m)(m1)!=my(1+O(m1)),m+,\frac{\Gamma(y+m)}{(m-1)!}=m^{y}(1+O(m^{-1})),\qquad m\to+\infty, (3.6)

where O(m1)O(m^{-1}) term is uniform as long as yy is uniformly bounded.

Proof of Theorem 3.5.

We show the convergence in distribution by verifying conditions of Lemma 2.3, namely, we show that the correlation kernel Kn(x1,u1;x2,u2)nx1x22K^{{}^{\prime}}_{n}(x_{1},u_{1};x_{2},u_{2})n^{\frac{x_{1}-x_{2}}{2}} of Xn\pazocal{X}^{{}^{\prime}}_{n}, where KnK^{{}^{\prime}}_{n} is given by (2.2), converges to KK^{{}^{\prime}} uniformly on compact subsets of x1,x22x_{1},x_{2}\in\mathbb{Z}_{\geq 2} and u1,u20u_{1},u_{2}\in\mathbb{R}_{\geq 0}. Note that the just introduced multiplication by nx1x22n^{\frac{x_{1}-x_{2}}{2}} does not change the value of the correlations functions det[Kn(xi,ui;xj,uj)]\det[K^{{}^{\prime}}_{n}(x_{i},u_{i};x_{j},u_{j})]. Let us fix arbitrary x1,x22x_{1},x_{2}\in\mathbb{Z}_{\geq 2}, strictly positive u1u_{1}, u2u_{2}, and analyze the asymptotic behavior of the double contour integral in (2.2).

The first step is to deform the contours of integration from the ones of Figure 4 to the ones of Figure 5. Using (2.3), the zz–dependent factors of the integrand are

(z+1)(z+2)(z+x2)i=1n(z+x22n+2i)(u2n)z1w+z+x2x1+1.\frac{(z+1)(z+2)\cdots(z+x_{2})}{\prod\limits_{i=1}^{n}(z+x_{2}-2n+2i)}\cdot\left(\frac{u_{2}}{\sqrt{n}}\right)^{z}\cdot\frac{1}{w+z+x_{2}-x_{1}+1}. (3.7)

The form of (3.7) implies that there are no zz–poles of the integrand to the right from the contour of Figure 4; hence, the value of the integral does not change in the deformation. For each fixed u2u_{2} and large enough nn, on the new contours of Figure 5, the expression (3.7) rapidly decays as z+\Re z\to+\infty (uniformly in nn). This observation allows us to control the part of the integral corresponding to large z\Re z, and it remains to study the nn\to\infty asymptotics of the integrand for finite zz and ww.

We use (3.6) to compute the asymptotic behavior of i=1n\prod_{i=1}^{n} in (3.7):

i=1n(z+x22n+2i)=(1)ni=1n(zx2+2n2i)\displaystyle\prod_{i=1}^{n}(z+x_{2}-2n+2i)=(-1)^{n}\prod_{i=1}^{n}(-z-x_{2}+2n-2i)
=(1)n2ni=1n(zx22+ni)=(1)n2nΓ(z+x22+n)Γ(z+x22)\displaystyle=(-1)^{n}2^{n}\prod_{i=1}^{n}\left(-\frac{z-x_{2}}{2}+n-i\right)=(-1)^{n}2^{n}\frac{\Gamma(-\frac{z+x_{2}}{2}+n)}{\Gamma(-\frac{z+x_{2}}{2})}
=(1)n2n(n1)!nz+x22Γ(z+x22)(1+O(n1)).\displaystyle={(-1)^{n}}2^{n}(n-1)!\frac{n^{-\frac{z+x_{2}}{2}}}{\Gamma(-\frac{z+x_{2}}{2})}\left(1+O(n^{-1})\right).

Similarly,

i=1n(x1w2n+2i1)=(1)n2n(n1)!nx1+w+12Γ(x1+w+12)(1+O(n1)).\prod_{i=1}^{n}(x_{1}-w-2n+2i-1)=(-1)^{n}2^{n}(n-1)!\frac{n^{\frac{-x_{1}+w+1}{2}}}{\Gamma(\frac{-x_{1}+w+1}{2})}\left(1+O(n^{-1})\right).

Thus, for λ=Δn\lambda=\Delta_{n}, we have

Gλ(z+x2n)Gλ(x1n1w)=Γ(z+x2+1)Γ(x1w)Γ(z+x22)Γ(x1+w+12)nx1+w+12nz+x22(1+O(n1)).\frac{G_{\lambda}(z+x_{2}-n)}{G_{\lambda}(x_{1}-n-1-w)}=\frac{\Gamma(z+x_{2}+1)}{\Gamma(x_{1}-w)}\frac{\Gamma(-\frac{z+x_{2}}{2})}{\Gamma(\frac{-x_{1}+w+1}{2})}n^{\frac{-x_{1}+w+1}{2}}n^{\frac{z+x_{2}}{2}}\left(1+O(n^{-1})\right).

Hence, the integrand in the double contour integral part of Kn(x1,u1;x2,u2)nx1x22K^{{}^{\prime}}_{n}(x_{1},u_{1};x_{2},u_{2})n^{\frac{x_{1}-x_{2}}{2}}, converges as nn\to\infty to

Γ(w)Γ(z+1)Γ(z+x2+1)Γ(x1w)Γ(z+x22)Γ(x1+w+12)u1wu2zw+z+x2x1+1.\frac{\Gamma(-w)}{\Gamma(z+1)}\frac{\Gamma(z+x_{2}+1)}{\Gamma(x_{1}-w)}\frac{\Gamma(-\frac{z+x_{2}}{2})}{\Gamma(\frac{-x_{1}+w+1}{2})}\frac{u_{1}^{w}u_{2}^{z}}{w+z+x_{2}-x_{1}+1}.

Combining with the straightforward asymptotics of the indicator term, we conclude that for all strictly positive u1u_{1} and u2u_{2},

limnKn(x1,u1;x2,u2)nx1x22=Kn(x1,u1;x2,u2).\lim_{n\to\infty}K^{{}^{\prime}}_{n}(x_{1},u_{1};x_{2},u_{2})n^{\frac{x_{1}-x_{2}}{2}}=K^{{}^{\prime}}_{n}(x_{1},u_{1};x_{2},u_{2}).

Clearly, the convergence is uniform over u1,u2u_{1},u_{2} in compact subsets of (0,+)(0,+\infty). It remains to show that there is no explosion for u1u_{1} or u2u_{2} near 0, which is done in exactly the same way as we did in Remark 2.7, i.e. by interpreting the integral as a sum of residues at simple zz- and ww-poles at non-negative integers. ∎

Proof of Theorem 3.6.

The proof is the same as for Theorem 3.5, with only some slight difference in calculation. This time, for λ=Δn(nk,k)\lambda=\Delta_{n}\setminus(n-k,k) we have

Gλ(z+x2n)=Γ(z+x2+1)i=1n(z+x22n+2i)z+x22kz+x22k+1,\displaystyle G_{\lambda}(z+x_{2}-n)=\frac{\Gamma(z+x_{2}+1)}{\prod\limits_{i=1}^{n}(z+x_{2}-2n+2i)}\cdot\frac{z+x_{2}-2k}{z+x_{2}-2k+1},
Gλ(x1n1w)=Γ(x1w)i=1n(x1w2n+2i1)x1w2k1x1w2k.\displaystyle G_{\lambda}(x_{1}-n-1-w)=\frac{\Gamma(x_{1}-w)}{\prod\limits_{i=1}^{n}(x_{1}-w-2n+2i-1)}\cdot\frac{x_{1}-w-2k-1}{x_{1}-w-2k}.

The additional fractions z+x22kz+x22k+1\frac{z+x_{2}-2k}{z+x_{2}-2k+1} and x1w2k1x1w2k\frac{x_{1}-w-2k-1}{x_{1}-w-2k} (as compared to λ=Δn\lambda=\Delta_{n} case) propagate to the final answer without changes. ∎

3.4 Connection to Anti-symmetric GUE

In this section we prove that the distribution of the limiting process of Theorem 3.5 coincides with the corners process of aGUE.

Proposition 3.7.

K(2k,u1;2k,u2)K^{{}^{\prime}}(2k,u_{1};2k,u_{2}) of Theorem 3.5 for u1,u20u_{1},u_{2}\in\mathbb{R}_{\geq 0}, k>0k\in\mathbb{Z}_{>0}, is equal to

2k+1i=0j=0k1\displaystyle 2^{k+1}\sum_{i=0}^{\infty}\sum_{j=0}^{k-1} (1)i+ku12j(2k2j1)!(2j)!Γ(j+12k)u22ia=1k(2i+2a1)i!12j+2i+1.\displaystyle(-1)^{i+k}\frac{u_{1}^{2j}}{(2k-2j-1)!(2j)!\,\Gamma(j+\frac{1}{2}-k)}\cdot\frac{u_{2}^{2i}\cdot\prod_{a=1}^{k}(2i+2a-1)}{i!}\cdot\frac{1}{2j+2i+1}. (3.8)
Remark 3.8.

When k=1k=1, we simplify K(2,u1;2,u2)=2πeu22K^{{}^{\prime}}(2,u_{1};2,u_{2})=\frac{2}{\sqrt{\pi}}e^{-u_{2}^{2}}. There’s only one particle on level {2}×0\{2\}\times\mathbb{R}_{\geq 0} which corresponds to the limit of the entry of the SYT at lower corner, with coordinate (n1,1)(n-1,1).

When k=2k=2, we simplify K(4,u1;4,u2)=1π[(12u12)(12u22)+2]eu22K^{{}^{\prime}}(4,u_{1};4,u_{2})=\frac{1}{\sqrt{\pi}}[(1-2u_{1}^{2})(1-2u_{2}^{2})+2]e^{-u_{2}^{2}}. There are two particles on level {4}×0\{4\}\times\mathbb{R}_{\geq 0} which correspond to the two entries of the SYT with coordinates (n3,1)(n-3,1) and (n2,2)(n-2,2).

Proof of Proposition 3.7.

K(2k,u1;2k,u2)K^{{}^{\prime}}(2k,u_{1};2k,u_{2}) of Theorem 3.5 is given by

1(2π𝐢)2Cz[0,)𝑑zCw[0,2k)𝑑wΓ(w)Γ(z+1)Γ(z+2k+1)Γ(2kw)Γ(z+2k2)Γ(2k+w+12)(u1)w(u2)zw+z+1,\displaystyle\frac{1}{(2\pi\mathbf{i})^{2}}\oint_{C_{z}[0,\infty)}dz\oint_{C_{w}[0,2k)}dw\frac{\Gamma(-w)}{\Gamma(z+1)}\frac{\Gamma(z+2k+1)}{\Gamma(2k-w)}\frac{\Gamma(-\frac{z+2k}{2})}{\Gamma(\frac{-2k+w+1}{2})}\frac{(u_{1})^{w}(u_{2})^{z}}{w+z+1},

The ww–integral is evaluated as the sum of residues at simple poles at even integers w=0,2,4,2k2w=0,2,4\dots,2k-2 and we have

Resw=2j[Γ(w)Γ(2kw)Γ(2k+w+12)(u1)ww+z+1]=Resw=2j[1Γ(2k+w+12)(w)(w+1)(w+2k1)(u1)ww+z+1]=[1Γ(2k+2j+12)(2k12j)!(2j)!(u1)2j2j+z+1],\mathrm{Res}_{w=2j}\left[\frac{\Gamma(-w)}{\Gamma(2k-w)\Gamma(\frac{-2k+w+1}{2})}\cdot\frac{(u_{1})^{w}}{w+z+1}\right]\\ =\mathrm{Res}_{w=2j}\left[\frac{1}{\Gamma(\frac{-2k+w+1}{2})(-w)(-w+1)\cdots(-w+2k-1)}\cdot\frac{(u_{1})^{w}}{w+z+1}\right]\\ =-\left[\frac{1}{\Gamma(\frac{-2k+2j+1}{2})\,(2k-1-2j)!(2j)!}\cdot\frac{(u_{1})^{2j}}{2j+z+1}\right],

which matches the jj–dependent factors in (3.8). The remaining zz–integral for the jj–th term becomes

1(2π𝐢)Cz[0,)Γ(z+2k+1)Γ(z+2k2)Γ(z+1)(u2)z2j+z+1𝑑z,\displaystyle\frac{1}{(2\pi\mathbf{i})}\oint_{C_{z}[0,\infty)}\frac{\Gamma(z+2k+1)\Gamma(-\frac{z+2k}{2})}{\Gamma(z+1)}\frac{(u_{2})^{z}}{2j+z+1}dz,

The last integral is evaluated as the sum of the residues at z=0,2,4,z=0,2,4,\dots. Using the fact that the residue of the Gamma function at a simple pole at (n)(-n), n=1,2,,n=1,2,\dots, is (1)nn!\tfrac{(-1)^{n}}{n!}, we get

Resz=2i[Γ(z+2k+1)Γ(z+2k2)Γ(z+1)(u2)z2j+z+1]=2(1)i+k(i+k)!(2i+2k)!(2i)!(u2)2i2j+2i+1=2k+1(1)i+ka=1k(2i+2a1)i!(u2)2i2j+2i+1,\mathrm{Res}_{z=2i}\left[\frac{\Gamma(z+2k+1)\Gamma(-\frac{z+2k}{2})}{\Gamma(z+1)}\frac{(u_{2})^{z}}{2j+z+1}\right]=-2\frac{(-1)^{i+k}}{(i+k)!}\cdot\frac{(2i+2k)!}{(2i)!}\cdot\frac{(u_{2})^{2i}}{2j+2i+1}\\ =-2^{k+1}(-1)^{i+k}\frac{\prod_{a=1}^{k}(2i+2a-1)}{i!}\cdot\frac{(u_{2})^{2i}}{2j+2i+1},

which matches the ii–dependent factors in (3.8). ∎

Proposition 3.9.

K(2k,u1;2k,u2)K^{{}^{\prime}}(2k,u_{1};2k,u_{2}) of Theorem 3.5 for u1,u20u_{1},u_{2}\in\mathbb{R}_{\geq 0}, k>0k\in\mathbb{Z}_{>0}, is also equal to

e(u2)2l=0k1H2l(u1)H2l(u2)22l1(2l)!π,e^{-(u_{2})^{2}}\sum_{l=0}^{k-1}\frac{H_{2l}(u_{1})H_{2l}(u_{2})}{2^{2l-1}(2l)!\sqrt{\pi}}, (3.9)

where Hi(u)H_{i}(u) are Hermite polynomials, as in (1.1) and Theorem 2.9.

Proof.

Our task is to show that (3.8) and (3.9) match. The proof is induction in kk. For k=1k=1, this is a content of Remark 3.8. Subtracting (3.8) at k+1k+1 and at kk, we get

2k+1i=0j=0k(1)i+ku12j(2k2j+1)!(2j)!Γ(j+12k)u22ia=1k(2i+2a1)i!12j+2i+1×[2(jk1/2)(2i+2k+1)(2k2j+1)(2k2j)]=2k+1j=0ku12j(2k2j)!(2j)!Γ(j+12k)i=0(1)i+ku22ia=1k(2i+2a1)i!2^{k+1}\sum_{i=0}^{\infty}\sum_{j=0}^{k}(-1)^{i+k}\frac{u_{1}^{2j}}{(2k-2j+1)!(2j)!\,\Gamma(j+\frac{1}{2}-k)}\cdot\frac{u_{2}^{2i}\cdot\prod_{a=1}^{k}(2i+2a-1)}{i!}\cdot\frac{1}{2j+2i+1}\\ \times\left[-2(j-k-1/2)(2i+2k+1)-(2k-2j+1)(2k-2j)\right]\\ =2^{k+1}\sum_{j=0}^{k}\frac{u_{1}^{2j}}{(2k-2j)!(2j)!\,\Gamma(j+\frac{1}{2}-k)}\sum_{i=0}^{\infty}(-1)^{i+k}\frac{u_{2}^{2i}\cdot\prod_{a=1}^{k}(2i+2a-1)}{i!} (3.10)

On the other hand, subtracting (3.9) at k+1k+1 and at kk, we get

e(u2)2H2k(u1)H2k(u2)22k1(2k)!π,e^{-(u_{2})^{2}}\frac{H_{2k}(u_{1})H_{2k}(u_{2})}{2^{2k-1}(2k)!\sqrt{\pi}}, (3.11)

The explicit expression for the Hermite polynomials [Sz, Chapter 5.5] is

H2k(u1)=j=0k(1)kj22j(2k)!(kj)!(2j)!u12jH_{2k}(u_{1})=\sum_{j=0}^{k}(-1)^{k-j}2^{2j}\frac{(2k)!}{(k-j)!(2j)!}u_{1}^{2j}

In order to match with u1u_{1}–dependent part of (3.10) we write

Γ(j+12k)=Γ(12)(121)(122)(12+jk)=(1)kj2kjπ135(2k2j1)=(1)kj22k2jπ(kj)!(2k2j)!\Gamma(j+\tfrac{1}{2}-k)=\frac{\Gamma(\tfrac{1}{2})}{(\tfrac{1}{2}-1)(\tfrac{1}{2}-2)\cdots(\tfrac{1}{2}+j-k)}=\frac{(-1)^{k-j}2^{k-j}\sqrt{\pi}}{1\cdot 3\cdot 5\cdots(2k-2j-1)}=\frac{(-1)^{k-j}2^{2k-2j}\sqrt{\pi}(k-j)!}{(2k-2j)!}

Hence, (3.10) gets transformed into

2k+1j=0ku12j(2j)!(1)kj22k2jπ(kj)!i=0(1)i+ku22ia=1k(2i+2a1)i!=21k(2k)!πH2k(u1)i=0(1)i+ku22ia=1k(2i+2a1)i!2^{k+1}\sum_{j=0}^{k}\frac{u_{1}^{2j}}{(2j)!(-1)^{k-j}2^{2k-2j}\sqrt{\pi}(k-j)!}\sum_{i=0}^{\infty}(-1)^{i+k}\frac{u_{2}^{2i}\cdot\prod_{a=1}^{k}(2i+2a-1)}{i!}\\ =\frac{2^{1-k}}{(2k)!\sqrt{\pi}}H_{2k}(u_{1})\sum_{i=0}^{\infty}(-1)^{i+k}\frac{u_{2}^{2i}\cdot\prod_{a=1}^{k}(2i+2a-1)}{i!}

Comparing with (3.11), it remains to show that:

e(u2)2H2k(u2)2k=?i=0(1)i+ku22ia=1k(2i+2a1)i!.e^{-(u_{2})^{2}}\frac{H_{2k}(u_{2})}{2^{k}}\stackrel{{\scriptstyle?}}{{=}}\sum_{i=0}^{\infty}(-1)^{i+k}\frac{u_{2}^{2i}\cdot\prod_{a=1}^{k}(2i+2a-1)}{i!}.

Or, equivalently,

H2k(u2)=?e(u2)22ki=0(1)i+ku22ia=1k(2i+2a1)i!.H_{2k}(u_{2})\stackrel{{\scriptstyle?}}{{=}}e^{(u_{2})^{2}}2^{k}\sum_{i=0}^{\infty}(-1)^{i+k}\frac{u_{2}^{2i}\cdot\prod_{a=1}^{k}(2i+2a-1)}{i!}. (3.12)

The identity (3.12) is a corollary of the Rodrigues formula for Hermite polynomials [Sz, Chapter 5.5]

Hn(u)=eu2(1)n(u)neu2H_{n}(u)=e^{u^{2}}(-1)^{n}(\partial_{u})^{n}e^{-u^{2}}

Indeed, we have

eu2(1)2k(u)2keu2=eu2(u)2k[i=0(u2)ii!]=eu2(u)2k[i=0(1)i+ku2i+2k(i+k)!]=eu2[i=0(1)i+ku2i(2i+1)(2i+2)(2i+2k)(i+k)!]=eu2[i=0(1)i+ku2i2k(2i+1)(2i+3)(2i+2k1)i!].e^{u^{2}}(-1)^{2k}(\partial_{u})^{2k}e^{-u^{2}}=e^{u^{2}}(\partial_{u})^{2k}\left[\sum_{i=0}^{\infty}\frac{(-u^{2})^{i}}{i!}\right]=e^{u^{2}}(\partial_{u})^{2k}\left[\sum_{i=0}^{\infty}(-1)^{i+k}\frac{u^{2i+2k}}{(i+k)!}\right]\\ =e^{u^{2}}\left[\sum_{i=0}^{\infty}(-1)^{i+k}\frac{u^{2i}(2i+1)(2i+2)\cdots(2i+2k)}{(i+k)!}\right]\\ =e^{u^{2}}\left[\sum_{i=0}^{\infty}(-1)^{i+k}\frac{u^{2i}2^{k}(2i+1)(2i+3)\cdots(2i+2k-1)}{i!}\right].\qed (3.13)
Theorem 3.10.

The point process X\pazocal{X}^{{}^{\prime}} in Theorem 3.5 has the same distribution as the point process X\pazocal{X} — the corner process of aGUE of Definition 1.2.

Proof.

Step 1. For each k=1,2,k=1,2,\dots, each of the processes X\pazocal{X}^{{}^{\prime}} and X\pazocal{X} has kk particles on level 2k2k (i.e. with the first coordinate 2k2k). Let us show the marginal distributions describing the joint law of these kk particles are the same. By Proposition 3.9, for X\pazocal{X}^{{}^{\prime}} these particles form a determinantal point process with kernel (3.9). By Theorem 2.9, for X\pazocal{X} these particles also form a determinantal point process with kernel given by plugging x=y=2kx=y=2k into (2.5). The kernels match and, hence, so do the point processes.

Step 2. Let us now fix kk and compare the joint distribution of particles in X\pazocal{X}^{{}^{\prime}} and in X\pazocal{X} on levels l=1,2,,2kl=1,2,\dots,2k. By part d in Proposition 2.8, the conditional law of the particles on levels l=1,2,,2k1l=1,2,\dots,2k-1 given kk particles on level kk is uniform (subject to interlacing conditions) for X\pazocal{X}. On the other hand, X\pazocal{X}^{{}^{\prime}} possesses the same conditional uniformity, because it was obtained by a scaling limit from uniformly random PYT\mathrm{PYT} TΔn\pazocal{T}_{\Delta_{n}} and conditional uniformity is preserved in limit transitions. Combining with Step 1, we conclude that the joint distributions of particles on levels l=1,2,,2kl=1,2,\dots,2k are the same for X\pazocal{X}^{{}^{\prime}} and X\pazocal{X}. Since kk is arbitrary, we are done. ∎

3.5 Proof of Theorem 3.1

Theorems 3.5 and 3.6 show that the point process Xn\pazocal{X}^{{}^{\prime}}_{n}, Xk,n\pazocal{X}^{{}^{\prime}}_{k,n} induced from n12(1T(l,m))n^{\frac{1}{2}}(1-\pazocal{T}(l,m)), the rescaled entries of PYT of shapes Δn\Delta_{n} and Δn(nk,k)\Delta_{n}\setminus(n-k,k), converge weakly as nn\to\infty to X\pazocal{X}^{{}^{\prime}} and Xk\pazocal{X}^{{}^{\prime}}_{k}, respectively. In particular, treating the entries of the PYT of these two shapes as infinite random vectors, they converges in the sense of finite dimensional distribution to the positions of the corresponding particles in Xn\pazocal{X}^{{}^{\prime}}_{n} and Xk,n\pazocal{X}^{{}^{\prime}}_{k,n}, respectively.

On the other hand, by Lemma 3.3, there’s a coupling of PYT T\pazocal{T} and SYT TT that, for any (l,m)(l,m),

n12[(1T(l,m)N)(1T(l,m))]n^{\frac{1}{2}}\left[\Bigl{(}1-\tfrac{T(l,m)}{N}\Bigr{)}-\Bigl{(}1-\pazocal{T}(l,m)\Bigr{)}\right]

converges to 0 in probability. Combining the above two results, we obtain the nn\to\infty convergence for finite dimensional distributions of

{n12(1T(l,m)N)}l2,  1ml2\left\{n^{\frac{1}{2}}\left(1-\frac{T(l,m)}{N}\right)\right\}_{l\geq 2,\,\,1\leq m\leq\lfloor\frac{l}{2}\rfloor}

to the limits given by particles of X\pazocal{X}^{{}^{\prime}} and Xk\pazocal{X}^{{}^{\prime}}_{k}. It remains to use Theorem 3.10 to match X\pazocal{X}^{{}^{\prime}} with X\pazocal{X}. ∎

4 Asymptotics for spacings

We start this section by studying general rotationally-invariant point processes on a circle, giving two different definitions of a spacing between particles in such processes, and proving that their distribution functions satisfy relations, which are discrete versions of (1.4) and (1.5). When specialized to the random sorting networks setting, these spacings are precisely the ones discussed in the introduction. After that we recall the Edelman-Greene bijection between standard Young tableaux and sorting networks. In the last subsection we combine all the ingredients to finish the proofs of Theorems 1.5, 1.11, 1.14, and 1.18.

4.1 Generalities on spacings

Consider a random point process 𝒫\mathscr{P} on a discrete circle of length K2K\in\mathbb{Z}_{\geq 2}. We index the possible positions of the particles by 11, 22, …, KK in the clockwise order and refer to the midpoint between 11 and KK as the origin, see Figure 6. Throughout this section we assume that the distribution of the point process is invariant under rotations of the circle; we also silently assume that almost surely 𝒫\mathscr{P} has at least one particle.

Refer to caption
Refer to caption
Figure 6: Two configurations of point process on discrete circle of length K=8K=8 described in Example 4.1. Particles are shown in black.

We use 𝒫\mathscr{P} to define several random variables with positive integer values. We define the waiting time W{1,2,,K}W\in\{1,2,...,K\} to be the smallest value of 1\ell\geq 1 such that there is a particle at position \ell. We let the spacing Sp1{1,2,,K}\mathrm{Sp}_{1}\in\{1,2,...,K\} to be the distance between two adjacent particles on the circle: one to the left from the origin and one to the right from the origin. By distance we mean here one plus the number of the holes between the particles (counted along the arc including the origin), so that if there are particles at positions 11 and KK, then the distance is 11.

Next, we define the conditional spacing Sp2{1,2,,K}\mathrm{Sp}_{2}\in\{1,2,...,K\} to be a random variable whose distribution is that of WW conditional on having a particle at position KK. Note that WW and Sp1\mathrm{Sp}_{1} are random variables (functions) on the probability space of 𝒫\mathscr{P}, but Sp2\mathrm{Sp}_{2} should be defined on a different probability space.

Example 4.1.

Figure 6 gives two possible configurations of a point process on the discrete circle of length K=8K=8. On the first configuration, Sp1=4\mathrm{Sp}_{1}=4 and W=1W=1. On the second configuration, Sp1=W=3\mathrm{Sp}_{1}=W=3. Since there is a particle at KK, we can also think of Sp2\mathrm{Sp}_{2} being equal to 33 for the second configuration.

Let f1()f_{1}(\cdot), f2()f_{2}(\cdot), and g()g(\cdot) be the probability mass functions of Sp1\mathrm{Sp}_{1}, Sp2\mathrm{Sp}_{2}, and WW, respectively:

f1()=[Sp1=],f2()=[Sp2=],g()=[W=],l=1,2,.f_{1}(\ell)=\mathbb{P}\left[\mathrm{Sp}_{1}=\ell\right],\qquad f_{2}(\ell)=\mathbb{P}\left[\mathrm{Sp}_{2}=\ell\right],\qquad g(\ell)=\mathbb{P}\left[W=\ell\right],\qquad l=1,2,\dots.

We also define ρ\rho to the the probability that there is a particle at position KK (in other words, this is the first correlation function or density of the point process 𝒫\mathscr{P}).

Proposition 4.2.

For any rotationally invariant point process 𝒫\mathscr{P}, we have

Δg()=f1(),=1,2,,-\Delta g(\ell)\cdot\ell=f_{1}(\ell),\qquad\ell=1,2,\dots, (4.1)
Δg()=ρf2(),=1,2,,-\Delta g(\ell)=\rho\cdot f_{2}(\ell),\qquad\ell=1,2,\dots, (4.2)

where Δ\Delta is the forward difference operator: Δg():=g(+1)g()\Delta g(\ell):=g(\ell+1)-g(\ell).

Proof.

We claim that

Δg()=[there are particles at positions K and , but not at 1,2,,1].-\Delta g(\ell)=\mathbb{P}\left[\text{there are particles at positions }K\text{ and }\ell,\text{ but not at }1,2,\dots,\ell-1\right]. (4.3)

Indeed, using rotational invariance of the law of 𝒫\mathscr{P}, we have

g()\displaystyle g(\ell) =[there is a particle at position , but not at 1,2,,1],\displaystyle=\mathbb{P}\left[\text{there is a particle at position }\ell,\text{ but not at }1,2,\dots,\ell-1\right],
g(+1)\displaystyle g(\ell+1) =[there is a particle at position , but not at K,1,2,,1].\displaystyle=\mathbb{P}\left[\text{there is a particle at position }\ell,\text{ but not at }K,1,2,\dots,\ell-1\right].

In order to prove (4.2), it is now sufficient to rewrite the probability in the right-hand side of (4.3) as the product of probability of having a particle at KK (which is ρ\rho) and conditional probability.

In order to prove (4.1), note that by definition

f1()=a=01[there are particles at Ka and a, but not at Ka+1,,a1].f_{1}(\ell)=\sum_{a=0}^{\ell-1}\mathbb{P}\left[\text{there are particles at }K-a\text{ and }\ell-a,\text{ but not at }K-a+1,\dots,\ell-a-1\right]. (4.4)

By rotational invariance all terms in the right-hand side of (4.4) are equal. There are \ell of them and each one is computed by (4.3), leading to (4.1). ∎

Remark 4.3.

A continuous version of Proposition 4.2 for translationally invariant point processes on the real line \mathbb{R} says that the following is true (under technical regularity assumptions on the point process, which we do not spell out, and which are needed to guarantee the existence of all the densities below): Suppose that Sp1\mathrm{Sp}_{1}, Sp2\mathrm{Sp}_{2} and WW are spacing (distance between closest particles to the left and to the right from the origin), conditional spacing, and waiting time for an arrival of a particle, and let f1(x)f_{1}(x), f2(x)f_{2}(x), g(x)g(x), x0x\geq 0 be probability densities of these random variables. Then we have

xxg(x)\displaystyle-x\cdot\partial_{x}g(x) =f1(x),\displaystyle=f_{1}(x),
xg(x)\displaystyle-\partial_{x}g(x) =ρf2(x).\displaystyle=\rho\cdot f_{2}(x).

where ρ\rho is equal to the first correlation function for the process.

4.2 Edelman-Greene bijection

The relation of the random Young tableaux (which we were studying in Sections 2 and 3) with the random sorting networks relies on the Edelman–Green bijection [EG], which we now present.

The bijection takes a standard Young tableau TT of shape Δn\Delta_{n} as an input and outputs a sorting network. This correspondence maps the uniform measure on standard Young tableaux of shape Δn\Delta_{n} to the uniform measure on sorting networks of size nn.

The bijection proceeds through the following algorithm, in which we use the standard (i,j)(i,j) coordinate system for Young tableaux, as in Section 2.1:

  1. 1.

    Given a standard Young tableau TT, find the box (n,)(n-\ell,\ell) which contains the largest entry of TT. Necessarily, this entry is N=n(n1)2N=\frac{n(n-1)}{2} and 1n11\leq\ell\leq n-1.

  2. 2.

    Set the first swap s1s_{1} of the corresponding sorting network to be \ell.

  3. 3.

    Define the sliding path in the following way: the first box of the path is (n,)(n-\ell,\ell). Compare the entries at (n1,)(n-\ell-1,\ell) and at (n,1)(n-\ell,\ell-1), and take the box with the larger entry as the second box (if only one of the boxes is inside the Young diagram, then take that one). Repeat this procedure (each time decreasing by 11 either the first of the second coordinate of the box and moving in the direction of the larger entry), until you arrive at the box (1,1)(1,1), which is the last box of the sliding path.

  4. 4.

    Slide the entries on the sliding path in such a way that the maximal entry NN is removed and all the remaining entries on the path are moved to the neighboring boxes on the path. No entry remains at (1,1)(1,1) after the sliding. Then we increase all entries by 11, and fill the box (1,1)(1,1) with new entry 11.

  5. 5.

    After transformations of steps 1-4, TT becomes a new standard Young tableau of shape Δn\Delta_{n}. Repeat steps 1-4 additional (N1)(N-1) times to get swaps s2s_{2},s3s_{3},…,sNs_{N} of the sorting network.

Using the Edelman-Green bijection, the random variable TFS,n(k)T_{FS,n}(k) of Theorem 1.5 — the first time swap kk occurs — gets recast in terms of the uniformly random Young tableau.

Lemma 4.4.

Let TΔn(l,m)T_{\Delta_{n}}(l,m) be a uniformly random standard Young tableau of shape Δn\Delta_{n} in the rotated coordinate system of Section 2.2. The following distributional identity holds:

TFS,n(k)=dN+1TΔn(2k,1)T_{FS,n}(k)\stackrel{{\scriptstyle d}}{{=}}N+1-T_{\Delta_{n}}(2k,1) (4.5)
Proof.

In each iteration of the Edelman–Greene algorithm the entry (nk,k)(n-k,k) (in the standard (i,j)(i,j) coordinate system for Young tableaux, as in Section 2.1) grows by 11 until it becomes NN, at which point the swap sks_{k} is added to the sorting network for the first time. The entry at (nk,k)(n-k,k) in the (i,j)(i,j) coordinate system is the same as the entry (2k,1)(2k,1) in the rotated (l,m)(l,m) coordinate system, leading to (4.5). ∎

We can also compute the distribution of the conditional spacing Sp^k,n\widehat{\mathrm{Sp}}_{k,n} of Definition 1.12.

Lemma 4.5.

Let TΔn(nk,k)(l,m)T_{\Delta_{n}\setminus(n-k,k)}(l,m) be a uniformly random standard Young tableau of shape Δn(nk,k){\Delta_{n}\setminus(n-k,k)} in the rotated coordinate system of Section 2.2. The following distributional identity holds:

Sp^k,n=d{Nmax(TΔn(nk,k)(2k1,1),TΔn(nk,k)(2k+1,1)),if  2kn2;NTΔn(nk,k)(3,1),ifk=1;NTΔn(nk,k)(2n3,1),ifk=n1.\widehat{\mathrm{Sp}}_{k,n}\stackrel{{\scriptstyle d}}{{=}}\begin{cases}N-\max\bigl{(}T_{\Delta_{n}\setminus(n-k,k)}(2k-1,1),T_{\Delta_{n}\setminus(n-k,k)}(2k+1,1)\bigr{)},&\text{if}\;\;2\leq k\leq n-2;\\ N-T_{\Delta_{n}\setminus(n-k,k)}(3,1),&\text{if}\;\;k=1;\\ N-T_{\Delta_{n}\setminus(n-k,k)}(2n-3,1),&\text{if}\;\;k=n-1.\end{cases} (4.6)
Proof.

In order to obtain the law of Sp^k,n\widehat{\mathrm{Sp}}_{k,n}, we need to condition on the first swap being kk; through the Edelman-Green bijection this is the same as conditioning on the largest entry NN in the tableau to be in the box (nk,k)(n-k,k) in the (i,j)(i,j) coordinate system. Note that if we condition a uniformly random standard Young tableau of shape Δn\Delta_{n} on the position of largest entry T(nk,k)=NT(n-k,k)=N, then the rest is a uniformly random standard Young tableau of shape Δn(nk,k)\Delta_{n}\setminus(n-k,k).

Once we do conditioning, the value of 1+Sp^k,n1+\widehat{\mathrm{Sp}}_{k,n} becomes the number of the iterations of the Edelman-Greene algorithm when the largest entry of the tableau is at (nk,k)(n-k,k) for the second time. After the first step of the algorithm, the entry at (nk,k)(n-k,k) is

1+max(TΔn(nk,k)(2k1,1),TΔn(nk,k)(2k+1,1))1+\max\bigl{(}T_{\Delta_{n}\setminus(n-k,k)}(2k-1,1),T_{\Delta_{n}\setminus(n-k,k)}(2k+1,1)\bigr{)}

and at each further step the entry grows by 11, until it reaches NN. Hence, we need Nmax(TΔn(nk,k)(2k1,1),TΔn(nk,k)(2k+1,1))N-\max\bigl{(}T_{\Delta_{n}\setminus(n-k,k)}(2k-1,1),T_{\Delta_{n}\setminus(n-k,k)}(2k+1,1)\bigr{)} additional iterations, matching the first case in (4.6). The second and the third cases correspond to the situations when (nk,k)(n-k,k) has only one neighboring box in the tableau. ∎

4.3 Proofs of the main theorems

Proof of Theorem 1.5.

By Lemma 4.4, we need to find the nn\to\infty asymptotics of

𝐓FS,n(k)=dN+1TΔn(2k,1)=Nn1/2n1/2(1N+1TΔn(2k,1)N).\mathbf{T}_{FS,n}(k)\stackrel{{\scriptstyle d}}{{=}}N+1-T_{\Delta_{n}}(2k,1)=\frac{N}{n^{1/2}}\cdot n^{1/2}\left(\frac{1}{N}+1-\frac{T_{\Delta_{n}}(2k,1)}{N}\right).

Recalling N12n2N\sim\tfrac{1}{2}n^{2} and using (3.1) in Theorem 3.1, we get (1.3) with the right-hand side being 12\tfrac{1}{2} times the law of the smallest eigenvalue in 2k×2k2k\times 2k aGUE. The law of the latter is given by the Fredholm determinant through Theorem 2.9 for the case x=y=2kx=y=2k and general properties of the determinantal point processes (cf.[B]). ∎

Remark 4.6.

When k=1k=1, there is only one particle at level x=2x=2 and 𝐓FS(1)\mathbf{T}_{FS}(1) equals in distribution to the coordinate of this particle. The density of this random variable is given by K(2,u;2,u)=2πeu2K(2,u;2,u)=\frac{2}{\sqrt{\pi}}e^{-u^{2}}, u0u\geq 0, as in Remark 3.8. The k=1k=1 result was first obtained in [R, Theorem 1.7].

Proof of Theorem 1.18.

This is a combination of Theorem 1.14, Lemma 4.5 with (3.2) in Theorem 3.1. ∎

Remark 4.7.

When k=1k=1, Z3Z^{\prime}_{3} is the coordinate of the unique particle of X1\pazocal{X}^{{}^{\prime}}_{1} on level x=3x=3. The density of this random variable is given by K1(3,u;3,u)=2ueu2K^{{}^{\prime}}_{1}(3,u;3,u)=2ue^{-u^{2}}, u0u\geq 0. Although without giving a formal definition, [R] used our second definition of spacing implicitly, and obtained this k=1k=1 result in [R, Theorem 1.7].

Proof of Theorem 1.11.

Let Λ=aX0\Lambda_{-}=a-X\in\mathbb{Z}_{\geq 0} and Λ+=Ya>0\Lambda_{+}=Y-a\in\mathbb{Z}_{>0}. We have Λ,Λ+0\Lambda_{-},\Lambda_{+}\geq 0, and Λ+Λ+=Spk,n\Lambda_{-}+\Lambda_{+}=\mathrm{Sp}_{k,n}. Note that Λ+=𝑑𝐓FS,n(k)\Lambda_{+}\overset{d}{=}\mathbf{T}_{FS,n}(k).

By translation invariance of Corollary 1.8,

[Λr,Λ+q]=[Λr,Λ+q],\mathbb{P}\left[\Lambda_{-}\geq r,\,\Lambda_{+}\geq q\right]=\mathbb{P}\left[\Lambda_{-}\geq r^{\prime},\,\Lambda_{+}\geq q^{\prime}\right],

if r,q,r,q0r,q,r^{\prime},q^{\prime}\geq 0 and r+q=r+qr+q=r^{\prime}+q^{\prime}. Hence, for any u,v0u,v\geq 0 by Theorem 1.5

[2Λn32u, 2Λ+n32v]=\displaystyle\mathbb{P}\left[2\frac{\Lambda_{-}}{n^{\frac{3}{2}}}\geq u,\,2\frac{\Lambda_{+}}{n^{\frac{3}{2}}}\geq v\right]= [Λun322,Λ+vn322]=[Λ0,Λ+un322+vn322]\displaystyle\mathbb{P}\left[\Lambda_{-}\geq u\frac{n^{\frac{3}{2}}}{2},\,\Lambda_{+}\geq v\frac{n^{\frac{3}{2}}}{2}\right]=\mathbb{P}\left[\Lambda_{-}\geq 0,\,\Lambda_{+}\geq u\frac{n^{\frac{3}{2}}}{2}+v\frac{n^{\frac{3}{2}}}{2}\right]
=\displaystyle= [2Λ+n32un322+vn322n322]nW(u+v),\displaystyle\mathbb{P}\left[2\frac{\Lambda_{+}}{n^{\frac{3}{2}}}\geq\frac{u\frac{n^{\frac{3}{2}}}{2}+v\frac{n^{\frac{3}{2}}}{2}}{\frac{n^{\frac{3}{2}}}{2}}\right]\stackrel{{\scriptstyle n\to\infty}}{{\longrightarrow}}W(u+v),

where W()W(\cdot) is the distribution function [𝐓(k)]\mathbb{P}\left[\mathbf{T}(k)\geq\cdot\right]. Smoothness of W()W(\cdot) and the above computation show that (2Λn32,2Λ+n32)\bigl{(}2\frac{\Lambda_{-}}{n^{\frac{3}{2}}},2\frac{\Lambda_{+}}{n^{\frac{3}{2}}}\bigr{)} converges in distribution as nn\to\infty to a random vector (Z1,Z2)(Z_{1},Z_{2}), whose probability density we denote f(x,y)f(x,y), x,y0x,y\geq 0. We also let F(a,b):=[Z1>a,Z2>b]F(a,b):=\mathbb{P}\left[Z_{1}>a,Z_{2}>b\right]; we already know that F(a,b)=W(a+b)F(a,b)=W(a+b), a,b0a,b\geq 0. Denoting F1F_{1} and F2F_{2} the partial derivatives of FF in the first and second coordinates, respectively, we have F1(a,b)=F2(a,b)=W(a+b)F_{1}(a,b)=F_{2}(a,b)=W^{\prime}(a+b). Recall that Z1+Z2Z_{1}+Z_{2} is the scaled limit of Spk,n\mathrm{Sp}_{k,n} that we are interested in. We have:

[Z1+Z2>b]\displaystyle\mathbb{P}\left[Z_{1}+Z_{2}>b\right] =0bbyf(x,y)𝑑x𝑑y+b0f(x,y)𝑑x𝑑y\displaystyle=\int_{0}^{b}\int_{b-y}^{\infty}f(x,y)dxdy+\int_{b}^{\infty}\int_{0}^{\infty}f(x,y)dxdy
=0b[F2(by,y)]𝑑y+0[F1(x,b)]𝑑x\displaystyle=\int_{0}^{b}\left[-F_{2}(b-y,y)\right]dy+\int_{0}^{\infty}\left[-F_{1}(x,b)\right]dx
=0bW(b)𝑑y0W(x+b)𝑑x\displaystyle=-\int_{0}^{b}W^{\prime}(b)dy-\int_{0}^{\infty}W^{\prime}(x+b)dx
=bW(b)bW(x)𝑑x\displaystyle=-bW^{\prime}(b)-\int_{b}^{\infty}W^{\prime}(x)dx

Differentiating the last identity in bb we get the desired (1.4). ∎

Proof of Theorem 1.14.

We fix kk and deal with periodic extension of the sorting network of Definition 1.7. Given a random sorting network (st)t(s_{t})_{t\in\mathbb{Z}}, we define a point process 𝒫\mathscr{P} on discrete circle of length 2N2N by declaring that a spot ii is occupied by a particle if and only if si=ks_{i}=k; in other words, this is the point process of times of swaps sks_{k}. By Corollary 1.8 the process 𝒫\mathscr{P} is rotationally invariant and, therefore, the results of Section 4.1 apply. Since we would like to find the distribution of the conditional spacing, we rely on (4.2) — the desired quantity is f2()f_{2}(\ell) in that formula, which is being connected to g()g(\ell) and ρ\rho. The asymptotics of g()g(\ell) is computed in Theorem 1.5. In order to find ρ\rho, we notice that by the Edelman–Greene bijection (cf.  [AHRV, Proposition 9]), it is computed as

ρ=#{Standard Young tableaux of shape Δn(nk,k)}#{Standard Young tableaux of shape Δn}.\rho=\frac{\#\{\text{Standard Young tableaux of shape }\Delta_{n}\setminus(n-k,k)\}}{\#\{\text{Standard Young tableaux of shape }\Delta_{n}\}}.

Both numerator and denominator are explicitly known (they can be computed by the hook length formula [FRT]). Applying the Stirling’s formula, we get as nn\to\infty:

1ρ=\displaystyle\frac{1}{\rho}= (n2)1j<k(2k2j)1j<k(2k12j)k<in(2i2k)k<in(2i2k1)\displaystyle\binom{n}{2}\frac{\prod_{1\leq j<k}(2k-2j)}{\prod_{1\leq j<k}(2k-1-2j)}\,\frac{\prod_{k<i\leq n}(2i-2k)}{\prod_{k<i\leq n}(2i-2k-1)}
=\displaystyle= (n2)(2n22k)!!(2n12k)!!(2k2)!!(2k1)!!π4n32(2k2)!!(2k1)!!.\displaystyle\binom{n}{2}\frac{(2n-2-2k)!!}{(2n-1-2k)!!}\frac{(2k-2)!!}{(2k-1)!!}\sim\frac{\sqrt{\pi}}{4}n^{\frac{3}{2}}\frac{(2k-2)!!}{(2k-1)!!}. (4.7)

Now take 0a<b0\leq a<b and sum (4.2) over an322bn322\lfloor a\frac{n^{\frac{3}{2}}}{2}\rfloor\leq\ell\leq\lfloor b\frac{n^{\frac{3}{2}}}{2}\rfloor. We get

an322bn322f2()=1ρ(g(an322)g(bn322+1)).\sum_{\lfloor a\frac{n^{\frac{3}{2}}}{2}\rfloor}^{\lfloor b\frac{n^{\frac{3}{2}}}{2}\rfloor}f_{2}(\ell)=\frac{1}{\rho}\left(g\left(\lfloor a\frac{n^{\frac{3}{2}}}{2}\rfloor\right)-g\left(\lfloor b\frac{n^{\frac{3}{2}}}{2}\rfloor+1\right)\right). (4.8)

We send nn\to\infty using Theorem 1.5, (4.7) and the following lemma, whose proof we leave as an exercise for the reader (the monotonicity condition is implied by (4.2)).

Lemma 4.8.

Let ξ1,ξ2,\xi_{1},\xi_{2},\dots be 0\mathbb{Z}_{\geq 0}–valued random variables, such that for each nn the probabilities [ξn=]\mathbb{P}\left[\xi_{n}=\ell\right] depend on =0,1,2\ell=0,1,2\dots in a monotone way. Suppose that we have a distributional limit limn2ξnn3/2=dξ\lim\limits_{n\to\infty}\frac{2\xi_{n}}{n^{3/2}}\stackrel{{\scriptstyle d}}{{=}}\xi, where ξ\xi is an absolutely continuous random variable with density h(x)h(x). Then for each a0a\geq 0

limnn3/22[ξn=an322]=h(a).\lim_{n\to\infty}\frac{n^{3/2}}{2}\mathbb{P}\left[\xi_{n}=\lfloor a\frac{n^{\frac{3}{2}}}{2}\rfloor\right]=h(a). (4.9)

Hence, in terms of Sp^k,n\widehat{\mathrm{Sp}}_{k,n} of Theorem 1.14, as nn\to\infty (4.8) implies

limn[a2Sp^k,nn32b]=π2(2k2)!!(2k1)!![b([𝐓FS(k)>b])a([𝐓FS(k)>a])].\lim_{n\to\infty}\mathbb{P}\left[a\leq 2\frac{\widehat{\mathrm{Sp}}_{k,n}}{n^{\frac{3}{2}}}\leq b\right]=\frac{\sqrt{\pi}}{2}\frac{(2k-2)!!}{(2k-1)!!}\left[\frac{\partial}{\partial b}\bigl{(}\mathbb{P}\left[\mathbf{T}_{\rm{FS}}(k)>b\right]\bigr{)}-\frac{\partial}{\partial a}\bigl{(}\mathbb{P}\left[\mathbf{T}_{\rm{FS}}(k)>a\right]\bigr{)}\right]. (4.10)

Thus, 2Sp^k,nn322\frac{\widehat{\mathrm{Sp}}_{k,n}}{n^{\frac{3}{2}}} converges in distribution to a random variable of density given by (1.5). ∎

Remark 4.9.

We expect that an alternative proof of Theorem 1.11 can be obtained by following the lines of the just presented argument, but using (4.1) instead of (4.2).

Remark 4.10.

Theorem 1.14 and Theorem 1.18 give two different expressions for the distribution of the same random variable Z^k\widehat{Z}_{k}. It would be interesting to find a direct (i.e. avoiding taking the limit from the sorting networks) proof that these expressions coincide.

References

  • [ADHV] O. Angel, D. Dauvergne, A. E. Holroyd, B. Virag, The Local Limit of Random Sorting Networks, Annales Institut Henri Poincare: Probability and Statististics 55, no. 1 (2019), 412–440. arXiv:1702.08368.
  • [AGH] O. Angel, V. Gorin, A. E. Holroyd, A pattern theorem for random sorting networks. Electronic Journal of Probability, 17, paper 99 (2012), 1–16. arXiv:1110.0160.
  • [AHRV] O. Angel, A. Holroyd, D. Romik, B. Virág, Random sorting networks, Advances in Mathematics 215, no. 2 (2007), 839–864, arXiv:0609538.
  • [B] A. Borodin, Determinantal point processes, Oxford Handbook of Random Matrix Theory, Oxford University Press, 2011, arXiv:0911.1153.
  • [DVe] D. Daley, D. Vere-Jones, An introduction to the theory of point processes: Vol. I. Elementary theory and methods, Springer–Verlag, New York, 2003.
  • [D] D. Dauvergne, The Archimedean limit of random sorting networks, to appear in Journal of the American Mathematical Society, arXiv:1802.08934.
  • [DVi] D. Dauvergne, B. Virag, Circular support in random sorting networks, Transactions of the American Mathematical Society 373 (2020), 1529-1553 arXiv:1802.08933.
  • [EG] P. Edelman, C. Greene, Balanced tableaux, Advances in Mathematics 63, no. 1 (1987), 42–99.
  • [GR] V. Gorin, M. Rahman, Random sorting networks: local statistics via random matrix laws, Probability Theory and Related Fields 175, no. 1 (2019), 45–96.
  • [F] P. J. Forrester, Log-Gases and Random Matrices, Princeton University Press, 2010.
  • [FRT] J. S. Frame, G. B. Robinson, and R. M. Thrall. The hook graphs of the symmetric groups. Canadian Journal of Mathematics, 6 (1954), 316–324.
  • [FN] P. J. Forrester, E. Nordenstam, The anti-symmetric GUE minor process, Moscow Mathematical Journal 9, no. 4 (2009), 749–774, arXiv:0804.3293.
  • [JKB] N. Johnson, S. Kotz, N. Balakrishnan, Continuous Univariate Distributions, vol. 2 (2nd edition). Wiley, 1995.
  • [K] M. Kotowski, Limits of random permuton processes and large deviations for the interchange process, PhD Thesis, University of Toronto, 2016.
  • [Le] A. Lenard, Correlation functions and the uniqueness of the state in classical statistical mechanics, Communications in Mathematical Physics 30 (1973), 35–44.
  • [M] M. L. Mehta, Random matrices, third edition, Pure and Applied Mathematics (Amsterdam), vol. 142, Elsevier/Academic Press, Amsterdam, 2004.
  • [P] L. Petrov, Asymptotics of random lozenge tilings via Gelfand-Tsetlin schemes, Probability Theory and Related Fields 160, np. 3 (2014), 429–487.
  • [RVV] M. Rahman, B. Virag, M. Vizer, Geometry of Permutation Limits, Combinatorica 39 (2019), 933–960. arXiv:1609.03891.
  • [R] A. Rozinov, Statistics of Random Sorting Networks, PhD Thesis, Courant Institute, NYU, 2016.
  • [S] R. P. Stanley, On the number of reduced decompositions of elements of Coxeter groups, European Journal of Combinatorics 5, no. 4 (1984), 359–372.
  • [Sz] G. Szego, Orthogonal polynomials. American Mathematical Society, 1939.