This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Validation tests of Gaussian boson samplers with photon-number resolving detectors

Alexander S. Dellios, Margaret D. Reid and Peter D. Drummond [email protected] Centre for Quantum Science and Technology Theory, Swinburne University of Technology, Melbourne 3122, Australia
Abstract

An important challenge with the current generation of noisy, large-scale quantum computers is the question of validation. Does the hardware generate correct answers? If not, what are the errors? This issue is often combined with questions of computational advantage, but it is a fundamentally distinct issue. In current experiments, complete validation of the output statistics is generally not possible because it is exponentially hard to do so. Here, we apply phase-space simulation methods to partially verify recent experiments on Gaussian boson sampling (GBS) implementing photon-number resolving (PNR) detectors. The positive-P phase-space distribution is employed, as it uses probabilistic sampling to reduce complexity. It is 101810^{18} times faster than direct classical simulation for experiments on 288288 modes where quantum computational advantage is claimed. When combined with binning and marginalization to improve statistics, multiple validation tests are efficiently computable, of which some tests can be carried out on experimental data. We show that the data as a whole shows discrepancies with theoretical predictions for perfect squeezing. However, a small modification of the GBS parameters greatly improves agreement. Hence, we suggest that such validation tests could form the basis of feedback methods to improve GBS quantum computer experiments.

I Introduction

Theoretical and experimental interest in photonic networks employed as quantum computers has grown considerably [1, 2]. These devices interfere nonclassical photons in a large scale interferometric network, then measure the resulting output photons using photodetectors. The computational task is sampling from a computationally hard probability distribution. Gaussian boson sampling (GBS) has proved scalable to the sizes needed for quantum advantage, which recent experiments have now claimed [3, 4, 5, 6]. In this case, photons are prepared in a squeezed state, with output probabilities in a #P-hard computational class [7, 8]. For photon-number resolving (PNR) detectors this is a Hafnian, while for threshold detectors it is a Torontonian. Neither function is computable in polynomial time, so that the random counts cannot be classically replicated at large scale.

In this paper, we use positive P-representation simulations to validate Gaussian boson sampling with photon number resolving (PNR) detectors, extending the tests that can be applied to experimental data. This is essential, as a number of large scale GBS experiments have claimed quantum advantage using either PNR [5] or threshold detectors [3, 4, 6]. To perform the tests, we generalize a binning method developed by Mandel [9, 10] for simulating photon statistics of classical light. The original formulation, which utilized the Glauber-Sudarshan P-representation [11, 12] to simulate moments for classical fields, is extended to the positive P-representation, which can treat any quantum state. An extension to binning in multiple dimensions is also provided, as these are more sensitive tests of quantum advantage than one-dimensional distributions.

We show that in the challenging prospect of using quantum computers to demonstrate quantum computational advantage, the GBS approach has a very important advantage. Unlike many other proposed quantum computing architectures, there are multiple quantitative statistical tests that are dependent on the experimental parameters, which can be used to validate the output distributions of the data. If experimental data fail such tests, it is invalid. This is an important criterion, used in comparable tests of classical random number generators and cryptographic security [13, 14]. Implementing such tests requires algorithms that compute the correlations or moments without needing to compute the full #P hard distribution.

An efficient and scalable method for validating GBS with threshold detectors, demonstrated for up to 16,00016,000 modes, was proposed in Ref.[15] and used to analyze data from a 100100-mode experiment [3]. The normally ordered positive P-representation was utilized to simulate grouped count probabilities (GCPs). This is a generalized binned photon counting distribution that includes all measurable marginal and binned correlations. The exponential sparsity of the detected count pattern probabilities means the full distribution cannot be either computed or measured. Therefore, validation tests based on marginals and binned probability distributions appear to be the most practical route to comprehensive validation.

Like phase-space methods for threshold detectors [15], the PNR method is efficient and scalable. Increased dimension of binning causes an increase in computation time, where in the limit of dimension equaling mode number the GCP converges to the Hafnian. PNR phase-space methods allow one to perform simulations of either the ground truth distribution or modifications such as decoherence, measurement errors and shot-to-shot fluctuations. Such errors are typically not analyzed, despite evidence suggesting that a thermalized, non-ideal ground truth distribution provides stronger claims of quantum advantage than the ideal ground truth distribution [16]. Unlike threshold detection, PNR detector binning has the advantage of not requiring the computation of multidimensional inverse discrete Fourier transforms that increase computation time for multidimensional GCPs with threshold detection.

The positive-P phase-space distribution is exact, non-singular and positive for any quantum state. The moments of the distribution are identical to normally ordered operator moments including intensity correlations [17], which are reproduced to any order. This property, coupled with probabilistic random sampling, makes the positive P-representation ideal for validating the statistics of nonclassical photon counting experiments such as GBS. It does not have the computational restraints imposed by other validation methods [18, 19, 20], since it is many orders of magnitude faster than direct classical simulation, with negligible rounding errors. While there are sampling errors, for current experimental parameters they are comparable to sampling errors in the measurement statistics. Both can be reduced in a controllable way.

Other validation methods involve comparing experimental count patterns to those obtained from classical algorithms that replicate the photo-detection measurements either exactly [20] or approximately [18, 19]. Exact algorithms generate count patterns by directly sampling the distributions, but are limited to small experiments only. Approximate algorithms typically neglect high-order correlations, and only use the low-order correlations. Despite the usefulness of these algorithms in defining classical simulation boundaries of quantum advantage and determining the correlations present in a GBS experiment, they are either computationally demanding or lack high-order correlations, which is necessary for claims of quantum advantage. This limits the application of these approaches when validating GBS experiments and investigating physical processes arising within networks.

In order to determine the validity of using the positive P-representation to simulate nonclassical photon counting experiments with PNR detectors, we compare simulated GCP moments to the exactly known photon counting distributions for special cases. Here, multi-mode pure squeezed states are transformed by a Haar random unitary matrix which is multiplied by a uniform amplitude loss coefficient. For losses comparable to current experiments, we see excellent agreement. In the less realistic lossless case, the resulting non-Gaussian number distribution oscillates strongly between even and odd total counts. While this is far from current experiments, the standard positive P-representation requires a much larger number of phase-space samples to converge in such cases. However, this limit can also be efficiently simulated if necessary using a symmetry-projected [21] phase-space method, which will be treated elsewhere.

Finally, using photon count data from the recent Borealis experiment implementing PNR detectors [5], we apply the simulated GCPs to validate the moments of measured photon counting distributions by comparing theory and experiment. Differences are quantified using chi-square and Z-statistic tests. As with earlier work [15, 16], we show that the experimental GCPs are far from the ground truth distribution for pure squeezed state inputs. Instead, they are described much better by a different target distribution which includes decoherence and measurement error corrections of the transmission matrix. Although we focus specifically on implementations to GBS, we emphasize that one can use the positive P-representation to simulate GCPs of any linear photon counting experiment using classical or nonclassical photons.

This paper is structured as follows, Section II introduces the necessary background and theoretical methods for simulating GCPs of any photon counting experiment using PNR detectors with nonclassical photons in phase-space. Section IV then applies this method to the exactly known multi-mode squeezed state distributions, while Section V simulates the theoretical distributions corresponding to experimental data from the Borealis GBS claiming quantum advantage.

II Grouped count probabilities

We first consider a standard GBS experiment where NN single-mode squeezed vacuum states are sent into a linear photonic network defined by an M×MM\times M transmission matrix 𝑻\boldsymbol{T}. If each input mode is independent, the full input density operator is defined as

ρ^(in)=jN|rjrj|,\hat{\rho}^{(\text{in})}=\prod_{j}^{N}\left|r_{j}\right\rangle\left\langle r_{j}\right|, (1)

where |rj=S^(rj)|0\left|r_{j}\right\rangle=\hat{S}(r_{j})\left|0\right\rangle denotes the squeezed vacuum state of the jj-th mode with squeezing parameter rjr_{j} and S^(rj)\hat{S}(r_{j}) is the squeezing operator [22].

Since squeezed states are Gaussian, they are characterized entirely by their quadrature variances

(Δx^j)2\displaystyle\left\langle\left(\Delta\hat{x}_{j}\right)^{2}\right\rangle =2(nj+mj)+1=e2rj\displaystyle=2\left(n_{j}+m_{j}\right)+1=e^{2r_{j}}
(Δy^j)2\displaystyle\left\langle\left(\Delta\hat{y}_{j}\right)^{2}\right\rangle =2(njmj)+1=e2rj,\displaystyle=2\left(n_{j}-m_{j}\right)+1=e^{-2r_{j}}, (2)

where the quadrature operators x^j=a^j(in)+a^j(in)\hat{x}_{j}=\hat{a}_{j}^{(\text{in})}+\hat{a}_{j}^{\dagger(\text{in})} and y^j=(a^j(in)a^j(in))/i\hat{y}_{j}=\left(\hat{a}_{j}^{(\text{in})}-\hat{a}_{j}^{\dagger(\text{in})}\right)/i satisfy the commutation relations [x^j,y^k]=2iδjk\left[\hat{x}_{j},\hat{y}_{k}\right]=2i\delta_{jk}, and the mean photon number and coherence per mode is defined as nj=a^j(in)a^j(in)=sinh2(rj)n_{j}=\left\langle\hat{a}_{j}^{\dagger(\text{in})}\hat{a}_{j}^{(\text{in})}\right\rangle=\sinh^{2}(r_{j}) and mj=a^j(in)a^j(in)=cosh(rj)sinh(rj)m_{j}=\left\langle\hat{a}_{j}^{(\text{in})}\hat{a}_{j}^{(\text{in})}\right\rangle=\cosh(r_{j})\sinh(r_{j}), respectively.

While most theoretical analyses of GBS consider pure squeezed state inputs, this is impractical, since decoherence effects such as mode mismatches [3], partial distinguishability [23], and many more [24] will always be present during experimental applications in quantum optics. Therefore, a realistic model of input states are Gaussian thermalized squeezed states for which a simple parameterization was introduced in [15]. Here, a thermalization component ϵ\epsilon alters the intensity of the input photons as measured by the coherence, which is reduced by a factor of 1ϵ1-\epsilon such that m~j=(1ϵ)mj\tilde{m}_{j}=(1-\epsilon)m_{j}. The quadrature variances are then altered as [15]

σxj2=(Δx^j)2\displaystyle\sigma_{x_{j}}^{2}=\left\langle\left(\Delta\hat{x}_{j}\right)^{2}\right\rangle =2(nj+m~j)\displaystyle=2\left(n_{j}+\tilde{m}_{j}\right)
σyj2=(Δy^j)2\displaystyle\sigma_{y_{j}}^{2}=\left\langle\left(\Delta\hat{y}_{j}\right)^{2}\right\rangle =2(njm~j),\displaystyle=2\left(n_{j}-\tilde{m}_{j}\right), (3)

where for ϵ=0\epsilon=0 the pure squeezed state definition Eq.(2) returns, while ϵ=1\epsilon=1 produces fully decoherent thermal states with σxj2=σyj2=2nj\sigma_{x_{j}}^{2}=\sigma_{y_{j}}^{2}=2n_{j}. More complex thermalized models are also possible, for example with mode dependent thermalization, ϵj\epsilon_{j}.

In the ideal GBS scenario, the transmission matrix is a Haar random unitary. However experimentally realistic transmission matrices contain losses, making them non-unitary. Throughout this paper, 𝑻\boldsymbol{T} is used to define a lossy transmission matrix, while 𝑼\boldsymbol{U} defines a Haar random unitary. When losses are present, the creation and annihilation operators for the input modes a^j(in),a^j(in)\hat{a}_{j}^{\dagger(\text{in})},\hat{a}_{j}^{(\text{in})} are converted to output modes following

a^k(out)=j=1NTkja^j(in)+j=1MBkjb^j(in),\hat{a}_{k}^{\dagger(\text{out})}=\sum_{j=1}^{N}T_{kj}\hat{a}_{j}^{\dagger(\text{in})}+\sum_{j=1}^{M}B_{kj}\hat{b}_{j}^{\dagger(\text{in})}, (4)

where 𝑩\boldsymbol{B} is a random noise matrix representing photon loss to jj-the environmental mode b^j(in)\hat{b}_{j}^{\dagger(\text{in})}. The resulting output photon number n^k=a^k(out)a^k(out)\hat{n}^{\prime}_{k}=\hat{a}_{k}^{\dagger(\text{out})}\hat{a}_{k}^{(\text{out})} is then measured by a series of photodetectors.

II.1 Photon counting projectors

Detectors with photon number resolution implement the normally ordered projection operator [25, 26]

p^j(cj)=1cj!:(n^j)cjen^j:,\hat{p}_{j}(c_{j})=\frac{1}{c_{j}!}:\left(\hat{n}^{\prime}_{j}\right)^{c_{j}}e^{-\hat{n}^{\prime}_{j}}:, (5)

where :::\cdots: denotes normal ordering and cj=0,1,2,3,c_{j}=0,1,2,3,\dots is the number of measured photon counts at the jj-th detector. Although theoretically there is no upper bound on cjc_{j}, practically PNR detectors saturate for some maximum number of counts cj(max)c_{j}^{(\text{max})}. This value is variable, typically depending on the physical implementation of the detector.

In any multi-mode photon counting experiment, a set of photon count patterns 𝒄=[c1,,cM]\boldsymbol{c}=[c_{1},\dots,c_{M}] are observed by measuring the output state ρ^(out)\hat{\rho}^{(\text{out})}. A straightforward extension of Eq.(5) allows one to define the projection operator for a specific count pattern as

P^(𝒄)=j=1Mp^j(cj).\hat{P}(\boldsymbol{c})=\bigotimes_{j=1}^{M}\hat{p}_{j}(c_{j}). (6)

For GBS, each pattern is a sample from the full output probability distribution and computing its expectation value corresponds to solving the matrix Hafnian [7, 27]

P^(𝒄)=1det(𝑸)|Haf(𝑩S)|2j=1cj!,\left\langle\hat{P}(\boldsymbol{c})\right\rangle=\frac{1}{\sqrt{\text{det}(\boldsymbol{Q})}}\frac{\left|\text{Haf}\left(\boldsymbol{B}_{S}\right)\right|^{2}}{\prod_{j=1}c_{j}!}, (7)

which is a #P-hard computational task due to its relationship to the matrix permanent. Here, Haf(𝑩S)\text{Haf}(\boldsymbol{B}_{S}) is the Hafnian of the square sub-matrix of 𝑩=𝑼(j=1Mtanh(rj))𝑼T\boldsymbol{B}=\boldsymbol{U}\left(\bigoplus_{j=1}^{M}\tanh(r_{j})\right)\boldsymbol{U}^{T} for pure state inputs with zero displacement, where the sub-matrix is formed from the output channels with recorded photon counts. The symmetric matrix 𝑩\boldsymbol{B} is related to the kernel matrix 𝑨=𝑩𝑩\boldsymbol{A}=\boldsymbol{B}\bigoplus\boldsymbol{B}^{*} which is formed from the 2M×2M2M\times 2M covariance matrix 𝑸\boldsymbol{Q} as [28]

𝑨=(0𝑰M𝑰M0)(𝑰M𝑸1),\boldsymbol{A}=\left(\begin{array}[]{cc}0&\boldsymbol{I}_{M}\\ \boldsymbol{I}_{M}&0\end{array}\right)\left(\boldsymbol{I}_{M}-\boldsymbol{Q}^{-1}\right), (8)

where 𝑰M\boldsymbol{I}_{M} denotes the M×MM\times M identity matrix. When the covariance matrix elements are non-negative, the kernel matrix and hence the symmetric matrix 𝑩\boldsymbol{B} are also non-negative [28] and the Hafnian can be estimated in polynomial time [29, 30].

In the current generation of GBS networks, photon loss is the dominant noise source. Although there is evidence that sampling from a lossy GBS distribution is also a #P-hard computational task [31], output photons from experiments implementing high loss networks become classical [32], allowing the output distribution to become efficiently sampled. There is currently no clear boundary between the computationally hard and efficiently sampled regimes, but a classical algorithm that exploits large experimental photon loss has recently been developed [33] and shown to produce samples that are closer to the ideal ground truth than recent experiments claiming quantum advantage. The effects of thermalization on the computational complexity of GBS is currently unknown, despite thermalization causing nonclassical photons to become classical at a faster rate than pure losses.

II.2 Grouped count probabilities for PNR detectors

Determining photon counting distributions by binning the photons arriving at photodetectors has a long history in quantum optics. Although binning methods can vary, e.g. binning continuous measurements within a time interval such as inverse bandwidth time [34, 35], such distributions are both experimentally accessible and theoretically useful. For pure squeezed states these can provide evidence of nonclassical behavior, as observed in the even-odd count oscillations [36]. For GBS there is also a statistical purpose, as the set of all possible observable patterns 𝒮P\mathcal{S}_{P} grows exponentially with mode number, causing the output probabilities to become exponentially sparse.

In this paper, we are interested in the total output photon number operator obtained from a subset of SjS_{j} efficient detectors:

n^Sj=iSjn^i.\hat{n}^{\prime}_{S_{j}}=\sum_{i\in S_{j}}\hat{n}^{\prime}_{i}. (9)

Our observable of interest is then the photon number distribution corresponding to counting the projectors onto eigenstates of n^Sj\hat{n}^{\prime}_{S_{j}}.

Following from the nomenclature introduced in [15], we define the dd-dimensional grouped photon counting distribution as the probability of observing 𝒎=[m1,,md]\boldsymbol{m}=[m_{1},\dots,m_{d}] grouped counts in jj distinct subsets of output modes Sj={1,,M/d}S_{j}=\{1,\dots,M/d\}, such that

𝒢𝑺(n)(𝒎)=j=1d[ci=mjP^Sj(𝒄)],\mathcal{G}_{\boldsymbol{S}}^{(n)}(\boldsymbol{m})=\left\langle\prod_{j=1}^{d}\left[\sum_{\sum c_{i}=m_{j}}\hat{P}_{S_{j}}(\boldsymbol{c})\right]\right\rangle, (10)

where n=j=1dMjMn=\sum_{j=1}^{d}M_{j}\leq M is the total correlation order and each grouped count is obtained from the summation mj=iSjcim_{j}=\sum_{i\in S_{j}}c_{i}. The general form of the pattern projector for any number of output modes within each subset is

P^Sj(𝒄)=iSj1ci!:(n^i)cien^i:.\hat{P}_{S_{j}}\left(\boldsymbol{c}\right)=\bigotimes_{i\in S_{j}}\frac{1}{c_{i}!}:(\hat{n}^{\prime}_{i})^{c_{i}}e^{-\hat{n}^{\prime}_{i}}:. (11)

Multi-dimensional probabilities are defined over a vector of disjoint subsets 𝑺=(S1,,Sd)\boldsymbol{S}=\text{$\left(S_{1},\dots,S_{d}\right)$} of mode numbers 𝑴=(M1,,Md)\boldsymbol{M}=\left(M_{1},\dots,M_{d}\right), in which case GCPs become joint probability distributions. This is readily illustrated when d=2d=2 and n=Mn=M, giving 𝒢S1,S2(M)(m1,m2)\mathcal{G}_{S_{1},S_{2}}^{(M)}(m_{1},m_{2}). Increasing the dimension to d=Md=M while keeping the correlation order constant at n=Mn=M causes the GCP to converge to the Hafnian.

Therefore, if the correlation order satisfies n=Mn=M, an increase in GCP dimension corresponds to observing the dd-th binned Glauber intensity correlation, hence high-order correlations become more statistically significant in the limit dMd\rightarrow M. If such correlations are present in the experimental data, multi-dimensional GCPs provide a more thorough test of quantum advantage as they allow one to directly compare these high-order correlations with their theoretical values [16]. If n<Mn<M, the GCP becomes a marginal probability as the remaining MnM-n output modes are ignored.

III Normally-ordered phase-space methods

Here we summarize phase-space methods, and explain the motivation for using them. The purpose is to transform an exponentially complex Hilbert space into a positive distribution that can be randomly sampled. Normally-ordering methods are the closest to experimental statistics using photon-counting. Representations that are not normally ordered, like the Wigner [37] or Q-function [38] approach, have a rapidly growing sampling error with system size.

III.1 Glauber-Sudarshan representation

In its original formulation, Mandel [9, 10] used the normally ordered Glauber-Sudarshan P-representation [11, 12] to evaluate the expectation value in Eq.(10). The P-representation expands the density operator as a diagonal sum of coherent states [11]

ρ^=P(𝜶)|𝜶𝜶|d2M𝜶,\hat{\rho}=\int P(\boldsymbol{\alpha})\left|\boldsymbol{\alpha}\right\rangle\left\langle\boldsymbol{\alpha}\right|\text{d}^{2M}\boldsymbol{\alpha}, (12)

hence its commonly called the diagonal P-representation, where |𝜶=j=1M|αj\left|\boldsymbol{\alpha}\right\rangle=\bigotimes_{j=1}^{M}\left|\alpha_{j}\right\rangle is a multi-mode coherent state vector with eigenvalues 𝜶=[α1,,αM]\boldsymbol{\alpha}=[\alpha_{1},\dots,\alpha_{M}] and P(𝜶)P(\boldsymbol{\alpha}) is the P-representation probability distribution. The diagonal P-representation can be used to evaluate expectation values of the product of normally ordered operators due to the relationship

j=1M:a^ja^j:\displaystyle\left\langle\prod_{j=1}^{M}:\hat{a}_{j}^{\dagger}\hat{a}_{j}:\right\rangle =limESj=1M|αj|2DP\displaystyle=\lim_{E_{S}\rightarrow\infty}\left\langle\prod_{j=1}^{M}\left|\alpha_{j}\right|^{2}\right\rangle_{DP}
=P(𝜶)(j=1M|αj|2)d2M𝜶,\displaystyle=\int P(\boldsymbol{\alpha})\left(\prod_{j=1}^{M}\left|\alpha_{j}\right|^{2}\right)\text{d}^{2M}\boldsymbol{\alpha}, (13)

where \left\langle\dots\right\rangle is a quantum expectation value, ESE_{S} is an ensemble of random phase-space samples and DP\left\langle\dots\right\rangle_{DP} is the diagonal-P phase-space ensemble mean.

Following the standard replacement of a^jαj\hat{a}_{j}\rightarrow\alpha_{j} and a^jαj\hat{a}_{j}^{\dagger}\rightarrow\alpha_{j}^{*}, where nj=|αj|2n^{\prime}_{j}=\left|\alpha_{j}\right|^{2} and n^SjnSj=iSj|αi|2\hat{n}^{\prime}_{S_{j}}\rightarrow n^{\prime}_{S_{j}}=\sum_{i\in S_{j}}\left|\alpha_{i}\right|^{2}, Mandel [9, 10] simulates Eq.(10) for d=1d=1 and n=Mn=M as

𝒢S(M)(m)=P(𝜶)[1m!(nS)menS]d2M𝜶,\mathcal{G}_{S}^{(M)}(m)=\int P(\boldsymbol{\alpha})\left[\frac{1}{m!}\left(n^{\prime}_{S}\right)^{m}e^{-n^{\prime}_{S}}\right]\text{d}^{2M}\boldsymbol{\alpha}, (14)

where S={1,,M}S=\{1,\dots,M\}. However, this is only valid for classical states such as coherent, thermal and squashed states [39] as they are defined entirely by their diagonal density matrix elements. Since the density operator of nonclassical states contains off-diagonal elements due to quantum superpositions, if the diagonal P-representation is used for simulations of nonclassical states P(𝜶)P(\boldsymbol{\alpha}) becomes singular and negative. From standard mathematical requirements that it should be real, normalizable, non-negative and non-singular, it is no longer a probability distribution.

III.2 Positive-P representation

To simulate the photon number distributions of nonclassical states one must instead use the normally ordered positive P-representation [40], which is part of a family of generalized P-representations that produce non-singular and exact phase-space distributions for any quantum state. The positive P-representation expands the density operator as a 4M4M-dimensional volume integral [40]

ρ^=P(𝜶,𝜷)|𝜶𝜷|𝜷|𝜶d2M𝜶d2M𝜷,\hat{\rho}=\iint P(\boldsymbol{\alpha},\boldsymbol{\beta})\frac{\left|\boldsymbol{\alpha}\right\rangle\left\langle\boldsymbol{\beta}\right|}{\left\langle\boldsymbol{\beta}|\boldsymbol{\alpha}\right\rangle}\text{d}^{2M}\boldsymbol{\alpha}\text{d}^{2M}\boldsymbol{\beta}, (15)

where the additional dimensions allow off-diagonal density matrix elements to be included in the coherent state basis |𝜷=j=1M|βj\left|\boldsymbol{\beta}\right\rangle=\bigotimes_{j=1}^{M}\left|\beta_{j}\right\rangle with eigenvalues 𝜷=[β1,,βM]\boldsymbol{\beta}=[\beta_{1},\dots,\beta_{M}]. The coherent amplitudes 𝜶\boldsymbol{\alpha}, 𝜷\boldsymbol{\beta} are independent and can vary along the entire complex plane. We can also take the real part of Eq.(15), since the density matrix is hermitian. The benefit of using the positive P-representation for photon counting experiments is that it includes the diagonal P-representation when the positive-P distribution satisfies P(𝜶,𝜷)=P(𝜶)δ(𝜶-𝜷)P(\boldsymbol{\alpha},\boldsymbol{\beta})=P(\boldsymbol{\alpha})\delta(\text{$\boldsymbol{\alpha}$-$\boldsymbol{\beta}$}). This allows us to define classical phase-space as having 𝜷=𝜶\boldsymbol{\beta}=\boldsymbol{\alpha}, while a nonclassical phase space requires non-vanishing probability of 𝜷𝜶\boldsymbol{\beta}\neq\boldsymbol{\alpha}.

Expectation values of products of normally ordered operators for nonclassical states can now be directly calculated from the moments of the positive-P distribution

j=1M:a^ja^j:\displaystyle\left\langle\prod_{j=1}^{M}:\hat{a}_{j}^{\dagger}\hat{a}_{j}:\right\rangle =limESj=1MαjβjP\displaystyle=\lim_{E_{S}\rightarrow\infty}\left\langle\prod_{j=1}^{M}\alpha_{j}\beta_{j}^{*}\right\rangle_{P}
=P(𝜶,𝜷)(j=1Mαjβj)d2M𝜶d2M𝜷,\displaystyle=\iint P(\boldsymbol{\alpha},\boldsymbol{\beta})\left(\prod_{j=1}^{M}\alpha_{j}\beta_{j}^{*}\right)\text{d}^{2M}\boldsymbol{\alpha}\text{d}^{2M}\boldsymbol{\beta}, (16)

where we define P\left\langle\dots\right\rangle_{P} as the positive-P ensemble mean and have used the c-number replacements a^jαj\hat{a}_{j}\rightarrow\alpha_{j} and a^jβj\hat{a}_{j}^{\dagger}\rightarrow\beta_{j}^{*}.

To simulate Eq.(10) for nonclassical states requires replacing the total output photon number Eq.(9) by the phase-space observable

nSj=iSjni=iSjαi(βi),n^{\prime}_{S_{j}}=\sum_{i\in S_{j}}n^{\prime}_{i}=\sum_{i\in S_{j}}\alpha^{\prime}_{i}\left(\beta^{\prime}_{i}\right)^{*}, (17)

such that the general multidimensional probability distribution is

𝒢𝑺(n)(𝒎)\displaystyle\mathcal{G}_{\boldsymbol{S}}^{(n)}(\boldsymbol{m}) =j=1d[ci=mjPSj(𝒄)]P,\displaystyle=\left\langle\prod_{j=1}^{d}\left[\sum_{\sum c_{i}=m_{j}}P_{S_{j}}(\boldsymbol{c})\right]\right\rangle_{P},
=ReP(𝜶,𝜷)[j=1d[1mj!(nSj)mjenSj]]dμ,\displaystyle=\text{Re}\int P(\boldsymbol{\alpha},\boldsymbol{\beta})\left[\prod_{j=1}^{d}\left[\frac{1}{m_{j}!}\left(n^{\prime}_{S_{j}}\right)^{m_{j}}e^{-n^{\prime}_{S_{j}}}\right]\right]\text{d}\mu, (18)

where dμd𝜶d𝜷\text{d}\mu\equiv\text{d}\boldsymbol{\alpha}\text{d}\boldsymbol{\beta} and the phase-space pattern projector is

PSj(𝒄)=iSj1ci!(ni)cieni.P_{S_{j}}(\boldsymbol{c})=\prod_{i\in S_{j}}\frac{1}{c_{i}!}(n^{\prime}_{i})^{c_{i}}e^{-n^{\prime}_{i}}. (19)

We have taken the real part as is required for the applications presented in the next section.

Equation 18 is a more general definition of GCPs than Mandel’s original formulation, since one can now simulate the marginal and multidimensional probabilities of any classical or nonclassical photon counting experiment. To simulate GCPs for input photons in a number state, it is more efficient to use another generalized P-representation: the complex P-representation. In this case P(𝜶,𝜷)P(\boldsymbol{\alpha},\boldsymbol{\beta}) includes a complex weight due to expanding the density operator as a contour integral [40]. Although the positive P-representation exists for Fock states, it has larger sampling errors than the complex version [41], meaning fewer stochastic samples are needed to perform accurate simulations.

Unlike methods for threshold detectors [15], Eq.(18) doesn’t require multidimensional inverse discrete Fourier transforms, which increase computation time significantly [16]. Since threshold detector outputs are always binary, the largest number of observable photon counts in a pattern corresponds to a detection event in every output mode, such that iMciM\sum_{i}^{M}c_{i}\leq M where the largest observable grouped count is m(M)m^{(M)}.

Since PNR detectors can resolve multiple photons arriving at each detector, an experiment or classical algorithm replicating photo-detection measurements can now produce count patterns where iMci>M\sum_{i}^{M}c_{i}>M, in which case m(M)m^{(M)} is no longer the largest observable grouped count. Instead of scaling with mode number, numerical simulations scale with photon count bins, where we define the maximum observable grouped count as m(max)m^{(\text{max})}. For experiments with small mean input photon numbers or high losses, one may have m(max)<m(M)m^{(\text{max})}<m^{(M)}, while low losses and large photon numbers will produce m(max)>m(M)m^{(\text{max})}>m^{(M)}. Typically, we choose m(max)m^{(\text{max})} to occur when probabilities are of the order 𝒢S(n)(m(max))=107\mathcal{G}_{S}^{(n)}(m^{(\text{max})})=10^{-7} or smaller, providing a standard cutoff for very small probabilities as was implemented in previous methods [15, 16].

Since our observable of interest is the total output photon number Eq.(9), this method requires a photon counting experiment to accurately measure the number of counts arriving at each detector, such that p^j(cj(max)+1)=0\left\langle\hat{p}_{j}(c_{j}^{(\text{max})}+1)\right\rangle=0. Failure to satisfy this requirement would mean output photons in the summations Eq.(9) would be larger than detected counts, leading to inaccurate comparisons. In some regards this is a benefit, as it allows one to determine whether the experimental counts are actually measuring all possible output photons, as some exact algorithms exploit these unmeasured photons to decrease GBS simulation time [20].

III.3 Squeezed states in phase-space

Initial coherent amplitudes are generated by randomly sampling the phase-space distribution of a specific state. For most states, analytical forms of this distribution exist, allowing one to derive an analytical expression for the initial samples of each state. Here we simply recall previously derived results for squeezed states [15], so that the theory presented is complete, since the form of the positive-P distribution doesn’t depend on the type of measurement, that is, PNR or not.

For applications to GBS, we first consider the ideal case with pure squeezed state inputs. In order to define the input density operator Eq.(1) in the positive P-representation we first expand each state as an integral over real coherent states [42]

|rj=Cjexp(αj2γj)|αjdαj,\left|r_{j}\right\rangle=\sqrt{C_{j}}\int\exp\left(\frac{-\alpha_{j}^{2}}{\gamma_{j}}\right)\left|\alpha_{j}\right\rangle\text{d}\alpha_{j}, (20)

with normalization constant Cj=γj+1/(πγj)C_{j}=\sqrt{\gamma_{j}+1}/(\pi\gamma_{j}) and γj=exp(2rj)1\gamma_{j}=\exp(2r_{j})-1. After some straightforward calculations, the input density operator becomes

ρ^(in)=ReP(𝜶,𝜷)|𝜶𝜷|𝜷|𝜶d𝜶d𝜷,\hat{\rho}^{(\text{in})}=\text{Re}\iint P(\boldsymbol{\alpha},\boldsymbol{\beta})\frac{\left|\boldsymbol{\alpha}\right\rangle\left\langle\boldsymbol{\beta}\right|}{\left\langle\boldsymbol{\beta}|\boldsymbol{\alpha}\right\rangle}\text{d}\boldsymbol{\alpha}\text{d}\boldsymbol{\beta}, (21)

where the positive-P distribution for pure squeezed states is [15]

P(𝜶,𝜷)=jCje(αj2+βj2)(γ1+1/2)+αjβj.P(\boldsymbol{\alpha},\boldsymbol{\beta})=\prod_{j}C_{j}e^{-\left(\alpha_{j}^{2}+\beta_{j}^{2}\right)\left(\gamma^{-1}+1/2\right)+\alpha_{j}\beta_{j}}. (22)

For small ϵ\epsilon, thermalized squeezed states remain nonclassical and produce a modified output distribution that more closely resembles recent experimental distributions using threshold detectors [15, 16]. Due to this non-classicality, it has been shown [16] that even if an experiments claim of quantum advantage against the ideal distribution becomes invalidated by classical algorithms that replicate photo-detection measurements [18, 19, 16], it may still be able to claim quantum advantage against this non-ideal thermalized distribution.

From this model of decoherence, one can introduce a general random sampling formalism to generate initial stochastic samples for any Gaussian input [15]

αj\displaystyle\alpha_{j} =12(σxjwj+iσyjwj+M)\displaystyle=\frac{1}{2}\left(\sigma_{x_{j}}w_{j}+i\sigma_{y_{j}}w_{j+M}\right)
βj\displaystyle\beta_{j}^{*} =12(σxjwjiσyjwj+M),\displaystyle=\frac{1}{2}\left(\sigma_{x_{j}}w_{j}-i\sigma_{y_{j}}w_{j+M}\right), (23)

where wjwkP=δjk\left\langle w_{j}w_{k}\right\rangle_{P}=\delta_{jk} are real Gaussian noises and the quadrature variances are defined in Eq.(3). For classical inputs, σyj\sigma_{y_{j}} is real and αj=βj\alpha_{j}=\beta_{j}. For quantum inputs, σyj\sigma_{y_{j}} is imaginary and αjβj\alpha_{j}\neq\beta_{j}. Following from Eq.(4), the output stochastic amplitudes used to perform simulations are then obtained via the transformations 𝜶=𝑻𝜶\boldsymbol{\alpha}^{\prime}=\boldsymbol{T}\boldsymbol{\alpha} and 𝜷=𝑻𝜷\boldsymbol{\beta}^{\prime}=\boldsymbol{T}\boldsymbol{\beta} where the amplitudes 𝜶\boldsymbol{\alpha}^{\prime}, 𝜷\boldsymbol{\beta}^{\prime} correspond to measurements of the density matrix ρ^(out)\hat{\rho}^{\text{(out)}}.

IV Exact models of GBS grouped count distributions

The photon statistics of pure squeezed states have been studied extensively in quantum optics [34, 35, 7, 31]. Assuming the input modes have equal squeezing parameters r=r1==rNr=r_{1}=\dots=r_{N} and hence, equal photon numbers n=n1==nNn=n_{1}=\dots=n_{N}, the one-dimensional GCP, or total count probability, can be computed exactly. For the case of pure squeezed states transformed by a matrix with uniform amplitude loss coefficient tt, applied as t𝑼t\boldsymbol{U}, the exact total count distribution takes the form [31]

𝒢S(M)(2m)\displaystyle\mathcal{G}_{S}^{(M)}(2m) =t4m𝒞pM/2(1p)mf1/2\displaystyle=t^{4m}\mathcal{C}p^{M/2}(1-p)^{m}f_{1/2}
𝒢S(M)(2m1)\displaystyle\mathcal{G}_{S}^{(M)}(2m-1) =2m(1t2)t4m2𝒞pM/2(1p)mf3/2,\displaystyle=2m(1-t^{2})t^{4m-2}\mathcal{C}p^{M/2}(1-p)^{m}f_{3/2}, (24)

which is defined for m>0m>0, where S={1,,M}S=\{1,\dots,M\}, p=1/(1+n)p=1/(1+n) is the success probability, 𝒞=(M2+m1m)\mathcal{C}=\binom{\frac{M}{2}+m-1}{m} and fc=F12(a,b;c;z)f_{c}={}_{2}F_{1}\left(a,b;c;z\right) is the Gauss hypergeometric function with a=m+12a=m+\frac{1}{2}, b=M2+mb=\frac{M}{2}+m, z=(1t2)2(1p)z=(1-t^{2})^{2}(1-p).

For a lossless network with t=1t=1, F12(a,b;c;0)=1{}_{2}F_{1}\left(a,b;c;0\right)=1 and Eq.(24) converges to the well known distribution [34, 35, 7]

𝒢S(M)(2m)\displaystyle\mathcal{G}_{S}^{(M)}(2m) =𝒞pM/2(1p)m\displaystyle=\mathcal{C}p^{M/2}(1-p)^{m}
𝒢S(M)(2m1)\displaystyle\mathcal{G}_{S}^{(M)}(2m-1) =0,\displaystyle=0, (25)

which contains distinct oscillations between even and odd counts. In the limit MM\rightarrow\infty, it reduces to a Poissonian for the pair counts [35]

𝒢S(M)(2m)\displaystyle\mathcal{G}_{S}^{(M)}(2m) =1m!eMn/2(Mn2)m\displaystyle=\frac{1}{m!}e^{-Mn/2}\left(\frac{Mn}{2}\right)^{m}
𝒢S(M)(2m1)\displaystyle\mathcal{G}_{S}^{(M)}(2m-1) =0,\displaystyle=0, (26)

while the full distribution is not Poissonian.

Such photon statistics are characteristic of photon bunching which, unlike photon anti-bunching, is not an exclusively nonclassical phenomena [25]. However as squeezed photons are always generated in highly correlated pairs, bunching is indicative of the non-classicality of the measured photons. The presence of even-odd oscillations [36], even for networks with decoherence and loss, provides an explicit test of non-classicality although their absence does not exclude non-classicality.

In this section, we focus on comparing positive-P moments to exact distributions of GBS networks whose size, mean squeezing parameters and loss rates correspond to those found in the PNR experiments of Madsen et al [5]. However, we note that when even-odd count oscillations are present in lossless and ultra-low loss photonic networks, the positive-P moments suffer from slow convergence and large sampling errors.

Since the even-odd count oscillations vanish as soon as one has a high chance of losing a photon, the lossless and ultra-low loss regimes are not relevant to any current or near future experimental network, which are both far from lossless and whose photon number distribution is approximately Gaussian for large system sizes with loss rates as small as t=0.95t=0.95. Therefore, a detailed theoretical analysis of the positive P-representation in the lossless and ultra-low loss regimes will not be performed here, as projected methods [21] may converge faster in such regimes.

IV.1 Comparisons with small and large sized networks

To test whether Eq.(18) can be utilized to simulate nonclassical photon counting experiments using PNR detectors with perfect squeezers, photodetectors and photonic networks with uniform photon loss, we first simulate small sized GBS networks M=N=16M=N=16 and M=N=72M=N=72 with uniform squeezing 𝒓=[0.89,,0.89]\boldsymbol{r}=[0.89,\dots,0.89] and amplitude loss rates t=0.6t=0.6 and t=0.56t=0.56, respectively.

Comparisons between positive-P moments and the exact distribution Eq.(24) are presented in Fig.(1a) and Fig.(1b) for these networks where simulations are performed for an ensemble size of ES=2.4×106E_{S}=2.4\times 10^{6}. In the limit of large ESE_{S} the positive-P moments converge to moments of a distribution (see Eq.(16)). The exact size of ESE_{S} required for a given accuracy varies. An ensemble size of the order ES106E_{S}\geq 10^{6} is generally needed for relative errors less than 10310^{-3}, typical of current experimental accuracy.

For both network sizes and loss rates, the positive-P moments converge to the required exact distribution with errors <103<10^{-3}. Increasing the network sizes to M=N=216M=N=216 and M=N=288M=N=288, positive-P moments simulated for loss rates t=0.57t=0.57 and t=0.6t=0.6 are presented in Fig.(1c) and Fig.(1d). In both cases, moments converge to the exact distribution with errors <103<10^{-3} such that they are not visible in the graphics. Therefore, in the loss regimes of the current experiments, the positive P-representation provides an accurate method to validate experimental networks. For larger sized networks within the ultra-low loss regime this is also true, as the even-odd count oscillations vanish for much smaller amplitude loss rates.

Refer to caption
Figure 1: Exact multi-mode squeezed state photon counting distributions (dashed black lines) for various uniform amplitude loss rates tt are compared to positive-P moments of the GCP Eq.(10) (solid blue lines) for various network sizes, squeezing parameters and loss rates corresponding to the various experimental networks implemented in Ref.[5]. a) A GBS set-up with M=N=16M=N=16 is considered where pure squeezed states with uniform squeezing parameter 𝒓=[0.89,,0.89]\boldsymbol{r}=[0.89,\dots,0.89] are transformed by a Haar random unitary matrix multiplied by a uniform amplitude loss coefficient t𝑼t\boldsymbol{U} with t=0.6t=0.6. Network sizes are now increased to b) M=N=72M=N=72, with 𝒓=[0.89,,0.89]\boldsymbol{r}=[0.89,\dots,0.89] and t=0.56t=0.56, c) M=N=216M=N=216 with 𝒓=[1.1,,1.1]\boldsymbol{r}=[1.1,\dots,1.1] and t=0.57t=0.57, and d) M=N=288M=N=288 mode network with 𝒓=[1,,1]\boldsymbol{r}=[1,\dots,1] and t=0.6t=0.6. In all comparisons, the even-odd grouped count oscillations characteristic of pure squeezed photon statistics are not present as losses are too large. Positive-P moments are obtained by averaging over sample ensembles of size ES=2.4×106E_{S}=2.4\times 10^{6} where upper and lower lines correspond to ±1σT,j\pm 1\sigma_{T,j}, where σT,j1/ES\sigma_{T,j}\propto 1/\sqrt{E_{S}} is the theoretical sampling errors for the jj-th photon count bin.

V Numerical simulations of experimental networks

V.1 Statistical tests

A number of statistical tests have been proposed to validate experimental GBS networks. These include total variation distance [18, 5], cross-entropy measures [18, 19, 33], two and higher-order correlation functions [43, 4], and Bayesian hypothesis testing [4, 39]. Although these tests provide useful insights into the behavior of specific correlations, they are limited to either small size systems or patterns with small numbers of detected photons, as they require computing the probabilities of specific count patterns, which is a #P-hard task.

If these patterns are closer to the ground truth distribution than classical algorithms replicating photo-detection measurements, then its assumed patterns from larger sized systems claiming quantum advantage will also pass these tests. The ground truth distribution is the target distribution GBS aims to sample from when losses are included in the transmission matrix. Generally its assumed to be close to the ideal lossless distribution, although this may not necessarily be the case.

In this work, the primary statistical test used to validate experimental data are chi-square tests [44], which in terms of GCPs are defined as [15]

χ2=i=1k(𝒢E,i𝒢¯S,i)2σi2.\chi^{2}=\sum_{i=1}^{k}\frac{\left(\mathcal{G}_{E,i}-\bar{\mathcal{G}}_{S,i}\right)^{2}}{\sigma_{i}^{2}}. (27)

Here we use the shorthand notation 𝒢¯i=𝒢𝑺(n)(mji)P=ES1e=1ES(𝒢𝑺(n)(mji))(e)\mathcal{\bar{G}}_{i}=\left\langle\mathcal{G}_{\boldsymbol{S}}^{(n)}(m_{j_{i}})\right\rangle_{P}=E_{S}^{-1}\sum_{e=1}^{E_{S}}\left(\mathcal{G}_{\boldsymbol{S}}^{(n)}(m_{j_{i}})\right)_{(e)} for the stochastic trajectory eESe\in E_{S} to denote the positive-P GCP ensemble means of the i=1,2,,ki=1,2,\dots,k photon count bin for the j=1,,dj=1,\dots,d grouped count, which coverages to the true theoretical GCP in the limit 𝒢S,i=limES𝒢¯S,i\mathcal{G}_{S,i}=\lim_{E_{S}\rightarrow\infty}\bar{\mathcal{G}}_{S,i} for any general input state SS. Simulated ensemble means are then compared to the experimental GCPs 𝒢E,i=NE1ici\mathcal{G}_{E,i}=N_{E}^{-1}\sum_{i}c_{i} formed from NE𝒮PN_{E}\subset\mathcal{S}_{P} photon count patterns.

Chi-square tests are used for two reasons: the first is that its a standard test applied to data from random number generators [14], where the summation is performed over kk count bins satisfying iSjci>10\sum_{i\in S_{j}}c_{i}>10. The second is that it contains an explicit dependence on both theoretical and experimental sampling errors as σi2=σE,i2+σT,i2\sigma_{i}^{2}=\sigma_{E,i}^{2}+\sigma_{T,i}^{2}, where σE,i1/NE\sigma_{E,i}\propto 1/\sqrt{N_{E}} is the estimated experimental sampling error. This provides a method of accurately determining whether observed differences are caused by large sampling error, as is the case for sparse data.

Output χ2\chi^{2} values follow a chi-squared distribution that is positively skewed for small kk. Performing a Wilson-Hilferty transformation [45], which transforms the chi-square result as χ2(χ2/k)1/3\chi^{2}\rightarrow\left(\chi^{2}/k\right)^{1/3}, the chi-square distribution converges to a Gaussian distribution (χ2/k)1/3𝒩(μ𝒩,σ𝒩2)\left(\chi^{2}/k\right)^{1/3}\rightarrow\mathcal{N}\left(\mu_{\mathcal{N}},\sigma_{\mathcal{N}}^{2}\right) with mean μ𝒩=(9k2)/9k\mu_{\mathcal{N}}=(9k-2)/9k and variance σ𝒩2=2/(9k)\sigma_{\mathcal{N}}^{2}=2/(9k) for k10k\geq 10 due to the central limit theorem [45, 46]. This allows one to convert chi-square results into measurable probability distances using approximate Z-statistic, or Z-score, tests [47]

Z=(χ2/k)1/3μ𝒩σ𝒩,Z=\frac{\left(\chi^{2}/k\right)^{1/3}-\mu_{\mathcal{N}}}{\sigma_{\mathcal{N}}}, (28)

which measures how many standard deviations away a test statistic is from its expected normally distributed mean.

The results presented in subsequent subsections use the notation introduced in [16]. Comparisons between theory and experiment for simulations with pure squeezed states input into the experimental M×MM\times M 𝑻\boldsymbol{T}-matrix are denoted by the subscript EIEI, which we call the ideal ground truth distribution. Our ideal ground truth corresponds to the ground truth distributions used in Ref. [5]. However they differ from the non-ideal, or modified, ground truth distribution that includes realistic experimental effects such as thermalization, and measurement corrections in the theoretical simulations, which are denoted by the subscript ETET.

In all cases, agreement within sampling errors produces test results of χ2/k1\chi^{2}/k\approx 1 and |Z|1\left|Z\right|\leq 1 when accounting for stochastic fluctuations in the output statistics. For Z-statistic tests we use a regime of validity with the limit Z6Z\leq 6 to define accurate comparisons. Above this limit test statistics are more than 6σ𝒩6\sigma_{\mathcal{N}} from their predicted means, corresponding to very small observable probabilities.

V.2 Grouped count probability comparisons

Due to claims of quantum advantage, in this section we focus predominantly on validating count patterns from the 216216-mode high squeezing (HS) parameter and 288288-mode experimental data sets. Summary tables of chi-square and Z-statistic test results for all Borealis data sets can be found in Appendix A.

First, we compare the theoretical one-dimensional GCPs of the ideal ground truth distribution to the experimental total counts. Similar comparisons were performed in [5], however it was found that at least for the 288288-mode data set the ideal ground truth total count distribution was incorrect [48], leading to inaccurate comparisons. Results of phase-space simulations are presented in Fig.(2a) for the 216216-mode HS data set and Fig.(3a) for the 288288-mode data set. In both cases, the simulated GCP ideal ground truth distributions appear to closely resemble the experimental data. However, upon closer inspection there are significant differences between the two distributions.

This is first observed numerically from statistical testing, where the 216216-mode HS experiment produces a chi-square result of χEI2/k54\chi_{EI}^{2}/k\approx 54 with ZEI70Z_{EI}\approx 70 and count patterns from the 288288-mode experiment output χEI2/k411\chi_{EI}^{2}/k\approx 411 with ZEI167Z_{EI}\approx 167. Although both experimental GCPs are far from their pure squeezed state distributions, the 288288-mode distribution is 98σ𝒩98\sigma_{\mathcal{N}} further from its expected normally distributed mean. These numerical differences are also reflected graphically by plotting the normalized difference between distributions Δ𝒢S(M)(m)/σi=(𝒢E,i𝒢¯I,i)/σi\Delta\mathcal{G}_{S}^{(M)}(m)/\sigma_{i}=\left(\mathcal{G}_{E,i}-\bar{\mathcal{G}}_{I,i}\right)/\sigma_{i}, which is plotted in Fig.(2b) for the 216216-mode HS data set and Fig.(3d) for the 288288-mode data set.

Refer to caption
Figure 2: Comparisons of one-dimensional GCPs 𝒢{1,,M}(M)(m)\mathcal{G}_{\{1,\dots,M\}}^{(M)}(m) versus grouped counts mm for experimental data obtained from the 216216-mode HS Borealis experiment. a) Data is compared to positive-P simulations with pure squeezed state inputs, where simulations are performed for an ensemble size of ES=1.2×106E_{S}=1.2\times 10^{6}, and the corresponding normalized difference between distributions is plotted in b). In c) experimental patterns are now compared to the simulated GCP using thermalized squeezed state inputs with ϵ=0.0510±0.0005\epsilon=0.0510\pm 0.0005. A transmission matrix measurement correction is also applied as t𝑻t\boldsymbol{T}, where t=0.9941±0.0005t=0.9941\pm 0.0005, to improve the fit. d) The normalized difference between thermalized and experimental GCPs, which display noticeable improvement when compared to pure squeezed state simulations.

Once a thermalization component ϵ\epsilon is added to the input states and a transmission matrix measurement correction tt is applied as t𝑻t\boldsymbol{T}, the positive-P simulated GCPs agree within sampling error to the experimental distributions. The 216216-mode HS data set requires ϵ=0.0510±0.0005\epsilon=0.0510\pm 0.0005 and t=0.9941±0.0005t=0.9941\pm 0.0005 to output χET2/k0.88±0.02\chi_{ET}^{2}/k\approx 0.88\pm 0.02 and |ZET|1±0.2\left|Z_{ET}\right|\approx 1\pm 0.2, while the best-fit theoretical GCP for 288288-mode data uses a similar thermalization component ϵ=0.0547±0.0005\epsilon=0.0547\pm 0.0005 and measurement correction t=0.9848±0.0005t=0.9848\pm 0.0005 to produce χET2/k0.87±0.02\chi_{ET}^{2}/k\approx 0.87\pm 0.02 and ZET|1.1|±0.2Z_{ET}\approx\left|1.1\right|\pm 0.2. This exact agreement between theory and experiment is reflected in the normalized difference plots as seen in Fig.(2d) and Fig.(3d) which have noticeably decreased from their pure squeezed state comparisons.

Similar results are obtained for the 7272-mode and 216216-mode low squeezing (LS) experiments. In these cases, comparisons with the ideal ground truth, which output ZEI=72Z_{EI}=72 and ZEI=66Z_{EI}=66 respectively, converge to a within sampling error agreement with a thermalized theory such that |ZET|0.2±0.2\left|Z_{ET}\right|\approx 0.2\pm 0.2 for the 7272-mode experiment and |ZET|0.89±0.1\left|Z_{ET}\right|\approx 0.89\pm 0.1 for the 216216-mode LS experiment.

Despite this excellent agreement between theoretical and experimental GCPs once decoherence and measurement corrections are accounted for, we note that stronger validation tests of quantum advantage based on multidimensional GCPs cannot be performed on these data sets. As the dimension increases, the number of bins generated scales as (Md1+1)d\left(Md^{-1}+1\right)^{d}. If there are too few counts per bin, i.e. the bins become more sparse, the experimental sampling errors of these bins increase dramatically, causing statistical tests to become dominated by sampling errors, outputting artificially small results.

Earlier works [16] on recent threshold detector experiments generating NE45×107N_{E}\approx 4\rightarrow 5\times 10^{7} count patterns showed these sampling error limitations arose for a GCP dimension of d=4d=4, such that reliable validation tests could only be performed for d<4d<4. However, the majority of data sets from Borealis generate NE1×106N_{E}\approx 1\times 10^{6} patterns, causing sampling errors to become dominant at only d=2d=2, limiting comparisons to only d=1d=1.

The exception to this is the 1616-mode experiment whose NE8.4×107N_{E}\approx 8.4\times 10^{7} samples were used benchmark the system architecture through total variation distance tests as the system is small enough that output photo-detection measurements can be replicated exactly. Interestingly, the one-dimensional GCP for this data set is further from its ideal ground truth distribution than most experiments performed using the Borealis quantum network, where χEI2/k2.5×103\chi_{EI}^{2}/k\approx 2.5\times 10^{3} and ZEI=163Z_{EI}=163. Unlike data sets claiming quantum advantage, including decoherence and 𝑻\boldsymbol{T}-matrix corrections is no longer enough to obtain a within sampling error agreement between theoretical and experimental GCPs, instead producing statistical test results of χET2/k14±2\chi_{ET}^{2}/k\approx 14\pm 2 and ZET18±1Z_{ET}\approx 18\pm 1.

Refer to caption
Figure 3: Description follows Fig.(2) except experimental count patterns are from the 288288-mode Borealis network. Input thermalized squeezed state simulations are performed using a decoherence component of ϵ=0.0547±0.0005\epsilon=0.0547\pm 0.0005 and matrix measurement error t=0.9848±0.0005t=0.9848\pm 0.0005.

Due to the large number of samples, statistical tests of multi-dimensional GCPs can be performed on this data set. A small increase in the GCP dimension to d=2d=2 shows the 1616-mode data is even further from both the ideal and non-ideal ground truth distributions once high-order correlations become more statistically prevalent, outputting ZEI=301Z_{EI}=301 and ZET=86Z_{ET}=86. An exponential number of comparison tests can be performed by randomly permuting the measured photon count patters [16], where each permutation allows one to compare different binned correlation moments. As some correlation moments may be closer to their theoretical values than others, Z-scores are likely to vary between permutations, as was observed in threshold detector experiment comparisons [16]. Focusing on comparisons with the non-ideal ground truth, a mean increase in the resulting Z-scores is observed from ten permutations, such that ZET𝒫139\left\langle Z_{ET}\right\rangle_{\mathcal{RP}}\approx 139, where 𝒫\left\langle\dots\right\rangle_{\mathcal{RP}} denotes a mean over random permutations. Although some permutations are closer to the non-ideal ground truth, with the closest being ZET73Z_{ET}\approx 73, the majority are further, where the largest observed Z-score was ZET235Z_{ET}\approx 235 from a single permutation.

Although this is not unexpected, such results indicate further systematic errors besides those tested are present. We conjecture that the presence of further systematic errors other than decoherence and measurement errors are also present in the 216216-mode HS and 288288-mode implementations. However the small sample size means such errors cannot be resolved in statistical testing for these data sets, limiting a thorough analysis.

V.3 Photon number moments

Finally, for completeness, we simulate the mean output photon number moments n^j\left\langle\hat{n}^{\prime}_{j}\right\rangle for all jj modes, where summary tables of statistical test results for all data sets is also presented in Appendix A. As with the GCP comparisons, count patterns from the 216216-mode HS and 288288-mode experiments are far from their theoretical first-order moments for pure squeezed state inputs with ZEI86Z_{EI}\approx 86 and ZEI192Z_{EI}\approx 192 for the 216216-mode HS and 288288-mode data sets respectively.

However, unlike the GCP comparisons, decoherence and transmission matrix corrections no longer cause the theoretical moments to be within error of the experimental moments, where ZET45Z_{ET}\approx 45 for the 216216-mode HS data set and the 288288-mode data set is ZETσ𝒩63σ𝒩Z_{ET}\sigma_{\mathcal{N}}\approx 63\sigma_{\mathcal{N}} standard deviations away from its expected mean. Similar trends are present for all data sets where, as was the case for GCPs, the 1616-mode moments are furthest from their expected first order moments than any other experiment. We emphasize that statistical testing is more accurate for this data set due to the large number of experimental samples, allowing one to resolve finer differences between theoretical and experimental photon number moments.

One noticeable feature in the experimental and theoretical moments is the decrease in n^j\left\langle\hat{n}^{\prime}_{j}\right\rangle as jMj\rightarrow M, with sharp decreases occurring for later modes as seen in Fig.(4) for 175<j<216175<j<216 and Fig.(5) for 250<j<288250<j<288, while distinct sharp spikes in the moments occur every 1616 modes. This clear structure is present in all data sets from this experiment but is not present in previous implementations with threshold detectors [3, 4] and is likely due to the time-domain multiplexing scheme used to form the linear network in Ref. [5]. The experimental set-up interferes a train of input squeezed photons on multiple fibre delay lines separated by temporal bins of various dimension. Output pulses are then sent into a 1-to-16 mode demultiplexer before measurements are performed by PNR detectors.

Refer to caption
Figure 4: a) Input thermalized squeezed state theoretical mean output photon numbers n^j\left\langle\hat{n}^{\prime}_{j}\right\rangle versus mode number jj are compared to their experimental counterparts for data from the 216216-mode HS Borealis experiment, where simulations are performed for an ensemble size of ES=1.2×106E_{S}=1.2\times 10^{6}. Decoherence and 𝑻\boldsymbol{T}-matrix parameters are ϵ=0.0510±0.0005\epsilon=0.0510\pm 0.0005 and t=0.9941±0.0005t=0.9941\pm 0.0005 which, although improving the fit with experimental data, no longer cause theory and experiment to agree within sampling error. b) Corresponding normalized difference between marginal moments.

Considering its input-output ratio the demultiplexer may be the cause of the periodic spikes. However the structure in the form of the overall decrease in n^j\left\langle\hat{n}^{\prime}_{j}\right\rangle for increasing mode number may be due to the use of fibre delay lines. Although providing a more compact architecture, it comes at the expense of reduced connectivity, i.e. photon interference between all modes isn’t possible. This has the effect of introducing a structure to the output photons that can be exploited by classical samplers such as the tree-width sampler [49].

Refer to caption
Figure 5: Figure description follows the same as Fig.(4), except data is now from the 288288-mode Borealis experiment where corresponding fitting parameters are ϵ=0.0547±0.0005\epsilon=0.0547\pm 0.0005 and t=0.9848±0.0005t=0.9848\pm 0.0005.

VI Conclusion

In this paper, we extended a binning method developed by Mandel [9, 10] to simulate nonclassical photon statistics of photon-number resolving detectors using the positive P-representation. We performed validation tests on quantum computers claiming quantum advantage with photon-number resolving detectors.

This method is computationally efficient, with computational run-times of a few minutes on a 12-core desktop at 1011\sim 10^{11} floating point operations/ sec (Linpack Flops) for one-dimensional GCP simulations of the largest sized networks implemented in the paper. Since the positive P-representation accurately computes correlation moments to any order generated in the photonic network, it is a time and resource efficient method of validation compared to classical algorithms that compute the Hafnian to replicate photo-detection measurements, or tensor network approximations [31]. Increasing the GCP dimension causes an increase in computation time, although this is still significantly faster than other validation methods. Directly calculating 100×100100\times 100 Hafnians are estimated [31] to take 1414 hours on Fugaku, one of the worlds largest supercomputers in 2024, at 1018\sim 10^{18} Flops. This would compute one probability, but rounding errors are unknown, as the calculation has not occurred. Generating enough classical counts for validation using this approach would take 1016s\sim 10^{16}s [31] on Fugaku, indicating an overall speed-up of >1018>10^{18} using the +P method at these network sizes.

We emphasize that our method does not generate classical counts, so it is strictly for validation of the computational task, and does not replicate the required random counts. Validation tests were performed on GBS data from a recent experiment claiming quantum advantage [5]. We showed that within sampling error agreement between theory and experiment for GCPs is obtained once decoherence is included in the input states, and a transmission matrix measurement error correction is applied. There are still large discrepancies in individual channel count probabilities, possibly because these can be calculated very accurately. This result agrees with a trend obtained previously for threshold detector experiments [15, 16].

These comparisons with threshold detector experiments are slightly misleading. Due to the smaller number of experimental samples from data sets claiming quantum advantage using PNR methods, only a small fraction of the possible tests could be performed. This was because experimental sampling errors were too large to apply more sensitive tests. Based on comparisons of the 1616-mode experimental data, whose sample size was large enough to perform accurate statistical testing, we conjecture that more experimental errors are present in the quantum advantage data sets than those tested for in this paper.

In the longer term, due to its computational efficiency and ability to generate any measurable moment, the positive-P based binning method could be used as a means to provide error correction, either by correcting parameter values as shown in this paper, or by giving feedback so they can be adjusted. We suggest that this could provide a more reasonable route to quantum computational advantage than simply using improved hardware, as is also thought to be the case for logic gate based quantum computers.

Finally, even though we focused here on GBS networks with pure or thermalized squeezed states, both the binning method and positive P-representation can be used to simulate photon statistics of any nonclassical state.

Acknowledgments

ASD would like to thank Nicholas Quesada, Fabian Laudenbach and Jonathan Lavoie for helpful discussions. This research was funded through grants from NTT Phi Laboratories and a Templeton Foundation grant ID 62843.

Appendix A: Summary of all experimental comparisons

The Borealis GBS implemented in [5] performed experiments for varying system sizes of M=16,72,216,and 288M=16,72,216,\>\text{and}\>288. The 216216-mode experiment was repeated for the mean squeezing parameters r¯=1.103\bar{r}=1.103 and r¯=0.533\bar{r}=0.533, hence they are distinguished as high squeezing (HS) and low squeezing (LS) experiments, respectively. The ideal ground truth distributions for each data set are formulated from their corresponding lossy transmission matrices 𝑻\boldsymbol{T} and simulated pure squeezed states using squeezing vectors 𝒓\boldsymbol{r}.

The graphs presented above have results of statistical tests for comparisons of theory and experiment predominantly for the 216216-mode HS and 288288-mode data sets, due to claims of quantum advantage. Here we summarize statistical test results for all data sets in [5], starting with one-dimensional GCPs for the ideal ground truth in Table.(1). We use the subscripts EIEI and ETET to denote comparisons between experimental data and the ideal and non-ideal ground truth distributions, respectively.

A summary of the non-ideal ground truth comparisons are presented in Table.(2), where the tt and ϵ\epsilon fitting parameters are found using a Nelder-Mead simplex algorithm which minimizes the distance between experimental and theoretical distributions. As stated earlier, experimental data agrees with their theoretical predictions within sampling error for all data sets bar the 1616-mode experiment. Meanwhile comparisons of the mean output photon number moments are presented in Table.(3) for both the ideal and non-ideal ground truth distributions.

Mode number χEI2/k\chi_{EI}^{2}/k kk ZEIZ_{EI}
1616 2.53×1032.53\times 10^{3} 3737 163163
7272 171171 5656 7272
216216-(LS) 165165 4848 6666
216216-(HS) 5454 140140 7070
288288 411411 149149 167167
Table 1: Summary of statistical tests for comparisons of one-dimensional GCPs 𝒢{1,,M}(M)(m)\mathcal{G}_{\{1,\dots,M\}}^{(M)}(m) for all experimental data sets obtained from [5]. Chi-square and Z-statistic tests are generated from positive-P simulated GCPs with an ensemble size of ES=1.2×106E_{S}=1.2\times 10^{6}. Although the 1616-mode and 288288-mode data are furthest from the ideal ground truth distribution, all experimental count patterns are at least 66σ𝒩66\sigma_{\mathcal{N}} from their expected normally distributed means.
Mode number tt ϵ\epsilon χET2/k\chi_{ET}^{2}/k kk |ZET|\left|Z_{ET}\right|
1616 0.98410.9841 0.06480.0648 14±214\pm 2 3636 18±118\pm 1
7272 0.98610.9861 0.05070.0507 1.02±0.031.02\pm 0.03 5454 0.2±0.20.2\pm 0.2
216216-(LS) 0.98820.9882 0.01050.0105 0.82±0.020.82\pm 0.02 4747 0.89±0.10.89\pm 0.1
216216-(HS) 0.99410.9941 0.05100.0510 0.88±0.020.88\pm 0.02 137137 1±0.21\pm 0.2
288288 0.98480.9848 0.05470.0547 0.87±0.020.87\pm 0.02 142142 1.1±0.21.1\pm 0.2
Table 2: Description follows those presented in Table.(1) except now experimental data is compared to a modified non-ideal ground truth distribution for the one-dimensional GCPs. Thermalisation component ϵ\epsilon and transmission matrix measurement correction tt are used to modify the ground truth distribution to best fit the experimental data, where ϵ,t\epsilon,t each have error bars of ±0.0005\pm 0.0005 for all data sets. The χET2/k\chi_{ET}^{2}/k and ZETZ_{ET} values are averages over ten simulation runs with corresponding standard deviations.
Mode number ZEIZ_{EI} ZETZ_{ET}
1616 264264 243243
7272 7878 1717
216216-(LS) 8585 3737
216216-(HS) 8686 4545
288288 192192 6363
Table 3: Summary of Z-statistic test results for comparisons of both the ideal and non-ideal ground truth distributions for the mean output photon number n^j\left\langle\hat{n}^{\prime}_{j}\right\rangle of each experimental data set. Positive-P simulations are performed for an ensemble size of ES=1.2×106E_{S}=1.2\times 10^{6} for all data sets except the 1616-mode experiment, where ES=3.36×107E_{S}=3.36\times 10^{7} is needed in order for σT,jσE,j\sigma_{T,j}\apprle\sigma_{E,j} which is required for accurate simulations. Fitting parameters applied to each data set are obtained from Table.(2).

Appendix B: Estimating experimental sampling errors of moments

Accurately estimating sampling errors is essential when comparing theoretical simulations with experimental data obtained via sampling. This is particularly important when using statistical tests that are dependent on sampling errors, as incorrect estimates will result in misleading test results.

Theoretical sampling errors are estimated numerically by subdividing the total phase-space ensemble as ES=NSNRE_{S}=N_{S}N_{R}. Breaking the full ensemble of phase-space samples into sub-ensembles is a common numerical procedure for applications of phase-space representations yielding two complementary benefits: efficient multi-core parallel computing, decreasing computation time depending on available resources, and estimates of theoretical sampling errors [16, 41].

The first sub-ensemble corresponds to sampling from the Gaussian phase-space distribution NSN_{S} times and NRN_{R} is the number of times the computation is repeated. Provided NR1N_{R}\gg 1, the second sub-ensemble allows one to accurately estimate the standard deviation in the ensemble mean as σT,i=σt,i/NR\sigma_{T,i}=\sigma_{t,i}/\sqrt{N_{R}} [47, 50], where σt\sigma_{t} is the computed sample standard deviation which takes the general form [16, 41]

σt=1NR1i=1NR(O¯(i)O¯).\sigma_{t}=\sqrt{\frac{1}{N_{R}-1}\sum_{i=1}^{N_{R}}\left(\bar{O}^{(i)}-\bar{O}\right)}. (29)

Here, O(𝜶,𝜷)O(\boldsymbol{\alpha},\boldsymbol{\beta}) is some general phase-space observable, O¯(i)=NS1k=1NS(O(k))(i)\bar{O}^{(i)}=N_{S}^{-1}\sum_{k=1}^{N_{S}}(O^{(k)})^{(i)} denotes the first sub-ensemble mean for the iNRi\in N_{R} trajectory and O¯=ES1e=1ESO(e)\bar{O}=E_{S}^{-1}\sum_{e=1}^{E_{S}}O_{(e)} is the full ensemble mean. In statistics literature [47, 51], σT,i\sigma_{T,i} is commonly called the standard error of the mean and its valid for both binned distributions and moments.

For experimental sampling errors defined as σE,i=σe,i/NE\sigma_{E,i}=\sigma_{e,i}/\sqrt{N_{E}}, we are interested in estimating the standard deviation of the experimental sample means σe,i\sigma_{e,i}. Unlike theoretical sampling errors this isn’t performed numerically. Instead, the variances of GCPs for both threshold and PNR detector experiments have Poissonian fluctuations, such that σe,i=𝒢S,i\sigma_{e,i}=\sqrt{\mathcal{G}_{S,i}} for some general input state SS. Since the true theoretical GCP 𝒢S,i\mathcal{G}_{S,i} is seldom known, we estimate it in phase-space as 𝒢S,i=limES𝒢¯S,i\mathcal{G}_{S,i}=\lim_{E_{S}\rightarrow\infty}\bar{\mathcal{G}}_{S,i}.

However, estimating σe,i\sigma_{e,i} for moments (either raw or central) and cumulants is somewhat more complicated as these aren’t probabilities but instead tell one about various properties of a probability distribution. In this appendix, we describe the method of estimating experimental sampling errors of raw moments used in the comparisons presented in the main text. Details on how to estimate sampling errors for central moments and cumulants can be found in [51, 52, 53].

Sampling errors of raw moments

Let YY be an i.i.d random variable with probability density function P(y)P(y), then the nn-th raw moment is defined as

μn=Yn=ynP(y)dy.\mu^{\prime}_{n}=\left\langle Y^{n}\right\rangle=\int y^{n}P(y)\text{d}y. (30)

Suppose one generates 𝒮\mathcal{S} samples denoted y1,,y𝒮y_{1},\dots,y_{\mathcal{S}} from the distribution P(y)P(y). The nn-th sample raw moment is then defined as [51, 52, 53]

mn=1𝒮i=1𝒮yin,m^{\prime}_{n}=\frac{1}{\mathcal{S}}\sum_{i=1}^{\mathcal{S}}y_{i}^{n}, (31)

the expected value of which is an unbiased estimator of μn\mu^{\prime}_{n}

mn=1𝒮i=1𝒮yin=μn.\left\langle m^{\prime}_{n}\right\rangle=\frac{1}{\mathcal{S}}\sum_{i=1}^{\mathcal{S}}\left\langle y_{i}^{n}\right\rangle=\mu^{\prime}_{n}. (32)

For our purposes, we are interested in the variance σmn2=(mn)2mn2\sigma_{m^{\prime}_{n}}^{2}=\left\langle(m^{\prime}_{n})^{2}\right\rangle-\left\langle m^{\prime}_{n}\right\rangle^{2}. Following from Ref. [51], the expectation value of the squared moment is simplified as

(mn)2\displaystyle\left\langle(m^{\prime}_{n})^{2}\right\rangle =1𝒮2i=1𝒮j=1𝒮yinyjn\displaystyle=\frac{1}{\mathcal{S}^{2}}\left\langle\sum_{i=1}^{\mathcal{S}}\sum_{j=1}^{\mathcal{S}}y_{i}^{n}y_{j}^{n}\right\rangle
=1𝒮2l=1𝒮yl2n+ij𝒮yinyjn,\displaystyle=\frac{1}{\mathcal{S}^{2}}\left\langle\sum_{l=1}^{\mathcal{S}}y_{l}^{2n}+\sum_{i\neq j}^{\mathcal{S}}y_{i}^{n}y_{j}^{n}\right\rangle, (33)

where the summations have been expanded to include instances of equal l=i=jl=i=j and non-equal iji\neq j samples. Utilizing the simplifications i=1𝒮yin=𝒮μn\sum_{i=1}^{\mathcal{S}}\left\langle y_{i}^{n}\right\rangle=\mathcal{S}\mu^{\prime}_{n} from Eq.(32) and ij𝒮=i𝒮1j𝒮\sum_{i\neq j}^{\mathcal{S}}=\sum_{i}^{\mathcal{S}-1}\sum_{j}^{\mathcal{S}}, the second-order sample raw moment is simplified as

(mn)2=μ2n+(𝒮1)(μn)2𝒮,\left\langle(m^{\prime}_{n})^{2}\right\rangle=\frac{\mu^{\prime}_{2n}+(\mathcal{S}-1)(\mu^{\prime}_{n})^{2}}{\mathcal{S}}, (34)

allowing one to estimate the variance in the sample raw moment as [51]

σmn2=μ2n(μn)2𝒮.\sigma_{m^{\prime}_{n}}^{2}=\frac{\mu^{\prime}_{2n}-(\mu^{\prime}_{n})^{2}}{\mathcal{S}}. (35)

Its possible to extend the above description for univariate statistics to the multivariate case. Let 𝒀=[Y1,Y2,]\boldsymbol{Y}=[Y_{1},Y_{2},\dots] be a vector of i.i.d random variables with multivariate probability density P(𝒚)P(\boldsymbol{y}), where 𝒚=[y1,y2,]\boldsymbol{y}=[y_{1},y_{2},\dots]. The nn-th raw moment is then defined as

μn1,n2,=Y1n1Y2n2=(y1n1y2n2)P(𝒚)d𝒚.\mu^{\prime}_{n_{1},n_{2},\dots}=\left\langle Y_{1}^{n_{1}}Y_{2}^{n_{2}}\dots\right\rangle=\int\left(y_{1}^{n_{1}}y_{2}^{n_{2}}\dots\right)P(\boldsymbol{y})\text{d}\boldsymbol{y}. (36)

As in the univariate case, suppose one generates 𝒮\mathcal{S} samples of the j=1,2,j=1,2,\dots-th random variable in 𝒀\boldsymbol{Y}, denoted y1,j,,y𝒮,jy_{1,j},\dots,y_{\mathcal{S},j}, then the multivariate nn-th sample raw moment is defined as [52]

mn1,n2,=1𝒮i=1𝒮yi,1n1yi,2n2,m^{\prime}_{n_{1},n_{2},\dots}=\frac{1}{\mathcal{S}}\sum_{i=1}^{\mathcal{S}}y_{i,1}^{n_{1}}y_{i,2}^{n_{2}}\cdots, (37)

whose expected value is an unbiased estimate of the multivariate raw moment [53]

mn1,n2,=1𝒮i=1𝒮yi,1n1yi,2n2=μn1,n2,.\left\langle m^{\prime}_{n_{1},n_{2},\dots}\right\rangle=\frac{1}{\mathcal{S}}\sum_{i=1}^{\mathcal{S}}\left\langle y_{i,1}^{n_{1}}y_{i,2}^{n_{2}}\cdots\right\rangle=\mu^{\prime}_{n_{1},n_{2},\dots}. (38)

Following from the derivation of the univariate case, the variance of the multivariate raw sample moment is obtained as [54]

σmn1,n2,2=μ2n1,2n2,(μn1,n2,)2𝒮.\sigma_{m^{\prime}_{n_{1},n_{2},\dots}}^{2}=\frac{\mu^{\prime}_{2n_{1},2n_{2},\dots}-(\mu^{\prime}_{n_{1},n_{2},\dots})^{2}}{\mathcal{S}}. (39)

By setting n1=n2==1n_{1}=n_{2}=\dots=1, sampling errors for higher photon number moments such as n^jn^k\left\langle\hat{n}^{\prime}_{j}\hat{n}^{\prime}_{k}\right\rangle, n^jn^kn^l\left\langle\hat{n}^{\prime}_{j}\hat{n}^{\prime}_{k}\hat{n}^{\prime}_{l}\right\rangle can now be estimated.

Accuracy of sampling error estimates

In this paper, the raw moments of interest are the mean output photon numbers per mode n^j\left\langle\hat{n}^{\prime}_{j}\right\rangle. Using Eq.(35), the experimental sampling errors can therefore be estimated as

σE,j=σe,jNE=n^j2n^j2NE,\sigma_{E,j}=\frac{\sigma_{e,j}}{\sqrt{N_{E}}}=\sqrt{\frac{\left\langle\hat{n}_{j}^{\prime 2}\right\rangle-\left\langle\hat{n}^{\prime}_{j}\right\rangle^{2}}{N_{E}}}, (40)

where, as was the case for GCPs, we estimate the true theoretical photon number moments in phase-space as n^j=limESn¯j\left\langle\hat{n}^{\prime}_{j}\right\rangle=\lim_{E_{S}\rightarrow\infty}\bar{n}^{\prime}_{j}. To determine the accuracy of this estimate, we must generate photon count patterns corresponding to PNR detectors from an exactly known model and compare the estimated sampling error with the standard deviation of the mean output photon number of the count patterns from their exact values. To do this, the simplest method is to use classical states.

Classical state PNR photo-detection algorithm

In the case of NN classical states input into a GBS linear network, one has rotated to a classical phase-space and we can use the coherent amplitudes corresponding to the diagonal P-representation to replicate photo-detection measurements of PNR detectors.

Generally, each stochastic trajectory in ESE_{S} corresponds to a single experimental shot. Therefore one can conserve the classical correlations generated within the network by computing the phase-space observable projector of the jj-th detector pjp_{j}, which corresponds to the operator Eq.(5), for the kk-th trajectory as

(pj(cjf))(k)=1cjf!((nj)cjfenj)(k),(p_{j}(c_{j}^{f}))^{(k)}=\frac{1}{c_{j}^{f}!}\left(\left(n^{\prime}_{j}\right)^{c_{j}^{f}}e^{-n^{\prime}_{j}}\right)^{(k)}, (41)

where kESk\in E_{S}, nj=|αj|2n^{\prime}_{j}=\left|\alpha_{j}\right|^{2} in the diagonal P-representation and cjf=0,1,2,,cj(max),fc_{j}^{f}=0,1,2,\dots,c_{j}^{(\text{max}),f} is a photon count of our fake experiment up to some maximum cut-off cj(max),fc_{j}^{(\text{max}),f}.

The phase-space projectors determine the probability of observing a specific count cjfc_{j}^{f}. Since each detector can distinguish between multiple oncoming photons, the actual computation involves calculating a vector of probabilities for each stochastic trajectory:

𝑷j(k)=[P0,j(k),P1,j(k),,Pmax,j(k)],\boldsymbol{P}_{j}^{(k)}=[P_{0,j}^{(k)},P_{1,j}^{(k)},\dots,P_{\text{max},j}^{(k)}], (42)

where P0,j(k)=(pj(0))(k)P_{0,j}^{(k)}=(p_{j}(0))^{(k)} is the success probability of observing no counts at the jj-th detector, P1,j(k)=(pj(1))(k)P_{1,j}^{(k)}=(p_{j}(1))^{(k)} is the success probability of observing a single count at the jj-th detector, and so on. The PNR photo-detection measurements are then replicated by randomly sampling Eq.(42), which outputs the fake photon count vector

(𝒄f)(k)=[Pc1,1(k),Pc2,2(k),,PcM,M(k)].(\boldsymbol{c}^{f})^{(k)}=[P_{c_{1},1}^{(k)},P_{c_{2},2}^{(k)},\dots,P_{c_{M},M}^{(k)}].

As the method scales with the total phase-space ensemble size, the number of faked patterns generated is NF=ESN_{F}=E_{S}. These are binned in the same way as the experimental data, where mjf=iSjcifm_{j}^{f}=\sum_{i\in S_{j}}c_{i}^{f} is the fake grouped count corresponding to the GCP 𝒢jf=NF1iSjcif\mathcal{G}_{j}^{f}=N_{F}^{-1}\sum_{i\in S_{j}}c_{i}^{f}.

Tests of estimate accuracy

The photon counting distributions for a number of classical states can be exactly computed, one of which is multi-mode thermal states. The one-dimensional GCP of a multi-mode thermal state is obtained from the single-mode distribution [25]

𝒢S(M)(m)=p(1p)m,\mathcal{G}_{S}^{(M)}(m)=p(1-p)^{m}, (43)

where p=1/(1+n)p=1/(1+n) is the success probability requiring a uniform squeezing parameter and hence equal photon numbers n=n1==nNn=n_{1}=\dots=n_{N}. The single-mode theory follows a geometric distribution which is a special case of the negative binomial distribution: if XX is an independent and geometrically distributed random variable with parameter pp, then Y=jMXjY=\sum_{j}^{M}X_{j} follows a negative binomial distribution with parameters MM and pp.

The one-dimensional GCP theory of multi-mode thermal state inputs therefore follows a negative binomial distribution

𝒢S(M)(m)=(m+M1m)pM(1p)m.\mathcal{G}_{S}^{(M)}(m)=\left(\begin{array}[]{c}m+M-1\\ m\end{array}\right)p^{M}(1-p)^{m}. (44)

For the case of thermal photons transformed by a Haar random unitary, the input and output mean photon numbers are equal n=n^=sinh2(r)n=\left\langle\hat{n}^{\prime}\right\rangle=\sinh^{2}(r) and the photon number variance can be computed exactly as σn2=n(n+1)=sinh2(r)(sinh2(r)+1)\sigma_{n}^{2}=n\left(n+1\right)=\sinh^{2}(r)\left(\sinh^{2}(r)+1\right) [55].

Now that we have a method of generating photon counts and an exactly known model with which to compare those counts, we calculate the accuracy of the estimate Eq.(40) for M=N=50M=N=50 thermal states with uniform squeezing 𝒓=[0.5,,0.5]\boldsymbol{r}=[0.5,\dots,0.5] sent into a lossless linear network described by a Haar random unitary matrix 𝑼\boldsymbol{U}. We generate NF=4×106N_{F}=4\times 10^{6} count patterns using a detector cut-off of cj(max),f=13c_{j}^{(\text{max}),f}=13 and bin them to compute the mean output photon number n¯jf\bar{n}_{j}^{\prime f}, which is compared to the exact value in Fig.(6).

Following from Eq.(40), the sampling error of our fake experiment is estimated by setting σe=σn=n(n+1)\sigma_{e}=\sigma_{n}=\sqrt{n(n+1)}, such that

σE=n(n+1)NF.\sigma_{E}=\sqrt{\frac{n(n+1)}{N_{F}}}. (45)

If this is an accurate estimate, it should converge to the actual error between moments

σA=1Mj=1M(n¯jfn)2,\sigma_{A}=\sqrt{\frac{1}{M}\sum_{j=1}^{M}\left(\bar{n}_{j}^{\prime f}-n\right)^{2}}, (46)

which is simply the standard deviation as nn is our expected mean value.

For the data in Fig.(6), our estimated sampling error is σE2.938×104\sigma_{E}\approx 2.938\times 10^{-4} while the actual error is σA3.095×104\sigma_{A}\approx 3.095\times 10^{-4}, corresponding to an 95%\approx 95\% agreement. If instead we assumed one could estimate the errors as Poissonian fluctuations, such as for GCPs, then σE=n/NF2.605×104\sigma_{E}=\sqrt{n/N_{F}}\approx 2.605\times 10^{-4} which is only 84%\approx 84\% of the actual error.

Therefore, Eq.(40) is an accurate method of estimating experimental sampling errors of moments. Naturally, the larger the number of experimental samples the better the estimate will be as the experimental moments will converge to their ground truth values more readily.

Refer to caption
Figure 6: The exactly computable mean output photon number (solid black line) of N=50N=50 thermal states sent into a lossless linear network 𝑼\boldsymbol{U} is compared to the mean output photon number (solid orange dots) generated from a classical algorithm that replicates PNR photo-detection measurements of classical states using the diagonal P-representation. The NF=4×106N_{F}=4\times 10^{6} fake count patterns are binned to form the observable moments, where the standard deviation of these moments from their exact values σA\sigma_{A} is the actual sampling error we aim to estimate using Eq.(45).

References

  • Knill et al. [2001] E. Knill, R. Laflamme, and G. J. Milburn, A scheme for efficient quantum computation with linear optics, Nature 409, 46 (2001).
  • Aaronson and Arkhipov [2013] S. Aaronson and A. Arkhipov, The Computational Complexity of Linear Optics, Theory of Computing 9, 143 (2013).
  • Zhong et al. [2020] H.-S. Zhong, H. Wang, Y.-H. Deng, M.-C. Chen, L.-C. Peng, Y.-H. Luo, J. Qin, D. Wu, X. Ding, Y. Hu, et al., Quantum computational advantage using photons, Science 370, 1460 (2020).
  • Zhong et al. [2021] H.-S. Zhong, Y.-H. Deng, J. Qin, H. Wang, M.-C. Chen, L.-C. Peng, Y.-H. Luo, D. Wu, S.-Q. Gong, H. Su, Y. Hu, P. Hu, X.-Y. Yang, W.-J. Zhang, H. Li, Y. Li, X. Jiang, L. Gan, G. Yang, L. You, Z. Wang, L. Li, N.-L. Liu, J. J. Renema, C.-Y. Lu, and J.-W. Pan, Phase-Programmable Gaussian Boson Sampling Using Stimulated Squeezed Light, Phys. Rev. Lett. 127, 180502 (2021).
  • Madsen et al. [2022] L. S. Madsen, F. Laudenbach, M. F. Askarani, F. Rortais, T. Vincent, J. F. F. Bulmer, F. M. Miatto, L. Neuhaus, L. G. Helt, M. J. Collins, A. E. Lita, T. Gerrits, S. W. Nam, V. D. Vaidya, M. Menotti, I. Dhand, Z. Vernon, N. Quesada, and J. Lavoie, Quantum computational advantage with a programmable photonic processor, Nature 606, 75 (2022).
  • Deng et al. [2023] Y.-H. Deng, Y.-C. Gu, H.-L. Liu, S.-Q. Gong, H. Su, Z.-J. Zhang, H.-Y. Tang, M.-H. Jia, J.-M. Xu, M.-C. Chen, J. Qin, L.-C. Peng, J. Yan, Y. Hu, J. Huang, H. Li, Y. Li, Y. Chen, X. Jiang, L. Gan, G. Yang, L. You, L. Li, H.-S. Zhong, H. Wang, N.-L. Liu, J. J. Renema, C.-Y. Lu, and J.-W. Pan, Gaussian boson sampling with pseudo-photon-number-resolving detectors and quantum computational advantage, Phys. Rev. Lett. 131, 150601 (2023).
  • Hamilton et al. [2017] C. S. Hamilton, R. Kruse, L. Sansoni, S. Barkhofen, C. Silberhorn, and I. Jex, Gaussian boson sampling, Phys. Rev. Lett. 119, 170501 (2017).
  • Quesada et al. [2018] N. Quesada, J. M. Arrazola, and N. Killoran, Gaussian boson sampling using threshold detectors, Physical Review A 98, 062322 (2018).
  • Mandel and Wolf [1965] L. Mandel and E. Wolf, Coherence Properties of Optical Fields, Rev. Mod. Phys. 37, 231 (1965).
  • Mandel and Wolf [1995] L. Mandel and E. Wolf, Optical Coherence and Quantum Optics (Cambridge University Press, Cambridge, 1995).
  • Glauber [1963a] R. J. Glauber, Coherent and Incoherent States of the Radiation Field, Phys. Rev. 131, 2766 (1963a).
  • Sudarshan [1963] E. C. G. Sudarshan, Equivalence of Semiclassical and Quantum Mechanical Descriptions of Statistical Light Beams, Phys. Rev. Lett. 10, 277 (1963).
  • Bassham III et al. [2010] L. E. Bassham III, A. L. Rukhin, J. Soto, J. R. Nechvatal, M. E. Smid, E. B. Barker, S. D. Leigh, M. Levenson, M. Vangel, D. L. Banks, et al., Sp 800-22 rev. 1a. a statistical test suite for random and pseudorandom number generators for cryptographic applications (2010).
  • Knuth [2014] D. E. Knuth, Art of computer programming, volume 2: Seminumerical algorithms (Addison-Wesley Professional, 2014).
  • Drummond et al. [2022] P. D. Drummond, B. Opanchuk, A. Dellios, and M. D. Reid, Simulating complex networks in phase space: Gaussian boson sampling, Phys. Rev. A 105, 012427 (2022).
  • Dellios et al. [2023] A. S. Dellios, B. Opanchuk, M. D. Reid, and P. D. Drummond, Validation tests of gbs quantum computers give evidence for quantum advantage with a decoherent target (2023), arXiv:2211.03480 [quant-ph] .
  • Glauber [1963b] R. J. Glauber, The Quantum Theory of Optical Coherence, Phys. Rev. 130, 2529 (1963b).
  • Villalonga et al. [2021] B. Villalonga, M. Y. Niu, L. Li, H. Neven, J. C. Platt, V. N. Smelyanskiy, and S. Boixo, Efficient approximation of experimental gaussian boson sampling, arXiv preprint arXiv:2109.11525  (2021).
  • Oh et al. [2023] C. Oh, L. Jiang, and B. Fefferman, Spoofing Cross-Entropy Measure in Boson Sampling, Phys. Rev. Lett. 131, 010401 (2023).
  • Bulmer et al. [2022] J. F. F. Bulmer, B. A. Bell, R. S. Chadwick, A. E. Jones, D. Moise, A. Rigazzi, J. Thorbecke, U.-U. Haus, T. Van Vaerenbergh, R. B. Patel, I. A. Walmsley, and A. Laing, The boundary for quantum advantage in Gaussian boson sampling, Sci. Adv. 8, eabl9236 (2022).
  • Drummond and Reid [2016] P. D. Drummond and M. D. Reid, Coherent states in projected hilbert spaces, Physical Review A 94, 063851 (2016).
  • Drummond and Ficek [2004] P. D. Drummond and Z. Ficek, eds., Quantum Squeezing (Springer-Verlag, Berlin, Heidelberg, New York, 2004).
  • Shi and Byrnes [2022] J. Shi and T. Byrnes, Effect of partial distinguishability on quantum supremacy in Gaussian Boson sampling, npj Quantum Inf 8, 54 (2022).
  • Drummond and Opanchuk [2020] P. D. Drummond and B. Opanchuk, Initial states for quantum field simulations in phase space, Physical Review Research 2, 033304 (2020).
  • Walls and Milburn [2008] D. Walls and G. Milburn, Quantum Optics (Springer, 2008).
  • Sperling et al. [2012] J. Sperling, W. Vogel, and G. S. Agarwal, True photocounting statistics of multiple on-off detectors, Phys. Rev. A 85, 023820 (2012).
  • Kruse et al. [2019] R. Kruse, C. S. Hamilton, L. Sansoni, S. Barkhofen, C. Silberhorn, and I. Jex, Detailed study of gaussian boson sampling, Physical Review A 100, 032326 (2019).
  • Quesada and Arrazola [2020] N. Quesada and J. M. Arrazola, Exact simulation of gaussian boson sampling in polynomial space and exponential time, Physical Review Research 2, 023005 (2020).
  • Barvinok [1999] A. Barvinok, Polynomial Time Algorithms to Approximate Permanents and Mixed Discriminants Within a Simply Exponential Factor, Random Struct. Alg. 14, 29 (1999).
  • Rudelson et al. [2016] M. Rudelson, A. Samorodnitsky, and O. Zeitouni, Hafnians, perfect matchings and Gaussian matrices, Ann. Probab. 4410.1214/15-AOP1036 (2016).
  • Deshpande et al. [2022] A. Deshpande, A. Mehta, T. Vincent, N. Quesada, M. Hinsche, M. Ioannou, L. Madsen, J. Lavoie, H. Qi, J. Eisert, D. Hangleiter, B. Fefferman, and I. Dhand, Quantum computational advantage via high-dimensional Gaussian boson sampling, Sci. Adv. 8, eabi7894 (2022).
  • Qi et al. [2020] H. Qi, D. J. Brod, N. Quesada, and R. García-Patrón, Regimes of classical simulability for noisy gaussian boson sampling, Physical review letters 124, 100502 (2020).
  • Oh et al. [2024] C. Oh, M. Liu, Y. Alexeev, B. Fefferman, and L. Jiang, Classical algorithm for simulating experimental gaussian boson sampling, Nature Physics , 1 (2024).
  • Huang and Kumar [1989] J. Huang and P. Kumar, Photon-counting statistics of multimode squeezed light, Phys. Rev. A 40, 1670 (1989).
  • Zhu and Caves [1990] C. Zhu and C. M. Caves, Photocount distributions for continuous-wave squeezed light, Phys. Rev. A 42, 6794 (1990).
  • Mehmet et al. [2010] M. Mehmet, H. Vahlbruch, N. Lastzka, K. Danzmann, and R. Schnabel, Observation of squeezed states with strong photon-number oscillations, Phys. Rev. A 81, 013814 (2010).
  • Wigner [1932] E. Wigner, On the Quantum Correction For Thermodynamic Equilibrium, Phys. Rev. 40, 749 (1932).
  • Husimi [1940] K. Husimi, Some formal properties of the density matrix, Proc. Phys. Math. Soc. Jpn. 22, 264 (1940).
  • Martínez-Cifuentes et al. [2023] J. Martínez-Cifuentes, K. M. Fonseca-Romero, and N. Quesada, Classical models may be a better explanation of the Jiuzhang 1.0 Gaussian Boson Sampler than its targeted squeezed light model, Quantum 7, 1076 (2023).
  • Drummond and Gardiner [1980] P. D. Drummond and C. W. Gardiner, Generalised p-representations in quantum optics, Journal of Physics A: Mathematical and General 13, 2353 (1980).
  • Opanchuk et al. [2018] B. Opanchuk, L. Rosales-Zárate, M. D. Reid, and P. D. Drummond, Simulating and assessing boson sampling experiments with phase-space representations, Physical Review A 97, 042304 (2018).
  • Adam et al. [1994] P. Adam, I. Földesi, and J. Janszky, Complete basis set via straight-line coherent-state superpositions, Physical Review A 49, 1281 (1994).
  • Phillips et al. [2019] D. Phillips, M. Walschaers, J. Renema, I. Walmsley, N. Treps, and J. Sperling, Benchmarking of gaussian boson sampling using two-point correlators, Physical Review A 99, 023836 (2019).
  • Pearson [1900] K. Pearson, X. on the criterion that a given system of deviations from the probable in the case of a correlated system of variables is such that it can be reasonably supposed to have arisen from random sampling, The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science 50, 157 (1900).
  • Wilson and Hilferty [1931] E. B. Wilson and M. M. Hilferty, The Distribution of Chi-Square, Proc. Natl. Acad. Sci. U.S.A. 17, 684 (1931).
  • Johnson [1970] N. L. Johnson, Continuous Univariate Distributions, Houghton Mifflin Series in Statistics (Houghton Mifflin, Boston, 1970).
  • Freund and Wilson [2003] R. J. Freund and W. J. Wilson, Statistical Methods (Elsevier, 2003).
  • [48] Private communication. Photon counting distribution for the ideal ground truth of the 288-mode data was independently cross-validated. The distribution presented in the Borealis paper is closer to the experimental data than the actual ideal ground truth presented here.
  • Oh et al. [2022] C. Oh, Y. Lim, B. Fefferman, and L. Jiang, Classical Simulation of Boson Sampling Based on Graph Structure, Phys. Rev. Lett. 128, 190501 (2022).
  • Kloeden and Platen [1992] P. E. Kloeden and E. Platen, Stochastic Differential Equations, in Numerical Solution of Stochastic Differential Equations (Springer Berlin Heidelberg, Berlin, Heidelberg, 1992) pp. 103–160.
  • Rao [2009] C. R. Rao, Linear Statistical Inference and Its Applications (John Wiley & Sons, 2009).
  • Fisher [1930] R. A. Fisher, Moments and Product Moments of Sampling Distributions, Proceedings of the London Mathematical Society s2-30, 199 (1930).
  • McCullagh [2018] P. McCullagh, Tensor Methods in Statistics: Monographs on Statistics and Applied Probability, first edition. ed., CRC Revivals (Chapman and Hall/CRC, Boca Raton, FL, 2018).
  • Kendall et al. [1987] M. G. M. G. Kendall, A. Stuart, and J. K. Ord, Kendall’s advanced theory of statistics, 5th ed. (C. Griffin, London, 1987).
  • Loudon [1983] R. Loudon, The Quantum Theory of Light, 2nd ed., Oxford Science Publications (Clarendon Press, Oxford, 1983).