This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Eberhard limit for photon-counting Bell tests and its utility in quantum key distribution

Thomas McDermott1,2,∗, Morteza Moradi1, Antoni Mikos-Nuszkiewicz1,
& Magdalena Stobińska1,2
1 Institute of Informatics, Faculty of Mathematics, Informatics and Mechanics, University of Warsaw, Banacha 2, 02-097 Warsaw, Poland
2 Institute of Theoretical Physics, Faculty of Physics, University of Warsaw, Pasteura 5, 02-093 Warsaw, Poland
*[email protected]
Abstract

Loophole-free Bell tests are essential if one wishes to perform device-independent quantum key distribution (QKD), since any loophole could be used by a potential adversary to undermine the security of the protocol. Crucial work by Eberhard demonstrated that weakly entangled two-qubit states have a far greater resistance to the detection loophole than maximally entangled states, allowing one to close the loophole with detection efficiency greater than 2/3. Here we demonstrate that this same limit holds for photon-counting CHSH Bell tests which can demonstrate non-locality for higher dimensional multiphoton states such as two-mode squeezed vacuum and generalized Holland-Burnett states. In fact, we show evidence that these tests are in some sense universal, allowing feasible detection loophole-free tests for any multiphoton bipartite state, as long as the two modes are well correlated in photon number. Additionally, by going beyond the typical two-input two-output Bell scenario, we show that there are also photon-counting CGLMP inequalities which can also match the Eberhard limit, paving the way for more exotic loophole-free Bell tests. Finally we show that by exploiting this increased loss tolerance of non maximally entangled states, one can increase the key rates and loss tolerances of QKD protocols based on photon-counting tests.

preprint: APS/123-QED

I Introduction

Loophole-free Bell tests are the gold standard of entanglement testing [1, 2, 3, 4, 5]. They provide definitive proof that multiple parties share quantum entanglement which can then be used for applications such as device-independent quantum key distribution (DI-QKD)[6, 7, 8, 9, 10, 11, 12, 13], quantum metrology[14, 15, 16, 17] and device-independent randomness generation[18, 19, 20]. If instead there exists some loopholes, the proof is not definitive but relies on some extra assumptions. For example, the detection loophole is opened by filtering out some measurement results (usually when a photon is lost) i.e. performing post-selection. The proof of entanglement is then not definitive but based on additional fair-sampling assumptions [21, 22]. In particular, QKD without loophole-free Bell tests are then vulnerable to attacks by exploiting these extra security assumptions, and commercial devices have been routinely hacked as a result [23, 24].

Closing the detection loophole is experimentally demanding and requires minimizing losses to increase the detection efficiency above some threshold. The crucial first steps towards experimental tests which closed the detection loophole were enabled by pivotal works by Eberhard, who demonstrated that using non-maximally entangled states can greatly lower the strict efficiency requirements [25]. Specifically, using a maximally entangled polarization Bell pair 12(|HV+|VH)\frac{1}{\sqrt{2}}(\ket{\textrm{HV}}+\ket{\textrm{VH}}), the detection loophole is closed only for detector efficiencies η>2(21)82.8%\eta>2(\sqrt{2}-1)\approx 82.8\%, whereas using a non-maximally entangled state 1g2|HV+g|VH\sqrt{1-g^{2}}\ket{\textrm{HV}}+g\ket{\textrm{VH}}, one can approach the limit η>2/3\eta>2/3 as g0g\to 0. Paradoxically, this is when the state approaches the separable |HV\ket{HV}, so that it becomes less tolerant to other types of noise. In practice, the optimal value of gg then depends on the relative importance of different sources of imperfection. The first experiments exploiting the Eberhard inequality chose a value of g=0.297g=0.297 [26].

Nowadays loophole-free Bell tests are becoming more common [1, 2, 3, 4], with the first fully DI-QKD demonstrations being performed [27]. However, until now all these Bell tests have been performed for 2-qubit polarization/spin singlet states and using the typical CHSH test based on polarization/spin measurements[28], precluding the possibility of testing the entanglement of some more exotic states, such as multiphoton/qudit states. Bell tests for multiphoton states of light based on photon-counting experiments were first theorised in the 1990s[29, 30, 31, 32] and have since been expanded on in a number of papers [33, 34, 35, 36], including a proposal for DI-QKD based on these tests [37]. These tests work for a variety of important states such as two-mode squeezed vacuum[33], generalized Holland-Burnett states[37] and hybrid-entangled Schrödinger cat states[38].

In this work we investigate whether these photon-counting tests form a universal test for entanglement in the photon-number degree of freedom. We demonstrate that any entangled state with high photon-number correlations can violate one of a number of Bell inequalities without employing any post-selection or filtering of results, thus having the possibility of closing the detection loophole. This requires low losses and high efficiency detectors, however we demonstrate that states which asymptotically approach the vacuum state have the maximum possible efficiency tolerance η>2/3\eta>2/3, matching that obtained by Eberhard for polarization/spin measurements. Moreover, this tolerance is not limited to the CHSH tests introduced in the original papers, but is also a property of the multiple output CGLMP inequalities[39], opening up loophole-free Bell tests to both more exotic states and more exotic inequalities. Finally, we demonstrate that the Eberhard inequality found in this paper has utility for DI-QKD protocols, allowing one to increase both the loss tolerance and key rates using less than maximally entangled states.

II Photon-counting Bell tests

Refer to caption
Figure 1: Setup for the photon-counting Bell tests. Alice and Bob share a two-mode state |Ψ\ket{\Psi} and wish to verify its entanglement. Each party interferes their mode with coherent states |α\ket{\alpha} and |β\ket{\beta}, on beam splitters with reflectivities rar_{a}, rbr_{b} respectively, and then measures the transmitted photon-numbers with photon-number-resolving (PNR) detectors. By splitting their measurements into binary outcomes such as zero/non-zero photons or even/odd numbers of photons, they can evaluate the CHSH inequality for this scenario. Ideally the beam splitters are in the ‘displacement limit’ with ra,rb0r_{a},r_{b}\to 0 and |α|2,|β|2|\alpha|^{2},|\beta|^{2}\to\infty such that they perform the coherent displacements D(δα)D(-\delta_{\alpha}) and D(δβ)D(-\delta_{\beta}), where δα=iαr\delta_{\alpha}=i\alpha\sqrt{r}, δβ=iβr\delta_{\beta}=i\beta\sqrt{r}.

Figure 1 depicts a Bell test for a two-mode state of light |Ψ\ket{\Psi}. The two modes are spatially separated, with one mode being sent to Alice and the other to Bob. Each party interferes their mode with a coherent state |α\ket{\alpha} and |β\ket{\beta}, on variable beam splitters with reflectivities rar_{a} and rbr_{b} respectively. Finally they use photon-number-resolving (PNR) detectors to measure the number of photons ii, jj in the two transmitted modes behind the beam splitters. The combination of the complex amplitude α\alpha and reflectivity rar_{a} form a ‘measurement setting’ for Alice, while β\beta and rbr_{b} form the corresponding setting for Bob. In the limits |α|,|β||\alpha|,|\beta|\to\infty and r0r\to 0 the beam splitter interactions are equivalent to displacement operators D(δα)D(-\delta_{\alpha}) and D(δβ)D(-\delta_{\beta}) where δα=iαarcsin(ra)iαra\delta_{\alpha}=i\alpha\arcsin(\sqrt{r_{a}})\approx i\alpha\sqrt{r_{a}} and δβ\delta_{\beta} defined similarly. By collecting measurement statistics for m2m\geq 2 settings per party, Alice and Bobs can prove entanglement of their shared state, either by showing violation of a Bell inequality or via linear programming methods.

Since the coherent states have a finite probability to contain any number of photons, i,ji,j can be any integer, and the number of measurement outcomes is unbounded. To make use of known Bell inequalities it is thus necessary to reduce these to a finite number of outcomes, Δ\Delta. The simplest case m=2,Δ=2m=2,\Delta=2 has been analyzed in the previous literature [29, 30, 31, 32]. One way to reduce the output to two outcomes is for Alice (Bob) to assign a measurement outcome of a=1a=1 (b=1)(b=1) when zero photons are transmitted and an outcome of a=1a=-1 (b=1)(b=-1) when non-zero photons are transmitted. The probability for a joint measurement of zero photons is then

p(i=0,j=0)=δα|aδβ|b|ΨΨ||δαa|δβb=Q(δα,δβ),p(i=0,j=0)=\bra{\delta_{\alpha}}_{a}\bra{\delta_{\beta}}_{b}\ket{\Psi}\bra{\Psi}\ket{\delta_{\alpha}}_{a}\ket{\delta_{\beta}}_{b}=Q(\delta_{\alpha},\delta_{\beta}), (1)

which is simply the two-mode Q-function of the state Q(δα,δβ)Q(\delta_{\alpha},\delta_{\beta}) (we neglect the usual factors of π\pi). Moreover, the marginal probability that Alice detects zero photons

p(i=0)=δα|Trb(|ΨΨ|)|δα=Qa(δα),p(i=0)=\bra{\delta_{\alpha}}\textrm{Tr}_{b}(\ket{\Psi}\bra{\Psi})\ket{\delta_{\alpha}}=Q_{a}(\delta_{\alpha}), (2)

is just the Q-function of her reduced state, Qa(δα)Q_{a}(\delta_{\alpha}), and Bob’s probability p(j=0)p(j=0) is defined similarly. Here Trb\textrm{Tr}_{b} is a trace over Bob’s modes. If Alice and Bob choose δα\delta_{\alpha} and δβ\delta_{\beta} randomly from the sets [δα1,δα2][\delta_{\alpha_{1}},\delta_{\alpha_{2}}] and [δβ1,δβ2][\delta_{\beta_{1}},\delta_{\beta_{2}}] respectively, the CHSH inequality[28] for this zero/non-zero test can then be expressed entirely in terms of Q-functions as follows

BQ24=Q11+Q12+Q21Q22Qa1Qb1,\frac{B_{Q}-2}{4}=Q_{11}+Q_{12}+Q_{21}-Q_{22}-Q_{a1}-Q_{b1}, (3)

|BQ|2|B_{Q}|\leq 2, where Qxy=Q(δαx,δβy)Q_{xy}=Q(\delta_{\alpha x},\delta_{\beta y}), Qax=Qa(δαx)Q_{ax}=Q_{a}(\delta_{\alpha x}) and Qby=Qb(δβy)Q_{by}=Q_{b}(\delta_{\beta y}).

Another simple way to obtain two outcomes is for Alice (Bob) to assign a measurement outcome of a=1a=1 (b=1)(b=1) when an even number of photons are transmitted and an outcome of a=1a=-1 (b=1)(b=-1) when an odd number are transmitted. This even/odd test can also be directly related to a quasi-probability distribution, since the quantum correlation ab\langle ab\rangle is given directly by the two-mode Wigner function

W(α,β)=Ψ|D(α)D(β)(1)aa(1)bbD(α)D(β)|Ψ.W(\alpha,\beta)=\bra{\Psi}D(\alpha)D(\beta)(-1)^{a^{\dagger}a}(-1)^{b^{\dagger}b}D(\alpha)^{\dagger}D(\beta)^{\dagger}\ket{\Psi}. (4)

The CHSH inequality for this test is then

BW=W11+W12+W21W22,\displaystyle B_{W}=W_{11}+W_{12}+W_{21}-W_{22}, (5)

|BW|2|B_{W}|\leq 2, where Wxy=W(δαx,δβy)W_{xy}=W(\delta_{\alpha x},\delta_{\beta y}).

Perhaps the most simple way to generalize the above tests is to extend the number of outcomes Δ\Delta beyond two. This requires the following general formula to measure arbitrary photon numbers i,ji,j

p(i,j|δαδβ)\displaystyle p(i,j|\delta_{\alpha}\delta_{\beta}) =|i|j|D(δα)D(δβ)|Ψ|2\displaystyle=\left|\bra{i}\bra{j}D(\delta_{\alpha})D(\delta_{\beta})\ket{\Psi}\right|^{2} (6)
=|m,n=0cmnDim(δα)Djn(δβ)|2,\displaystyle=\left|\sum_{m,n=0}^{\infty}c_{mn}D_{im}(\delta_{\alpha})D_{jn}(\delta_{\beta})\right|^{2},

where Dim(δα)=i|D(δα)|mD_{im}(\delta_{\alpha})=\bra{i}D(\delta_{\alpha})\ket{m} is the transition element of the displacement operator between two Fock states

Dim\displaystyle D_{im} (δα)=m!i!δαime|δα|22\displaystyle(\delta_{\alpha})=\sqrt{m!i!}\hskip 3.99994pt\delta_{\alpha}^{i-m}\hskip 3.99994pte^{-\frac{|\delta_{\alpha}|^{2}}{2}} (7)
×p=max(0,mi)m(|α|2)pp!(mp)!(pm+i)!.\displaystyle\times\sum_{p=\text{max}(0,m-i)}^{m}\frac{(-|\alpha|^{2})^{p}}{p!(m-p)!(p-m+i)!}.

Using Eq. (6) we can evaluate inequalities in the scenario m=2m=2, Δ=3\Delta=3. The facet-defining inequalities here are the CGLMP inequalities[39], plus the usual projections onto CHSH. Instead of separating into zero/non-zero outcomes as before, we can now separate into zero photons, single photons and two or more photons, forming our three outcomes per party. The CGLMP inequality for this scenario may be written as

I=G(δα1,δβ1)G(δα2,δβ1)+G(δα2,δβ2)+G(δβ2,δα1),I=G(\delta_{\alpha 1},\delta_{\beta 1})-G(\delta_{\alpha 2},\delta_{\beta 1})+G(\delta_{\alpha 2},\delta_{\beta 2})+G(\delta_{\beta 2},\delta_{\alpha 1}), (8)

I2I\leq 2, where

G(δα,δβ)=\displaystyle G(\delta_{\alpha},\delta_{\beta})= 3[p(i=0,j=0|δα,δβ)\displaystyle 3\Big{[}p(i=0,j=0|\delta_{\alpha},\delta_{\beta}) (9)
+p(i=1,j=0|δα,δβ)\displaystyle+p(i=1,j=0|\delta_{\alpha},\delta_{\beta})
+p(i=1,j=1|δα,δβ)]\displaystyle+p(i=1,j=1|\delta_{\alpha},\delta_{\beta})\Big{]}
2p(i=1|δα)2p(j=0|δβ)\displaystyle-2p(i=1|\delta_{\alpha})-2p(j=0|\delta_{\beta})
p(i=0|δα)p(j=1|δβ)+1.\displaystyle-p(i=0|\delta_{\alpha})-p(j=1|\delta_{\beta})+1.

Notice that all of the described photon-counting Bell tests take into account every possible experimental outcome, including cases where photons are lost, so that no post-selection or fair sampling assumptions are required. This allows the possibility for these tests to close the detection loophole without any modifications.

Refer to caption
Figure 2: Photon-counting CHSH tests for highly correlated states of the form C00|00+C11|11+C22|22C_{00}\ket{00}+C_{11}\ket{11}+C_{22}\ket{22}, where C22=+1C002C112C_{22}=+\sqrt{1-C_{00}^{2}-C_{11}^{2}} is fixed by normalization. Figures (a), (b) and (c) show for each value C00,C11C_{00},C_{11} the maximum violation for the zero/non-zero test, the even/odd test, and their combination, when numerically optimized over all measurement settings. The white areas show regions where there was no violation i.e. CHSH2\textrm{CHSH}\leq 2. The greatest violation of 2.69 is found for the states (±|00±|11)/2(\pm\ket{00}\pm\ket{11})/\sqrt{2}. Figures (d), (e) and (f) shows the maximum loss before the CHSH test fails (B<2B<2), for the tests in (a), (b) and (c) respectively. The zero/non-zero test is significantly more loss tolerant than the even/odd test, particularly in the vicinity of the points |C00|1|C_{00}|\approx 1 where the state approaches vacuum. Here there is an Eberhard level of loss tolerance 33%33\%, although the loss tolerance is significant even outside of this extreme.
Refer to caption
Figure 3: Photon-counting Bell tests on a two-mode squeezed vacuuum state as a function of parametric gain. (a) Bell inequality value for three different tests. (b) Maximum loss tolerance for these tests, showing Eberhard level loss tolerance η>2/3\eta>2/3 as the parametric gain tends to zero. All three tests satisfy this requirement although the even/odd and CGLMP tests degrade much faster with increasing gain.

To determine whether such tests work as a universal test for entanglement-verification in the photon number regime, we consider states of the form m,n=0dCmn|m|n\sum_{m,n=0}^{d}C_{mn}\ket{m}\ket{n}. We first considered the perfectly correlated states Cmn=CnnδmnC_{mn}=C_{nn}\delta_{mn}, and performed an exhaustive search over all complex values CnnC_{nn} for d4d\leq 4. By numerically maximizing the CHSH violation over all four complex displacements, we found that the test worked to prove entanglement for the vast majority of states, the only exception being states with very weak levels of entanglement i.e. those that are close to separable states. Thus we may conjecture that the photon-counting Bell tests work for all perfectly correlated pure entangled states, and since these tests do not require any post-selection or filtering of results, they have the possibility of closing the detection loophole. To illustrate this point, we plot in Fig. 2 the results for states of up to d=2d=2 photons, with real coefficients CnnC_{nn}. C00C_{00} and C11C_{11} are varied while C22C_{22} is fixed by the normalization condition C22=+1C002C112C_{22}=+\sqrt{1-C^{2}_{00}-C^{2}_{11}}. The positive square root can be taken without loss of generality since the negative square root is equivalent to flipping the sign of all other coefficients, as the state would change only by a global phase.

Fig. 2(a) shows the CHSH value for the zero/non-zero test. One observes that the test succeeds mostly everywhere, notably except along the line C00=0C_{00}=0, i.e. states which have no vacuum component. The test also fails for states like C00+ϵC22C_{00}+\epsilon C_{22}, ϵ1\epsilon\ll 1 i.e. states that are close to vacuum with no single photon component. Everywhere else this test works well, and can yield fairly high violations of local realism. The greatest violation of 2.69 is found for the states (±|00±|11)/2(\pm\ket{00}\pm\ket{11})/\sqrt{2}. Although these states are maximally entangled in 2 dimensions, the basis states are not qubits but Fock states which are infinite dimensional. Fig. 2(b) shows the CHSH value for the even/odd test, where one observes that it is much weaker than the zero/non-zero test. One advantage however is that, unlike the zero/non-zero test, it can succeed for states with no vacuum component C00=0C_{00}=0. This test instead fails for states with a definite photon number parity e.g. along the line C11=0C_{11}=0, as could be expected. Performing both of these tests can then eliminate the weaknesses of the individual tests, leading to fairly large CHSH violations for the vast majority of photon-number correlated states, as shown in Fig. 2(c). In contrast, states without perfect correlation, Cmn0C_{mn}\neq 0 for mnm\neq n, cannot always be verified by these tests. In general, as the levels of entanglement and photon-number correlations increase, the more likely the test will succeed.

III Loss Tolerance

The amount of CHSH violation does not give a complete picture, however. A crucial factor for experimental tests is how tolerant the Bell inequality violation is to photonic losses. These losses may be modelled by interfering the state with vacuum on a fictional beamsplitter of reflectivity 1η1-\eta and tracing out the reflected beam, equivalent to a photon being successfully transmitted with probability η\eta. With these losses considered solely to occur after the displacement operation, η\eta can be interpreted as the detection efficiency of the two detectors. Fig. 2(d) shows the loss tolerance of the zero/non-zero test for the 2-photon states. Interestingly, one observes that as one approaches the vacuum state C001C_{00}\to 1, the loss tolerance approaches η>2/3\eta>2/3, exactly the Eberhard limit. Placing the losses in the source i.e. before the displacement operation, leads to very similar conclusions with only a slightly more strict limit η>0.71\eta>0.71. From a mathematical point of view the fact that we can reach the Eberhard limit is not too surprising, since the state in this limit can be approximated as something like 1g2|00+g|11\sqrt{1-g^{2}}\ket{00}+g\ket{11} which is mathematically identical to the 2-qubit state originally considered by Eberhard. However, it is interesting that this limit can be approached through this completely different measurement scheme, with routine optical experimental devices[40, 41], and with entanglement in photon-number rather than polarization or spin. The even/odd test has a much worse loss tolerance, as seen in Fig. 2(e). Although it also displays Eberhard level loss tolerance in the same limit, it drops rapidly as we leave the vicinity of the vacuum state |00\ket{00}, so that it is not really visible in this figure.

For a concrete example we can consider the two-mode squeezed vacuum state |ψ=1coshgn=0(eiϕtanhg)n|n|n\ket{\psi}=\frac{1}{\cosh g}\sum_{n=0}^{\infty}(-e^{i\phi}\tanh g)^{n}\ket{n}\ket{n} which is a qudit state of in principle infinite dimension. The photon-counting Bell tests have been considered for these states in the early papers[29, 30, 31, 32], althought they didn’t find the greatest violation of the Bell inequalities and didn’t realise its large loss tolerance in the low squeezing limit g0g\to 0.

The relevant Q-functions for the TMSV state are

Q(α,β)=1cosh2gexp[|α|2|β|22Re{αβeiϕ}tanhg],Q(\alpha,\beta)=\frac{1}{\cosh^{2}g}\exp[-|\alpha|^{2}-|\beta|^{2}-2\textrm{Re}\{\alpha\beta e^{-i\phi}\}\tanh g], (10)
Qa(α)=1cosh2gexp[|α|2(1tanh2g)],Q_{a}(\alpha)=\frac{1}{\cosh^{2}g}\exp[-|\alpha|^{2}(1-\tanh^{2}g)], (11)

with Qb(β)Q_{b}(\beta) defined analogously, while the two-mode Wigner function is

W(α,β)=exp[\displaystyle W(\alpha,\beta)=\exp[ 2(|α|2+|β|2)cosh(2g)\displaystyle-2(|\alpha|^{2}+|\beta|^{2})\cosh(2g) (12)
4Re{αβeiϕ}sinh(2g)].\displaystyle-4\textrm{Re}\{\alpha\beta e^{-i\phi}\}\sinh(2g)]. (13)

The CHSH parameters for both tests, |BQ||B_{Q}| and |BW||B_{W}| were numerically maximized by a combination of differential evolution, Nelder-mead and simulated annealing methods, using Eqs. (3) and (5). For the zero/non-zero test, the optimal choice of measurement settings [δα1,δα2],[δβ1,δβ2][\delta_{\alpha_{1}},\delta_{\alpha_{2}}],[\delta_{\beta_{1}},\delta_{\beta_{2}}] maximizing |BQ||B_{Q}| were found to have the form [x,y],[xeiϕ,yeiϕ][x,-y],[xe^{i\phi},-ye^{i\phi}], where xx and yy are positive real numbers which depend on the amount of squeezing gg. This choice of relative phases can be understood as those maximizing Q(δα1,δβ2)Q(\delta_{\alpha 1},\delta_{\beta 2}) and Q(δα2,δβ1)Q(\delta_{\alpha 2},\delta_{\beta 1}) while minimizing Q(δα2,δβ2)Q(\delta_{\alpha 2},\delta_{\beta 2}). The relative phase between δα1\delta_{\alpha_{1}} and δα2\delta_{\alpha_{2}} is then fixed, forcing Q(δα1,δβ1)Q(\delta_{\alpha 1},\delta_{\beta 1}) to be minimized. For the even/odd test, the optimal settings maximizing |BW||B_{W}| have the same form, with similar reasoning, in this case maximizing/minimizing the various Wigner functions rather than the Q functions.

The results are shown in Fig. 3(a) as a function of parametric gain, gg. One sees that the zero/non-zero test is optimal for the more experimentally relevant low gain regime, shown by the fact that |BQ|>|BW||B_{Q}|>|B_{W}|, reaching an optimal value of |BQ|=2.45|B_{Q}|=2.45 at g=0.74g=0.74. Increasing gg further, the zero/non-zero test weakens due to the fact that zero photon events become increasing unlikely. The even/odd test dominates here |BW|>|BQ||B_{W}|>|B_{Q}|, and reaches a maximum value |BW|=2.32|B_{W}|=2.32 in the limit of high gain gg\to\infty. Since the TMSV state becomes the original EPR state in this limit, in theory the even/odd test becomes a solution to the EPR paradox as originally proposed. However, as pointed out in the original analysis [30], there is difficulty both reaching, and performing the relevant measurements in, this limit. These results further highlight the general features we noticed before. The zero/non-zero test works better for low photon-number states i.e. low squeezing gg, and the even/odd test works better for high photon-number states i.e. high squeezing gg. Additionally, we also consider the CGLMP inequality with 3 outputs, distinguishing between 0, 1 and 2 or more photons. This performs similarly to the zero/non-zero test, although the Bell inequality violation is slightly lower. Furthermore from Fig. 3(b), one sees that all three tests reach the Eberhard level loss tolerance η>2/3\eta>2/3, although the zero/non-zero test has a loss tolerance which decays much less rapidly than the other two, so is more practical for testing the non-locality of the two-mode squeezed vacuum state. The reason for this is that losing a photon is catastrophic for the even/odd test since it changes an even result to an odd one and vice-versa. In contrast the zero/non-zero test is more robust since a non-zero result can remain non-zero. The fact that the CGLMP inequality also displays Eberhard level loss tolerace in the low squeezing regime, opens up the possibility of loophole-free Bell inequality tests using higher dimensional inequalities. Unfortunuately, the loss tolerance degrades faster with parametric gain than the zero/non-zero CHSH test, so that it appears to have limited practical use for the two-mode squeezed vacuum state in particular.

The Eberhard limit is reached when we consider detector losses i.e. with the losses occurring after the displacement operation, however we have also considered losses in the source i.e. with losses occurring before the displacement operation. The loss tolerance in this case is shown as the dashed lines in Fig. 3(b). The conclusions are not much changed, except that we reach a slightly worse maximum loss tolerance of η>0.71\eta>0.71.

IV Utility for QKD

As an example of the utility of this Eberhard limit for photon-counting, we can consider a DI-QKD protocol based upon these tests, similar to what has been considered in[37]. There, a proposal was considered in which generalized Holland-Burnett states (i.e. the output of two Fock states interfering on a beam splitter) are distributed between two parties, and the zero/non-zero Bell test was used to establish a shared secret key. We now consider the non-maximally entangled state 1ϵ|00+ϵ|11\sqrt{1-\epsilon}\ket{00}+\sqrt{\epsilon}\ket{11} and demonstrate that using non-maximal entanglement can yield increased key rates and loss tolerances for key distribution. For ϵ=1/2\epsilon=1/2 this state becomes 12(|00+|11)\frac{1}{\sqrt{2}}(\ket{00}+\ket{11}) which leads to the same Bell inequality violation 2.69~{}2.69 and loss tolerance 15%~{}15\% as the state considered in [37]. We have already seen that lowering ϵ\epsilon from this symmetrical value can increase the loss tolerance of the Bell tests, approaching the Eberhard limit, but it remains to be seen whether it can also increase the loss tolerance of the DI-QKD protocol as a whole, since in general the loss tolerance of DI-QKD is even more strict than the loophole-free Bell test itself.

We consider a simple DI-QKD scenario based on CHSH violation as outlined in[6, 7, 8, 11]. In addition to the two measurement settings per party used in the CHSH test [δα1,δα2][\delta_{\alpha_{1}},\delta_{\alpha_{2}}] and [δβ1,δβ2][\delta_{\beta_{1}},\delta_{\beta_{2}}], Alice now uses a third setting δα0\delta_{\alpha_{0}}. Each time the entangled state is distributed to them, Alice performs a measurement randomly choosing between her three settings, and Bob randomly chooses between his two settings. After a number of trials NN\to\infty, they publicly communicate the settings used for each trial. A number of results using settings from the sets [δα1,δα2][\delta_{\alpha_{1}},\delta_{\alpha_{2}}] and [δβ1,δβ2][\delta_{\beta_{1}},\delta_{\beta_{2}}] are also communicated publicly and used to evaluate the CHSH inequality, while a small sample, NN^{\prime} of results using the combination δα0\delta_{\alpha_{0}} and δβ1\delta_{\beta_{1}} are kept secret and form the raw key shared between Alice and Bob. Violation of the CHSH inequality guarantees that Eve cannot obtain full information about the key, providing information-theoretic security of communications where they key is used in a one-time-pad.

After performing classical privacy amplification and error correction on their raw key of length NN^{\prime}, they obtain the final key of reduced length KNKN^{\prime} where the key rate KK is given by[42]

K=I(A0:B1)I(B1:E).K=I(A_{0}:B_{1})-I(B_{1}:E). (14)

Here I(A0:B1)I(A_{0}:B_{1}) is the mutual information between Alice and Bob (when using settings δα0,δβ1\delta_{\alpha_{0}},\delta_{\beta_{1}}), specifying the fraction of bits that are kept after performing error correction, while I(B1:E)I(B_{1}:E) is the mutual information between Bob and Eve, giving the number of bits that are sacrificed for privacy amplification. Assuming Eve performs optimal collective attacks against the protocol, the key rate can be lower bounded[8, 9, 10] by

K=1h(1+B2/412)H(A0|B1),K=1-h\left(\frac{1+\sqrt{B^{2}/4-1}}{2}\right)-H(A_{0}|B_{1}), (15)

where h(x)h(x) is the binary entropy function, BB is the obtained CHSH inequality value, and H(A0|B1)H(A_{0}|B_{1}) is the conditional entropy between Alice and Bob. There are also methods for upgrading security against collective attacks to security against the most general coherent attacks [11]. Having loophole-free Bell inequality violation B>2B>2 is required, but not sufficient, to obtain a finite key rate K>0K>0, since the entropy H(A0|B1)H(A_{0}|B_{1}) must also be minimized. For this reason, we cannot obtain a loss tolerance of η>2/3\eta>2/3 for DI-QKD, but by using non-maximally entangled states, the loss tolerance can still be improved.

By numerically maximizing Eq. (15) with respect to the three settings of Alice and two settings of Bob, we have calculated the key rate that Alice and Bob may obtain by distributing the state 1ϵ|00+ϵ|11\sqrt{1-\epsilon}\ket{00}+\sqrt{\epsilon}\ket{11}, for varying levels of photonic loss. The results for the maximally entangled state ϵ=1/2\epsilon=1/2 are shown in the blue line Fig. 4, where one observes that after 7% loss, the key rate drops to zero, indicating a rather strict efficiency requirement of 93%. However, this can be improved considerably by instead distributing non-maximally entangled states. For each value of loss we have found the optimal value of ϵ\epsilon and calculated the key rates, which are shown as the orange line in Fig. 4. Compared to maximally entangled states, the loss tolerance is increased from 7% to 10.5%, a 50% improvement.

Refer to caption
Figure 4: Demonstration of the utility of the increased loss tolerance of less maximally entangled states for photon-counting QKD demonstrations. Both the key rate and loss tolerance can be increased by using non-maximally entangled states 1ϵ|00+ϵ|11\sqrt{1-\epsilon}\ket{00}+\sqrt{\epsilon}\ket{11} rather than the maximally entangled states (in two-dimensions).

V Conclusion

We have demonstrated the Eberhard inequality for photon-counting Bell tests, showing that detection loophole-free Bell tests are possible with efficiencies as low as 2/3, for photon-number entangled states which are generally multiphoton. We have shown this for two types of CHSH tests, based on discriminating between zero/non-zero photons, and even/odd numbers of photons respectively, but we have also shown it for CGLMP inequalities with more than two outputs. This opens up loophole-free Bell tests, and applications such as QKD, to more exotic states and types of entanglement than typical polarization/spin singlet states. The Eberhard limit for CGLMP inequalities could also allow detection loophole-free tests other than CHSH for the first time. The CHSH tests considered have been shown to work for states that are perfectly correlated in photon number, although for states without perfect correlation it can sometimes fail. Generally the more correlated in photon number states are, the more likely they are to violate these photon counting Bell tests, although a complete set of criteria to understand which states should violate these tests is still sought after.

To highlight the utility of the Eberhard level of loss tolerance for photon-counting Bell tests, we have shown that the loss tolerances and key rates of DI-QKD protocols can be drastically improved by the distribution of photon-number entangled states which are not maximally entangled. DI-QKD represents the ultimate level of cryptography which is immune to any eavesdropping at either the source or detection, and DI-QKD proposals based on photon number entanglement is an active area of research[37] since it has been shown to yield high key rates and immunity to large transmission losses. Thus our work represents a key step toward realization of this ambitious goal.

Acknowledgements.
T.M. and M.S. were supported by the Foundation for Polish Science “First Team” project No. POIR.04.04.00-00- 220E/16-00 (originally FIRST TEAM/2016-2/17). A.M.-N. and M.S. were supported by the National Science Centre “Sonata Bis” project No. 2019/34/E/ST2/00273. M.M. and M.S. were supported by the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie project “AppQInfo” No. 956071. T.M. and M.S. were supported by the National Science Centre within the QuantERA II Programme that has received funding from the European Union’s Horizon 2020 research and innovation programme under Grant Agreement No 101017733, project “PhoMemtor” No. 2021/03/Y/ST2/00177.

References

  • Hensen et al. [2015] B. Hensen, H. Bernien, A. E. Dréau, A. Reiserer, N. Kalb, M. S. Blok, J. Ruitenberg, R. F. L. Vermeulen, R. N. Schouten, C. Abellán, W. Amaya, V. Pruneri, M. W. Mitchell, M. Markham, D. J. Twitchen, D. Elkouss, S. Wehner, T. H. Taminiau, and R. Hanson, Experimental loophole-free violation of a Bell inequality using entangled electron spins separated by 1.3 km (2015), arXiv:1508.05949 [quant-ph] .
  • Giustina et al. [2015a] M. Giustina, M. A. M. Versteegh, S. Wengerowsky, J. Handsteiner, A. Hochrainer, K. Phelan, F. Steinlechner, J. Kofler, J.-A. Larsson, C. Abellán, W. Amaya, V. Pruneri, M. W. Mitchell, J. Beyer, T. Gerrits, A. E. Lita, L. K. Shalm, S. W. Nam, T. Scheidl, R. Ursin, B. Wittmann, and A. Zeilinger, Significant-loophole-free test of bell’s theorem with entangled photons, Phys. Rev. Lett. 115, 250401 (2015a).
  • Giustina et al. [2015b] M. Giustina, M. A. M. Versteegh, S. Wengerowsky, J. Handsteiner, A. Hochrainer, K. Phelan, F. Steinlechner, J. Kofler, J.-A. Larsson, C. Abellán, W. Amaya, V. Pruneri, M. W. Mitchell, J. Beyer, T. Gerrits, A. E. Lita, L. K. Shalm, S. W. Nam, T. Scheidl, R. Ursin, B. Wittmann, and A. Zeilinger, Significant-loophole-free test of bell’s theorem with entangled photons, Phys. Rev. Lett. 115, 250401 (2015b).
  • Shalm et al. [2015] L. K. Shalm, E. Meyer-Scott, B. G. Christensen, P. Bierhorst, M. A. Wayne, M. J. Stevens, T. Gerrits, S. Glancy, D. R. Hamel, M. S. Allman, K. J. Coakley, S. D. Dyer, C. Hodge, A. E. Lita, V. B. Verma, C. Lambrocco, E. Tortorici, A. L. Migdall, Y. Zhang, D. R. Kumor, W. H. Farr, F. Marsili, M. D. Shaw, J. A. Stern, C. Abellán, W. Amaya, V. Pruneri, T. Jennewein, M. W. Mitchell, P. G. Kwiat, J. C. Bienfang, R. P. Mirin, E. Knill, and S. W. Nam, Strong loophole-free test of local realism, Phys. Rev. Lett. 115, 250402 (2015).
  • Brunner et al. [2014] N. Brunner, D. Cavalcanti, S. Pironio, V. Scarani, and S. Wehner, Bell nonlocality, Rev. Mod. Phys. 86, 419 (2014).
  • Acín et al. [2006] A. Acín, S. Massar, and S. Pironio, Efficient quantum key distribution secure against no-signalling eavesdroppers, New Journal of Physics 8, 126 (2006).
  • Acín et al. [2007] A. Acín, N. Brunner, N. Gisin, S. Massar, S. Pironio, and V. Scarani, Device-independent security of quantum cryptography against collective attacks, Phys. Rev. Lett. 98, 230501 (2007).
  • Pironio et al. [2009] S. Pironio, A. Acín, N. Brunner, N. Gisin, S. Massar, and V. Scarani, Device-independent quantum key distribution secure against collective attacks, New Journal of Physics 11, 045021 (2009).
  • Masanes et al. [2011] L. Masanes, S. Pironio, and A. Acín, Secure device-independent quantum key distribution with causally independent measurement devices, Nature Communications 210.1038/ncomms1244 (2011).
  • Woodhead et al. [2021] E. Woodhead, A. Acín, and S. Pironio, Device-independent quantum key distribution with asymmetric CHSH inequalities, Quantum 5, 443 (2021).
  • Vazirani and Vidick [2014] U. Vazirani and T. Vidick, Fully device-independent quantum key distribution, Phys. Rev. Lett. 113, 140501 (2014).
  • Primaatmaja et al. [2022] I. W. Primaatmaja, K. T. Goh, E. Y. Z. Tan, J. T. F. Khoo, S. Ghorai, and C. C. W. Lim, Security of device-independent quantum key distribution protocols: a review (2022), arXiv:2206.04960 .
  • Zapatero et al. [2022] V. Zapatero, T. van Leent, R. Arnon-Friedman, W.-Z. Liu, Q. Zhang, H. Weinfurter, and M. Curty, Advances in device-independent quantum key distribution (2022), arXiv:2208.12842 .
  • Giovannetti et al. [2011] V. Giovannetti, S. Lloyd, and L. Maccone, Advances in quantum metrology, Nature Photonics 5, 222 (2011).
  • Colombo et al. [2022] S. Colombo, E. Pedrozo-Peñafiel, A. F. Adiyatullin, Z. Li, E. Mendez, C. Shu, and V. Vuletić, Time-reversal-based quantum metrology with many-body entangled states, Nature Physics 18, 925 (2022).
  • Liu et al. [2021a] J. Liu, M. Zhang, H. Chen, L. Wang, and H. Yuan, Optimal scheme for quantum metrology, Advanced Quantum Technologies 5, 2100080 (2021a).
  • Thekkadath et al. [2020] G. S. Thekkadath, M. E. Mycroft, B. A. Bell, C. G. Wade, A. Eckstein, D. S. Phillips, R. B. Patel, A. Buraczewski, A. E. Lita, T. Gerrits, S. W. Nam, M. Stobińska, A. I. Lvovsky, and I. A. Walmsley, Quantum-enhanced interferometry with large heralded photon-number states, npj Quantum Information 610.1038/s41534-020-00320-y (2020).
  • Pironio et al. [2010] S. Pironio, A. Acín, S. Massar, A. B. de la Giroday, D. N. Matsukevich, P. Maunz, S. Olmschenk, D. Hayes, L. Luo, T. A. Manning, and C. Monroe, Random numbers certified by bell’s theorem, Nature 464, 1021 (2010).
  • Liu et al. [2021b] W.-Z. Liu, M.-H. Li, S. Ragy, S.-R. Zhao, B. Bai, Y. Liu, P. J. Brown, J. Zhang, R. Colbeck, J. Fan, Q. Zhang, and J.-W. Pan, Device-independent randomness expansion against quantum side information, Nature Physics 17, 448 (2021b).
  • Avesani et al. [2021] M. Avesani, H. Tebyanian, P. Villoresi, and G. Vallone, Semi-device-independent heterodyne-based quantum random-number generator, Phys. Rev. Applied 15, 034034 (2021).
  • Pearle [1970] P. M. Pearle, Hidden-variable example based upon data rejection, Phys. Rev. D 2, 1418 (1970).
  • Clauser and Horne [1974] J. F. Clauser and M. A. Horne, Experimental consequences of objective local theories, Phys. Rev. D 10, 526 (1974).
  • Makarov [2009] V. Makarov, Controlling passively quenched single photon detectors by bright light, New Journal of Physics 11, 065003 (2009).
  • Lydersen et al. [2010] L. Lydersen, C. Wiechers, C. Wittmann, D. Elser, J. Skaar, and V. Makarov, Hacking commercial quantum cryptography systems by tailored bright illumination, Nature Photonics 4, 686 (2010).
  • Eberhard [1993] P. H. Eberhard, Background level and counter efficiencies required for a loophole-free einstein-podolsky-rosen experiment, Phys. Rev. A 47, R747 (1993).
  • Giustina et al. [2013] M. Giustina, A. Mech, S. Ramelow, B. Wittmann, J. Kofler, J. Beyer, A. Lita, B. Calkins, T. Gerrits, S. W. Nam, R. Ursin, and A. Zeilinger, Bell violation using entangled photons without the fair-sampling assumption, Nature 497, 227 (2013).
  • Zhang et al. [2022] W. Zhang, T. van Leent, K. Redeker, R. Garthoff, R. Schwonnek, F. Fertig, S. Eppelt, W. Rosenfeld, V. Scarani, C. C.-W. Lim, and H. Weinfurter, A device-independent quantum key distribution system for distant users, Nature 607, 687 (2022).
  • Clauser et al. [1969] J. F. Clauser, M. A. Horne, A. Shimony, and R. A. Holt, Proposed Experiment to Test Local Hidden-Variable Theories, Physical Review Letters 23, 880 (1969).
  • Banaszek and Wódkiewicz [1996] K. Banaszek and K. Wódkiewicz, Direct Probing of Quantum Phase Space by Photon Counting, Physical Review Letters 76, 4344 (1996).
  • Banaszek and Wódkiewicz [1998] K. Banaszek and K. Wódkiewicz, Nonlocality of the einstein-podolsky-rosen state in the wigner representation, Phys. Rev. A 58, 4345 (1998).
  • Banaszek and Wódkiewicz [1999] K. Banaszek and K. Wódkiewicz, Testing Quantum Nonlocality in Phase Space, Physical Review Letters 82, 2009 (1999).
  • Banaszek et al. [2002] K. Banaszek, A. Dragan, K. Wódkiewicz, and C. Radzewicz, Direct measurement of optical quasidistribution functions: Multimode theory and homodyne tests of bell’s inequalities, Phys. Rev. A 66, 043803 (2002).
  • Jeong et al. [2003] H. Jeong, W. Son, M. S. Kim, D. Ahn, and i. c. v. Brukner, Quantum nonlocality test for continuous-variable states with dichotomic observables, Phys. Rev. A 67, 012106 (2003).
  • Lee and Jeong [2011] S.-W. Lee and H. Jeong, High-dimensional bell test for a continuous-variable state in phase space and its robustness to detection inefficiency, Phys. Rev. A 83, 022103 (2011).
  • Dastidar and Sarbicki [2022] M. G. Dastidar and G. Sarbicki, Detecting entanglement between modes of light, Phys. Rev. A 105, 062459 (2022).
  • Brask and Chaves [2012] J. B. Brask and R. Chaves, Robust nonlocality tests with displacement-based measurements, Phys. Rev. A 86, 010103 (2012).
  • Mycroft et al. [2022] M. E. Mycroft, T. McDermott, A. Buraczewski, and M. Stobińska, Proposal for distribution of multi-photon entanglement with TF-QKD rate-distance scaling (2022), arXiv:1812.10935 [quant-ph] .
  • Ketterer [2016] A. Ketterer, Modular Variables in Quantum Information, Ph.D. thesis, Université Paris 7, Sorbonne Paris Cité (2016).
  • Collins et al. [2002] D. Collins, N. Gisin, N. Linden, S. Massar, and S. Popescu, Bell inequalities for arbitrarily high-dimensional systems, Phys. Rev. Lett. 88, 040404 (2002).
  • Banaszek et al. [1999] K. Banaszek, C. Radzewicz, K. Wódkiewicz, and J. S. Krasiński, Direct measurement of the wigner function by photon counting, Phys. Rev. A 60, 674 (1999).
  • Nehra et al. [2019] R. Nehra, A. Win, M. Eaton, R. Shahrokhshahi, N. Sridhar, T. Gerrits, A. Lita, S. W. Nam, and O. Pfister, State-independent quantum state tomography by photon-number-resolving measurements, Optica 6, 1356 (2019).
  • Devetak and Winter [2005] I. Devetak and A. Winter, Distillation of secret key and entanglement from quantum states, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 461, 207 (2005).