This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\typearea

14

Faster Amplitude Estimation

Kouhei Nakaji Department of Applied Physics and Physico-Informatics & Quantum Computing Center, Keio University, Hiyoshi 3-14-1, Kohoku, Yokohama, 223-8522, Japan
Abstract

In this paper, we introduce an efficient algorithm for the quantum amplitude estimation task which is tailored for near-term quantum computers. The quantum amplitude estimation is an important problem which has various applications in fields such as quantum chemistry, machine learning, and finance. Because the well-known algorithm for the quantum amplitude estimation using the phase estimation does not work in near-term quantum computers, alternative approaches have been proposed in recent literature. Some of them provide a proof of the upper bound which almost achieves the Heisenberg scaling. However, the constant factor is large and thus the bound is loose. Our contribution in this paper is to provide the algorithm such that the upper bound of query complexity almost achieves the Heisenberg scaling and the constant factor is small.

1 Introduction

The application of near-term quantum computers has been attracting significant interest recently. In the near-term quantum computers, the depth of the circuit and the number of qubits are constrained for reducing the noise. Under the constraints, most of the quantum algorithms which realize quantum speed-up in the ideal quantum computers are not available in the near-term quantum computers. The development of algorithms tailored for near-term quantum computers is demanded.

In this paper, we focus on the problem of quantum amplitude estimation in near-term quantum computers. Quantum amplitude estimation is the problem of estimating the value of sinθ\sin\theta in the following equation: 𝒜|0n|0=sinθ|Ψ~1n|1+cosθ|Ψ~0n|0\mathcal{A}|0\rangle_{n}|0\rangle=\sin\theta|\tilde{\Psi}_{1}\rangle_{n}|1\rangle+\cos\theta|\tilde{\Psi}_{0}\rangle_{n}|0\rangle. It is well known that the amplitude estimation can be applied to quantum chemistry, finance and machine learning [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]. In these applications, the cost of execution 𝒜\mathcal{A} is high, thus how to reduce the number of calling 𝒜\mathcal{A} while estimating θ\theta with required accuracy, is the heart of the problem.

The efficient quantum amplitude estimation algorithm which uses the phase estimation has been well documented in previous literature[11, 12]. The algorithm achieves Heisenberg scaling, that is, if we demand that the estimation error of sin2θ\sin^{2}\theta is within ϵ\epsilon with the probability larger than 1/21/2, then the query complexity (required number of the call of the operator 𝒜\cal{A}) is O(1/ϵ)O(1/\epsilon). However, the phase estimation requires a lot of noisy gates of two qubits and therefore is not suitable for near-term quantum computers. Thus, it is fair to inquire if there are any quantum amplitude estimation algorithms which work with less resources and still achieve Heisenberg scaling.

Recently, quantum amplitude estimation algorithms without phase estimation have been suggested in some literature [13, 14, 15, 16]. Suzuki et al [13] suggests an algorithm which uses maximum likelihood estimation. It shows that the algorithm achieves Heisenberg scaling numerically but there is no rigorous proofs. Wie [14] also studies the problem in the context of quantum counting. Still, it is not rigorously proved that the algorithm achieves Heisenberg scaling. Aaronson et al[15] suggests an algorithm which achieves Heisenberg scaling and give rigorous proof. However, the constant factor proportional to 1ϵln(1δ)\frac{1}{\epsilon}\ln\left(\frac{1}{\delta}\right) is large where δ\delta is the probability that the error is less than ϵ\epsilon. Grinko et al [16] also suggests an algorithm which achieves Heisenberg scaling, but the constant factor is still large in the worst case even though it is shown numerically that the constant is smaller in most of the cases.

In this paper, we propose a quantum amplitude estimation algorithm without phase estimation which acheives the Heisenberg scaling and the constant factor proportional to 1ϵln(1δ)\frac{1}{\epsilon}\ln\left(\frac{1}{\delta}\right) is smaller than previous literature[15, 16]. The structure of the paper is as follows. In Section 2, we discuss our proposed algorithm. The validity of the algorithm is verified by numerical experiments in Section 3. In Section Acknowledgement, we conclude with a discussion on future works. The complexity upper bound is proved in the Appendix.

2 Algorithm

In this section, we show our proposed algorithm. Before going into the detail, let us summarize the algorithm briefly.

Similar to the previous literature, we use the quantum amplitude amplification technique[11]. Given the amplitude sinθ\sin\theta that we want to estimate, the quantum amplitude amplification enables us to estimate the values of cos(2(2m+1)θ)\cos(2(2m+1)\theta) for each non-negative integer mm directly by measurements, and the resulting estimation errors of cos(2(2m+1)θ)\cos(2(2m+1)\theta) are of the order O(s)O(\sqrt{s}) with high probability when ss measurements are executed for each mm.

Our algorithm estimates the values of cos(2(2m+1)θ)\cos(2(2m+1)\theta) for each m=2j1(j=1)m=2^{j-1}(j=1\dots\ell) iteratively. As in Kitaev’s iterative phase estimation [17] (see also [18]), if 2(2j+1)θ|mod2π(j=1)2(2^{j}+1)\theta|_{\bmod 2\pi}(j=1\dots\ell) are estimated and those estimation errors are within π/2\sim\pi/2, then the value of θ\theta can be iteratively estimated with error O(1/2)O(1/2^{\ell}), meaning that the Heisenberg scaling is achieved. However, the value of 2(2j+1)θ|mod2π2(2^{j}+1)\theta|_{\bmod 2\pi} is generally not determinable only by the estimate of cos(2(2j+1)θ)\cos(2(2^{j}+1)\theta) because there is ambiguity whether 2(2j+1)θ|mod2π[0,π]2(2^{j}+1)\theta|_{\bmod 2\pi}\in[0,\pi] or 2(2j+1)θ|mod2π[π,2π]2(2^{j}+1)\theta|_{\bmod 2\pi}\in[\pi,2\pi]. Our algorithm solves this ambiguity by taking two-stages method. The algorithm is in the first stage when 2(2j+1)θ<π2(2^{j}+1)\theta<\pi. In this stage, 2(2j+1)θ|mod2π2(2^{j}+1)\theta|_{\bmod 2\pi} can be obtained from the estimate of cos(2(2j+1)θ)\cos(2(2^{j}+1)\theta) without ambiguity by using the inverse cosine function. When the estimate of 2(2j0+1)θπ/22(2^{j_{0}}+1)\theta\sim\pi/2 at the iteration j=j0j=j_{0}, the algorithm moves to the second stage. In the second stage, 2(2j+1)θ2(2^{j}+1)\theta might be larger than π\pi, hence 2(2j+1)θ|mod2π2(2^{j}+1)\theta|_{\bmod 2\pi} cannot be determined only by the measurements of cos(2(2j+1)θ)\cos(2(2^{j}+1)\theta) because of the above mentioned ambiguity. However, by combining the measurements of cos(2(2j+2j0+1)θ)\cos(2(2^{j}+2^{j_{0}}+1)\theta) with those of cos(2(2j+1)θ)\cos(2(2^{j}+1)\theta), the value of sin(2(2j+1)θ)\sin(2(2^{j}+1)\theta) can be estimated by using the trigonometric addition formula, and accordingly 2(2j+1)θ|mod2π2(2^{j}+1)\theta|_{\bmod 2\pi} can be determined without the ambiguity. As a result, the algorithm can estimate the value of θ\theta with the error less than O(1/2)O(1/2^{\ell}).

The algorithm in the reference [16] suggests a different approach for solving the ambiguity. However their method requires precise measurements of cosine in the worst case and therefore the complexity upper bound becomes loose. On the other hand, our proposed algorithm works with relatively rough estimates of cosine even in the worst case. Thus, as we will see in Section 2.3 and Appendix A, the complexity upper bound of our algorithm is tighter than existing method.

In the following discussion in this section, we first define the problem in Section 2.1. Next we look into the detail of the algorithm and finally we show the complexity upper bound of our algorithm.

2.1 Preliminary

The quantum amplitude estimation is the problem of estimating the value of 𝐚{\bf a} in the following equation:

|Ψ𝒜|0n|0=𝐚|Ψ~1n|1+1𝐚2|Ψ~0n|0,\displaystyle|\Psi\rangle\equiv\mathcal{A}|0\rangle_{n}|0\rangle={\bf a}|\tilde{\Psi}_{1}\rangle_{n}|1\rangle+\sqrt{1-{\bf a}^{2}}|\tilde{\Psi}_{0}\rangle_{n}|0\rangle, (1)

where 𝐚[0,1]{\bf a}\in[0,1]. In applications, it often takes cost to execute 𝒜\mathcal{A}, thus reducing the number of calling 𝒜\mathcal{A} while estimating 𝐚{\bf a} with required accuracy is the heart of the problem.

As we see later, our proposed algorithm works correctly if the amplitude is less than or equals to 1/41/4. However, imposing the condition on 𝐚{\bf a} is not necessary because the amplitude can be attenuated by adding an extra ancilla qubit as follows:

|ΨX|0n|00\displaystyle|\Psi^{\prime}\rangle\equiv X|0\rangle_{n}|00\rangle =𝐚4|Ψ~1n|11+15𝐚4|Ψ~1|10+1𝐚24|Ψ~0|01+15(1𝐚2)4|Ψ~0|00.\displaystyle=\frac{{\bf a}}{4}|\tilde{\Psi}_{1}\rangle_{n}|11\rangle+\frac{\sqrt{15}{{\bf a}}}{4}|\tilde{\Psi}_{1}\rangle|10\rangle+\frac{\sqrt{1-{{\bf a}}^{2}}}{4}|\tilde{\Psi}_{0}\rangle|01\rangle+\frac{\sqrt{15(1-{{\bf a}}^{2})}}{4}|\tilde{\Psi}_{0}\rangle|00\rangle.
=sinθ|Ψ~1n|11+cosθ|,\displaystyle=\sin\theta|\tilde{\Psi}_{1}\rangle_{n}|11\rangle+\cos\theta|{\perp}\rangle, (2)

where X=𝒜X=\cal{A}\otimes\cal{R} and \cal{R} operates as follows:

R|0=14|1+154|0.\displaystyle R|0\rangle=\frac{1}{4}|1\rangle+\frac{\sqrt{15}}{4}|0\rangle. (3)

In the last line of (2.1), we define sinθ𝐚/4\sin\theta\equiv{\bf a}/4 and ||\perp\rangle as a state orthogonal to |Ψ~1n|11|\tilde{\Psi}_{1}\rangle_{n}|11\rangle. As expected, the amplitude is attenuated as sinθ[0,1/4]\sin\theta\in[0,1/4] and therefore

0θ<0.252.\displaystyle 0\leq\theta<0.252. (4)

Thus, instead of estimating the value of 𝐚{\bf a} directly, we estimate the value of θ\theta and convert it to 𝐚{\bf a}. The condition (4) is utilized as the initial bound in our proposed algorithm.

Similar to the original amplitude amplification[11], we define an operator 𝐐{\bf Q} as

𝐐\displaystyle\mathbf{Q} X(In+22|0n+20|n+2)X(In+22In|1111|),\displaystyle\equiv X(I_{n+2}-2|0\rangle_{n+2}\langle 0|_{n+2})X^{{\dagger}}(I_{n+2}-2I_{n}\otimes|11\rangle\langle 11|), (5)

where InI_{n} is the identity operator in nn dimension. It is worth mentioning that

𝐐m|Ψ=sin(2m+1)θ|Ψ~1n|11+cos(2m+1)θ|.\displaystyle\mathbf{Q}^{m}|\Psi^{\prime}\rangle=\sin(2m+1)\theta|\tilde{\Psi}_{1}\rangle_{n}|11\rangle+\cos(2m+1)\theta|\perp\rangle. (6)

We get the estimates of cos(2(2m+1)θ)\cos(2(2m+1)\theta) by measuring the state (6) for multiple mm; the following defined cmc_{m} readily computable from the measurement result, is an estimate of cos(2(2m+1)θ)\cos(2(2m+1)\theta):

cm12N11Nshot,\displaystyle c_{m}\equiv 1-2\frac{N_{11}}{N_{\rm shot}}, (7)

where N11N_{11} is the number of the results of the measurements in which the last two qubits in (6) are both one and NshotN_{\rm shot} is the total number of measurements of the state (6). The estimation error of cmc_{m} can be evaluated by using the Chernoff bound for the Bernoulli distribution as discussed in [17], i.e., given the confidence interval of cmc_{m} as [cmmin,cmmax][c^{\rm min}_{m},c^{\rm max}_{m}], the bounds of the interval are computed as

cmmax=min[1,cm+ln(2δc)12Nshot],cmmin=max[1,cmln(2δc)12Nshot].\displaystyle c^{\max}_{m}=\min\left[1,c_{m}+\sqrt{\ln\left(\frac{2}{\delta_{c}}\right)\frac{12}{N_{\rm shot}}}\right],\qquad c^{\min}_{m}=\max\left[-1,c_{m}-\sqrt{\ln\left(\frac{2}{\delta_{c}}\right)\frac{12}{N_{\rm shot}}}\right]. (8)

where δc\delta_{c} is the probability that the true value of cmc_{m} (i.e. cos(2(2m+1)θ)\cos(2(2m+1)\theta)) is out of the interval. For later purpose, we define three functions: 𝐂𝐎𝐒(m,Nshot){\bf COS}(m,N_{\rm shot}), 𝐂𝐇𝐄𝐑𝐍𝐎𝐅𝐅(cm,Nshot,δc){\bf CHERNOFF}(c_{m},N_{\rm shot},\delta_{c}) and atan(s,c){\rm atan}(s,c). The function 𝐂𝐎𝐒(m,Nshot){\bf COS}(m,N_{\rm shot}) returns cmc_{m} as a result of NshotN_{\rm shot} times measurements of the state (6). The function 𝐂𝐇𝐄𝐑𝐍𝐎𝐅𝐅(cm,Nshot,δc){\bf CHERNOFF}(c_{m},N_{\rm shot},\delta_{c}) returns the confidence interval [cmmin,cmmax][c^{\rm min}_{m},c^{\rm max}_{m}] of cmc_{m} computed from the parameters: cmc_{m}, NshotN_{\rm shot} and δc\delta_{c}. The function atan(s,c){\rm atan}(s,c) is an extended arctangent function defined in the realm c,s[1,1]c,s\in[-1,1] as

atan(s,c)={arctan(s/c)(c>0)π/2(c=0,s>0)0(c=0,s=0)π/2(c=0,s<0)π+arctan(s/c)(c<0,s0)π+arctan(s/c)(c<0,s<0).\displaystyle{\rm atan}(s,c)=\begin{cases}\arctan(s/c)&({\rm}c>0)\\ \pi/2&(c=0,s>0)\\ 0&(c=0,s=0)\\ -\pi/2&(c=0,s<0)\\ \pi+\arctan(s/c)&({c<0,s\geq 0})\\ -\pi+\arctan(s/c)&({c<0,s<0}).\\ \end{cases} (9)

Finally, we define NoracN_{\rm orac} as the number of calls of 𝐐\mathbf{Q} required for estimating θ\theta. Our objective in this paper is providing an algorithm to estimate θ\theta with required accuracy while reducing the number of NoracN_{\rm orac}.

2.2 Proposed Algorithm

In this subsection, we show our proposed algorithm. Our procedure is shown in Algorithm 1111The source code of the algorithm is shown in https://github.com/quantum-algorithm/faster-amplitude-estimation. Given [θminj,θmaxj][\theta_{\min}^{j},\theta_{\max}^{j}] as the confidence interval of θ\theta in jj-th iteration, the algorithm updates the values of θmaxj\theta_{\max}^{j} and θminj\theta_{\min}^{j} so that θmaxjθminj\theta_{\max}^{j}-\theta_{\min}^{j} becomes smaller in each iteration. The total iteration count \ell is a parameter given by users of the algorithm and it is chosen so that the final result satisfies the required accuracy. As we see later, given acceptable error of the amplitude ϵ\epsilon, ϵ1/2\epsilon\sim 1/2^{\ell}. Therefore, it is suffice to take \ell as log2(1/ϵ)\ell\sim\log_{2}(1/\epsilon).

Algorithm 1 Faster Amplitude Estimation (δc\delta_{c} and \ell as the parameters)
1:#θminj\theta^{j}_{\min} and θmaxj\theta^{j}_{\max}: the confidence interval of θ\theta in jj-th iteration.
2:Set θmin0\theta^{0}_{\min} to 0 and θmax0\theta^{0}_{\max} to 0.2520.252.
3:Set Nshot1st=1944ln(2δc)N_{\rm shot}^{1st}=1944\ln\left(\frac{2}{\delta_{c}}\right) and Nshot2nd=972ln(2δc)N_{\rm shot}^{2nd}=972\ln\left(\frac{2}{\delta_{c}}\right).
4:Set FIRST_STAGE{\rm FIRST\_STAGE} to true{\rm true}.
5:Set j0j_{0} to \ell.
6:for j=1j=1 to \ell do
7:   if FIRST_STAGE then
8:      Set c2j1c_{2^{j-1}} to 𝐂𝐎𝐒(2j1,Nshot1st)\mathbf{COS}(2^{j-1},N_{\rm shot}^{\rm 1st}).
9:      Set c2j1min,c2j1maxc^{\min}_{2^{j-1}},c^{\max}_{2^{j-1}} to 𝐂𝐇𝐄𝐑𝐍𝐎𝐅𝐅(c2j1,Nshot1st,δc)\mathbf{CHERNOFF}(c_{2^{j-1}},N_{\rm shot}^{\rm 1st},\delta_{c}).
10:      Set θmaxj=arccos(c2j1min)/(2j+1+2)\theta_{\max}^{j}=\arccos(c^{\min}_{2^{j-1}})/(2^{j+1}+2) and θminj=arccos(c2j1max)/(2j+1+2)\theta_{\min}^{j}=\arccos(c^{\max}_{2^{j-1}})/(2^{j+1}+2).
11:      if 2j+1θmaxj3π82^{j+1}\theta_{\max}^{j}\geq\frac{3\pi}{8} and j<j<\ell then
12:         Set j0j_{0} to jj.
13:         Set ν=2j0(θmaxj0+θminj0)\nu=2^{j_{0}}(\theta_{\max}^{j_{0}}+\theta_{\min}^{j_{0}}) # the estimate of 2j0+1θ2^{j_{0}+1}\theta
14:         Set FIRST_STAGE{\rm FIRST\_STAGE} to false{\rm false}.
15:      end if
16:   else
17:      Set c2j1c_{2^{j-1}} to 𝐂𝐎𝐒(2j1,Nshot2nd)\mathbf{COS}(2^{j-1},N_{\rm shot}^{\rm 2nd}).
18:      Set s2j1s_{2^{j-1}} to (c2j1cosν𝐂𝐎𝐒(2j1+2j01,Nshot2nd))/sinν(c_{2^{j-1}}\cos\nu-\mathbf{COS}(2^{j-1}+2^{j_{0}-1},N_{\rm shot}^{\rm 2nd}))/\sin\nu.
19:      Set ρj=atan(s2j1,c2j1)\rho_{j}={\rm atan}\left(s_{2^{j-1}},c_{2^{j-1}}\right).
20:      Set njn_{j} to [12π((2j+1+2)θmaxj1ρj+π/3)][\frac{1}{2\pi}\left((2^{j+1}+2)\theta_{\max}^{j-1}-\rho_{j}+\pi/3\right)] where [x][x] is the largest integer which does not exceed xx.
21:      Set θminj=(2πnj+ρjπ/3)/(2j+1+2)\theta_{\min}^{j}=(2\pi n_{j}+\rho_{j}-\pi/3)/(2^{j+1}+2) and θmaxj=(2πnj+ρj+π/3)/(2j+1+2)\theta_{\max}^{j}=(2\pi n_{j}+\rho_{j}+\pi/3)/(2^{j+1}+2).
22:   end if
23:end for
24:(θmin+θmax)/2(\theta_{\min}^{\ell}+\theta_{\max}^{\ell})/2, estimate of θ\theta where the probability that θ[θminj,θmaxj]\theta\in[\theta_{\min}^{j},\theta_{\max}^{j}] is larger than 1(2j0)δc1-(2\ell-j_{0})\delta_{c}.

Even though θ\theta is not always inside the confidence interval: [θminj,θmaxj][\theta_{\min}^{j},\theta_{\max}^{j}], the probability is bounded and exponentially decreases as Nshot1stN_{\rm shot}^{1st} and Nshot2ndN_{\rm shot}^{2nd} increases. Thus, for simplicity, we discuss only the case when θ[θminj,θmaxj]\theta\in[\theta_{\min}^{j},\theta_{\max}^{j}] holds for all jjs in this subsection. As we show later, the probability that θ[θminj,θmaxj]\theta\in[\theta_{\min}^{j},\theta_{\max}^{j}] holds for all jj is larger than 12δc1-2\ell\delta_{c}.

In the following, we show how our algorithm works. As we see in Algorithm 1, there are two stages and the estimation methods are different in each stage. At the beginning of the iteration (j=1j=1), the algorithm is in the first stage and later the algorithm may change into the second stage if a condition is satisfied. We show the detail in the following.

First Stage

The algorithm is in the first stage when j=1j=1 or when j>1j>1 and all 2k+1θmaxk(k=1j1)2^{k+1}\theta_{\max}^{k}(k=1\dots j-1) satisfy 2k+1θmaxk<3π82^{k+1}\theta_{\max}^{k}<\frac{3\pi}{8}. In this stage, θminj,θmaxj\theta_{\min}^{j},\theta_{\max}^{j} is gotten by inverting c2j1minc^{\rm min}_{2^{j-1}} and c2j1maxc^{\rm max}_{2^{j-1}} as

θmaxj\displaystyle\theta_{\max}^{j} =arccos(c2j1min)2j+1+2,θminj=arccos(c2j1max)2j+1+2\displaystyle=\frac{\arccos(c^{\min}_{2^{j-1}})}{2^{j+1}+2},\qquad\theta_{\min}^{j}=\frac{\arccos(c^{\max}_{2^{j-1}})}{2^{j+1}+2} (10)

because (2j+1+2)θ<π(2^{j+1}+2)\theta<\pi is guaranteed as the following argument; if j=1j=1, the bound (4) leads to (21+1+2)θ<1.52<π(2^{1+1}+2)\theta<1.52<\pi, and if j>1j>1 and 2k+1θmaxk<3π82^{k+1}\theta_{\max}^{k}<\frac{3\pi}{8} for (k=1j1)(k=1\dots j-1) then

(2j+1+2)θ<2(2jθmaxj1)+2θ<3/4π+0.504<π.\displaystyle(2^{j+1}+2)\theta<2(2^{j}\theta_{\max}^{j-1})+2\theta<3/4\pi+0.504<\pi. (11)

The algorithm changes into the second stage if 2j+1θmaxj3π/82^{j+1}\theta_{\max}^{j}\geq 3\pi/8. The two values are memorized for the purpose of our utilizing them in the second stage. One is j0j_{0} defined as the last value of jj in the first stage. Another is ν\nu defined as

ν=2j0+1×θmaxj0+θminj02.\displaystyle\nu=2^{j_{0}+1}\times\frac{\theta_{\max}^{j_{0}}+\theta_{\min}^{j_{0}}}{2}. (12)

Note that above defined ν\nu is an estimate of 2j0+1θ2^{j_{0}+1}\theta and the confidence interval is obtainable from the Chernoff bound.

In case that 2j+1θmaxj2^{j+1}\theta_{\max}^{j} is less than 3π/83\pi/8 for all j(<){j}(<\ell), the algorithm finishes without going to the second stage and the final result is (θmax+θmin)/2(\theta_{\max}^{\ell}+\theta_{\min}^{\ell})/{2}. The value of j0j_{0} is set to \ell. In the case, the error of the final result is at most Δθ(θmaxθmin)/2=(arccos(c21min)arccos(c21max))/(2+2+4)\Delta\theta\equiv(\theta_{\max}^{\ell}-\theta_{\min}^{\ell})/{2}=(\arccos(c^{\min}_{2^{\ell-1}})-\arccos(c^{\max}_{2^{\ell-1}}))/(2^{\ell+2}+4). Thus, the error of the amplitude is bounded as

ϵ\displaystyle\epsilon =4(sin(θ+Δθ)sinθ)<4Δθ<arccos(c21min)arccos(c21max)2.\displaystyle=4\left(\sin(\theta+\Delta\theta)-\sin\theta\right)<4\Delta\theta<\frac{\arccos(c^{\min}_{2^{\ell-1}})-\arccos(c^{\max}_{2^{\ell-1}})}{2^{\ell}}. (13)

The probability that θ\theta is inside the confidence interval is (1δc)>1δc(=1(2j0)δc)(1-\delta_{c})^{\ell}>1-\ell\delta_{c}(=1-(2\ell-j_{0})\delta_{c}).

Second Stage

In the second stage, (2j+1+2)θ(2^{j+1}+2)\theta may be larger than π\pi. Thus, the value of (2j+1+2)θ(2^{j+1}+2)\theta can not be estimated by inverting c2j1c_{2^{j-1}}. However, it is still possible to estimate the value of (2j+1+2)θ(2^{j+1}+2)\theta by utilizing the results of measurements in other angle: (2j+1+2j0+1+2)θ(2^{j+1}+2^{j_{0}+1}+2)\theta, in addition to the bounds of θ\theta gotten in the previous iteration. Here, firstly we show how to estimate (2j+1+2)θ|mod2π(2^{j+1}+2)\theta|_{{\rm mod}2\pi}, next we show how to estimate (2j+1+2)θ(2^{j+1}+2)\theta without mod(2π){\rm mod}(2\pi) ambiguity.

(i)The estimate of (2j+1+2)θ|mod2π(2^{j+1}+2)\theta|_{{\rm mod}2\pi}

To estimate (2j+1+2)θ|mod2π(2^{j+1}+2)\theta|_{{\rm mod}2\pi}, not only the estimate of cos((2j+1+2)θ)\cos((2^{j+1}+2)\theta) (i.e., c2j1c_{2^{j-1}}) but also the estimate of sin((2j+1+2)θ)\sin((2^{j+1}+2)\theta) are necessary. The estimate of sin((2j+1+2)θ)\sin((2^{j+1}+2)\theta) is not directly obtainable from measurements but can be computed by the following procedure. From the trigonometric addition formula:

cos((2j+1+2j0+1+2)θ)=cos((2j+1+2)θ)cos(2j0+1θ)sin((2j+1+2)θ)sin(2j0+1θ),\displaystyle\cos((2^{j+1}+2^{j_{0}+1}+2)\theta)=\cos((2^{j+1}+2)\theta)\cos(2^{j_{0}+1}\theta)-\sin((2^{j+1}+2)\theta)\sin(2^{j_{0}+1}\theta), (14)

if sin(2j0+1θ)\sin(2^{j_{0}+1}\theta) is not zero,

sin((2j+1+2)θ)=cos((2j+1+2)θ)cos(2j0+1θ)cos((2j+1+2j0+1+2)θ)sin(2j0+1θ).\displaystyle\sin((2^{j+1}+2)\theta)=\frac{\cos((2^{j+1}+2)\theta)\cos(2^{j_{0}+1}\theta)-\cos((2^{j+1}+2^{j_{0}+1}+2)\theta)}{\sin(2^{j_{0}+1}\theta)}. (15)

Replacing cos((2j+1+2)θ\cos((2^{j+1}+2)\theta as c2j1c_{2^{j-1}}, 2j0+1θ2^{j_{0}+1}\theta as ν\nu and cos((2j+1+2j0+1+2)θ\cos((2^{j+1}+2^{j_{0}+1}+2)\theta as c2j1+2j0+1c_{2^{j-1}+2^{j_{0}+1}} in the right hand side of the above formula, we can define s2j1s_{2^{j-1}} as

s2j1=c2j1cosνc2j1+2j01sinν,\displaystyle s_{2^{j-1}}=\frac{c_{2^{j-1}}\cos\nu-c_{2^{j-1}+2^{j_{0}-1}}}{\sin\nu}, (16)

then s2j1s_{2^{j-1}} becomes the estimate of sin((2j+1+2)θ)\sin((2^{j+1}+2)\theta). The estimation error of s2j1s_{2^{j-1}} reflects the estimation errors of c2j1c_{2^{j-1}}, c2j1+2j01c_{2^{j-1}+2^{j_{0}-1}} and ν\nu, which is discussed in Appendix A. It is straightforward to get the estimate of (2j+1+2)θ|mod2π(2^{j+1}+2)\theta|_{{\rm mod}2\pi} from s2j1s_{2^{j-1}} and c2j1c_{2^{j-1}}; if we define ρj\rho_{j} [π,π]\in[-\pi,\pi] as

ρj=atan(s2j1,c2j1),\displaystyle\rho_{j}={\rm atan}\left(s_{2^{j-1}},c_{2^{j-1}}\right), (17)

then ρj\rho_{j} is an estimate of (2j+1+2)θ|mod2π(2^{j+1}+2)\theta|_{{\rm mod}2\pi}.

Refer to caption
Figure 1: The overview of the definition of Δρj\Delta\rho_{j}.

The confidence interval of ρj\rho_{j} can be derived from those of c2j1c_{2^{j-1}}, c2j1+2j01c_{2^{j-1}+2^{j_{0}-1}} and ν\nu as in the case of s2j1s_{2^{j-1}}. Note that there are two types of the confidence interval. One is the connected confidence interval, meaning that there is no discontinuities in the confidence interval, e.g., [π/3,π/4][-\pi/3,\pi/4]. Another is disconnected confidence interval, meaning that the confidence interval is separated to an interval containing π-\pi and an interval containing π\pi, e.g., [π,2π/3][-\pi,-2\pi/3] and [3π/4,π][3\pi/4,\pi], which is realized when the confidence interval of c2j1c_{2^{j-1}} contains 1-1 and that of s2j1s_{2^{j-1}} contains 0222If both the confidence intervals of c2j1c_{2^{j-1}} and s2j1s_{2^{j-1}} contain 0, the confidence interval of ρj\rho_{j} has discontinuity in ρj=π/2,π/2\rho_{j}=\pi/2,-\pi/2. However, as long as Nshot1stN_{\rm shot}^{1st} and Nshot2ndN_{\rm shot}^{2nd} takes the upper bound value derived in Appendix A, the estimation errors are bounded so that either the confidence interval of c2j1c_{2^{j-1}} or that of s2j1s_{2^{j-1}} does not contain 0 because (c2j1)2+(s2j1)21(c_{2^{j-1}})^{2}+(s_{2^{j-1}})^{2}\simeq 1 holds. Thus, we do not discuss the type of discontinuity in the following.. In the connected confidence interval case, given interval as [a,b][a,b], we define Δρj=max(ρja,bρj)\Delta\rho_{j}=\max(\rho_{j}-a,b-\rho_{j}). On the other hand, in the disconnected confidence interval case, given intervals as [π,c][-\pi,c] and [d,π][d,\pi], we define Δρj\Delta\rho_{j} as

Δρj={max(2π+ρjd,cρj)(ifρj[π,c])max(ρjd,2π+cρj)(ifρj[d,π]).\displaystyle\Delta\rho_{j}=\begin{cases}\max(2\pi+\rho_{j}-d,c-\rho_{j})&({\rm if\ }\rho_{j}\in[-\pi,c])\\ \max(\rho_{j}-d,2\pi+c-\rho_{j})&({\rm if\ }\rho_{j}\in[d,\pi])\end{cases}. (18)

The overview of the definition of Δρj\Delta\rho_{j} is shown in Figure 1. The above defined Δρj\Delta\rho_{j} can be interpreted as the estimation error of ρj\rho_{j} in a sense that

2πnj+ρjΔρj(2j+1+2)θ2πnj+ρj+Δρj\displaystyle 2\pi n_{j}+\rho_{j}-\Delta\rho_{j}\leq(2^{j+1}+2)\theta\leq 2\pi n_{j}+\rho_{j}+\Delta\rho_{j} (19)

holds with an unknown integer njn_{j} as long as the true value of ρj\rho_{j} (i.e. (2j+1+2)θ|mod2π(2^{j+1}+2)\theta|_{\rm\mod 2\pi}) is inside the confidence interval.

(ii)The estimate of (2j+1+2)θ(2^{j+1}+2)\theta

Now let us show how (2j+1+2)θ(2^{j+1}+2)\theta is estimated from ρj\rho_{j}. Using (19) and the inequality,

(2j+1+2)θminj1(2j+1+2)θ(2j+1+2)θmaxj1,\displaystyle(2^{j+1}+2)\theta^{j-1}_{\min}\leq(2^{j+1}+2)\theta\leq(2^{j+1}+2)\theta^{j-1}_{\max}, (20)

it can be shown that

(2j+1+2)θminj1ρjΔρj2πnj(2j+1+2)θmaxj1ρj+Δρj.\displaystyle(2^{j+1}+2)\theta^{j-1}_{\min}-\rho_{j}-\Delta\rho_{j}\leq 2\pi n_{j}\leq(2^{j+1}+2)\theta^{j-1}_{\max}-\rho_{j}+\Delta\rho_{j}. (21)

Thus, if

(2j+1+2)(θmaxj1θminj1)+2Δρj<2π\displaystyle(2^{j+1}+2)(\theta^{j-1}_{\max}-\theta^{j-1}_{\min})+2\Delta\rho_{j}<2\pi (22)

then njn_{j} can be uniquely determined as

nj=12π[(2j+1+2)θmaxj1ρj+Δρj]\displaystyle n_{j}=\frac{1}{2\pi}[(2^{j+1}+2)\theta_{\max}^{j-1}-\rho_{j}+\Delta\rho_{j}] (23)

where [x][x] is the largest integer that does not exceed xx. By using (20) and (26), it can be inductively shown that if all ρk(k=j0+1j1)\rho_{k}(k=j_{0}+1\dots j-1) are determined with the precision of Δρkπ/3\Delta\rho_{k}\leq\pi/3 then the condition (22) is satisfied.

Although (19) with njn_{j} in (26) gives the upper/lower bounds of (2j+1+2)θ(2^{j+1}+2)\theta, a complicated procedure is necessary for evaluating Δρj\Delta\rho_{j} in the algorithm. Thus, in our algorithm, instead of estimating Δρj\Delta\rho_{j}, we set the upper/lower bounds of θ\theta at the jj-th iteration as

θminj=2πnj+ρjπ/32j+1+2,θmaxj=2πnj+ρj+π/32j+1+2,\displaystyle\theta_{\min}^{j}=\frac{2\pi n_{j}+\rho_{j}-\pi/3}{2^{j+1}+2},\qquad\theta_{\max}^{j}=\frac{2\pi n_{j}+\rho_{j}+\pi/3}{2^{j+1}+2}, (24)

and

nj=12π[(2j+1+2)θmaxj1ρj+π/3],\displaystyle n_{j}=\frac{1}{2\pi}[(2^{j+1}+2)\theta_{\max}^{j-1}-\rho_{j}+\pi/3], (25)

which are correct as far as Δρjπ/3\Delta\rho_{j}\leq\pi/3. In Appendix A, we show that for all j(>j0)j(>j_{0}), Δρjπ/3\Delta\rho_{j}\leq\pi/3 holds and (19) is satisfied with the probability larger than 1(2j0)δc1-(2\ell-j_{0})\delta_{c} when at least

Nshot1st=1944ln(2δc),Nshot2nd=972ln(2δc).\displaystyle N_{\rm shot}^{1st}=1944\ln\left(\frac{2}{\delta_{c}}\right),\qquad N_{\rm shot}^{2nd}=972\ln\left(\frac{2}{\delta_{c}}\right). (26)

In the \ell-th iteration, the final result is set to (θmax+θmin)/2(\theta_{\max}^{\ell}+\theta_{\min}^{\ell})/2. Then, the error of the final result Δθ\Delta\theta is less than Δθ=(θmaxθmin)/2π/(32+1)\Delta\theta=(\theta_{\max}^{\ell}-\theta_{\min}^{\ell})/2\leq\pi/(3\cdot 2^{\ell+1}). Thus, the error of the amplitude is

ϵ\displaystyle\epsilon =4(sin(θ+Δθ)sinθ)<4Δθ<π321.\displaystyle=4\left(\sin(\theta+\Delta\theta)-\sin\theta\right)<4\Delta\theta<\frac{\pi}{3\cdot 2^{\ell-1}}. (27)

We show the overview of our algorithm when =5\ell=5 and j0=3j_{0}=3 in Figure 2.

Refer to caption
Figure 2: The overview of our algorithm when =5\ell=5 and j0=3j_{0}=3.

2.3 Complexity Upper Bound

As we show in Appendix, by using our proposed algorithm, the required query complexity (NoracN_{\rm orac}) with which the estimation error of 𝐚{\bf a} is less than ϵ\epsilon with the probability less than δ\delta is bounded as

Norac<4.1103ϵln(4log2(2π/3ϵ)δ).\displaystyle\qquad N_{\rm orac}<\frac{4.1\cdot 10^{3}}{\epsilon}\ln\left(\frac{4\log_{2}(2\pi/3\epsilon)}{\delta}\right). (28)

The worst case is realized when the algorithm moves to the second stage at the first iteration (when j=1j=1). We see that the upper bound of NoracN_{\rm orac} almost achieves Heisenberg scaling: (Norac1/ϵN_{\rm orac}\propto 1/\epsilon) because the dependency of the factor ln(log2(π/ϵ))\ln(\log_{2}(\pi/\epsilon)) on ϵ\epsilon is small, e.g., even when ϵ=1020\epsilon=10^{-20}, the factor is at most 66. The tightest upper bound in previous literature is give by [16] as Norac<1.15106ϵln(2δlog3(3π20ϵ))N_{\rm orac}<\frac{1.15\cdot 10^{6}}{\epsilon}\ln\left(\frac{2}{\delta}\log_{3}\left(\frac{3\pi}{20\epsilon}\right)\right) in our notation. We see that the constant factor is O(102)O(10^{2}) times smaller in our algorithm.

Although detail discussion is made in Appendix, here we briefly show why the upper bound is proportional to 1/ϵ1/\epsilon. In order for ϵ\epsilon to be bounded as (27), it is suffice that the errors of all c2j1c_{2^{j-1}}s used in our algorithm are less than 1/201/20, which is realized if NshotO(1000log(1/δ))N_{\rm shot}\sim O(1000\log\left(1/\delta\right)) measurements for each jj. The number of oracle call in each jj is about 2j12^{j-1} for each measurement. Thus, NoracNshotj=1j=2j1=Nshot2Nshot/ϵN_{\rm orac}\sim N_{\rm shot}\sum_{j=1}^{j=\ell}2^{j-1}=N_{\rm shot}2^{\ell}\propto N_{\rm shot}/\epsilon as we expected.

Refer to caption
Figure 3: Estimation error ϵ\epsilon vs NoracN_{\rm orac} for 𝐚=0.1{\bf a}=0.1(left top), 𝐚=0.2{\bf a}=0.2(right top), 𝐚=0.3{\bf a}=0.3(left bottom) and 𝐚=0.4{\bf a}=0.4(right bottom). The green dots are plotted so that the estimation errors in 1000 trials are equals to or smaller than the plotted value. The green dots are fitted with log10Norac=log10(ϵ)+b\log_{10}N_{\rm orac}=-\log_{10}(\epsilon)+b and shown as blue lines. The value of j0j_{0} is also shown for each data point.

3 Numerical Experiment

In this section, we verify the validity of the algorithm introduced in Section 2 by numerical experiments. We choose 𝐚=0.1,0.2,0.3,{\bf a}=0.1,0.2,0.3, and 0.40.4 as the amplitudes estimated. δc\delta_{c} is taken as 0.010.01. We compute NoracN_{\rm orac} and the estimation error ϵ\epsilon with changing the total number of algorithm steps \ell. In each parameter set (𝐚,)({\bf a},\ell), we execute 1000 trials of the algorithm.

The computation results are shown in Fig. 3. For each NoracN_{\rm orac}, we plot the estimation errors(green dots) so that 95%95\% of the estimation errors in 1000 trials are equals to or smaller than the plotted value. In the same figure, we also show j0j_{0}. For data points where the algorithm does not go to the second stage, we write “First Stage Only” instead of writing the value of j0j_{0}. The data points are fitted with log10(Norac)=log10(ϵ)+b\log_{10}(N_{\rm orac})=-\log_{10}(\epsilon)+b (blue lines) where the fitting parameter bb is determined by the least-squares.

Here is the list of notable points:

  • As expected, the Heisenberg scaling NoracC×1/ϵN_{\rm orac}\leq C\times 1/\epsilon is almost achieved.

  • In “First Stage Only” cases, ϵ\epsilon tends to be below the blue line, i.e., required NoracN_{\rm orac} is small for fixed 1/ϵ1/\epsilon compared with the case when the algorithm goes to the second stage. Because in “First Stage Only” cases, the cause of the error is limited; only cos(2j+1+2)θ\cos(2^{j+1}+2)\theta is needed to estimate 2j+1θ2^{j+1}\theta.

  • As 𝐚{\bf a} increases, j0j_{0} decreases as long as the algorithm goes to the second stage. Because, as 𝐚{\bf a} increases, 2j+1θmax3/8π2^{j+1}\theta_{\rm max}\geq 3/8\pi is satisfied with smaller jj.

4 Conclusion

The quantum amplitude estimation is an important problem that can be applied in various applications. Recently, the way of solving the problem without the phase estimation has been studied. Some of them suggest algorithms which achieve Heisenberg scaling (NoracC×1/ϵN_{\rm orac}\leq C\times 1/\epsilon) and they give rigorous proof. However the constant factor CC in each algorithm is large. Our contribution in this paper is providing an algorithm which almost achieves Heisenberg scaling and the constant factor is smaller than previous methods. We also give proof of the upper bound.

In a practical usage of the algorithm, some improvements might be possible. Although we determine the values of Nshot1stN_{\rm shot}^{1st} and Nshot2ndN_{\rm shot}^{2nd} at the beginning of the algorithm for simplicity, we can reduce those values by iteratively determining them. For example, in the second stage, Nshot2ndN_{\rm shot}^{2nd} can be smaller than that in (26) as long as njn_{j} in (21) is uniquely determined. Investigating those possible improvements are left for future works.

The effect of noise should also be examined. Although our algorithm can reduce the depth of the circuit compared to the quantum phase estimation algorithm, the required depth is still O(1/ϵ)O(1/\epsilon) and the effect of noise is not neglectable. Thus, studying how to tailor noise in our algorithm would be important for discussing the practicability of our algorithm, which is also left for future works.

Acknowledgement

We acknowledge Naoki Yamamoto for insightful discussions and constructive comments. We also thank the two anonymous reviewers whose comments/suggestions helped improve and clarify this manuscript.

References

  • [1] E. Knill, G. Ortiz, and R.D. Somma. Optimal Quantum Measurements of Expectation Values of Observables, 2006; arXiv:quant-ph/0607019. DOI: 10.1103/PhysRevA.75.012328.
  • [2] I. Kassal, S. P. Jordan, P. J. Love, M. Mohseni, and A. Aspuru-Guzik. Polynomial-time quantum algorithm for the simulation of chemical dynamics, 2008, Proc. Natl. Acad. Sci. 105, 18681(2008); arXiv:0801.2986. DOI: 10.1073/pnas.0808245105.
  • [3] P. Rebentrost, B. Gupt, and T. R. Bromley. Quantum computational finance: Monte Carlo pricing of financial derivatives, 2018, Phys. Rev. A 98, 022321 (2018); arXiv:1805.00109. DOI: 10.1103/PhysRevA.98.022321.
  • [4] A. Montanaro. Quantum speedup of Monte Carlo methods, 2015, Proc. Roy. Soc. Ser. A, vol. 471 no. 2181, 20150301, 2015; arXiv:1504.06987. DOI: 10.1098/rspa.2015.0301.
  • [5] S. Woerner and D. J. Egger. Quantum Risk Analysis, 2018, npj Quantum Inf 5, 15 (2019); arXiv:1806.06893. DOI: 10.1038/s41534-019-0130-6.
  • [6] K. Miyamoto and K. Shiohara. Reduction of Qubits in Quantum Algorithm for Monte Carlo Simulation by Pseudo-random Number Generator, 2019; arXiv:1911.12469.
  • [7] N. Wiebe, A. Kapoor, and K. Svore. Quantum Algorithms for Nearest-Neighbor Methods for Supervised and Unsupervised Learning, 2014, Quantum Information & Computation 15(3 & 4): 0318-0358 (2015); arXiv:1401.2142.
  • [8] N. Wiebe, A. Kapoor, and K. M. Svore. Quantum Deep Learning, 2014; Quantum Information & Computation 16, 541–587 (2016) arXiv:1412.3489.
  • [9] N. Wiebe, A. Kapoor, and K. M. Svore. Quantum Perceptron Models, 2016; Advances in Neural Information Processing Systems 29 3999-4007 (2016) arXiv:1602.04799.
  • [10] I. Kerenidis, J. Landman, A. Luongo, and A. Prakash. q-means: A quantum algorithm for unsupervised machine learning (2018); Advances in Neural Information Processing Systems 32 4136-4146 (2019) arXiv:1812.03584.
  • [11] G. Brassard, P. Hoyer, M. Mosca, and A. Tapp. Quantum Amplitude Amplification and Estimation, 2000, Quantum Computation and Quantum Information, Samuel J. Lomonaco, Jr. (editor), AMS Contemporary Mathematics, 305:53-74, 2002; arXiv:quant-ph/0005055.
  • [12] A. Y. Kitaev. Quantum measurements and the Abelian Stabilizer Problem, 1995; Electronic Colloquium on Computational Complexity 3 (1996) arXiv:quant-ph/9511026.
  • [13] Y. Suzuki, S. Uno, R. Raymond, T. Tanaka, T. Onodera, and N. Yamamoto. Amplitude Estimation without Phase Estimation, 2019; arXiv:1904.10246.
  • [14] C. R. Wie. Simpler Quantum Counting, 2019, Quantum Information & Computation, Vol.19, No.11&12, pp0967-0983, (2019); arXiv:1907.08119. DOI: 10.26421/QIC19.11-12.
  • [15] S. Aaronson and P. Rall. Quantum Approximate Counting, Simplified, 2019; arXiv:1908.10846.
  • [16] D. Grinko, J. Gacon, C. Zoufal, and S. Woerner. Iterative Quantum Amplitude Estimation, 2019; arXiv:1912.05559.
  • [17] A. Y. Kitaev, A. H. Shen, and M. N. Vyalyi. 2002. Classical and Quantum Computation. American Mathematical Society, Boston, MA, USA.
  • [18] K. M. Svore, M. B. Hastings, and M. Freedman. Faster Phase Estimation, 2013, Quant. Inf. Comp. Vol. 14, No. 3&4, pp. 306-328 (2013); arXiv:1304.0741.

Appendix A Proof of Complexity Upper Bound

In this appendix, we provide a proof of the complexity upper bound.

Theorem 1.

The following upper bound holds for NoracN_{\rm orac}:

Norac<4.1103ϵln(4log2(2π/3ϵ)δ).\displaystyle N_{\rm orac}<\frac{4.1\cdot 10^{3}}{\epsilon}\ln\left(\frac{4\log_{2}(2\pi/3\epsilon)}{\delta}\right). (29)
Proof.

Our strategy to obtain the upper bound is calculating the required number of Nshot1stN_{\rm shot}^{1st} and Nshot2ndN_{\rm shot}^{2nd} for the algorithm to work correctly with the probability 1δ1-\delta. Both upper bounds of Nshot1stN_{\rm shot}^{1st} and Nshot2ndN_{\rm shot}^{2nd} can be derived from the condition that our algorithm works correctly in the second stage because even though the condition from the first stage also bounds Nshot1stN_{\rm shot}^{1st} loosely, the most strict upper bound of Nshot1stN_{\rm shot}^{1st} can be gotten from the condition that the estimation error of ν\nu is small enough. Thus, in the following, we only discuss the condition from the second stage.

In the second stage, as we mention in Section 2.2, the algorithm works correctly as long as Δρjπ/3\Delta\rho_{j}\leq\pi/3. Even though the conditions for atan(s2j1,c2j1){\rm atan}\left(s_{2^{j-1}},c_{2^{j-1}}\right) derived from Δρjπ/3\Delta\rho_{j}\leq\pi/3 are different depending on whether the confidence interval of ρj\rho_{j} is the connected confidence interval or the disconnected confidence interval, the required precisions for s2j1s_{2^{j-1}} and c2j1c_{2^{j-1}} do not change depending on the interval type. Therefore, in the following, we discuss only the case of the connected confidence interval. Then, the condition Δρjπ/3\Delta\rho_{j}\leq\pi/3 can be converted to

|atan(s2j1,c2j1)atan(s2j1,c2j1)|π3\displaystyle\left|{\rm atan}\left(s_{2^{j-1}},c_{2^{j-1}}\right)-{\rm atan}\left(s^{\ast}_{2^{j-1}},c^{\ast}_{2^{j-1}}\right)\right|\leq\frac{\pi}{3} (30)

where s2j1,c2j1s^{\ast}_{2^{j-1}},c^{\ast}_{2^{j-1}} are the true values of s2j1s_{2^{j-1}} and c2j1c_{2^{j-1}} respectively. Given Δc2j1=|c2j1c2j1|\Delta c_{2^{j-1}}=|c_{2^{j-1}}-c^{\ast}_{2^{j-1}}|, Δs2j1=|s2j1s2j1|\Delta s_{2^{j-1}}=|s_{2^{j-1}}-s^{\ast}_{2^{j-1}}|, from (46) in Appendix B, the following inequality holds for the left hand side of (30):

|atan(s2j1,c2j1)atan(s2j1,c2j1)|<max(2Δc2j1+2Δs2j1,3Δc2j1)\displaystyle\left|{\rm atan}\left(s_{2^{j-1}},c_{2^{j-1}}\right)-{\rm atan}\left(s^{\ast}_{2^{j-1}},c^{\ast}_{2^{j-1}}\right)\right|<\max(2\Delta c_{2^{j-1}}+2\Delta s_{2^{j-1}},3\Delta c_{2^{j-1}}) (31)

as long as Δc2j1<1/4\Delta c_{2^{j-1}}<1/4 and Δs2j1<1/3\Delta s_{2^{j-1}}<1/3. On the other hand, from (16), it holds that

Δs2j1\displaystyle\Delta s_{2^{j-1}} =|s2j1(sinνsin(νΔν))+c2j1(cosνcos(νΔν))+Δc2j1cosν+Δc2j1+2j01sinν|\displaystyle=\left|\frac{-s^{\ast}_{2^{j-1}}(\sin\nu-\sin(\nu-\Delta\nu))+c^{\ast}_{2^{j-1}}(\cos\nu-\cos(\nu-\Delta\nu))+\Delta c_{2^{j-1}}\cos\nu+\Delta c_{2^{j-1}+2^{j_{0}-1}}}{\sin\nu}\right| (32)
22cos(Δν)+|Δc2j1cosν|+|Δc2j1+2j01|sinν\displaystyle\leq\frac{\sqrt{2-2\cos(\Delta\nu)}+|\Delta c_{2^{j-1}}\cos\nu|+|\Delta c_{2^{j-1}+2^{j_{0}-1}}|}{\sin\nu} (33)

where Δν=ν2j0+1θ\Delta\nu=\nu-2^{j_{0}+1}\theta and 3π/8|Δν|ν3π/4|Δν|3\pi/8-|\Delta\nu|\leq\nu\leq 3\pi/4-|\Delta\nu|. Thus, if at least the estimation errors are bounded as

Δc2j1\displaystyle\Delta c_{2^{j-1}} 19\displaystyle\leq\frac{1}{9} (34)
Δc2j1+2j01\displaystyle\quad\Delta c_{2^{j-1}+2^{j_{0}-1}} 19,\displaystyle\leq\frac{1}{9}, (35)
|Δν|\displaystyle|\Delta\nu| <π60\displaystyle<\frac{\pi}{60} (36)

then it holds

Δs2j1<22cos(π60)+19|cos(3π4π60)|+19sin(3π4π60)<13.\displaystyle\Delta s_{2^{j-1}}<\frac{\sqrt{2-2\cos(\frac{\pi}{60})}+\frac{1}{9}|\cos(\frac{3\pi}{4}-\frac{\pi}{60})|+\frac{1}{9}}{\sin(\frac{3\pi}{4}-\frac{\pi}{60})}<\frac{1}{3}. (37)

As a result,

|atan(s2j1,c2j1)atan(s2j1,c2j1)|<max(219+213,319)<π3\displaystyle\left|{\rm atan}\left(s_{2^{j-1}},c_{2^{j-1}}\right)-{\rm atan}\left(s^{\ast}_{2^{j-1}},c^{\ast}_{2^{j-1}}\right)\right|<\max(2\cdot\frac{1}{9}+2\cdot\frac{1}{3},3\cdot\frac{1}{9})<\frac{\pi}{3} (38)

is satisfied. Thus, by using (8), if

Nshot2nd=972ln(2δc)\displaystyle N_{\rm shot}^{2nd}=972\ln\left(\frac{2}{\delta_{c}}\right) (39)

then both the conditions (34) and (35) is satisfied with the probability 12δc1-2\delta_{c}. On the other hand, (36) is achieved if at least

Δc2j1<192,\displaystyle\Delta c_{2^{j-1}}<\frac{1}{9\sqrt{2}}, (40)

holds in the first stage because

Δν\displaystyle\Delta\nu =12(arccos(c2j01min)arccos(c2j01max))\displaystyle=\frac{1}{2}\left(\arccos\left(c^{\min}_{2^{j_{0}-1}}\right)-\arccos\left(c^{\max}_{2^{j_{0}}-1}\right)\right)
<12(arccos(cos(3π4))arccos(cos(3π4)+192))\displaystyle<\frac{1}{2}\left(\arccos\left(\cos\left(\frac{3\pi}{4}\right)\right)-\arccos\left(\cos\left(\frac{3\pi}{4}\right)+\frac{1}{9\sqrt{2}}\right)\right)
<π60.\displaystyle<\frac{\pi}{60}. (41)

Thus, by using (8) again, it is shown that if

Nshot1st=1944ln(2δc).\displaystyle N_{\rm shot}^{1st}=1944\ln\left(\frac{2}{\delta_{c}}\right). (42)

then (36) is satisfied with the probability 1δc1-\delta_{c}. In summary, as far as (39) and (42) are satisfied, for all j(>j0)j(>j_{0}), Δρjπ/3\Delta\rho_{j}\leq\pi/3 holds and (19) is satisfied as long as all the estimates of cosines are inside the confidence interval; the probability is (1δc)j0+2(j0)>1(2j0)δc(1-\delta_{c})^{j_{0}+2(\ell-j_{0})}>1-(2\ell-j_{0})\delta_{c}.

Finally, we evaluate the query complexity in the worst case. The worst case is that the algorithm moves to the second stage at the first iteration(j=1j=1). In the case, the number of oracle call is

Norac<Nshot1st+j=2(2Nshot2nd×2j1)=1944ln(2δc)+1944(22)ln(2δc).\displaystyle N_{\rm orac}<N_{\rm shot}^{1st}+\sum_{j=2}^{\ell}(2N_{\rm shot}^{2nd}\times 2^{j-1})=1944\ln\left(\frac{2}{\delta_{c}}\right)+1944(2^{\ell}-2)\ln\left(\frac{2}{\delta_{c}}\right). (43)

and the success probability of the algorithm is 1(21)δc1-(2\ell-1)\delta_{c}. Thus, if we demand that the success probability is more than 1δ1-\delta then δc<δ/2\delta_{c}<\delta/2\ell and

Norac<19442ln(4δ).\displaystyle N_{\rm orac}<1944\cdot 2^{\ell}\ln\left(\frac{4\ell}{\delta}\right). (44)

By combining with (27)

Norac<4.1103ϵln(4log2(2π/3ϵ)δ).\displaystyle N_{\rm orac}<\frac{4.1\cdot 10^{3}}{\epsilon}\ln\left(\frac{4\log_{2}(2\pi/3\epsilon)}{\delta}\right). (45)

Appendix B Theorem for atan function

Theorem 2.

When c,c,s[1,1]c,c^{*},s\in[-1,1], ss^{*} takes one of the value of ±1c2\pm\sqrt{1-c^{\ast 2}}, Δc=|cc|\Delta c=|c-c^{\ast}| and Δs=|ss|\Delta s=|s-s^{\ast}|, the following inequality holds:

|atan(s,c)atan(s,c)|<max(2Δc+2Δs,3Δc)\displaystyle|{\rm atan}(s,c)-{\rm atan}(s^{\ast},c^{\ast})|<{\rm max}(2\Delta c+2\Delta s,3\Delta c) (46)

if Δs<1/2\Delta s<1/2 and Δc<1/4\Delta c<1/4 and if there is no discontinuity of atan(s,c){\rm atan}(s,c) in the intervals: sΔsss+Δss^{\ast}-\Delta s\leq s\leq s^{\ast}+\Delta s and cΔccc+Δcc^{\ast}-\Delta c\leq c\leq c^{\ast}+\Delta c.

Proof.

It is suffice to prove in following three cases: (i) cc>0cc^{\ast}>0 (ii) cc<0cc^{\ast}<0 and (iii) cc=0cc^{\ast}=0. In case (i) cc>0cc^{\ast}>0, using trigonometric addition formulas for arctan\arctan, it holds that

|atan(s,c)atan(s,c)|\displaystyle|{\rm atan}(s,c)-{\rm atan}(s^{\ast},c^{\ast})| =|arctan(s,c)arctan(s,c)|\displaystyle=|{\rm arctan}(s,c)-{\rm arctan}(s^{\ast},c^{\ast})|
=|arctan(sΔccΔs1+cΔc+sΔs)|\displaystyle=\left|\arctan\left(\frac{s^{\ast}\Delta c-c^{\ast}\Delta s}{1+c^{\ast}\Delta c+s^{\ast}\Delta s}\right)\right|
||s|Δc+|c|Δs1|c|Δc|s|Δs|\displaystyle\leq\left|\frac{|s^{\ast}|\Delta c+|c^{\ast}|\Delta s}{1-|c^{\ast}|\Delta c-|s^{\ast}|\Delta s}\right|
<2Δc+2Δs.\displaystyle<2\Delta c+2\Delta s. (47)

To show the last inequality, we use 1|c|Δc|s|Δs>1(1/3)2+(1/4)2>1/21-|c^{\ast}|\Delta c-|s^{\ast}|\Delta s>1-\sqrt{(1/3)^{2}+(1/4)^{2}}>1/2 .

In case (ii) cc<0cc^{\ast}<0,

|atan(s,c)atan(s,c)|\displaystyle|{\rm atan}(s,c)-{\rm atan}(s^{\ast},c^{\ast})| =limη0(|arctan(s,c)arctan(s,η)|\displaystyle=\lim_{\eta\to 0}\left(\left|\arctan(s,c)-\arctan(s,\eta)\right|\right.
+|arctan(s,η)arctan(s,η)|)\displaystyle\qquad+\left.\left|\arctan(s,\eta)-\arctan(s^{\ast},-\eta)\right|\right)
+|arctan(s,η)arctan(s,c)|)\displaystyle\qquad+\left.\left|\arctan(s^{\ast},-\eta)-\arctan(s^{\ast},c^{\ast})\right|\right) (48)

where the sign of η\eta is same as that of cc. The first term in (B) can be bounded as

limη0|arctan(s,c)arctan(s,η)|\displaystyle\lim_{\eta\to 0}\left|\arctan(s,c)-\arctan(s,\eta)\right| =limη0|carctan(sc)|c=c0(cη)|\displaystyle=\lim_{\eta\to 0}\left|\frac{\partial}{\partial c}\arctan\left(\frac{s}{c}\right)|_{c=c_{0}}(c-\eta)\right|
=limη0|sc02+s2(cη)|\displaystyle=\lim_{\eta\to 0}\left|\frac{-s}{c_{0}^{2}+s^{2}}(c-\eta)\right|
|sc02+s2c|\displaystyle\leq\left|\frac{s}{c_{0}^{2}+s^{2}}c\right|
|1(c(cc0))2+(s(ss))2c|\displaystyle\leq\left|\frac{1}{(c_{\ast}-(c_{\ast}-c_{0}))^{2}+(s_{\ast}-(s_{\ast}-s))^{2}}c\right|
|1(3514)2+(4513)2c|\displaystyle\leq\left|\frac{1}{\left(\frac{3}{5}-\frac{1}{4}\right)^{2}+\left(\frac{4}{5}-\frac{1}{3}\right)^{2}}c\right|
<3|c|\displaystyle<3|c| (49)

where c0c_{0} take the value between η\eta and cc, and we use the mean value theorem for showing the first equality. Similary,

limη0|arctan(s,η)arctan(s,c)|\displaystyle\lim_{\eta\to 0}\left|\arctan(s_{\ast},-\eta)-\arctan(s_{\ast},c_{\ast})\right| <3|c|.\displaystyle<3|c_{\ast}|. (50)

By substituting (B), (50) and limη0|arctan(s,η)arctan(s,η)|=0\lim_{\eta\to 0}\left|\arctan(s,\eta)-\arctan(s^{\ast},-\eta)\right|=0 (that follows from no-discontinuity condition) to the right-hand side of (B), it follows

|atan(s,c)atan(s,c)|<3(|c|+|c|)=3Δc.\displaystyle|{\rm atan}(s,c)-{\rm atan}(s^{\ast},c^{\ast})|<3(|c|+|c_{\ast}|)=3\Delta c. (51)

The last equality holds because the signs of cc and cc_{\ast} are different.

In case (iii) cc=0cc_{\ast}=0, when c=0c_{\ast}=0,

|atan(s,c)atan(s,c)|\displaystyle|{\rm atan}(s,c)-{\rm atan}(s^{\ast},c^{\ast})| =limη0(|±π2arctan(s,η)|+|arctan(s,η)arctan(s,η)|\displaystyle=\lim_{\eta\to 0}\left(\left|\pm\frac{\pi}{2}-\arctan(s_{\ast},\eta)\right|+\left|\arctan(s_{\ast},\eta)-\arctan(s,\eta)\right|\right.
+|arctan(s,η)arctan(s,c)|)\displaystyle\qquad\qquad+\left.\left|\arctan(s,\eta)-\arctan(s,c)\right|\right) (52)

where the sign ±\pm is the same as the sign of ss and the sign of η\eta is same as that of cc. The values of the first line go to 0 and the value of the second line can be evaluated by the same arguments as (B). Thus, it follows

|atan(s,c)atan(s,c)|<3(|c|)=3Δc.\displaystyle|{\rm atan}(s,c)-{\rm atan}(s^{\ast},c^{\ast})|<3(|c|)=3\Delta c. (53)

By the same discussion, when c=0c=0

|atan(s,c)atan(s,c)|<3(|c|)=3Δc.\displaystyle|{\rm atan}(s,c)-{\rm atan}(s^{\ast},c^{\ast})|<3(|c_{\ast}|)=3\Delta c. (54)

In all of the three cases, (46) is proved. ∎