This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Computational Complexity of Learning Efficiently Generatable Pure States

Taiga Hiroka Yukawa Institute for Theoretical Physics, Kyoto University, Kyoto, Japan
[email protected]
Min-Hsiu Hsieh Hon Hai Research Institute, Taipei, Taiwan
[email protected]
Abstract

Understanding the computational complexity of learning efficient classical programs in various learning models has been a fundamental and important question in classical computational learning theory. In this work, we study the computational complexity of quantum state learning, which can be seen as a quantum generalization of distributional learning introduced by Kearns et.al [STOC94]. Previous works by Chung and Lin [TQC21], and Bădescu and O’Donnell [STOC21] study the sample complexity of the quantum state learning and show that polynomial copies are sufficient if unknown quantum states are promised efficiently generatable. However, their algorithms are inefficient, and the computational complexity of this learning problem remains unresolved.

In this work, we study the computational complexity of quantum state learning when the states are promised to be efficiently generatable. We show that if unknown quantum states are promised to be pure states and efficiently generateable, then there exists a quantum polynomial time algorithm 𝒜\mathcal{A} and a language 𝐏𝐏\mathcal{L}\in\mathbf{PP} such that 𝒜\mathcal{A}^{\mathcal{L}} can learn its classical description. We also observe the connection between the hardness of learning quantum states and quantum cryptography. We show that the existence of one-way state generators with pure state outputs is equivalent to the average-case hardness of learning pure states. Additionally, we show that the existence of EFI implies the average-case hardness of learning mixed states.

1 Introduction

Since Valiant [Val84] formalized the notion of PAC learning, understanding the computational complexity of learning efficient classical programs in various learning models has been a fundamental and important question in classical computational learning theory. It is shown that we can learn arbitrary efficient classical programs in the PAC learning model if we can solve an arbitrary problem in 𝐍𝐏\mathbf{NP} [BEHW87, BP92, Sch90]. The 𝐍𝐏\mathbf{NP} hardness of learning efficient classical programs in the proper PAC learning model is also shown [PV88]. While showing the 𝐍𝐏\mathbf{NP} hardness of improper PAC learning has been an open question, several works [DLSS14, DSS16, Vad17, DV21] show its hardness assuming the specific average-case hardness assumptions of 𝐍𝐏\mathbf{NP} (See also [HN22, Hir22]). Understanding the computational complexity of learning efficient classical programs is also crucial for constructing classical cryptographic primitives. In fact, it is known that the existence of fundamental classical cryptographic primitives such as one-way functions (OWFs) is equivalent to the average-case hardness of learning efficient classical programs in the various learning models [IL90, BFKL94, NR06, OS17, NE19, HN23]. Furthermore, it has been shown that assuming the hardness of specific classical learning problems, such as the Learning with Errors (LWE) problem, various useful classical cryptographic protocols can be constructed [Reg09]. Given the importance of understanding the computational complexity of learning efficient classical programs, it is an interesting direction to study computational complexity of learning efficient quantum processes in the various learning models.

In this work, we study the computational complexity of quantum state learning, which can be seen as a quantum generalization of distributional learning introduced by [KMR+94]. In quantum state learning, a learner is given many copies of an unknown quantum state which is promised within some family of quantum states, and the task is to find a hypothetical quantum circuit hh whose output is close to the given unknown quantum state. When there is no promise on the input states, it is known that an exponential number of copies relative to the number of qubits is required [OW16, HHJ+16]. On the other hand, because many quantum states of interest can be efficiently generated, it is natural to ask the sample and computational complexities of quantum state learning with the promise that an unknown quantum state is efficiently generatable. In previous works, Chung and Lin [CL21], and Bǎdescu and O’Donell [BO21] independently showed that polynomial copies are sufficient when an unknown quantum state is promised to be efficiently generatable. However, their algorithms are not time-efficient, and it is natural to ask the computational complexity of the learning task.

Currently, it is known that if there exist pseudorandom state generators (PRSGs) [JLS18], then there exist pure states that can be generated by quantum polynomial-time algorithms, but hard to learn its description [ZLK+23, CLS24]. This implies that it seems difficult to learn arbitrary quantum states generated by quantum polynomial-time algorithms even if we can query 𝐐𝐌𝐀\mathbf{QMA} oracle because it is known that PRSGs exist relative to an oracle 𝒪\mathcal{O} such that 𝐁𝐐𝐏𝒪=𝐐𝐌𝐀𝒪\mathbf{BQP}^{\mathcal{O}}=\mathbf{QMA}^{\mathcal{O}} [Kre21]. On the other hand, we do not know whether a QPT algorithm can learn quantum states generated by quantum polynomial-time algorithms even if the algorithm can query 𝐏𝐒𝐀𝐏𝐂𝐄\mathbf{PSAPCE} oracle. In particular, in quantum state learning, instances are themselves quantum states, and hence it is unclear how to reduce quantum state learning to traditional classical languages, where instances are represented by classical strings. Therefore, we believe that the following is an interesting open question:

Which oracle does help us to learn quantum states generated by an arbitrary quantum polynomial-time algorithm?

We also ask connection between quantum state learning and quantum cryptography. In the classical case, there exists an equivalence between the average-case hardness of learning in the various learning models and the existence of cryptography [IL90, BFKL94, NR06, OS17, NE19, HN23]. For example, it is known that the average-case hardness of distirutional learning is equivalent to the existence of OWFs [HN23]. It is natural to wonder whether a similar relationship holds in the quantum case. Therefore, we ask

Is the average-case hardness of learning quantum states equivalent to the existence of quantum cryptographic primitives?

1.1 Our Results

In this work, we study two questions above. Our contributions are summarized as follows.

Types of Problem Computational Complexity
Upper Bound Lower Bound
Learning Pure States generated in Polynomial Time 𝐏𝐏\mathbf{PP} (Theorem 4.1) OWSGs with Pure State Outputs
Average-Case Learning Pure States generated in Polynomial Time OWSGs with Pure State Outputs (Theorem 5.1) OWSGs with Pure State Outputs (Theorem 5.1)
Learning Mixed State generated in Polynomial Time N/A EFI (Theorem 5.3)
Table 1: Computational Complexities in Quantum State Learning
  • In column "Computational Complexity", the upper bound means that the learning tasks can be achieved with the listed class of algorithms, and the lower bound means that the listed class of algorithms is necessary to achieve the learning task. 𝐏𝐏\mathbf{PP} stands for QPT algorithm with PP oracle, OWSGs with pure state outputs stand for QPT algorithm which can query an algorithm that breaks any OWSGs with pure state outputs, and EFI stands for QPT algorithm which can query an algorithm that breaks any EFI.

  1. 1.

    In Section 3, we formally define a quantum state learning model, which can be seen as a quantum generalization of the distributional learning model [KMR+94]. The distributional learning model can be categorized into two flavors: proper distributional learning and improper distributional learning. In this work, we focus on the proper one and generalize it to a quantum setting. In our proper quantum state learning model, as a setup, a learner 𝒜\mathcal{A} is given a description of quantum polynomial-time algorithm 𝒞\mathcal{C}, which takes as input 1n1^{n} and x𝒳n={0,1}poly(n)x\in\mathcal{X}_{n}=\{0,1\}^{{\mathrm{poly}}(n)} and outputs a quantum state ψx\psi_{x}. Then, it receives polynomially many copies of ψx\psi_{x} generated by 𝒞\mathcal{C}. We say that the leaner 𝒜\mathcal{A} can properly learn 𝒞\mathcal{C} if, for all x𝒳nx\in\mathcal{X}_{n}, ϵ\epsilon and δ\delta, it can find y𝒳ny\in\mathcal{X}_{n} such that 𝖳𝖣(ψx,ψy)1/ϵ\mathsf{TD}(\psi_{x},\psi_{y})\leq 1/\epsilon with probability at least 11/δ1-1/\delta, where ψy𝒞(1n,y)\psi_{y}\leftarrow\mathcal{C}(1^{n},y)111 The other possible learning model is improper quantum state learning. In the improper quantum state learning model, a learner’s task is to find an arbitrary quantum algorithm hh whose output is close to |ψn\ket{\psi_{n}} instead of finding y𝒳ny\in\mathcal{X}_{n} such that |ψy\ket{\psi_{y}} is close to |ψx\ket{\psi_{x}}. Remark that if we can properly learn 𝒞\mathcal{C}, then we can also improperly learn 𝒞\mathcal{C}. On the other hand, the other direction may not hold. .

  2. 2.

    In Section 4, we show that, for all quantum polynomial-time algorithms 𝒞\mathcal{C} with pure state outputs, there exists a quantum polynomial-time algorithm 𝒜\mathcal{A} and a language 𝐏𝐏\mathcal{L}\in\mathbf{PP} such that 𝒜\mathcal{A}^{\mathcal{L}} can properly learn 𝒞\mathcal{C} given poly(n,log(|𝒳n|),ϵ,log(δ)){\mathrm{poly}}(n,\log(\absolutevalue{\mathcal{X}_{n}}),\epsilon,\log(\delta))-copies of |ψx\ket{\psi_{x}} with unknown xx 222Our QPT algorithm with 𝐏𝐏\mathbf{PP} oracle can be applied to improper learning.. This implies that, if a novel algorithm is found to solve any 𝐏𝐏\mathbf{PP} problems, then it will be useful to learn arbitrary pure states generated by any QPT algorithms. Let us stress that our learning algorithm with 𝐏𝐏\mathbf{PP} oracle cannot be applied to learning mixed state generated by any QPT algorithms.

  3. 3.

    In Section 5, we study the connection between quantum state learning and quantum cryptography. We observe that the average-case hardness of learning pure states in quantum state learning model is equivalent to the existence of one-way state generators (OWSGs) with pure state outputs [MY22b]. In OWSGs with pure state outputs, the adversary receives many copies of |ψx\ket{\psi_{x}}, where xx is randomly sampled and |ψx\ket{\psi_{x}} is prepared by some QPT algorithm 𝒞(x)\mathcal{C}(x). What the adversary needs to do is to find hh such that the measurement {|ψhψh|,I|ψhψh|}\{\ket{\psi_{h}}\bra{\psi_{h}},I-\ket{\psi_{h}}\bra{\psi_{h}}\} on |ψx\ket{\psi_{x}} yields |ψhψh|\ket{\psi_{h}}\bra{\psi_{h}} with high probability. This is very similar to what the learner needs to do in our learning model. We observe that we can break arbitrary OWSGs with pure state outputs if and only if we can learn pure states in the average-case version of our learning model. Intuitively, in average-case learning, the learner needs to find hh such that |ψh\ket{\psi_{h}} is close to |ψx\ket{\psi_{x}} for a large fraction of 𝒳n\mathcal{X}_{n} instead of for all x𝒳nx\in\mathcal{X}_{n}. Therefore, the learning requirement is relaxed compared to worst-case learning.

    We also observe that the existence of EFI [Yan22, BCQ23] implies that there are mixed states generated by quantum polynomial-time algorithms, which is hard to properly learn. Because [LMW24] conjectured that EFI may exist relative to a random oracle and any classical oracle, our observation implies that any classical oracles seem not help to properly learn arbitrary mixed states generated by quantum polynomial-time algorithms 333Remark that the argument cannot be applied to the hardness of improperly mixed state learning, and thus there exists the possibility that we can improperly learn mixed states by querying classical oracles.. Therefore, this implies that QPT algorithms with classical oracles do not seem applicable to learning any efficiently generatable mixed states.

The upper and lower bounds of the computational complexity of the quantum state learning problem are summarized in the Table 1.

1.2 Technical Overview

In this section, we explain how a quantum polynomial-time algorithm with 𝐏𝐏\mathbf{PP}-oracle can learn pure states generated by a quantum polynomial-time algorithm.

Our Learning Algorithm.

Our technical contribution is, given 𝐏𝐏\mathbf{PP} oracle, showing how to efficiently implement Chung and Lin’s algorithm (Given in Section4 in [CL21]). First, let us briefly recall their strategy in our language.

Description of CL algorithm 𝒜\mathcal{A}:

  1. 1.

    As a setup, 𝒜\mathcal{A} receives a quantum algorithm 𝒞\mathcal{C} that takes x𝒳n={0,1}poly(n)x\in\mathcal{X}_{n}=\{0,1\}^{{\mathrm{poly}}(n)} and outputs |ψx\ket{\psi_{x}}.

  2. 2.

    𝒜\mathcal{A} receives {|ψxi}i[T]\{\ket{\psi_{x}}_{i}\}_{i\in[T]} with sufficiently large TT and an unknown x𝒳nx\in\mathcal{X}_{n}.

  3. 3.

    For all i[T]i\in[T], 𝒜\mathcal{A} measures |ψxi\ket{\psi_{x}}_{i} with Haar random POVM measurements \mathcal{M}, and obtains ziz_{i}. Note that this process is possibly inefficient. Let 𝒞()\mathcal{C}^{*}(\cdot) be a quantum algorithm that takes xx, generates |ψx\ket{\psi_{x}}, measure it with \mathcal{M}, and outputs the measurement outcome.

  4. 4.

    𝒜\mathcal{A} outputs h𝒳nh\in\mathcal{X}_{n} such that

    maxx𝒳n{Πi[T]Pr[zi𝒞(x)]}=Πi[T]Pr[zi𝒞(h)].\displaystyle\max_{x\in\mathcal{X}_{n}}\{\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}(x)]\}=\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}(h)]. (1)

    Note that this step is possibly inefficient.

Let us explain an intuitive reason why this algorithm works. First, it is known that Haar random measurements do not decrease the trace distance between quantum states if measured states are pure states [Sen06]. From this property, the measurements in the third step of 𝒜\mathcal{A} does not decrease the trace distance between quantum states. In other words, there is a constant cc such that, for all a,b𝒳na,b\in\mathcal{X}_{n},

𝒞(a)𝒞(b)1c|ψa|ψb1.\displaystyle\norm{\mathcal{C}^{*}(a)-\mathcal{C}^{*}(b)}_{1}\geq c\norm{\ket{\psi_{a}}-\ket{\psi_{b}}}_{1}. (2)

This implies that, for finding hh such that |ψh|ψx1\norm{\ket{\psi_{h}}-\ket{\psi_{x}}}_{1} is small, it is sufficient to find a h𝒳nh\in\mathcal{X}_{n}, such that 𝒞(h)𝒞(x)1\norm{\mathcal{C}^{*}(h)-\mathcal{C}^{*}(x)}_{1} is small, given polynomially many copies of z𝒞(x)z\leftarrow\mathcal{C}^{*}(x). In the fourth step, 𝒜\mathcal{A} finds a maximum-likelihood h𝒳nh\in\mathcal{X}_{n} such that 𝒞(h)\mathcal{C}^{*}(h) is most likely to output {zi}i[T]\{z_{i}\}_{i\in[T]}. It is natural to expect that a maximum-likelihood algorithm 𝒞(h)\mathcal{C}^{*}(h) seems to approximate 𝒞(x)\mathcal{C}^{*}(x) well with high probability. [CL21] shows that if TO(ϵ2(log(|𝒳n|)+log(δ)))T\geq O(\epsilon^{2}\cdot(\log(\absolutevalue{\mathcal{X}_{n}})+\log(\delta))), then we have 𝒞(h)𝒞(x)11/ϵ\norm{\mathcal{C}^{*}(h)-\mathcal{C}^{*}(x)}_{1}\leq 1/\epsilon with probability 11/δ1-1/\delta, where the probability is taken over {zi}i[T]𝒞(x)\{z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}^{*}(x).

Although their algorithm learns pure states, their algorithm does not work efficiently. The first bottleneck is the third step of 𝒜\mathcal{A}. In the third step, 𝒜\mathcal{A} needs Haar random measurements, which is hard to implement. Fortunately, it is shown that we have the same property if we implement 44-design measurements instead of Haar random measurements [AE07]. Therefore, we can efficiently implement the third step by running 44-design measurements instead of Haar random measurements. The second bottleneck is the fourth step of 𝒜\mathcal{A}. In the fourth step, given {zi}i[T]\{z_{i}\}_{i\in[T]}, 𝒜\mathcal{A} needs to find a maximum-likelihood h𝒳nh\in\mathcal{X}_{n} such that

maxx𝒳n{Πi[T]Pr[zi𝒞(x)]}=Πi[T]Pr[zi𝒞(h)],\displaystyle\max_{x\in\mathcal{X}_{n}}\{\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}(x)]\}=\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}(h)], (3)

which we currently do not know how to implement efficiently. In our work, we show that we can achieve an approximated version of the task by querying 𝐏𝐏\mathbf{PP} oracle. More precisely, we show that for arbitrary quantum polynomial-time algorithms 𝒞\mathcal{C}^{*} with classical inputs and classical outputs, there exists a language 𝐏𝐏\mathcal{L}\in\mathbf{PP} and an oracle-aided PPT algorithm 𝒜\mathcal{A}^{\mathcal{L}} that can find an approximated maximum-likelihood h𝒳nh\in\mathcal{X}_{n} such that

maxx𝒳n{Πi[T]Pr[zi𝒞(x)]}(1+1/ϵ(n))Πi[T]Pr[zi𝒞(h)]maxx𝒳n{Πi[T]Pr[zi𝒞(x)]},\displaystyle\frac{\max_{x\in\mathcal{X}_{n}}\left\{\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}(x)]\right\}}{\left(1+1/\epsilon(n)\right)}\leq\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}(h)]\leq\max_{x\in\mathcal{X}_{n}}\left\{\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}(x)]\right\}, (4)

where ϵ\epsilon is arbitrary polynomial.

A subtle but important point is that an “approximated” maximum-likelihood algorithm 𝒞(h)\mathcal{C}^{*}(h) might not statistically approximate 𝒞(x)\mathcal{C}^{*}(x) even if we take TT sufficiently large. Actually, we do not know how to prove an approximated maximum-likelihood algorithm 𝒞(h)\mathcal{C}^{*}(h) approximates 𝒞(x)\mathcal{C}^{*}(x) for sufficiently large TT. To overcome the issue, inspired by [AW92], we slightly modify the learning algorithm 𝒜\mathcal{A} as follows. In a modified learning algorithm \mathcal{B}, instead of 𝒞()\mathcal{C}^{*}(\cdot), \mathcal{B} considers a quantum algorithm 𝒞1/2()\mathcal{C}^{*}_{1/2}(\cdot), where 𝒞1/2()\mathcal{C}^{*}_{1/2}(\cdot) takes as input x𝒳nx\in\mathcal{X}_{n}, and outputs the output of 𝒞(x)\mathcal{C}^{*}(x) with probability 1/21/2, and outputs a uniformly random string zz with probability 1/21/2. Given many copies of |ψxi\ket{\psi_{x}}_{i}, \mathcal{B} measures |ψxi\ket{\psi_{x}}_{i} with Haar random measurement and sets the measurement outcome as ziz_{i} with probability 1/21/2, and uniform random string as ziz_{i} with probability 1/21/2. \mathcal{B} outputs hh such that

maxx𝒳n{Πi[T]Pr[zi𝒞1/2(x)]}(1+1/ϵ(n))Πi[T]Pr[zi𝒞1/2(h)]maxx𝒳n{Πi[T]Pr[zi𝒞1/2(x)]}\displaystyle\frac{\max_{x\in\mathcal{X}_{n}}\left\{\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(x)]\right\}}{\left(1+1/\epsilon(n)\right)}\leq\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(h)]\leq\max_{x\in\mathcal{X}_{n}}\left\{\Pi_{i\in[T]}\Pr[z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(x)]\right\} (5)

This modification makes it possible to prove that for sufficiently large TT, 𝒞1/2(x)𝒞1/2(h)1\norm{\mathcal{C}^{*}_{1/2}(x)-\mathcal{C}^{*}_{1/2}(h)}_{1} is small with high probability. Furthermore, we have

𝒞1/2(x)𝒞1/2(h)1=12𝒞(x)𝒞(h)1c|ψx|ψh1.\displaystyle\norm{\mathcal{C}^{*}_{1/2}(x)-\mathcal{C}^{*}_{1/2}(h)}_{1}=\frac{1}{2}\norm{\mathcal{C}^{*}(x)-\mathcal{C}^{*}(h)}_{1}\geq c\norm{\ket{\psi_{x}}-\ket{\psi_{h}}}_{1}. (6)

This means that |ψh\ket{\psi_{h}} statistically approximates |ψx\ket{\psi_{x}}. Therefore, the remaining part is showing how to find an approximated maximum-likelihood algorithm 𝒞(h)\mathcal{C}^{*}(h) given 𝐏𝐏\mathbf{PP} oracle.

Probabilistic Polynomial-Time Algorithm with PP oracle for Quantum Maximum-Likelihood Problem.

In the rest of the section, we explain how to solve the following quantum maximum-likelihood problem using 𝐏𝐏\mathbf{PP} oracle.

Quantum Maximum-Likelihood Problem (QMLH).

  • Input:

    A classical string zz and a quantum algorithm 𝒞\mathcal{C}, which takes as input 1n1^{n} and a classical string x𝒳n={0,1}poly(n)x\in\mathcal{X}_{n}=\{0,1\}^{{\mathrm{poly}}(n)}, and outputs a classical string z𝒵nz\in\mathcal{Z}_{n}. (In the following, we often omit 1n1^{n} of z𝒞(1n,x)z\leftarrow\mathcal{C}(1^{n},x).)

  • Output:

    Output h𝒳nh\in\mathcal{X}_{n} such that

    maxx𝒳n{Pr[z𝒞(x)]}(1+1/ϵ(n))Pr[z𝒞(h)]maxx𝒳n{Pr[z𝒞(x)]}.\displaystyle\frac{\max_{x\in\mathcal{X}_{n}}\{\Pr[z\leftarrow\mathcal{C}(x)]\}}{\left(1+1/\epsilon(n)\right)}\leq\Pr[z\leftarrow\mathcal{C}(h)]\leq\max_{x\in\mathcal{X}_{n}}\{\Pr[z\leftarrow\mathcal{C}(x)]\}. (7)

As explained later, a PPT algorithm 𝒜\mathcal{A} with 𝐏𝐏\mathbf{PP} oracle can exactly simulate an arbitrary probability distribution generated by a quantum polynomial-time algorithm even if it does post-selection. Therefore, it is sufficient to construct a quantum algorithm 𝒬^[z]\widehat{\mathcal{Q}}[z] with post-selection, which solves the quantum maximum likelihood problem. Let us explain how 𝒬^[z]\widehat{\mathcal{Q}}[z] works.

  1. 1.

    For sufficiently large TT, 𝒬^[z]\widehat{\mathcal{Q}}[z] generates the following quantum state

    UQ|0W,Y1V1,YTVT\displaystyle U_{Q}\ket{0}_{W,Y_{1}V_{1}\cdots,Y_{T}V_{T}} 1|𝒳n|x𝒳n|xWi[T]|𝒞(x)YiVi\displaystyle\coloneqq\frac{1}{\sqrt{\absolutevalue{\mathcal{X}_{n}}}}\sum_{x\in\mathcal{X}_{n}}\ket{x}_{W}\bigotimes_{i\in[T]}\ket{\mathcal{C}(x)}_{Y_{i}V_{i}} (8)
    =1|𝒳n|x𝒳n|xWi[T](y𝒵nPr[y𝒞(x)]|yYi|ϕx,yVi).\displaystyle=\frac{1}{\sqrt{\absolutevalue{\mathcal{X}_{n}}}}\sum_{x\in\mathcal{X}_{n}}\ket{x}_{W}\bigotimes_{i\in[T]}\left(\sum_{y\in\mathcal{Z}_{n}}\sqrt{\Pr[y\leftarrow\mathcal{C}(x)]}\ket{y}_{Y_{i}}\ket{\phi_{x,y}}_{V_{i}}\right). (9)

    Here, we write |𝒞(x)YiVi\ket{\mathcal{C}(x)}_{Y_{i}V_{i}} to mean a quantum state generated by a quantum algorithm 𝒞(x)\mathcal{C}(x) before measuring the state with the computational basis, where the YiY_{i} register corresponds to the outcome register and the ViV_{i} register corresponds to the auxiliary register.

  2. 2.

    It post-selects all of Y1YTY_{1}\cdots Y_{T} registers to zz. Let |ψ[z]WY1V1YTVT\ket{\psi[z]}_{WY_{1}V_{1}\cdots Y_{T}V_{T}} be the post-selected quantum state.

  3. 3.

    It measures the WW register of |ψ[z]WY1V1YTVT\ket{\psi[z]}_{WY_{1}V_{1}\cdots Y_{T}V_{T}} in the computational basis, and outputs the measurement outcome hh.

From the construction of 𝒬^[z]\widehat{\mathcal{Q}}[z],|ψ[z]WY1V1YTVT\ket{\psi[z]}_{WY_{1}V_{1}\cdots Y_{T}V_{T}} satisfies

|ψ[z]WY1V1YTVT=1x𝒳n(Pr[z𝒞(x)]T)h𝒳nPr[z𝒞(h)]T|hWi[T]|zYi|ϕh,zVi.\displaystyle\ket{\psi[z]}_{WY_{1}V_{1}\cdots Y_{T}V_{T}}=\frac{1}{\sqrt{\sum_{x\in\mathcal{X}_{n}}(\Pr[z\leftarrow\mathcal{C}(x)]^{T})}}\sum_{h\in\mathcal{X}_{n}}\sqrt{\Pr[z\leftarrow\mathcal{C}(h)]^{T}}\ket{h}_{W}\bigotimes_{i\in[T]}\ket{z}_{Y_{i}}\ket{\phi_{h,z}}_{V_{i}}. (10)

Therefore, it is natural to expect that, if we measure the WW register of |ψ[z]WY1V1YTVT\ket{\psi[z]}_{WY_{1}V_{1}\cdots Y_{T}V_{T}}, then, we obtain a measurement outcome hh such that Pr[z𝒞(h)]\Pr[z\leftarrow\mathcal{C}(h)] is high. We can prove the intuition by taking TT sufficiently large as follows. If hh is not a good answer for the quantum maximum likelihood problem (i.e. Pr[z𝒞(h)]<maxx𝒳nPr[z𝒞(x)](1+1/ϵ(n))\Pr[z\leftarrow\mathcal{C}(h)]<\frac{\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]}{\left(1+1/\epsilon(n)\right)}), then

Pr[h𝒬^[z]]=Pr[z𝒞(h)]Tx𝒳nPr[z𝒞(x)]TPr[z𝒞(h)]Tmaxx𝒳nPr[z𝒞(x)]T<(11+1/ϵ)T.\displaystyle\Pr[h\leftarrow\widehat{\mathcal{Q}}[z]]=\frac{\Pr[z\leftarrow\mathcal{C}(h)]^{T}}{\sum_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]^{T}}\leq\frac{\Pr[z\leftarrow\mathcal{C}(h)]^{T}}{\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]^{T}}<\left(\frac{1}{1+1/\epsilon}\right)^{T}. (11)

This implies that

Prh𝒬^[z][Pr[z𝒞(h)]<maxx𝒳nPr[z𝒞(x)](1+1/ϵ(n))]|𝒳n|(11+1/ϵ)T.\displaystyle\Pr_{h\leftarrow\widehat{\mathcal{Q}}[z]}\left[\Pr[z\leftarrow\mathcal{C}(h)]<\frac{\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]}{\left(1+1/\epsilon(n)\right)}\right]\leq\absolutevalue{\mathcal{X}_{n}}\cdot\left(\frac{1}{1+1/\epsilon}\right)^{T}. (12)

Therefore, by taking TT sufficiently large, 𝒬^[z]\widehat{\mathcal{Q}}[z] outputs hh such that

maxx𝒳nPr[z𝒞(x)]1+1/ϵPr[z𝒞(h)]maxx𝒳nPr[z𝒞(x)]\displaystyle\frac{\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]}{1+1/\epsilon}\leq\Pr[z\leftarrow\mathcal{C}(h)]\leq\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)] (13)

with high probability.

Although 𝒬^[z]\widehat{\mathcal{Q}}[z] solves the quantum maximum likelihood problem with high probability, it is possibly inefficient since it does post-selection. In the following, we explain how a PPT algorithm 𝒜\mathcal{A} with 𝐏𝐏\mathbf{PP} oracle exactly simulates the probability distribution of 𝒬^[z]\widehat{\mathcal{Q}}[z]. As shown in [FR99], even a deterministic polynomial time algorithm \mathcal{B} with 𝐏𝐏\mathbf{PP} oracle can compute Pr[z𝒞]\Pr[z\leftarrow\mathcal{C}] for an arbitrary QPT algorithm 𝒞\mathcal{C} and a classical string zz. Our PPT algorithm 𝒜𝐏𝐏\mathcal{A}^{\mathbf{PP}} simulates 𝒬^[z]\widehat{\mathcal{Q}}[z] by using 𝐏𝐏\mathcal{B}^{\mathbf{PP}} as follows:

  • For each i[|x|]i\in[\absolutevalue{x}], 𝒜𝐏𝐏\mathcal{A}^{\mathbf{PP}} sequentially works as follows:

    1. 1.

      For each b{0,1}b\in\{0,1\}, it computes the probability Pr[b,xi1,,x1,z]\Pr[b,x_{i-1},\cdots,x_{1},z] and Pr[xi1,,x1,z]\Pr[x_{i-1},\cdots,x_{1},z]. Here,

      Pr[xi1,,x1,z]Tr((|xi1x1xi1x1|W[i1]i[T]|zz|Yi)U𝒬|0W,Y1V1,,YTVT),\displaystyle\Pr[x_{i-1},\cdots,x_{1},z]\coloneqq\Tr(\left(\ket{x_{i-1}\cdots x_{1}}\bra{x_{i-1}\cdots x_{1}}_{W[i-1]}\bigotimes_{i\in[T]}\ket{z}\bra{z}_{Y_{i}}\right)U_{\mathcal{Q}}\ket{0}_{W,Y_{1}V_{1},\cdots,Y_{T}V_{T}}), (14)

      where W[i1]W[i-1] is the register in the first (i1)(i-1)-bits of WW. Note that, in this step, 𝒜𝐏𝐏\mathcal{A}^{\mathbf{PP}} uses 𝐏𝐏\mathcal{B}^{\mathbf{PP}} given in [FR99].

    2. 2.

      Set xibx_{i}\coloneqq b with probability

      Pr[b,xi1,,x1,z]Pr[xi1,,x1,z].\displaystyle\frac{\Pr[b,x_{i-1},\cdots,x_{1},z]}{\Pr[x_{i-1},\cdots,x_{1},z]}. (15)
  • Output hx|x|,,x1h\coloneqq x_{\absolutevalue{x}},\cdots,x_{1}.

From the construction of 𝒜𝐏𝐏\mathcal{A}^{\mathbf{PP}}, we have

Pr[x𝒜𝐏𝐏]Pr[x|x|,,x1,z]Pr[z].\displaystyle\Pr[x\leftarrow\mathcal{A}^{\mathbf{PP}}]\coloneqq\frac{\Pr[x_{\absolutevalue{x}},\cdots,x_{1},z]}{\Pr[z]}. (16)

for all x𝒳nx\in\mathcal{X}_{n}. From the definition, Pr[x|x|,,x1,z]Pr[z]\frac{\Pr[x_{\absolutevalue{x}},\cdots,x_{1},z]}{\Pr[z]} is exactly equal to the probability that Q^[z]\widehat{Q}[z] outputs xx. This means that the output distribution of 𝒜𝐏𝐏\mathcal{A}^{\mathbf{PP}} is equivalent to that of Q^[z]\widehat{Q}[z].

1.3 More on Related works

On Quantum Cryptography.

Ji, Liu, and Song [JLS18] introduce a notion of PRSGs, and show that it can be constructed from OWFs. Kretchmer [Kre21] shows that PRSGs exist relative to an oracle 𝒪\mathcal{O} such that 𝐁𝐐𝐏𝒪=𝐐𝐌𝐀𝒪\mathbf{BQP}^{\mathcal{O}}=\mathbf{QMA}^{\mathcal{O}}, and that PRSGs does not exist if its adversary can query 𝐏𝐏\mathbf{PP} oracle. The subsequent works [ZLK+23, CLS24] observe that the existence of PRSGs implies that there exists a quantum state, which can be generated by quantum algorithms but hard to learn. Morimae and Yamakawa [MY22b] introduce the notion of OWSGs, and show how to construct them from PRSGs. In their original definition of OWSGs, the output quantum states are restricted to pure states, and its definition is generalized to mixed states by [MY22a]. In this work, we focus on the original pure-state version. Based on [KT24], Cavalar et. al. [CGG+23] shows that OWSGs with pure state outputs can be broken if its adversary can query 𝐏𝐏\mathbf{PP} oracle. As pointed out in our work, the existence of OWSGs with pure state outputs is equivalent to the average-case hardness of learning pure states generated by a quantum polynomial-time algorithm. Therefore, to rephrase their results in terms of learning, a quantum polynomial-time algorithm with 𝐏𝐏\mathbf{PP} oracle can learn an arbitrary pure state generated by quantum polynomial-time algorithms in the average sense. Remark that their algorithm works only for average-case learning, and it is unclear whether their QPT algorithm with PP oracle can learn pure states in the worst-case sense.

Brakerski, Canetti, and Qian [BCQ23] introduce the notion of EFI. (See also [Yan22]). It is shown that the existence of EFI is equivalent to that of quantum bit commitments, oblivious transfer, and multi-party computation [GLSV21, BCKM21, Yan22, BCQ23]. It is conjectured that EFI may exist relative to a random oracle and any classical oracle [LMW24].

On Classical Learning Theory.

Valiant [Val84] introduced the PAC (probabilistically approximately correct) learning model to capture the notion of computationally efficient learning. It is shown that PAC learning is equivalent to Occam learning [BEHW87, BP92, Sch90], which can be formulated as a search problem in NP. Inspired by Valiant’s formalization, Kearns et. al. [KMR+94] formalizes a distributional learning model. Note that our quantum state learning model can be seen as a quantum generalization of the distributional learning model. It is shown that a polynomial time learnability of a class of probabilistic automata 𝒞\mathcal{C} with respect to Kullback–Leibler divergence is equivalent to that of the maximum likelihood problem for the same class 𝒞\mathcal{C} [AW92].

Valiant observed that the existence of pseudorandom functions, which is equivalent to the existence of OWFs [GGM86, HILL99], implies the hardness of learning for P/poly in the PAC model [Val84]. Impagliazzo and Levin [IL90] initiate the study of the reverse direction, i.e. from the average-case hardness of learning to cryptography. Subsequent work [BFKL94, NR06, OS17, NE19, HN23] studies how to construct cryptographic primitives from the average-case hardness of learning in more general learning models. Among them, our result (i.e. the equivalence between OWSGs and average-case hardness of learning pure states) can be seen as a quantum analog of [HN23] in the following sense. In one of their results, they show the equivalence between the existence of OWFs and the average-case hardness of distributional learning [KMR+94]. In our work, we observe that the existence of OWSGs with pure state outputs is equivalent to the average case hardness of learning pure states.

More on Quantum Learning.

In quantum state tomography, a learner is given many independent copies of an unknown nn-qubit quantum state ρ\rho, and outputs a classical description ρ^\widehat{\rho} such that 𝖳𝖣(ρ,ρ^)1/ϵ\mathsf{TD}(\rho,\widehat{\rho})\leq 1/\epsilon. It is known that a learner needs O(22nϵ2)O(2^{2n}\cdot\epsilon^{2})-copies if there is no promise on the unknown quantum state [OW16, HHJ+16]. Given the negative results, a natural question is which subclass of quantum states can be learned using only polynomially many copies of unknown quantum states. [CL21, BO21] shows that if given quantum states ρk\rho_{k} are promised contained in some restricted families 𝒞\mathcal{C} of quantum states, then a learner can find ρh𝒞\rho_{h}\in\mathcal{C} such that 𝖳𝖣(ρk,ρh)1/ϵ\mathsf{TD}(\rho_{k},\rho_{h})\leq 1/\epsilon with probability 11/δ1-1/\delta given only poly(log(|𝒞|),log(δ),ϵ){\mathrm{poly}}(\log(\absolutevalue{\mathcal{C}}),\log(\delta),\epsilon)-many copies of ρk\rho_{k}. This means that all quantum states generated by quantum polynomial-time algorithms can be learned in a sample-efficient way.

Beyond sample efficiency, these lines of work ([Mon17, GIKL23a, GIKL23b, HLB+24]) study which subclass of quantum states can be learned in polynomial-time. It is shown that stabilizer quantum circuits can be learned in a time-efficient way [Mon17]. Grewal, Iyer, Kretschmer, and Liang [GIKL23a, GIKL23b] generalize their results and show that low-rank stabilizer pure states can be learned in a time-efficient way. Huang, Liu, Broughton, Kim, Anshu, Landau, and McClean [HLB+24] show that a family of pure states generated by geometrically local shallow quantum circuits can be learned in a time-efficient way. Compared to their unconditional results, although our learning algorithm needs to query 𝐏𝐏\mathbf{PP} oracle, our learning algorithm can learn a classical description of arbitrary efficiently generatable pure states.

These lines of work ([Aar07, CHY16, Roc18, ACH+19, GL22, CHL+22, BJ95, AS07, AdW17, GKZ19, SSG23, CL21, Aar18, HKP20, BO21]) study how to learn quantum processes in other learning models. Aaronson [Aar07] considers one of quantum generalization of the PAC learning model [Val84], and the subsequent works [Roc18, ACH+19, GL22, CHL+22] study the learning model. In their model, a learner is given examples (Ei,Tr(Eiρ))(E_{i},\Tr(E_{i}\rho)), where ρ\rho is an unknown quantum state and EiE_{i} is a POVM operator sampled according to some distribution DD, to output a quantum state σ\sigma such that PrE𝒟[Tr(Eρ)Tr(Eσ)ϵ]1γ\Pr_{E\leftarrow\mathcal{D}}[\Tr(E\rho)-\Tr(E\sigma)\leq\epsilon]\geq 1-\gamma for small ϵ\epsilon and γ\gamma. Bshouty and Jackson [BJ95] consider another quantum generalization of the PAC learning model. Roughly speaking, the model is the same as standard Valiant’s PAC model except that the learner can receive xD(x)|x|c(x)\sum_{x}\sqrt{D(x)}\ket{x}\ket{c(x)} instead of (x,c(x))(x,c(x)), where xx is sampled according to the distribution DD. The subsequent works [AS07, AdW17, GKZ19, SSG23] study the learning model. Aaronson introduces another learning model called shadow tomography [Aar18]. In the model, a learner is given many copies of ρ\rho, and POVM operators E1,,EmE_{1},\cdots,E_{m}. The learner needs to estimate Tr(ρE1),Tr(ρEm)\Tr(\rho E_{1}),\cdots\Tr(\rho E_{m}) up to additive error ϵ\epsilon. The subsequent works [HKP20, BO21] also study the model.

1.4 Future Directions

We raise several open questions that remain to be explored. Below, we highlight the following two questions.

Show the 𝐏𝐏\mathbf{PP} hardness of properly learning pure states.

We show that 𝐏𝐏\mathbf{PP} oracle helps to properly learn pure states generated by quantum polynomial time algorithms. We leave the reverse direction as an open problem. In other words, can we decide any language in 𝐏𝐏\mathbf{PP} efficiently assuming that we can properly learn arbitrary pure states generated by quantum polynomial-time algorithms?

Remark that in the classical learning theory, the hardness of proper PAC learning is characterized by 𝐍𝐏\mathbf{NP}. It is shown that PAC learning can be reduced to Occam learning [BEHW87, BP92, Sch90], which can be formulated as a search problem in 𝐍𝐏\mathbf{NP}. Hence, 𝐍𝐏\mathbf{NP} oracle is sufficient to achieve proper PAC learning for arbitrary classical polynomial time algorithms. On the other hand, it is shown that if we can properly learn arbitrary classical polynomial-time algorithms in the proper PAC model [PV88], then we can decide any language in 𝐍𝐏\mathbf{NP}. Does a similar relationship hold between 𝐏𝐏\mathbf{PP} and proper pure state learning as the relationship between 𝐍𝐏\mathbf{NP} and proper PAC learning?

Characterize the hardness of learning mixed states.

In this work, we observe that the existence of EFI implies that there exist mixed states generated by quantum polynomial-time algorithms, which is hard to properly learn. This implies that it is unlikely to reduce proper mixed-state learning to classical language because it is conjectured that any classical oracles do not help to break the security of EFI [LMW24]. Note that the argument cannot be applied to improper mixed-state learning, and there exists the possibility that we can reduce improper mixed-state learning to classical language. We leave as an open problem whether we can reduce improper mixed-state learning to classical language or not. We also leave as an open problem whether we can characterize the hardness of proper mixed-state learning in terms of unitary complexities [RY22, MY23, BEM+23] instead of classical languages.

2 Preliminaries

2.1 Notations

Here, we introduce basic notations we will use in this paper. xXx\leftarrow X denotes selecting an element xx from a finite set XX uniformly at random, and y𝒜(x)y\leftarrow\mathcal{A}(x) denotes assigning to yy the output of a quantum or probabilistic or deterministic algorithm 𝒜\mathcal{A} on an input xx. Let [n]{1,,n}[n]\coloneqq\{1,\cdots,n\}. QPT stands for quantum polynomial time. A function f:f:\mathbb{N}\rightarrow\mathbb{R} is a negligible function if, for any constant cc, there exists λ0\lambda_{0}\in\mathbb{N} such that for any λ>λ0\lambda>\lambda_{0}, f(λ)<1/λcf(\lambda)<1/\lambda^{c}. We write f(λ)𝗇𝖾𝗀𝗅(λ)f(\lambda)\leq{\mathsf{negl}}(\lambda) to denote f(λ)f(\lambda) being a negligible function. When we refer to polynomials, we mean functions p:p:\mathbb{N}\rightarrow\mathbb{N}, where there exists a constant cc such that for all nn\in\mathbb{N}, it holds that p(n)cncp(n)\leq cn^{c}.

For simplicity, we often write |0\ket{0} to mean |00\ket{0\cdots 0}. For any two quantum states ρ1\rho_{1} and ρ2\rho_{2}, F(ρ1,ρ2)F(\rho_{1},\rho_{2}) is the fidelity between them, and 𝖳𝖣(ρ1,ρ2)\mathsf{TD}(\rho_{1},\rho_{2}) is the trace distance between them.

Sampleable Distribution.

We say that a family of distributions 𝒟={𝒟n}n\mathcal{D}=\{\mathcal{D}_{n}\}_{n\in\mathbb{N}} is efficiently sampleable distribution if there exists a uniform QPT algorithm DD such that, for each nn\in\mathbb{N}, the distribution of D(1n)D(1^{n}) is statistically identical to 𝒟n\mathcal{D}_{n}.

Definition 2.1 (Kullback–Leibler Divergence).

Let PP and QQ be a distribution over 𝒵n={0,1}n\mathcal{Z}_{n}=\{0,1\}^{n}.

DKL(P||Q)𝔼zP[log(Pr[zP]Pr[zQ])]=z𝒵nPr[zP]log(Pr[zP]Pr[zQ]).\displaystyle D_{KL}(P||Q)\coloneqq\mathbb{E}_{z\leftarrow P}\left[\log(\frac{\Pr[z\leftarrow P]}{\Pr[z\leftarrow Q]})\right]=\sum_{z\in\mathcal{Z}_{n}}\Pr[z\leftarrow P]\log(\frac{\Pr[z\leftarrow P]}{\Pr[z\leftarrow Q]}). (17)

2.2 Quantum Cryptographic Primitives

In this section, we introduce quantum cryptographic primitives, which we will use in this work. Note that throughout this work, as adversaries, we consider uniform QPT algorithms instead of non-uniform QPT algorithms.

One-Way State Generators (OWSGs)

Definition 2.2 (OWSGs with Pure State Outputs [MY22b]).

A one-way state generator with pure state outputs is a set of uniform QPT algorithms Σ(𝖪𝖾𝗒𝖦𝖾𝗇,𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇)\Sigma\coloneqq(\mathsf{KeyGen},\mathsf{StateGen}) such that:

  • 𝖪𝖾𝗒𝖦𝖾𝗇(1λ)\mathsf{KeyGen}(1^{\lambda}):

    It takes a security parameter 1λ1^{\lambda}, and outputs a classical string kk.

  • 𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1λ,k)\mathsf{StateGen}(1^{\lambda},k):

    It takes a security parameter 1λ1^{\lambda} and kk, and outputs a pure quantum state |ψk\ket{\psi_{k}}.

Security.

We consider the following security experiment 𝖤𝗑𝗉Σ,𝒜(1λ)\mathsf{Exp}_{\Sigma,\mathcal{A}}(1^{\lambda}).

  1. 1.

    The challenger samples k𝖪𝖾𝗒𝖦𝖾𝗇(1λ)k\leftarrow\mathsf{KeyGen}(1^{\lambda}).

  2. 2.

    𝒜\mathcal{A} can receive arbitrary polynomially many copies of |ψk\ket{\psi_{k}}.

  3. 3.

    𝒜\mathcal{A} outputs kk^{*}.

  4. 4.

    The challenger measures |ψk\ket{\psi_{k}} with {|ψkψk|,I|ψkψk|}\{\ket{\psi_{k^{*}}}\bra{\psi_{k^{*}}},I-\ket{\psi_{k^{*}}}\bra{\psi_{k^{*}}}\}.

  5. 5.

    The experiment outputs 11 if the result is |ψkψk|\ket{\psi_{k^{*}}}\bra{\psi_{k^{*}}}, and 0 otherwise.

We say that an OWSG with pure state outputs satisfies security if, for any uniform QPT adversaries 𝒜\mathcal{A},

Pr[𝖤𝗑𝗉Σ,𝒜(1λ)=1]𝗇𝖾𝗀𝗅(λ).\displaystyle\Pr[\mathsf{Exp}_{\Sigma,\mathcal{A}}(1^{\lambda})=1]\leq{\mathsf{negl}}(\lambda). (18)
Remark 2.3.

We sometimes omit the security parameter 1λ1^{\lambda} of 𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1λ,k)\mathsf{StateGen}(1^{\lambda},k) when it is clear from the context.

Definition 2.4 (Secretly-Verifiable Statistically-Invertible OWSGs (SV-SI-OWSG) [MY22a]).

A SV-SI-OWSG is a set of uniform QPT algorithms Σ(𝖪𝖾𝗒𝖦𝖾𝗇,𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇)\Sigma\coloneqq(\mathsf{KeyGen},\mathsf{StateGen}) such that:

𝖪𝖾𝗒𝖦𝖾𝗇(1λ)\mathsf{KeyGen}(1^{\lambda}):

It takes a security parameter 1λ1^{\lambda}, and outputs a classical string kk.

𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1λ,k)\mathsf{StateGen}(1^{\lambda},k):

It takes a security parameter 1λ1^{\lambda} and kk, and outputs ψλ,k\psi_{\lambda,k}.

Statistically Invertibility:

For any (k,k)(k,k^{*}) with kkk\neq k^{*},

𝖳𝖣(ψλ,k,ψλ,k)1𝗇𝖾𝗀𝗅(λ).\displaystyle\mathsf{TD}(\psi_{\lambda,k},\psi_{\lambda,k^{*}})\geq 1-{\mathsf{negl}}(\lambda). (19)

Computational Non-Invertibility:

For any uniform QPT adversaries 𝒜\mathcal{A} and any polynomials tt,

Pr[k𝒜(ψλ,kt(λ)):k𝖪𝖾𝗒𝖦𝖾𝗇(1λ),ψλ,k𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1λ,k)]𝗇𝖾𝗀𝗅(λ).\displaystyle\Pr[k\leftarrow\mathcal{A}(\psi_{\lambda,k}^{\otimes t(\lambda)}):k\leftarrow\mathsf{KeyGen}(1^{\lambda}),\psi_{\lambda,k}\leftarrow\mathsf{StateGen}(1^{\lambda},k)]\leq{\mathsf{negl}}(\lambda). (20)

Note that [MY22a] shows that the existence of EFI is equivalent to that of SV-SI-OWSG.

3 Model of Quantum State Learning

Definition 3.1 (Quantum State Learnability).

Let 𝒞\mathcal{C} be a quantum algorithm that takes 1n1^{n} with nn\in\mathbb{N}, x𝒳n={0,1}poly(n)x\in\mathcal{X}_{n}=\{0,1\}^{{\mathrm{poly}}(n)}, and outputs a poly(n){\mathrm{poly}}(n)-qubit quantum state ψx\psi_{x}. We say that an algorithm 𝒜\mathcal{A} can learn 𝒞\mathcal{C} if, for all polynomials ϵ\epsilon and δ\delta, there exists a polynomial tt such that for all x𝒳nx\in\mathcal{X}_{n}, we have

Pr[𝖳𝖣(ψx,ψy)1/ϵ(n):ψx𝒞(x,1n)y𝒜(𝒞,ψxt(n),1n)ψy𝒞(y,1n)]11/δ(n)\displaystyle\Pr[\begin{array}[]{ll}\mathsf{TD}(\psi_{x},\psi_{y})\leq 1/\epsilon(n)\end{array}:\begin{array}[]{ll}\psi_{x}\leftarrow\mathcal{C}(x,1^{n})\\ y\leftarrow\mathcal{A}(\mathcal{C},\psi_{x}^{\otimes t(n)},1^{n})\\ \psi_{y}\leftarrow\mathcal{C}(y,1^{n})\end{array}]\geq 1-1/\delta(n) (25)

for all sufficiently large nn\in\mathbb{N}.

Remark 3.2.

Throughout this work, we consider proper quantum state learning, where the learner’s output is promised within 𝒳n\mathcal{X}_{n}. The other variant of quantum state learning model is an improper one, where the learner’s task is not to output yy such that 𝒞(y)\mathcal{C}(y) approximates 𝒞(x)\mathcal{C}(x), but output arbitrary polynomial-size quantum circuits hh, whose output approximates 𝒞(x)\mathcal{C}(x). Note that if we can properly learn 𝒞\mathcal{C}, then we can improperly learn 𝒞\mathcal{C}. On the other hand, the other direction does not follow in general.

In this work, we also consider the average-case hardness of quantum state learning, which is defined as follows:

Definition 3.3 (Average-Case Hardness of Quantum State Learning).

Let 𝒞\mathcal{C} be a quantum algorithm that takes 1n1^{n} with nn\in\mathbb{N}, x𝒳={0,1}poly(n)x\in\mathcal{X}=\{0,1\}^{{\mathrm{poly}}(n)}, and outputs a poly(n){\mathrm{poly}}(n)-qubit quantum state ψx\psi_{x}. We say that 𝒞\mathcal{C} is hard on average if there exists an efficiently sampleable distribution 𝒮n\mathcal{S}_{n} over 𝒳n\mathcal{X}_{n} and polynomials ϵ\epsilon and δ\delta such that for all polynomials tt and all QPT algorithms 𝒜\mathcal{A}, we have

Pr[𝖳𝖣(ψx,ψy)1/ϵ(n):x𝒮nψx𝒞(x,1n)y𝒜(𝒞,ψxt(n),1n)ψy𝒞(y,1n)]1/δ(n)\displaystyle\Pr[\begin{array}[]{ll}\mathsf{TD}(\psi_{x},\psi_{y})\leq 1/\epsilon(n)\end{array}:\begin{array}[]{ll}x\leftarrow\mathcal{S}_{n}\\ \psi_{x}\leftarrow\mathcal{C}(x,1^{n})\\ y\leftarrow\mathcal{A}(\mathcal{C},\psi_{x}^{\otimes t(n)},1^{n})\\ \psi_{y}\leftarrow\mathcal{C}(y,1^{n})\end{array}]\leq 1/\delta(n) (31)

for all sufficiently large nn\in\mathbb{N}.

4 Quantum State Learning Algorithm with Classical Oracle

In this section, we prove the following main theorem.

Theorem 4.1 (Quantum State Learning Algorithm with 𝐏𝐏\mathbf{PP} oracle).

For all quantum polynomial-time algorithms 𝒞\mathcal{C} with pure state outputs, there exists a language 𝐏𝐏\mathcal{L}\in\mathbf{PP} and an oracle-aided quantum polynomial-time algorithm 𝒜\mathcal{A}^{\mathcal{L}} that can learn 𝒞\mathcal{C} in the sense of Definition 3.1.

To prove Theorem 4.1, we use the following Lemmata 4.2, 4.4 and 4.3, and the following Proposition 4.5.

Lemma 4.2 (Algorithm for Quantum Maximum Likelihood Problem).

For all quantum polynomial-time algorithms 𝒞\mathcal{C} that takes 1n1^{n} and x𝒳n={0,1}poly(n)x\in\mathcal{X}_{n}=\{0,1\}^{{\mathrm{poly}}(n)} as input and outputs a classical string z𝒵n={0,1}poly(n)z\in\mathcal{Z}_{n}=\{0,1\}^{{\mathrm{poly}}(n)}, there exists a language 𝐏𝐏\mathcal{L}\in\mathbf{PP} and an oracle-aided PPT algorithm 𝒜\mathcal{A}^{\mathcal{L}} such that for all z𝒵nz\in\mathcal{Z}_{n} with 0<maxx𝒳nPr[z𝒞(1n,x)]0<\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(1^{n},x)] and all polynomial ϵ\epsilon and δ\delta,

Prh𝒜(1n,𝒞,z)[maxx𝒳nPr[z𝒞(x,1n)](1+1/ϵ(n))Pr[z𝒞(h,1n)]maxx𝒳nPr[z𝒞(x,1n)]]12δ(n)\displaystyle\Pr_{h\leftarrow\mathcal{A}^{\mathcal{L}}(1^{n},\mathcal{C},z)}\left[\frac{\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x,1^{n})]}{\left(1+1/\epsilon(n)\right)}\leq\Pr[z\leftarrow\mathcal{C}(h,1^{n})]\leq\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x,1^{n})]\right]\geq 1-2^{-\delta(n)} (32)

for all sufficiently large nn\in\mathbb{N}.

We give the proof of Lemma 4.2 in Section 4.1.

Lemma 4.3 (Random Measurement preserves Distance [AE07]).

There exists constants AA and BB such that, for any nn-qubit mixed states ρ1,ρ2\rho_{1},\rho_{2} with 2n/3<Aρ1ρ2F42^{-n/3}<A\norm{\rho_{1}-\rho_{2}}_{F}^{4}, we have

Bρ0ρ1F(ρ1)(ρ2)1,\displaystyle B\norm{\rho_{0}-\rho_{1}}_{F}\leq\norm{\mathcal{M}(\rho_{1})-\mathcal{M}(\rho_{2})}_{1}, (33)

where \mathcal{M} is a POVM with respect to an 2n/32^{-n/3}-approximate (4, 4)-design.

We refer the proof of Lemma 4.3 to [AE07].

Lemma 4.4 (Followed by Hoeffding Inequality).

Let nn\in\mathbb{N}, 𝒵={0,1}poly(n)\mathcal{Z}=\{0,1\}^{{\mathrm{poly}}(n)}, and let 𝒳\mathcal{X} be a family of function such that X:𝒵[0,M]X:\mathcal{Z}\rightarrow[0,M] for all X𝒳X\in\mathcal{X}, where MM\in\mathbb{R}. Let 𝒞\mathcal{C} be an arbitrary distribution over 𝒵\mathcal{Z}. If

TM2ϵ2|log(|𝒳|)+log(δ)|,\displaystyle T\geq\frac{M^{2}}{\epsilon^{2}}\absolutevalue{\log{\absolutevalue{\mathcal{X}}}+\log(\delta)}, (34)

then, for all X𝒳X\in\mathcal{X}, we have

|1mi[T]X(zi)𝔼z𝒞[X(z)]|ϵ\displaystyle\absolutevalue{\frac{1}{m}\sum_{i\in[T]}X(z_{i})-\mathbb{E}_{z\leftarrow\mathcal{C}}[X(z)]}\leq\epsilon (35)

with probability greater than 11/δ1-1/\delta, where the probability is taken over {zi}i[T]𝒞T\{z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}^{T}.

Proposition 4.5.

When ρ0\rho_{0} and ρ1\rho_{1} are pure states,

ρ0ρ11=2ρ0ρ1F\displaystyle\norm{\rho_{0}-\rho_{1}}_{1}=\sqrt{2}\norm{\rho_{0}-\rho_{1}}_{F} (36)

We describe the proof of Theorem 4.1.

Proof of Theorem 4.1.

Let \mathcal{L} be a language in 𝐏𝐏\mathbf{PP} and 𝒜\mathcal{A}^{\mathcal{L}} given in Lemma 4.2. Let us describe our notations, and our oracle-aided learning algorithm \mathcal{B}^{\mathcal{L}}.

Notations:

  • Let 𝒞\mathcal{C} be a quantum polynomial-time algorithm that takes as input 1n1^{n} and x𝒳n={0,1}poly(n)x\in\mathcal{X}_{n}=\{0,1\}^{{\mathrm{poly}}(n)}, and outputs a |ψx\ket{\psi_{x}}.

  • Let 𝒞α\mathcal{C}^{*}_{\alpha} be a quantum polynomial-time algorithm that takes as input 1n1^{n} and x𝒳nx\in\mathcal{X}_{n}, generates |ψx\ket{\psi_{x}} by running 𝒞(1n,x)\mathcal{C}(1^{n},x), measures |ψx\ket{\psi_{x}} with \mathcal{M}, and obtains the measurement outcome z𝒵n={0,1}(n)z\in\mathcal{Z}_{n}=\{0,1\}^{\ell(n)}, where \mathcal{M} is a POVM measurement given in Lemma 4.3. 𝒞α\mathcal{C}^{*}_{\alpha} outputs the measurement outcome zz with probability 1α1-\alpha, and, with probability α\alpha, outputs a uniformly random Z𝒵nZ\leftarrow\mathcal{Z}_{n}.

  • Let 𝒞αT\mathcal{C}^{*T}_{\alpha} be a TT-repetition version of 𝒞α\mathcal{C}^{*}_{\alpha}. More formally, it takes as input 1n1^{n} and x𝒳nx\in\mathcal{X}_{n}, and generates {|ψxi}i[T]\{\ket{\psi_{x}}_{i}\}_{i\in[T]}. Then, for each i[T]i\in[T], it measures |ψxi\ket{\psi_{x}}_{i} with \mathcal{M}, obtains the measurement outcome ziz_{i}, and sets Zi=ziZ_{i}=z_{i} with probability 1α1-\alpha and Zi𝒵nZ_{i}\leftarrow\mathcal{Z}_{n} with probability α\alpha, where \mathcal{M} is a POVM measurement given in Lemma 4.3. Finally, it outputs {Zi}i[T]\{Z_{i}\}_{i\in[T]}.

  • Let ϵ(n)=4ϵ(n)2B2\epsilon^{*}(n)=\frac{4\epsilon(n)^{2}}{B^{2}}, where BB is a constant given in Lemma 4.3.

  • Let Tϵ2(n)((n)+1)2(log(|𝒳n|)+log(8δ))T\geq\epsilon^{*2}(n)\left(\ell(n)+1\right)^{2}\left(\log(\absolutevalue{\mathcal{X}_{n}})+\log(8\delta)\right) be a number of copies of |ψ\ket{\psi}, which \mathcal{B} uses.

The description of \mathcal{B}^{\mathcal{L}}:

  1. 1.

    It takes as input 1n1^{n}, a quantum polynomial-time algorithm 𝒞\mathcal{C}, and {|ψxi}i[T]\{\ket{\psi_{x}}_{i}\}_{i\in[T]}, which is generated by 𝒞(1n,x)\mathcal{C}(1^{n},x) with unknown xx.

  2. 2.

    For all i[T]i\in[T], it measures |ψxi\ket{\psi_{x}}_{i} with the POVM \mathcal{M} given in Lemma 4.3, obtains a classical string ziz_{i}, and sets Zi=ziZ_{i}=z_{i} with probability 1/21/2 and Zi𝒵nZ_{i}\leftarrow\mathcal{Z}_{n} with probability 1/21/2.

  3. 3.

    For all i[T]i\in[T], it runs 𝒜(1n,𝒞1/2T,{Zi}i[T])\mathcal{A}^{\mathcal{L}}(1^{n},\mathcal{C}^{*T}_{1/2},\{Z_{i}\}_{i\in[T]}) given in Lemma 4.2, and obtains h𝒳nh\in\mathcal{X}_{n} such that

    maxx𝒳n{Pr[{Zi}i[T]𝒞1/2T(1n,x)]}Pr[{Zi}i[T]𝒞1/2T(1n,h)](1+1/p(n))2(T2ϵ(n))\displaystyle\frac{\max_{x\in\mathcal{X}_{n}}\left\{\Pr[\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}^{*T}_{1/2}(1^{n},x)]\right\}}{\Pr[\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}^{*T}_{1/2}(1^{n},h)]}\leq\left(1+1/p(n)\right)\leq 2^{\left(\frac{T}{2\epsilon^{*}(n)}\right)} (37)

    with some polynomial pp.

Now, we show that

12|ψx|ψh11/ϵ(n)\displaystyle\frac{1}{2}\norm{\ket{\psi_{x}}-\ket{\psi_{h}}}_{1}\leq 1/\epsilon(n) (38)

with probability 12δ(n)1-2^{-\delta(n)}, where the probability is taken over {Zi}i[T]𝒞1/2T\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}_{1/2}^{*T} and the internal randomness of 𝒜\mathcal{A}.

First, for all y𝒳ny\in\mathcal{X}_{n}, the output hh of \mathcal{B}^{\mathcal{L}} satisfies

𝔼Z𝒞1/2(x)[log(Pr[Z𝒞1/2(y)]Pr[Z𝒞1/2(h)])]1/ϵ(n)\displaystyle\mathbb{E}_{Z\leftarrow\mathcal{C}^{*}_{1/2}(x)}\left[\log(\frac{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(y)]}{\Pr[Z\leftarrow\mathcal{C}_{1/2}^{*}(h)]})\right]\leq 1/\epsilon^{*}(n) (39)

with probability 12δ(n)1-2^{-\delta(n)}, where the probability is taken over {Zi}i[T]𝒞1/2T\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}_{1/2}^{*T} and the internal randomness of 𝒜\mathcal{A}. We will prove the inequality above later. Once we have obtained the inequality above, the remaining part straightforwardly follows. First, we have

DKL(𝒞1/2(x)||𝒞1/2(h))\displaystyle D_{KL}(\mathcal{C}_{1/2}^{*}(x)||\mathcal{C}_{1/2}^{*}(h)) (40)
𝔼Z𝒞1/2(x)[log(Pr[Z𝒞1/2(x)]Pr[Z𝒞1/2(h)])]1/ϵ(n)\displaystyle\coloneqq\mathbb{E}_{Z\leftarrow\mathcal{C}^{*}_{1/2}(x)}\left[\log(\frac{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(x)]}{\Pr[Z\leftarrow\mathcal{C}_{1/2}^{*}(h)]})\right]\leq 1/\epsilon^{*}(n) (41)

with probability 12δ1-2^{-\delta}. Furthermore, from Pinsker’s inequality, we have

𝒞1/2(x)𝒞1/2(h)12DKL(𝒞1/2(x)||𝒞1/2(h))2ϵ(n).\displaystyle\norm{\mathcal{C}_{1/2}^{*}(x)-\mathcal{C}_{1/2}^{*}(h)}_{1}\leq\sqrt{2D_{KL}(\mathcal{C}_{1/2}^{*}(x)||\mathcal{C}_{1/2}^{*}(h))}\leq\sqrt{\frac{2}{\epsilon^{*}(n)}}. (42)

On the other hand, for all 0α10\leq\alpha\leq 1, we have

𝒞α(x)𝒞α(h)1\displaystyle\norm{\mathcal{C}_{\alpha}^{*}(x)-\mathcal{C}_{\alpha}^{*}(h)}_{1} (43)
12Z|Pr[Z𝒞α(x)]Pr[Z𝒞α(y)]|\displaystyle\coloneqq\frac{1}{2}\sum_{Z}\absolutevalue{\Pr[Z\leftarrow\mathcal{C}_{\alpha}^{*}(x)]-\Pr[Z\leftarrow\mathcal{C}_{\alpha}^{*}(y)]} (44)
=12Z|(1α)Pr[Z𝒞0(x)](1α)Pr[Z𝒞0(y)]|\displaystyle=\frac{1}{2}\sum_{Z}\absolutevalue{(1-\alpha)\Pr[Z\leftarrow\mathcal{C}_{0}^{*}(x)]-(1-\alpha)\Pr[Z\leftarrow\mathcal{C}_{0}^{*}(y)]} (45)
=(1α)12Z|Pr[Z𝒞0(x)]Pr[Z𝒞0(y)]|\displaystyle=(1-\alpha)\cdot\frac{1}{2}\sum_{Z}\absolutevalue{\Pr[Z\leftarrow\mathcal{C}_{0}^{*}(x)]-\Pr[Z\leftarrow\mathcal{C}_{0}^{*}(y)]} (46)
=(1α)𝒞0(x)𝒞0(h)1.\displaystyle=(1-\alpha)\norm{\mathcal{C}_{0}^{*}(x)-\mathcal{C}_{0}^{*}(h)}_{1}. (47)

Hence,

𝒞0(x)𝒞0(h)12𝒞1/2(x)𝒞1/2(h)122ϵ(n).\displaystyle\norm{\mathcal{C}_{0}^{*}(x)-\mathcal{C}_{0}^{*}(h)}_{1}\leq 2\norm{\mathcal{C}_{1/2}^{*}(x)-\mathcal{C}_{1/2}^{*}(h)}_{1}\leq 2\sqrt{\frac{2}{\epsilon^{*}(n)}}. (48)

Suppose that

|ψx|ψh1(42n/3A)1/4,\displaystyle\norm{\ket{\psi_{x}}-\ket{\psi_{h}}}_{1}\leq\left(\frac{4\cdot 2^{-n/3}}{A}\right)^{1/4}, (49)

where AA is a constant given in Lemma 4.3. Then, for all polynomial ϵ\epsilon, we have

|ψx|ψh11/ϵ(n)\displaystyle\norm{\ket{\psi_{x}}-\ket{\psi_{h}}}_{1}\leq 1/\epsilon(n) (50)

for all sufficiently large nn\in\mathbb{N}. On the other hand, if

(42n/3A)1/4<|ψx|ψh1,\displaystyle\left(\frac{4\cdot 2^{-n/3}}{A}\right)^{1/4}<\norm{\ket{\psi_{x}}-\ket{\psi_{h}}}_{1}, (51)

then Lemmata 4.3 and 4.5 imply that

12|ψx)|ψh1=12|ψx|ψhF1B2𝒞0(x)𝒞0(h)12Bϵ(n)1ϵ(n).\displaystyle\frac{1}{2}\norm{\ket{\psi_{x}})-\ket{\psi_{h}}}_{1}=\frac{1}{\sqrt{2}}\norm{\ket{\psi_{x}}-\ket{\psi_{h}}}_{F}\leq\frac{1}{B\sqrt{2}}\norm{\mathcal{C}^{*}_{0}(x)-\mathcal{C}^{*}_{0}(h)}_{1}\leq\frac{2}{B\sqrt{\epsilon^{*}(n)}}\coloneqq\frac{1}{\epsilon(n)}. (52)

Now, we prove the remaining part, that is, for all y𝒳ny\in\mathcal{X}_{n},

𝔼Z𝒞1/2(x)[log(Pr[Z𝒞1/2(y)]Pr[Z𝒞1/2(h)])]1/ϵ(n)\displaystyle\mathbb{E}_{Z\leftarrow\mathcal{C}^{*}_{1/2}(x)}\left[\log(\frac{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(y)]}{\Pr[Z\leftarrow\mathcal{C}_{1/2}^{*}(h)]})\right]\leq 1/\epsilon^{*}(n) (53)

with 12δ(n)1-2^{-\delta(n)}, where the probability is taken over {Zi}i[T]𝒞1/2T\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}_{1/2}^{*T} and the internal randomness of 𝒜\mathcal{A}. From the construction of \mathcal{B}^{\mathcal{L}}, for all {Zi}i[T]\{Z_{i}\}_{i\in[T]}, \mathcal{B} outputs hh such that

maxs𝒳n{Pr[{Zi}i[T]𝒞1/2T(s)]}Pr[{Zi}i[T]𝒞1/2T(h)]\displaystyle\frac{\max_{s\in\mathcal{X}_{n}}\left\{\Pr[\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}^{*T}_{1/2}(s)]\right\}}{\Pr[\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}^{*T}_{1/2}(h)]} (54)
=maxs𝒳n{Πi[T]Pr[Zi𝒞1/2(s)]}Πi[T]Pr[Zi𝒞1/2(h)](1+1/p(n))2(T2ϵ(n))\displaystyle=\frac{\max_{s\in\mathcal{X}_{n}}\left\{\Pi_{i\in[T]}\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(s)]\right\}}{\Pi_{i\in[T]}\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(h)]}\leq\left(1+1/p(n)\right)\leq 2^{\left(\frac{T}{2\epsilon^{*}(n)}\right)} (55)

with probability 12δ(n)/21-2^{-\delta(n)}/2, where the probability is taken over the internal randomness of 𝒜\mathcal{A}. By taking logarithmic and dividing with TT, we have

(1Ti[T]log(1Pr[Zi𝒞1/2(h)])+maxs𝒳n1Ti[T]log(Pr[Zi𝒞1/2(s)]))12ϵ(n).\displaystyle\left(\frac{1}{T}\sum_{i\in[T]}\log(\frac{1}{\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(h)]})+\max_{s\in\mathcal{X}_{n}}\frac{1}{T}\sum_{i\in[T]}\log(\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(s)])\right)\leq\frac{1}{2\epsilon^{*}(n)}. (56)

This implies that for all y𝒳ny\in\mathcal{X}_{n},

(1Ti[T]log(1Pr[Zi𝒞1/2(h)])1Ti[T]log(1Pr[Zi𝒞1/2(y)]))\displaystyle\left(\frac{1}{T}\sum_{i\in[T]}\log(\frac{1}{\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(h)]})-\frac{1}{T}\sum_{i\in[T]}\log(\frac{1}{\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(y)]})\right) (57)
=(1Ti[T]log(1Pr[Zi𝒞1/2(h)])+1Ti[T]log(Pr[Zi𝒞1/2(y)]))\displaystyle=\left(\frac{1}{T}\sum_{i\in[T]}\log(\frac{1}{\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(h)]})+\frac{1}{T}\sum_{i\in[T]}\log(\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(y)])\right) (58)
(1Ti[T]log(1Pr[Zi𝒞1/2(h)])+maxs𝒳n1Ti[T]log(Pr[Zi𝒞1/2(s)]))12ϵ(n)\displaystyle\leq\left(\frac{1}{T}\sum_{i\in[T]}\log(\frac{1}{\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(h)]})+\max_{s\in\mathcal{X}_{n}}\frac{1}{T}\sum_{i\in[T]}\log(\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(s)])\right)\leq\frac{1}{2\epsilon^{*}(n)} (59)

with probability 12δ(n)/21-2^{-\delta(n)}/2, where the probability is taken over the internal randomness 𝒜\mathcal{A}. Furthermore, because

0log(1Pr[Z𝒞1/2(x)])(n)+1\displaystyle 0\leq\log(\frac{1}{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(x)]})\leq\ell(n)+1 (60)

for all Z𝒵n={0,1}(n)Z\in\mathcal{Z}_{n}=\{0,1\}^{\ell(n)} and x𝒳nx\in\mathcal{X}_{n}, we have

(𝔼Z𝒞1/2(x)[log(1Pr[Z𝒞1/2(h)])]1Ti[T]log(1Pr[Zi𝒞1/2(h)]))14ϵ(n)\displaystyle\left(\mathbb{E}_{Z\leftarrow\mathcal{C}_{1/2}^{*}(x)}\left[\log(\frac{1}{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(h)]})\right]-\frac{1}{T}\sum_{i\in[T]}\log(\frac{1}{\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(h)]})\right)\leq\frac{1}{4\epsilon^{*}(n)} (61)

and

(1Ti[T]log(1Pr[Zi𝒞1/2(y)])𝔼Z𝒞1/2(x)[log(1Pr[Z𝒞1/2(y)])])14ϵ(n)\displaystyle\left(\frac{1}{T}\sum_{i\in[T]}\log(\frac{1}{\Pr[Z_{i}\leftarrow\mathcal{C}^{*}_{1/2}(y)]})-\mathbb{E}_{Z\leftarrow\mathcal{C}_{1/2}^{*}(x)}\left[\log(\frac{1}{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(y)]})\right]\right)\leq\frac{1}{4\epsilon^{*}(n)} (62)

with probability 12δ(n)/21-2^{-\delta(n)}/2, where the probability is taken over {Zi}i[T]𝒞1/2T(x)\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}_{1/2}^{*T}(x). Note that, here, we use Lemma 4.4.

By summing up three inequalities above, we have

(𝔼Z𝒞1/2(x)[log(1Pr[Z𝒞1/2(h)])]𝔼Z𝒞1/2(x)[log(1Pr[Z𝒞1/2(y)])])1/ϵ(n)\displaystyle\left(\mathbb{E}_{Z\leftarrow\mathcal{C}_{1/2}^{*}(x)}\left[\log(\frac{1}{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(h)]})\right]-\mathbb{E}_{Z\leftarrow\mathcal{C}_{1/2}^{*}(x)}\left[\log(\frac{1}{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(y)]})\right]\right)\leq 1/\epsilon^{*}(n) (63)

with probability 12δ(n)1-2^{-\delta(n)}, where the probability is taken over {Zi}i[T]𝒞1/2T(x)\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}_{1/2}^{*T}(x) and the internal randomness of 𝒜\mathcal{A}. This implies that

𝔼Z𝒞1/2(x)[log(Pr[Z𝒞1/2(y)]Pr[Z𝒞1/2(h)])]1/ϵ(n)\displaystyle\mathbb{E}_{Z\leftarrow\mathcal{C}^{*}_{1/2}(x)}\left[\log(\frac{\Pr[Z\leftarrow\mathcal{C}^{*}_{1/2}(y)]}{\Pr[Z\leftarrow\mathcal{C}_{1/2}^{*}(h)]})\right]\leq 1/\epsilon^{*}(n) (64)

with probability 12δ(n)1-2^{-\delta(n)}, where the probability is taken over {Zi}i[T]𝒞1/2T(x)\{Z_{i}\}_{i\in[T]}\leftarrow\mathcal{C}_{1/2}^{*T}(x) and the internal randomness of 𝒜\mathcal{A}. ∎

4.1 Algorithm for Quantum Maximum-Likelihood Problem

In this section, we prove Lemma 4.2. To prove Lemma 4.2, we use the following Lemma 4.6.

Lemma 4.6 ([FR99]).

There exists a language 𝐏𝐏\mathcal{L}\in\mathbf{PP} and an oracle-aided deterministic polynomial-time algorithm \mathcal{B}^{\mathcal{L}} such that the following holds.

(1n,𝒞,z)=Pr[z𝒞(1n)].\displaystyle\mathcal{B}^{\mathcal{L}}(1^{n},\mathcal{C},z)=\Pr[z\leftarrow\mathcal{C}(1^{n})]. (65)

Here, 𝒞\mathcal{C} is a quantum polynomial-time algorithm that takes 1n1^{n} as input, and outputs zz.

We refer the proof to [FR99]. Now, we prove Lemma 4.2.

Proof of Lemma 4.2.

Let 𝐏𝐏\mathcal{L}\in\mathbf{PP} and \mathcal{B} be a deterministic polynomial time algorithm given in Lemma 4.6. We describe our notations and our PPT algorithm 𝒜\mathcal{A}^{\mathcal{L}}.

Notations:

  • Let 𝒞\mathcal{C} be a QPT algorithm that takes as input 1n1^{n} and x𝒳nx\in\mathcal{X}_{n}, and outputs z𝒵nz\in\mathcal{Z}_{n}. Let \ell be the length of x𝒳nx\in\mathcal{X}_{n}.

  • Without loss of generality, we can assume that the QPT algorithm 𝒞(1n,x)\mathcal{C}(1^{n},x) works as follows:

    1. 1.

      It prepares a quantum state

      U𝒞(x)|0YV|𝒞(x)YV=z𝒵Pr[z𝒞(x)]|zY|ϕx,zV.\displaystyle U_{\mathcal{C}(x)}\ket{0}_{YV}\coloneqq\ket{\mathcal{C}(x)}_{YV}=\sum_{z\in\mathcal{Z}}\sqrt{\Pr[z\leftarrow\mathcal{C}(x)]}\ket{z}_{Y}\ket{\phi_{x,z}}_{V}. (66)
    2. 2.

      It measures the YY register, and outputs the measurement outcome.

  • Let Tδ+log(|𝒳n|)log(1+1/ϵ)T\geq\frac{\delta+\log(\absolutevalue{\mathcal{X}_{n}})}{\log(1+1/\epsilon)} so that

    |𝒳n|(11+1/ϵ)T2δ.\displaystyle\absolutevalue{\mathcal{X}_{n}}\left(\frac{1}{1+1/\epsilon}\right)^{T}\leq 2^{-\delta}. (67)
  • Let 𝒬^[z]\widehat{\mathcal{Q}}[z] be a quantum algorithm with post-selection working as follows:

    1. 1.

      It prepares

      U𝒬|0WY1V1YTVT\displaystyle U_{\mathcal{Q}}\ket{0}_{WY_{1}V_{1}\cdots Y_{T}V_{T}} 1|𝒳n|x𝒳n|xWi[T](|𝒞(x)YiVi)\displaystyle\coloneqq\frac{1}{\sqrt{\absolutevalue{\mathcal{X}_{n}}}}\sum_{x\in\mathcal{X}_{n}}\ket{x}_{W}\bigotimes_{i\in[T]}\left(\ket{\mathcal{C}(x)}_{Y_{i}V_{i}}\right) (68)
      =1|𝒳n|x𝒳n|xWi[T](z𝒵nPr[z𝒞(x)]|zYi|ϕx,zVi),\displaystyle=\frac{1}{\sqrt{\absolutevalue{\mathcal{X}_{n}}}}\sum_{x\in\mathcal{X}_{n}}\ket{x}_{W}\bigotimes_{i\in[T]}\left(\sum_{z\in\mathcal{Z}_{n}}\sqrt{\Pr[z\leftarrow\mathcal{C}(x)]}\ket{z}_{Y_{i}}\ket{\phi_{x,z}}_{V_{i}}\right), (69)

      measures the Y1YTY_{1}\cdots Y_{T} in the computational basis, and obtains {yi}i[T]\{y_{i}\}_{i\in[T]}. Repeat this step until yi=zy_{i}=z for all i[T]i\in[T]. Let |ψ[z]WY1V1YTVT\ket{\psi[z]}_{WY_{1}V_{1}\cdots Y_{T}V_{T}} be the post-selected quantum state. Note that we have

      |ψ[z]WY1V1YTVT=1x𝒳nPr[z𝒞(x)]Th𝒳nPr[z𝒞(h)]T|hWi[T](|zYi|ϕh,zVi).\displaystyle\ket{\psi[z]}_{WY_{1}V_{1}\cdots Y_{T}V_{T}}=\frac{1}{\sqrt{\sum_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]^{T}}}\sum_{h\in\mathcal{X}_{n}}\sqrt{\Pr[z\leftarrow\mathcal{C}(h)]^{T}}\ket{h}_{W}\bigotimes_{i\in[T]}\left(\ket{z}_{Y_{i}}\ket{\phi_{h,z}}_{V_{i}}\right). (70)
    2. 2.

      Measure the WW register and output the measurement outcome hh.

Our oracle aided PPT algorithm 𝒜\mathcal{A}^{\mathcal{L}} simulates 𝒬^[z]\widehat{\mathcal{Q}}[z] by working as follows.

Description of 𝒜\mathcal{A}^{\mathcal{L}}:

  • As an input of 𝒜\mathcal{A}^{\mathcal{L}}, it receives 1n1^{n}, a QPT algorithm 𝒞\mathcal{C}, and a classical string z𝒵nz\in\mathcal{Z}_{n}.

  • For i[]i\in[\ell], 𝒜\mathcal{A}^{\mathcal{L}} sequentially works as follows:

    • For each b{0,1}b\in\{0,1\}, it computes

      Pr[b|xi1,,x1,z]=Pr[b,xi1,,x1,z]Pr[xi1,,x1,z].\displaystyle\Pr[b|x_{i-1},\cdots,x_{1},z]=\frac{\Pr[b,x_{i-1},\cdots,x_{1},z]}{\Pr[x_{i-1},\cdots,x_{1},z]}. (71)

      Here,

      Pr[xi1,,x1,z]=Tr((|xi1,,x1xi1,,x1|W[i1]i[T]|zz|Yi)U𝒬|0WY1V1YTVT),\displaystyle\Pr[x_{i-1},\cdots,x_{1},z]=\Tr(\left(\ket{x_{i-1},\cdots,x_{1}}\bra{x_{i-1},\cdots,x_{1}}_{W[i-1]}\bigotimes_{i\in[T]}\ket{z}\bra{z}_{Y_{i}}\right)U_{\mathcal{Q}}\ket{0}_{WY_{1}V_{1}\cdots Y_{T}V_{T}}), (72)

      where W[i1]W[i-1] is the register in the first (i1)(i-1)-bits of WW. For this procedure, we use the algorithm \mathcal{B} and 𝐏𝐏\mathcal{L}\in\mathbf{PP} given in Lemma 4.6.

    • Set xi=bx_{i}=b with probability

      Pr[b|xi1,,x1,z]=Pr[b,xi1,,x1,z]Pr[xi1,,x1,z].\displaystyle\Pr[b|x_{i-1},\cdots,x_{1},z]=\frac{\Pr[b,x_{i-1},\cdots,x_{1},z]}{\Pr[x_{i-1},\cdots,x_{1},z]}. (73)
  • Output hx,,x1h\coloneqq x_{\ell},\cdots,x_{1}.

From the construction of 𝒜\mathcal{A}^{\mathcal{L}}, we have

Pr[x𝒜(1n,𝒞,z)]=Pr[x,,x1,z]Pr[z]\displaystyle\Pr[x\leftarrow\mathcal{A}^{\mathcal{L}}(1^{n},\mathcal{C},z)]=\frac{\Pr[x_{\ell},\cdots,x_{1},z]}{\Pr[z]} (74)

for all x𝒳nx\in\mathcal{X}_{n}. From the definition, Pr[x,,x1,z]Pr[z]\frac{\Pr[x_{\ell},\cdots,x_{1},z]}{\Pr[z]} is exactly equal to the probability that Q^[z]\widehat{Q}[z] outputs xx. This means that the output distribution of 𝒜(1n,𝒞,z)\mathcal{A}^{\mathcal{L}}(1^{n},\mathcal{C},z) is equivalent to that of Q^[z]\widehat{Q}[z]. Therefore, it is sufficient to show that

Prh𝒬^[z](1n)[maxx𝒳nPr[z𝒞(x)]1+1/ϵPr[z𝒞(h)]]12δ.\displaystyle\Pr_{h\leftarrow\widehat{\mathcal{Q}}[z](1^{n})}\left[\frac{\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]}{1+1/\epsilon}\leq\Pr[z\leftarrow\mathcal{C}(h)]\right]\geq 1-2^{-\delta}. (75)

Let 𝖡𝖺𝖽z\mathsf{Bad}_{z} be a subset of 𝒳n\mathcal{X}_{n} such that

𝖡𝖺𝖽z{h𝒳n:Pr[z𝒞(h)]<maxx𝒳nPr[z𝒞(x)]1+1/ϵ}.\displaystyle\mathsf{Bad}_{z}\coloneqq\left\{h\in\mathcal{X}_{n}:\Pr[z\leftarrow\mathcal{C}(h)]<\frac{\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]}{1+1/\epsilon}\right\}. (76)

From the construction of 𝒬^[z]\widehat{\mathcal{Q}}[z], we have

Pr[h𝒬^[z]]=Pr[z𝒞(h)]Tx𝒳nPr[z𝒞(x)]T\displaystyle\Pr[h\leftarrow\widehat{\mathcal{Q}}[z]]=\frac{\Pr[z\leftarrow\mathcal{C}(h)]^{T}}{\sum_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]^{T}} (77)

for all h𝒳nh\in\mathcal{X}_{n}. Then, for all h𝖡𝖺𝖽zh\in\mathsf{Bad}_{z}, we have

Pr[h𝒬^[z]]=Pr[z𝒞(h)]Tx𝒳nPr[z𝒞(x)]TPr[z𝒞(h)]Tmaxx𝒳nPr[z𝒞(x)]T<(11+1/ϵ)T.\displaystyle\Pr[h\leftarrow\widehat{\mathcal{Q}}[z]]=\frac{\Pr[z\leftarrow\mathcal{C}(h)]^{T}}{\sum_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]^{T}}\leq\frac{\Pr[z\leftarrow\mathcal{C}(h)]^{T}}{\max_{x\in\mathcal{X}_{n}}\Pr[z\leftarrow\mathcal{C}(x)]^{T}}<\left(\frac{1}{1+1/\epsilon}\right)^{T}. (78)

Therefore, we have

h𝖡𝖺𝖽zPr[h𝒬^[z]]<h𝖡𝖺𝖽z(11+1/ϵ)T=|𝖡𝖺𝖽z|(11+1/ϵ)T|𝒳n|(11+1/ϵ)T2δ,\displaystyle\sum_{h\in\mathsf{Bad}_{z}}\Pr[h\leftarrow\widehat{\mathcal{Q}}[z]]<\sum_{h\in\mathsf{Bad}_{z}}\left(\frac{1}{1+1/\epsilon}\right)^{T}=\absolutevalue{\mathsf{Bad}_{z}}\left(\frac{1}{1+1/\epsilon}\right)^{T}\leq\absolutevalue{\mathcal{X}_{n}}\left(\frac{1}{1+1/\epsilon}\right)^{T}\leq 2^{-\delta}, (79)

which completes the proof.

5 Hardness of Quantum State Learning

5.1 Equivalence between Average-Case Hardness of Learning Pure States and OWSGs

Theorem 5.1.

The existence of OWSGs with pure state outputs is equivalent to that of a polynomial-size quantum circuit with pure state outputs, which is hard on average in the sense of Definition 3.3.

Remark 5.2.

This result works only for "proper" quantum state learning, and it is unclear how to show the hardness of improper quantum state learning from OWSGs as far as we understand. On the other hand, we can show the hardness of improper quantum state learning from the existence of pseudorandom state generators.

Theorem 5.1 relatively straightforwardly follows from the definition. We describe the proof for clarity.

Proof of Theorem 5.1.

First, we show that if there exist OWSGs with pure state outputs, then there exists a quantum polynomial-time algorithm 𝒞\mathcal{C} with pure state outputs, which is hard on average.

OWSGs \rightarrow Average-Case Hardness of Quantum State Learning.

It is sufficient to construct a quantum polynomial-time algorithm 𝒞\mathcal{C} such that there exists a sampleable distribution 𝒮n\mathcal{S}_{n} over 𝒳n\mathcal{X}_{n} and polynomials ϵ\epsilon and δ\delta such that for all polynomials tt,

Pr[F(|ψx,|ψx)11/ϵ(n):x𝒮n|ψx𝒞(1n,x)x𝒜(|ψxt(n),1n)|ψx𝒞(1n,x)]11/δ(n)\displaystyle\Pr[\begin{array}[]{ll}F(\ket{\psi_{x}},\ket{\psi_{x^{*}}})\geq 1-1/\epsilon(n)\end{array}:\begin{array}[]{ll}x\leftarrow\mathcal{S}_{n}\\ \ket{\psi_{x}}\leftarrow\mathcal{C}(1^{n},x)\\ x^{*}\leftarrow\mathcal{A}(\ket{\psi_{x}}^{\otimes t(n)},1^{n})\\ \ket{\psi_{x^{*}}}\leftarrow\mathcal{C}(1^{n},x^{*})\end{array}]\leq 1-1/\delta(n) (85)

for all sufficiently large nn\in\mathbb{N}.

We define 𝒞\mathcal{C} and 𝒮\mathcal{S} as follows:

The description of 𝒮n\mathcal{S}_{n}:

  • 𝒮n\mathcal{S}_{n} is the same as the distribution of 𝖪𝖾𝗒𝖦𝖾𝗇(1n)\mathsf{KeyGen}(1^{n}).

The description of 𝒞\mathcal{C}:

  • 𝒞\mathcal{C} runs 𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1n,x)\mathsf{StateGen}(1^{n},x), and outputs its output.

For contradiction, assume that for all polynomials ϵ\epsilon and δ\delta there exists a polynomial tt, and a QPT learner \mathcal{B} such that

Pr[F(|ψx,|ψx)11/ϵ(n):x𝒮n|ψx𝒞(1n,x)x(|ψxt(n),1n)|ψx𝒞(1n,x)]>11/δ(n)\displaystyle\Pr[\begin{array}[]{ll}F(\ket{\psi_{x}},\ket{\psi_{x^{*}}})\geq 1-1/\epsilon(n)\end{array}:\begin{array}[]{ll}x\leftarrow\mathcal{S}_{n}\\ \ket{\psi_{x}}\leftarrow\mathcal{C}(1^{n},x)\\ x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)},1^{n})\\ \ket{\psi_{x^{*}}}\leftarrow\mathcal{C}(1^{n},x^{*})\end{array}]>1-1/\delta(n) (91)

for infinitely many nn\in\mathbb{N}. Now, by using \mathcal{B}, we construct a QPT adversary 𝒜\mathcal{A} that breaks the security of OWSGs Σ\Sigma. We describe the 𝒜\mathcal{A} in the following.

  1. 1.

    𝒜\mathcal{A} receives sufficiently many copies of |ψx\ket{\psi_{x}}, where x𝖪𝖾𝗒𝖦𝖾𝗇(1n)x\leftarrow\mathsf{KeyGen}(1^{n}) and |ψx𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1n,x)\ket{\psi_{x}}\leftarrow\mathsf{StateGen}(1^{n},x).

  2. 2.

    𝒜\mathcal{A} passes them to \mathcal{B}, and receives xx^{*}.

  3. 3.

    𝒜\mathcal{A} sends xx^{*} to the challenger.

  4. 4.

    The challenger measures |ψx\ket{\psi_{x}} with {|ψxψx|,I|ψxψx|}\{\ket{\psi_{x^{*}}}\bra{\psi_{x^{*}}},I-\ket{\psi_{x^{*}}}\bra{\psi_{x^{*}}}\}.

  5. 5.

    The challenger outputs 11 if the measurement result is |ψxψx|\ket{\psi_{x^{*}}}\bra{\psi_{x^{*}}} is obtained, and 0 otherwise.

Let us define

𝖦𝗈𝗈𝖽{(x,x):|ψx|ψx|211/ϵ(n)}.\displaystyle\mathsf{Good}\coloneqq\left\{(x,x^{*}):\absolutevalue{\bra{\psi_{x}}\ket{\psi_{x^{*}}}}^{2}\geq 1-1/\epsilon(n)\right\}. (92)

Now, we compute the probability that the challenger outputs 11.

Pr[1Challenger:x𝖪𝖾𝗒𝖦𝖾𝗇(1n),|ψx𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1λ,x),x𝒜(|ψxt(n))]\displaystyle\Pr[1\leftarrow\mbox{Challenger}:x\leftarrow\mathsf{KeyGen}(1^{n}),\ket{\psi_{x}}\leftarrow\mathsf{StateGen}(1^{\lambda},x),x^{*}\leftarrow\mathcal{A}(\ket{\psi_{x}}^{\otimes t(n)})] (93)
=Pr[1Challenger:x𝒮n|ψx𝒞(1n,x)x(|ψxt(λ),1n)|ψx𝒞(1n,x)]\displaystyle=\Pr[1\leftarrow\mbox{Challenger}:\begin{array}[]{ll}x\leftarrow\mathcal{S}_{n}\\ \ket{\psi_{x}}\leftarrow\mathcal{C}(1^{n},x)\\ x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(\lambda)},1^{n})\\ \ket{\psi_{x^{*}}}\leftarrow\mathcal{C}(1^{n},x^{*})\end{array}] (98)
=x,xPr[x𝒮n]Pr[x(|ψxt(n))]|ψx|ψx|2\displaystyle=\sum_{x,x^{*}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]\absolutevalue{\bra{\psi_{x^{*}}}\ket{\psi_{x}}}^{2} (99)
x,x𝖦𝗈𝗈𝖽Pr[x𝒮n]Pr[x(|ψxt(n))]|ψx|ψx|2\displaystyle\geq\sum_{x,x^{*}\in\mathsf{Good}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]\absolutevalue{\bra{\psi_{x}}\ket{\psi_{x^{*}}}}^{2} (100)
x,x𝖦𝗈𝗈𝖽Pr[x𝒮n]Prx(|ψxt(n))(11/ϵ(n))\displaystyle\geq\sum_{x,x^{*}\in\mathsf{Good}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})](1-1/\epsilon(n)) (101)
(11/δ(n))(11/ϵ(n))\displaystyle\geq(1-1/\delta(n))(1-1/\epsilon(n)) (102)

for infinitely many nn\in\mathbb{N}. This contradicts that Σ\Sigma is OWSG with pure state outputs, which completes the proof.

Average-case hardness of quantum state learning \rightarrow OWSGs

Let 𝒞\mathcal{C} be a quantum polynomial-time algorithm that takes 1n1^{n} and x𝒳n={0,1}poly(n)x\in\mathcal{X}_{n}=\{0,1\}^{{\mathrm{poly}}(n)} and outputs |ψx\ket{\psi_{x}}, and let 𝒮n\mathcal{S}_{n} be a sampleable distribution over 𝒳n\mathcal{X}_{n} and let ϵ\epsilon and δ\delta be polynomials such that for all polynomials tt and all QPT algorithms 𝒜\mathcal{A},

Pr[F(|ϕx,|ψx)1/ϵ(n):x𝒮n|ψx𝒞(1n,x)x𝒜(|ψxt(n),1n)|ψx𝒞(1n,x)]1/δ(n)\displaystyle\Pr[\begin{array}[]{ll}F(\ket{\phi_{x^{*}}},\ket{\psi_{x}})\geq 1/\epsilon(n)\end{array}:\begin{array}[]{ll}x\leftarrow\mathcal{S}_{n}\\ \ket{\psi_{x}}\leftarrow\mathcal{C}(1^{n},x)\\ x^{*}\leftarrow\mathcal{A}(\ket{\psi_{x}}^{\otimes t(n)},1^{n})\\ \ket{\psi_{x^{*}}}\leftarrow\mathcal{C}(1^{n},x^{*})\end{array}]\leq 1/\delta(n) (108)

for infinitely many nn\in\mathbb{N}. Let 𝖲𝖺𝗆𝗉\mathsf{Samp} be an algorithm that takes 1n1^{n} and perfectly approximates 𝒮n\mathcal{S}_{n}.

We describe the construction of OWSGs (𝖪𝖾𝗒𝖦𝖾𝗇,𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇)(\mathsf{KeyGen},\mathsf{StateGen}):

𝖪𝖾𝗒𝖦𝖾𝗇(1n)\mathsf{KeyGen}(1^{n}):

  • 𝖪𝖾𝗒𝖦𝖾𝗇(1n)\mathsf{KeyGen}(1^{n}) runs 𝖲𝖺𝗆𝗉(1n)\mathsf{Samp}(1^{n}) and obtains x𝒳nx\in\mathcal{X}_{n}.

𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1n,x)\mathsf{StateGen}(1^{n},x):

  • 𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1n,x)\mathsf{StateGen}(1^{n},x) outputs the output of 𝒞(1n,x)\mathcal{C}(1^{n},x).

For contradiction, assume that there exist polynomials tt, and pp and a QPT algorithm \mathcal{B} such that

1δ(n)+1ϵ(n)1δ(n)ϵ(n)=1/p(n)<x,xPr[x𝒮n]Pr[x(|ψxt(n))]|ψx|ψx|2\displaystyle\frac{1}{\delta(n)}+\frac{1}{\epsilon(n)}-\frac{1}{\delta(n)\cdot\epsilon(n)}=1/p(n)<\sum_{x,x^{*}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]\absolutevalue{\bra{\psi_{x^{*}}}\ket{\psi_{x}}}^{2} (109)

for infinitely many nn\in\mathbb{N}. For some polynomial, let us define

𝖦𝗈𝗈𝖽{(x,x):1ϵ(n)<|ψx|ψx|21}\displaystyle\mathsf{Good}\coloneqq\left\{(x,x^{*}):\frac{1}{\epsilon(n)}<\absolutevalue{\bra{\psi_{x}}\ket{\psi_{x^{*}}}}^{2}\leq 1\right\} (110)
𝖡𝖺𝖽{(x,x):|ψx|ψx|21ϵ(n)}\displaystyle\mathsf{Bad}\coloneqq\left\{(x,x^{*}):\absolutevalue{\bra{\psi_{x}}\ket{\psi_{x^{*}}}}^{2}\leq\frac{1}{\epsilon(n)}\right\} (111)

Now, we have

1/p(n)\displaystyle 1/p(n)\leq x,xPr[x𝒮n]Pr[x(|ψxt(n))]|ψx|ψx|2\displaystyle\sum_{x,x^{*}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]\absolutevalue{\bra{\psi_{x^{*}}}\ket{\psi_{x}}}^{2} (112)
=x,x𝖦𝗈𝗈𝖽Pr[x𝒮n]Pr[x(|ψxt(n))]|ψx|ψx|2+x,x𝖡𝖺𝖽Pr[x𝒮n]Pr[x(|ψxt(n))]|ψx|ψx|2\displaystyle=\sum_{x,x^{*}\in\mathsf{Good}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]\absolutevalue{\bra{\psi_{x^{*}}}\ket{\psi_{x}}}^{2}+\sum_{x,x^{*}\in\mathsf{Bad}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]\absolutevalue{\bra{\psi_{x^{*}}}\ket{\psi_{x}}}^{2} (113)
x,x𝖦𝗈𝗈𝖽Pr[x𝒮n]Pr[x(|ψxt(n))]+x,x𝖡𝖺𝖽Pr[x𝒮n]Pr[x(|ψxt(n))]1ϵ(n)\displaystyle\leq\sum_{x,x^{*}\in\mathsf{Good}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]+\sum_{x,x^{*}\in\mathsf{Bad}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]\cdot\frac{1}{\epsilon(n)} (114)
=1ϵ(n)+(11ϵ(n))x,x𝖦𝗈𝗈𝖽Pr[x𝒮n]Pr[x(|ψxt(n))]\displaystyle=\frac{1}{\epsilon(n)}+\left(1-\frac{1}{\epsilon(n)}\right)\sum_{x,x^{*}\in\mathsf{Good}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})] (115)

This implies that

ϵ(n)p(n)p(n)(ϵ(n)1)x,x𝖦𝗈𝗈𝖽Pr[x𝒮n]Pr[x(|ψxt(n))].\displaystyle\frac{\epsilon(n)-p(n)}{p(n)\cdot(\epsilon(n)-1)}\leq\sum_{x,x^{*}\in\mathsf{Good}}\Pr[x\leftarrow\mathcal{S}_{n}]\Pr[x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)})]. (116)

This means that

Pr[F(|ψx,|ψx)1ϵ(n):x𝒮n|ψx𝒞(1n,x)x(|ψxt(n),1n)|ψx𝒞(1n,x)]>ϵ(n)p(n)p(n)(ϵ(n)1)=1/δ(n)\displaystyle\Pr[\begin{array}[]{ll}F(\ket{\psi_{x^{*}}},\ket{\psi_{x}})\geq\frac{1}{\epsilon(n)}\end{array}:\begin{array}[]{ll}x\leftarrow\mathcal{S}_{n}\\ \ket{\psi_{x}}\leftarrow\mathcal{C}(1^{n},x)\\ x^{*}\leftarrow\mathcal{B}(\ket{\psi_{x}}^{\otimes t(n)},1^{n})\\ \ket{\psi_{x^{*}}}\leftarrow\mathcal{C}(1^{n},x^{*})\end{array}]>\frac{\epsilon(n)-p(n)}{p(n)\cdot(\epsilon(n)-1)}=1/\delta(n) (122)

for infinitely many nn\in\mathbb{N}, which is a contradiction. Therefore, (𝖪𝖾𝗒𝖦𝖾𝗇,𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇)(\mathsf{KeyGen},\mathsf{StateGen}) satisfies

k,kPr[k𝖪𝖾𝗒𝖦𝖾𝗇(1n)]Pr[k𝒜(|ψkt(n))]|ψk|ψk|21/p(n)\displaystyle\sum_{k,k^{*}}\Pr[k\leftarrow\mathsf{KeyGen}^{*}(1^{n})]\Pr[k^{*}\leftarrow\mathcal{A}(\ket{\psi_{k}}^{\otimes t(n)})]\absolutevalue{\bra{\psi_{k}}\ket{\psi_{k^{*}}}}^{2}\leq 1/p(n) (123)

for all polynomials tt and infinitely many nn\in\mathbb{N}. As shown in [MY22a], from (𝖪𝖾𝗒𝖦𝖾𝗇,𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇)(\mathsf{KeyGen},\mathsf{StateGen}), we can construct (𝖪𝖾𝗒𝖦𝖾𝗇,𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇)(\mathsf{KeyGen}^{*},\mathsf{StateGen}^{*}) such that for all polynomials tt,

k,kPr[k𝖪𝖾𝗒𝖦𝖾𝗇(1n)]Pr[k𝒜(|ψkt(n))]|ψk|ψk|2𝗇𝖾𝗀𝗅(n),\displaystyle\sum_{k,k^{*}}\Pr[k\leftarrow\mathsf{KeyGen}^{*}(1^{n})]\Pr[k^{*}\leftarrow\mathcal{A}(\ket{\psi_{k}}^{\otimes t(n)})]\absolutevalue{\bra{\psi_{k}}\ket{\psi_{k^{*}}}}^{2}\leq{\mathsf{negl}}(n), (124)

which completes the proof. ∎

5.2 Average-Case Hardness of Learning Mixed State from EFI

Theorem 5.3.

If there exists EFI, then there exists a quantum polynomial-time algorithm with mixed states, which is hard on average in the sense of Definition 3.3.

Remark 5.4.

This result works only for "proper" quantum state learning, and it is unclear how to show the hardness of improperly learning mixed states from EFI as far as we understand.

Theorem 5.3 straightforwardly follows from the definition. We describe the proof for clarity.

Proof of Theorem 5.3.

Let (𝖪𝖾𝗒𝖦𝖾𝗇,𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇)(\mathsf{KeyGen},\mathsf{StateGen}) be a SV-SI-OWSGs such that 𝖳𝖣(ψn,x,ψn,x)1𝗇𝖾𝗀𝗅(n)\mathsf{TD}(\psi_{n,x},\psi_{n,x^{*}})\geq 1-{\mathsf{negl}}(n) for any xxx\neq x^{*}, where ψn,x𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1n,x)\psi_{n,x}\leftarrow\mathsf{StateGen}(1^{n},x). It is known that EFI is equivalent to SV-SI-OWSGs. Therefore, it is sufficient from SV-SI-OWSGs (𝖪𝖾𝗒𝖦𝖾𝗇,𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇)(\mathsf{KeyGen},\mathsf{StateGen}) to construct a quantum polynomial-time algorithm 𝒞\mathcal{C} such that there exists a sampleable distribution 𝒮n\mathcal{S}_{n} over 𝒳n\mathcal{X}_{n} and polynomials ϵ\epsilon and δ\delta such that for all polynomials tt and all QPT algorithms 𝒜\mathcal{A},

Pr[𝖳𝖣(ψx,ψy)1/ϵ(n):x𝒮nψx𝒞(x,1n)y𝒜(𝒞,ψxt(n),1n)ψy𝒞(y,1n)]1/δ(n)\displaystyle\Pr[\begin{array}[]{ll}\mathsf{TD}(\psi_{x},\psi_{y})\leq 1/\epsilon(n)\end{array}:\begin{array}[]{ll}x\leftarrow\mathcal{S}_{n}\\ \psi_{x}\leftarrow\mathcal{C}(x,1^{n})\\ y\leftarrow\mathcal{A}(\mathcal{C},\psi_{x}^{\otimes t(n)},1^{n})\\ \psi_{y}\leftarrow\mathcal{C}(y,1^{n})\end{array}]\leq 1/\delta(n) (130)

for all sufficiently large nn\in\mathbb{N}.

Let us describe 𝒞\mathcal{C} and 𝒮n\mathcal{S}_{n}.

The description of 𝒮n\mathcal{S}_{n}:

  • 𝒮n\mathcal{S}_{n} outputs the output of 𝖪𝖾𝗒𝖦𝖾𝗇(1n)\mathsf{KeyGen}(1^{n}).

The description of 𝒞\mathcal{C}:

  • 𝒞(1n,x)\mathcal{C}(1^{n},x) outputs the output of 𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(x,1n)\mathsf{StateGen}(x,1^{n}).

For contradiction, assume that for all polynomials ϵ\epsilon and δ\delta, there exists a QPT learner \mathcal{B} and a polynomial tt such that

Pr[𝖳𝖣(ψy,ψx)1/ϵ(n)hn𝒮n:x𝒮nψx𝒞(x,1n)y(𝒞,ψxt(n),1n)ψy𝒞(y,1n)]>1/δ(n)\displaystyle\Pr[\begin{array}[]{ll}&\mathsf{TD}(\psi_{y},\psi_{x})\leq 1/\epsilon(n)\\ &h_{n}\in\mathcal{S}_{n}\end{array}:\begin{array}[]{ll}&x\leftarrow\mathcal{S}_{n}\\ &\psi_{x}\leftarrow\mathcal{C}(x,1^{n})\\ &y\leftarrow\mathcal{B}(\mathcal{C},\psi_{x}^{\otimes t(n)},1^{n})\\ &\psi_{y}\leftarrow\mathcal{C}(y,1^{n})\end{array}]>1/\delta(n) (137)

for infinitely many nn\in\mathbb{N}. Now, by using \mathcal{B}, we construct a QPT adversary 𝒜\mathcal{A} that breaks the security of the SV-SI-OWSG scheme Σ\Sigma as follows.

  1. 1.

    𝒜\mathcal{A} receives sufficiently many copies of ψx\psi_{x}, where x𝖪𝖾𝗒𝖦𝖾𝗇(1λ)x\leftarrow\mathsf{KeyGen}(1^{\lambda}) and ψx𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(x)\psi_{x}\leftarrow\mathsf{StateGen}(x).

  2. 2.

    𝒜\mathcal{A} passes them to \mathcal{B}, and receives xx^{*}.

  3. 3.

    𝒜\mathcal{A} sends xx^{*} to the challenger.

Let us define a family 𝖦𝗈𝗈𝖽\mathsf{Good}

𝖦𝗈𝗈𝖽{(x,x):𝖳𝖣(ψx,ψx)1/ϵ(n)}.\displaystyle\mathsf{Good}\coloneqq\{(x,x^{*}):\mathsf{TD}(\psi_{x},\psi_{x^{*}})\leq 1/\epsilon(n)\}. (138)

From the definition of \mathcal{B}, we have

1/δ(n)<\displaystyle 1/\delta(n)< (x,x)𝖦𝗈𝗈𝖽Pr[x𝖪𝖾𝗒𝖦𝖾𝗇(1n)]Pr[x(ψxt(n))]\displaystyle\sum_{(x,x^{*})\in\mathsf{Good}}\Pr[x\leftarrow\mathsf{KeyGen}(1^{n})]\Pr[x^{*}\leftarrow\mathcal{B}(\psi_{x}^{\otimes t(n)})] (139)
=xPr[x𝖪𝖾𝗒𝖦𝖾𝗇(1n)]Pr[x(ψxt(n))]\displaystyle=\sum_{x}\Pr[x\leftarrow\mathsf{KeyGen}(1^{n})]\Pr[x\leftarrow\mathcal{B}(\psi_{x}^{\otimes t(n)})] (140)

for infinitely many nn\in\mathbb{N}. Here, in the second equation, we have used that for all xxx\neq x^{*}

𝖳𝖣(ψn,x,ψn,x)1𝗇𝖾𝗀𝗅(n).\displaystyle\mathsf{TD}(\psi_{n,x},\psi_{n,x^{*}})\geq 1-{\mathsf{negl}}(n). (141)

Now, we compute the winning probability of the adversary 𝒜\mathcal{A} as follows.

Pr[x𝒜(ψxt(n)):x𝖪𝖾𝗒𝖦𝖾𝗇(1n),ψx𝖲𝗍𝖺𝗍𝖾𝖦𝖾𝗇(1n,x)]\displaystyle\Pr[x\leftarrow\mathcal{A}(\psi_{x}^{\otimes t(n)}):x\leftarrow\mathsf{KeyGen}(1^{n}),\psi_{x}\leftarrow\mathsf{StateGen}(1^{n},x)] (142)
=xPr[x𝖪𝖾𝗒𝖦𝖾𝗇(1n)]Pr[x𝒜(ψxt(n))]\displaystyle=\sum_{x}\Pr[x\leftarrow\mathsf{KeyGen}(1^{n})]\Pr[x\leftarrow\mathcal{A}(\psi_{x}^{\otimes t(n)})] (143)
=xPr[x𝖪𝖾𝗒𝖦𝖾𝗇(1n)]Pr[x(ψxt(n))]>1/δ(n)\displaystyle=\sum_{x}\Pr[x\leftarrow\mathsf{KeyGen}(1^{n})]\Pr[x\leftarrow\mathcal{B}(\psi_{x}^{\otimes t(n)})]>1/\delta(n) (144)

for infinitely many nn\in\mathbb{N}. This contradicts that Σ\Sigma satisfies the security of SV-SI-OWSGs, which completes the proof. ∎

Acknowledgements. We appreciate Tomoyuki Morimae and Yuki Shirakawa for illuminating discussions. TH is supported by JSPS research fellowship and by JSPS KAKENHI No. JP22J21864.

References

  • [Aar07] Scott Aaronson. The learnability of quantum states. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 463(2088):3089–3114, September 2007.
  • [Aar18] Scott Aaronson. Shadow tomography of quantum states. In Ilias Diakonikolas, David Kempe, and Monika Henzinger, editors, 50th ACM STOC, pages 325–338. ACM Press, June 2018.
  • [ACH+19] Scott Aaronson, Xinyi Chen, Elad Hazan, Satyen Kale, and Ashwin Nayak. Online learning of quantum states. Journal of Statistical Mechanics: Theory and Experiment, 2019(12):124019, December 2019.
  • [AdW17] Srinivasan Arunachalam and Ronald de Wolf. Optimal quantum sample complexity of learning algorithms. In Proceedings of the 32nd Computational Complexity Conference, CCC ’17, Dagstuhl, DEU, 2017. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik.
  • [AE07] Andris Ambainis and Joseph Emerson. Quantum t-designs: t-wise independence in the quantum world. In Twenty-Second Annual IEEE Conference on Computational Complexity (CCC’07), pages 129–140, 2007.
  • [AS07] Alp Atıcı and Rocco A. Servedio. Quantum algorithms for learning and testing juntas. Quantum Information Processing, 6(5):323–348, September 2007.
  • [AW92] Naoki Abe and Manfred K Warmuth. On the computational complexity of approximating distributions by probabilistic automata. Machine Learning, 9, 1992.
  • [BCKM21] James Bartusek, Andrea Coladangelo, Dakshita Khurana, and Fermi Ma. On the round complexity of secure quantum computation. In Tal Malkin and Chris Peikert, editors, CRYPTO 2021, Part I, volume 12825 of LNCS, pages 406–435, Virtual Event, August 2021. Springer, Heidelberg.
  • [BCQ23] Zvika Brakerski, Ran Canetti, and Luowen Qian. On the computational hardness needed for quantum cryptography. In Yael Tauman Kalai, editor, 14th Innovations in Theoretical Computer Science Conference, ITCS 2023, January 10-13, 2023, MIT, Cambridge, Massachusetts, USA, volume 251 of LIPIcs, pages 24:1–24:21. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2023.
  • [BEHW87] Anselm Blumer, Andrzej Ehrenfeucht, David Haussler, and Manfred K. Warmuth. Occam’s razor. Information Processing Letters, 24(6):377–380, 1987.
  • [BEM+23] John Bostanci, Yuval Efron, Tony Metger, Alexander Poremba, Luowen Qian, and Henry Yuen. Unitary complexity and the uhlmann transformation problem. arXiv:2306.13073, 2023. https://arxiv.org/abs/2306.13073.
  • [BFKL94] Avrim Blum, Merrick L. Furst, Michael J. Kearns, and Richard J. Lipton. Cryptographic primitives based on hard learning problems. In Douglas R. Stinson, editor, CRYPTO’93, volume 773 of LNCS, pages 278–291. Springer, Heidelberg, August 1994.
  • [BJ95] Nader H. Bshouty and Jeffrey C. Jackson. Learning dnf over the uniform distribution using a quantum example oracle. In Proceedings of the Eighth Annual Conference on Computational Learning Theory, COLT ’95, page 118–127, New York, NY, USA, 1995. Association for Computing Machinery.
  • [BO21] Costin Bădescu and Ryan O’Donnell. Improved quantum data analysis. STOC 2021, page 1398–1411, New York, NY, USA, 2021. Association for Computing Machinery.
  • [BP92] Raymond Board and Leonard Pitt. On the necessity of occam algorithms. Theoretical Computer Science, 100(1):157–184, 1992.
  • [CGG+23] Bruno Cavalar, Eli Goldin, Matthew Gray, Peter Hall, Yanyi Liu, and Angelos Pelecanos. On the computational hardness of quantum one-wayness. arXiv:2312.08363, 2023.
  • [CHL+22] Xinyi Chen, Elad Hazan, Tongyang Li, Zhou Lu, Xinzhao Wang, and Rui Yang. Adaptive online learning of quantum states. arXiv:2206.00220, 2022.
  • [CHY16] Hao-Chung Cheng, Min-Hsiu Hsieh, and Ping-Cheng Yeh. The learnability of unknown quantum measurements. Quantum Info. Comput., 16(7–8):615–656, May 2016.
  • [CL21] Kai-Min Chung and Han-Hsuan Lin. Sample Efficient Algorithms for Learning Quantum Channels in PAC Model and the Approximate State Discrimination Problem. In Min-Hsiu Hsieh, editor, 16th Conference on the Theory of Quantum Computation, Communication and Cryptography (TQC 2021), volume 197 of Leibniz International Proceedings in Informatics (LIPIcs), pages 3:1–3:22, Dagstuhl, Germany, 2021. Schloss Dagstuhl – Leibniz-Zentrum für Informatik.
  • [CLS24] Nai-Hui Chia, Daniel Liang, and Fang Song. Quantum state learning implies circuit lower bounds. arXiv:2405.10242, 2024.
  • [DLSS14] Amit Daniely, Nati Linial, and Shai Shalev-Shwartz. From average case complexity to improper learning complexity. In Proceedings of the Forty-Sixth Annual ACM Symposium on Theory of Computing, STOC ’14, page 441–448, New York, NY, USA, 2014. Association for Computing Machinery.
  • [DSS16] Amit Daniely and Shai Shalev-Shwartz. Complexity theoretic limitations on learning dnf’s. In Vitaly Feldman, Alexander Rakhlin, and Ohad Shamir, editors, 29th Annual Conference on Learning Theory, volume 49 of Proceedings of Machine Learning Research, pages 815–830, Columbia University, New York, New York, USA, 23–26 Jun 2016. PMLR.
  • [DV21] Amit Daniely and Gal Vardi. From local pseudorandom generators to hardness of learning. In Mikhail Belkin and Samory Kpotufe, editors, Proceedings of Thirty Fourth Conference on Learning Theory, volume 134 of Proceedings of Machine Learning Research, pages 1358–1394. PMLR, 15–19 Aug 2021.
  • [FR99] Lance Fortnow and John Rogers. Complexity limitations on quantum computation. Journal of Computer and System Sciences, 59(2):240–252, 1999.
  • [GGM86] Oded Goldreich, Shafi Goldwasser, and Silvio Micali. How to construct random functions. Journal of the ACM, 33(4):792–807, 1986.
  • [GIKL23a] Sabee Grewal, Vishnu Iyer, William Kretschmer, and Daniel Liang. Efficient learning of quantum states prepared with few non-clifford gates. arXiv:2305.13409, 2023.
  • [GIKL23b] Sabee Grewal, Vishnu Iyer, William Kretschmer, and Daniel Liang. Efficient learning of quantum states prepared with few non-clifford gates ii: Single-copy measurements. arXiv:2308.07175, 2023.
  • [GKZ19] Alex B. Grilo, Iordanis Kerenidis, and Timo Zijlstra. Learning-with-errors problem is easy with quantum samples. Physical Review A, 99(3), March 2019.
  • [GL22] Aravind Gollakota and Daniel Liang. On the Hardness of PAC-learning Stabilizer States with Noise. Quantum, 6:640, February 2022.
  • [GLSV21] Alex B. Grilo, Huijia Lin, Fang Song, and Vinod Vaikuntanathan. Oblivious transfer is in MiniQCrypt. In Anne Canteaut and François-Xavier Standaert, editors, EUROCRYPT 2021, Part II, volume 12697 of LNCS, pages 531–561. Springer, Heidelberg, October 2021.
  • [HHJ+16] Jeongwan Haah, Aram W. Harrow, Zhengfeng Ji, Xiaodi Wu, and Nengkun Yu. Sample-optimal tomography of quantum states. In Proceedings of the Forty-Eighth Annual ACM Symposium on Theory of Computing, STOC ’16, page 913–925, New York, NY, USA, 2016. Association for Computing Machinery.
  • [HILL99] Johan Håstad, Russell Impagliazzo, Leonid A. Levin, and Michael Luby. A pseudorandom generator from any one-way function. SIAM Journal on Computing, 28(4):1364–1396, 1999.
  • [Hir22] Shuichi Hirahara. Np-hardness of learning programs and partial mcsp. 2022 IEEE 63rd Annual Symposium on Foundations of Computer Science (FOCS), pages 968–979, 2022.
  • [HKP20] Hsin-Yuan Huang, Richard Kueng, and John Preskill. Predicting many properties of a quantum system from very few measurements. Nature Physics, 16(10):1050–1057, June 2020.
  • [HLB+24] Hsin-Yuan Huang, Yunchao Liu, Michael Broughton, Isaac Kim, Anurag Anshu, Zeph Landau, and Jarrod R. McClean. Learning shallow quantum circuits. arXiv:2401.10095, 2024.
  • [HN22] Shuichi Hirahara and Mikito Nanashima. On worst-case learning in relativized heuristica. In 2021 IEEE 62nd Annual Symposium on Foundations of Computer Science (FOCS), pages 751–758, 2022.
  • [HN23] Shuichi Hirahara and Mikito Nanashima. Learning in pessiland via inductive inference. In 2023 IEEE 64th Annual Symposium on Foundations of Computer Science (FOCS), pages 447–457, 2023.
  • [IL90] R. Impagliazzo and L.A Levin. No better ways to generate hard NP instances than picking uniformly at random. In Proceedings [1990] 31st Annual Symposium on Foundations of Computer Science, pages 812–821 vol.2, 1990.
  • [JLS18] Zhengfeng Ji, Yi-Kai Liu, and Fang Song. Pseudorandom quantum states. In Hovav Shacham and Alexandra Boldyreva, editors, CRYPTO 2018, Part III, volume 10993 of LNCS, pages 126–152. Springer, Heidelberg, August 2018.
  • [KMR+94] Michael Kearns, Yishay Mansour, Dana Ron, Ronitt Rubinfeld, Robert E. Schapire, and Linda Sellie. On the learnability of discrete distributions. In Proceedings of the Twenty-Sixth Annual ACM Symposium on Theory of Computing, STOC ’94, page 273–282, New York, NY, USA, 1994. Association for Computing Machinery.
  • [Kre21] William Kretschmer. Quantum Pseudorandomness and Classical Complexity. In Min-Hsiu Hsieh, editor, 16th Conference on the Theory of Quantum Computation, Communication and Cryptography (TQC 2021), volume 197 of Leibniz International Proceedings in Informatics (LIPIcs), pages 2:1–2:20, Dagstuhl, Germany, 2021. Schloss Dagstuhl – Leibniz-Zentrum für Informatik.
  • [KT24] Dakshita Khurana and Kabir Tomer. Commitments from quantum one-wayness. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, STOC 2024, page 968–978, New York, NY, USA, 2024. Association for Computing Machinery.
  • [LMW24] Alex Lombardi, Fermi Ma, and John Wright. A one-query lower bound for unitary synthesis and breaking quantum cryptography. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, STOC 2024, page 979–990, New York, NY, USA, 2024. Association for Computing Machinery.
  • [Mon17] Ashley Montanaro. Learning stabilizer states by bell sampling. arXiv:1707.04012, 2017.
  • [MY22a] Tomoyuki Morimae and Takashi Yamakawa. One-wayness in quantum cryptography. Cryptology ePrint Archive, Paper 2022/1336, 2022.
  • [MY22b] Tomoyuki Morimae and Takashi Yamakawa. Quantum commitments and signatures without one-way functions. In Yevgeniy Dodis and Thomas Shrimpton, editors, Advances in Cryptology – CRYPTO 2022, pages 269–295, Cham, 2022. Springer Nature Switzerland.
  • [MY23] Tony Metger and Henry S. Yuen. stateQIP = statePSPACE. 2023 IEEE 64th Annual Symposium on Foundations of Computer Science (FOCS), pages 1349–1356, 2023.
  • [NE19] Moni Naor and Yogev Eylon. Bloom filters in adversarial environments. ACM Trans. Algorithms, 15(3), jun 2019.
  • [NR06] Moni Naor and Guy N. Rothblum. Learning to impersonate. In Proceedings of the 23rd International Conference on Machine Learning, ICML ’06, page 649–656, New York, NY, USA, 2006. Association for Computing Machinery.
  • [OS17] Igor C. Carboni Oliveira and Rahul Santhanam. Conspiracies Between Learning Algorithms, Circuit Lower Bounds, and Pseudorandomness. In Ryan O’Donnell, editor, 32nd Computational Complexity Conference (CCC 2017), volume 79 of Leibniz International Proceedings in Informatics (LIPIcs), pages 18:1–18:49, Dagstuhl, Germany, 2017. Schloss Dagstuhl – Leibniz-Zentrum für Informatik.
  • [OW16] Ryan O’Donnell and John Wright. Efficient quantum tomography. In Proceedings of the Forty-Eighth Annual ACM Symposium on Theory of Computing, STOC ’16, page 899–912, New York, NY, USA, 2016. Association for Computing Machinery.
  • [PV88] Leonard Pitt and Leslie G. Valiant. Computational limitations on learning from examples. J. ACM, 35(4):965–984, oct 1988.
  • [Reg09] Oded Regev. On lattices, learning with errors, random linear codes, and cryptography. J. ACM, 56(6):34:1–34:40, 2009.
  • [Roc18] Andrea Rocchetto. Stabiliser states are efficiently pac-learnable. Quantum Info. Comput., 18(7–8):541–552, jun 2018.
  • [RY22] Gregory Rosenthal and Henry Yuen. Interactive Proofs for Synthesizing Quantum States and Unitaries. In Mark Braverman, editor, 13th Innovations in Theoretical Computer Science Conference (ITCS 2022), volume 215 of Leibniz International Proceedings in Informatics (LIPIcs), pages 112:1–112:4, Dagstuhl, Germany, 2022. Schloss Dagstuhl – Leibniz-Zentrum für Informatik.
  • [Sch90] Robert E Schapire. The strength of weak learnability. Machine Learning, 5(2):197–227, 1990.
  • [Sen06] Pranab Sen. Random measurement bases, quantum state distinction and applications to the hidden subgroup problem. In 21st Annual IEEE Conference on Computational Complexity (CCC’06), pages 14 pp.–287, 2006.
  • [SSG23] Wilfred Salmon, Sergii Strelchuk, and Tom Gur. Provable advantage in quantum pac learning. arXiv:2309.10887, 2023.
  • [Vad17] Salil Vadhan. On learning vs. refutation. In Satyen Kale and Ohad Shamir, editors, Proceedings of the 2017 Conference on Learning Theory, volume 65 of Proceedings of Machine Learning Research, pages 1835–1848. PMLR, 07–10 Jul 2017.
  • [Val84] L. G. Valiant. A theory of the learnable. Commun. ACM, 27(11):1134–1142, nov 1984.
  • [Yan22] Jun Yan. General properties of quantum bit commitments (extended abstract). In Shweta Agrawal and Dongdai Lin, editors, Advances in Cryptology – ASIACRYPT 2022, pages 628–657, Cham, 2022. Springer Nature Switzerland.
  • [ZLK+23] Haimeng Zhao, Laura Lewis, Ishaan Kannan, Yihui Quek, Hsin-Yuan Huang, and Matthias C. Caro. Learning quantum states and unitaries of bounded gate complexity. arXiv:2310.19882, 2023.