This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Parallel Quantum Annealing

Elijah Pelofske Georg Hahn Harvard T.H. Chan School of Public Health, Boston, MA 02115, USA [email protected] Hristo N. Djidjev Los Alamos National Laboratory CCS-3 Information Sciences, Los Alamos, NM 87545, USA Institute of Information and Communication Technologies, Bulgarian Academy of Sciences, Sofia, Bulgaria
Abstract

Quantum annealers of D-Wave Systems, Inc., offer an efficient way to compute high quality solutions of NP-hard problems. This is done by mapping a problem onto the physical qubits of the quantum chip, from which a solution is obtained after quantum annealing. However, since the connectivity of the physical qubits on the chip is limited, a minor embedding of the problem structure onto the chip is required. In this process, and especially for smaller problems, many qubits will stay unused. We propose a novel method, called parallel quantum annealing, to make better use of available qubits, wherein either the same or several independent problems are solved in the same annealing cycle of a quantum annealer, assuming enough physical qubits are available to embed more than one problem. Although the individual solution quality may be slightly decreased when solving several problems in parallel (as opposed to solving each problem separately), we demonstrate that our method may give dramatic speed-ups in terms of the Time-to-Solution (TTS) metric for solving instances of the Maximum Clique problem when compared to solving each problem sequentially on the quantum annealer. Additionally, we show that solving a single Maximum Clique problem using parallel quantum annealing reduces the TTS significantly.

1 Introduction

Quantum annealers manufactured by D-Wave Systems, Inc. are designed to compute approximate high-quality solutions of NP-hard problems using a process called quantum annealing. In particular, the annealers of D-Wave Systems are specialized quantum machines to minimize discrete functions that can be expressed in the form

f(x1,,xn)=i=1nhixi+i<jJijxixj,\displaystyle f(x_{1},\ldots,x_{n})=\sum_{i=1}^{n}h_{i}x_{i}+\sum_{i<j}J_{ij}x_{i}x_{j}, (1)

where the linear weights hih_{i}\in\mathbb{R} and the quadratic couplers JijJ_{ij}\in\mathbb{R}, i,j{1,,n}i,j\in\{1,\ldots,n\}, are specified by the user and define the problem under investigation. The variables xix_{i} are unknown and take two values only: if xi{0,1}x_{i}\in\{0,1\} the function in eq. (1) is called a quadratic binary optimization problem (QUBO), and if xi{1,+1}x_{i}\in\{-1,+1\} it is called an Ising problem. Both formulations are computationally equivalent.

The time evolution of any quantum system is characterized by an operator called the Hamiltonian. For the D-Wave quantum chip, it is given by

H(s)=A(s)2i=1nσix+B(s)2(i=1nhiσiz+ijJijσizσjz).\displaystyle H(s)=-\frac{A(s)}{2}\sum_{i=1}^{n}\sigma^{x}_{i}+\frac{B(s)}{2}\left(\sum_{i=1}^{n}h_{i}\sigma^{z}_{i}+\sum_{i\leq j}J_{ij}\sigma^{z}_{i}\sigma^{z}_{j}\right). (2)

This operator is interpreted as follows. The first term encodes an equal superposition of all states (making each bit string equally likely). The second term of eq. (2) encodes the problem to be solved, given in eq. (1), which is fully determined through its linear and quadratic couplers hih_{i} and JijJ_{ij}, respectively. During annealing, the quantum system slowly transitions from an equal superposition of all states to one whose ground state encodes the implemented problem to be solved. This is realized with the help of a so-called anneal path, given by the functions A(s)A(s) and B(s)B(s). At the start of the anneal, A(s)A(s) is large and B(s)B(s) is small, and during annealing A(s)A(s) decreases to zero while B(s)B(s) increases to some maximal value. During the annealing process, it is expected that the system stays in a ground state, thus allowing to read off a low-energy (optimal or near-optimal) solution of the implemented problem upon termination.

\begin{overpic}[width=173.44534pt]{figures/embeddings/LANL_2000Q_20.pdf} \put(46.0,-3.5){{(a)}} \end{overpic}\begin{overpic}[width=173.44534pt]{figures/embeddings/Advantage1.1_20.pdf} \put(46.0,-3.5){{(b)}} \end{overpic}
Figure 1: Clique of size 2020 embedded 1212 times into the hardware of DW 2000Q, which uses Chimera graph topology (a), and 6868 times into DW Advantage, which has a Pegasus hardware graph (b). The coloring is arbitrary and used to highlight the separation of different embeddings, with unused qubits and couplers in grey. Note that the actual embeddings do not make use of all of the physical couplers between the qubits used in the embedding; in theses figures we are showing the subgraphs induced by the qubits used in each embedding.

To minimize eq. (1) on the D-Wave quantum annealer, the following steps are typically required:

(i) After expressing the problem to be solved as a minimization problem of the type of eq. (1), one can represent eq. (1) as a problem graph PP itself having nn vertices, one for each variable xix_{i}, i{1,,n}i\in\{1,\ldots,n\}. Each vertex ii has a vertex weight hih_{i}. Each non-zero term JijxixjJ_{ij}x_{i}x_{j} becomes an edge between vertices ii and jj with edge weight JijJ_{ij}.

(ii) The problem graph PP is mapped onto the graph of qubit connectivity of the quantum annealer. Since the D-Wave annealers offer a fixed number of physical qubits whose connectivity usually does not match the one of the problem graph PP, a minor embedding of PP onto the qubit hardware graph is typically needed. In the embedding, it is usually the case that some logical qubits are represented by a set of physical qubits on the D-Wave hardware graph, called a chain. Chained qubits require the specification of a coupler weight as well, called the chain strength or chain weight, to incentivize them to take the same value during annealing. The number of chained qubits per logical qubit is called the chain length. Embedding the problem graph PP onto the hardware graph results in a subgraph of the hardware graph, denoted as PP^{\prime}.

(iii) D-Wave starts the annealing process by initializing all the qubits used in the embedding of PP^{\prime} in an equal superposition, from which the system slowly transitions to the user-specified QUBO or Ising problem while aiming to keep all qubits in the ground state. Thus a solution can be read off once the system is fully transitioned to the problem Hamiltonian.

(iv) Though they represent one logical qubit, chained hardware qubits are not guaranteed to take the same value after annealing (we call those broken chains). To arrive at a consistent solution, we need to unembed chains, meaning that each broken chain has to be assigned one value after annealing. There is no unique way to achieve this, though D-Wave offers several default methods for unembedding.

Since computing a minor embedding is computationally intensive, the annealer is oftentimes used with a fixed precomputed embedding of a complete graph KsK_{s} of size ss, which is the maximum size that can fit onto the qubit hardware. This has the advantage that any problem graph PP of ss vertices can be easily embedded into the annealer as it is a subgraph of KsK_{s}. However, using a fixed complete graph embedding also has disadvantages. First, using a fixed complete embeddding might lead to a poorer solution quality than using tailored embeddings [1]. This is due to increased problem complexity, as well as longer chain lengths that distort the original problem. Second, if the problem to be solved on D-Wave does not require all (or the majority) of the hardware qubits, the D-Wave hardware is used suboptimally since qubits remain unused during annealing that could be employed to solve other problems. This second case is aggravated by the fact that the largest complete embedding which fits onto the hardware does not use all of the available qubits.

In this contribution, we show that multiple instances of either the same or different problems can be embedded and subsequently solved in parallel on a single quantum annealing chip, while suffering from only small decrease in individual solution quality (measured by the probability of finding a ground state solution). Specifically, we propose a method called parallel quantum annealing for solving multiple problems simultaneously on a quantum annealer, which works as follows. Given a limit on the number of variables allowed per QUBO problem to be solved simultaneously, say ss\in\mathbb{N}, we compute a maximum number of embeddings, say kk\in\mathbb{N}, of complete graphs of order ss that can be placed on the quantum chip (see Figure 1). This will allow any kk QUBO problems with no more than ss variables each to be easily embedded into the QPU (quantum processing unit), since any problem graph of ss vertices can be embedded into a complete graph of order ss. We can then solve the resulting composite QUBO in a single call on the quantum annealer and decompose the returned solution into solutions of the individual QUBOs. This is justified since, when adding together independent QUBOs (that is when adding together functions of the form of eq. (1) that do not share any unknown variables), the solution of the composite QUBO will be the union of all individual ground states, and the bitstring yielding the minimum will be the concatenation of all individual bitstring solutions.

We apply our proposed parallel quantum annealing methodology to the Maximum Clique problem, an important NP-hard problem with applications in network analysis, bioinformatics, and computational chemistry [2, 3, 4]. For a graph G=(V,E)G=(V,E) with vertex set VV and edge set EE, a clique in GG is any subgraph CC of GG that is complete, i.e., there is an edge between each pair of vertices of CC. The Maximum Clique problem asks us to find a clique of maximum size. A formulation of the Maximum Clique problem in the form of eq. (1) is available in the literature [4, 5]. In this article, we will consider both the D-Wave 2000Q at Los Alamos National Laboratory (referred to as DW 2000Q), and the newer D-Wave Advantage_System 1.1 (referred to as DW Advantage) which we accessed through D-Wave Leap.

Multiple problems of the same type may need to be solved in various scenarios. For instance, decomposition methods [6] for solving optimization problems such as Maximum Clique allow one to divide an input problem that is too large to fit on the QPU into many smaller problems from which a solution of the original problem can be constructed. Methods in computer vision [7], genomics [8], and protein structure analysis [9] have been shown to lead to solving multiple Maximum Clique problems. Also, a hard instance of the same problem (e.g., Maximum Clique) may need to be solved multiple times on the quantum annealer in order to find an optimal or high-quality solution, and all such executions are independent and can be run in parallel with our proposed method.

The article is structured as follows. We start with a brief literature review in Section 1.1. Experimental results are reported in Section 2, followed by a discussion in Section 3. In Section 4 we highlight the rationale behind solving problems in parallel on the D-Wave hardware architecture, in particular the choice of the embedding. We also provide a generalization of the TTS (Time-To-Solution) formula we employ to measure the TTS metric for parallel and sequential problem solving.

1.1 Literature review

Ising spin glass models have been extensively studied in the literature, for instance with respect to the stability of the system, its phase transitions, and its magnetization distribution [10]. To find the ground state of such spin glass models, as well as for other problem Hamiltonians, quantum annealing has been shown to find ground states with higher probability than classical (thermal) simulated annealing [11].

The fact that a quantum system will stay near its ground state if its Hamiltonian varies slowly enough led to adiabatic quantum algorithms [12], for which it was hypothesized early that they might be able to outperform classical algorithms on instances of NP-complete problems. In particular, quantum annealing can be shown to converge to the ground state energy at a faster rate than its classical counterpart [13]. A comprehensive review on combinatorial optimization problems (such as the ones studied in the present article), spin glasses, and quantum annealing via adiabatic evolution is available in the literature [14].

There are very few published results yet concerning parallelism and quantum annealing, and none of them discusses the type of parallelism we propose. Some papers consider pairing quantum computing systems with modern HPC infrastructure [15]. The focus there lays on the design of macroarchitecture, microarchitecture, and programming models needed to integrate near-term quantum computers with HPC systems, and the near-term feasibility of such systems. They do not consider, however, parallelism at the level of a quantum device.

In a different context [16], the authors study the problem of simulating dynamical systems on a quantum annealer. Such systems are intrinsically sequential in the time component, which makes them difficult to simulate on a quantum computer. They propose to use the parallel in time idea from classical computing to reformulate a problem so that it can be solved as the task of finding a ground state of an Ising model. Such an Ising model is then solved on a quantum annealer. However, unlike our work, there is only one Ising problem solved at a time.

Another approach considers a framework built upon Field Programmable Gate Array chips (FPGA) in connection with probabilistic bits to simulate Gibbs samplers[17]. The authors use their algorithm on blocks of conditionally independent nodes of the input graph, which are updated in parallel using a quantum annealer. The authors remark that technically, all blocks have to be completed before some synchronization step is carried out; however, if not all blocks are completed, the network is still able to find exact ground states. However, the parallel architecture proposed by the authors [17] is not quantum.

To the best of our knowledge, we are the first to propose solving multiple independent problems in parallel on a single quantum device. In a previous contribution [18], the parallel quantum annealing methodology has been used in a specific manner as a way to solve 255255 small QUBOs (with a maximum of 44 variables each) simultaneously on the D-Wave 2000Q at Los Alamos National Laboratory with a relatively small number of anneals. In contrast to the previous work [18], the present article expands on the parallel quantum annealing idea by investigating larger problem sizes, considering the NP-hard Maximum Clique problem, and applying it to the new DW Advantage generation that uses a Pegasus qubit connection topology [19].

2 Experiments

In this section, we assess the proposed parallel quantum annealing technique by solving several Maximum Clique problems in the same D-Wave call. First, after two brief notes on the parameter setting (Section 2.1) and the Time-To-Solution metric (Section 2.2), we investigate in Section 2.3 the behavior of both the ground state probability and the TTS measure as a function of the problem size for various graph densities, and for both DW 2000Q and DW Advantage. Next, in Section 2.4, we compare parallel quantum annealing and sequential quantum annealing, again with respect to both ground state probability and the TTS measure. A comparison of parallel quantum annealing with the classical FMC solver [20] on the Maximum Clique problem is given in Section 2.5. We conclude with an assessment of how our method can help to solve a single hard problem by implementing it multiple times on the hardware in Section 2.6.

2.1 Parameter tuning

The D-Wave annealers offer a variety of tuning parameters that can be chosen by the user before each anneal. Those encompass the annealing time, the chain strength, anneal offsets, etc. We perform a grid search on the DW 2000Q in order to find reasonable values for both the chain strength and the annealing time, with the goal to minimize the QPU time while maximizing the probability of finding the ground state of each of the embedded problems.

In all experiments, we set the annealing time to 5050 microseconds, and compute the chain strength using the uniform torque compensation method included in the D-Wave Ocean SDK [21] with a UTC prefactor of 0.20.2. Lastly, We set the programming thermalization time to 0 microseconds, and the readout thermalization time to 0 microseconds. We use the same parameters for the experiments with DW Advantage as we do for the ones with DW 2000Q. However, the performance of DW Advantage might be potentially improved by additional parameter optimization.

2.2 Time-To-Solution for an ensemble of problems

Since quantum annealers are stochastic (and heuristic) solvers, they are not guaranteed to find the ground state (the global minimum) of a problem of the type of eq. (1) in each anneal. Therefore, to have a meaningful metric of accuracy, Time-To-Solution is used to quantify the expected time it takes to reach an optimum solution with a 9999 percent confidence [22, 23, 24, 25]. It is defined, in the sequential case, as

TTSseq=1A(TQPU+U)log(0.01)log(1p),\displaystyle\text{TTS}_{seq}=\frac{1}{A}(T_{\text{QPU}}+U)\frac{\log(0.01)}{\log(1-p)}, (3)

where pp is the probability of finding the ground state within a single anneal, TQPUT_{\text{QPU}} is the total D-Wave QPU time (specifically the qpu-access-time), and UU is the total CPU process time used to compute the unembedded solutions in seconds. Note that in eq. (3), we do not include the embedding time because we compute full clique embeddings, and then save and re-use those embeddings.

In the context of the present work, a generalization of the standard TTS measure is required since we solve KK problems of size NN simultaneously on the same D-Wave chip, using a total QPU processing time of TQPUT_{\text{QPU}} and AA anneals.

For each problem i{1,,K}i\in\{1,\ldots,K\} we solve simultaneously, we record the proportion of correct solutions pip_{i} among the AA anneals. We weigh those using the formula

pK=1Ki=1Kpi,p_{K}=\frac{1}{K}\sum_{i=1}^{K}p_{i},

where every pip_{i} must be non-zero. We generalize eq. (3) as

TTSens(K)=1A(TQPUK+U)log(0.01)log(1pK),\displaystyle\text{TTS}_{ens}(K)=\frac{1}{A}\left(\frac{T_{\text{QPU}}}{K}+U\right)\frac{\log(0.01)}{\log(1-p_{K})}, (4)

where UU is the total CPU unembedding time used to unembed the solutions of each of the KK problems. We refer to TTSens\text{TTS}_{ens} as the ensemble Time-To-Solution, since it captures the time to reach an optimal solution for each problem in a group of problems that are solved simultaneously.

2.3 Accuracy of parallel quantum annealing

\begin{overpic}[width=212.47617pt]{figures/LANL_2000Q/GSP_parallel_0_0_50_0.2.pdf} \put(46.0,-1.0){{(a)}} \end{overpic}\begin{overpic}[width=212.47617pt]{figures/Advantage1.1/GSP_parallel_0_0_50_0.2.pdf} \put(46.0,-1.0){{(b)}} \end{overpic}
Figure 2: Ground state probability (GSP) achieved by parallel quantum annealing for DW 2000Q (a) and for DW Advantage (b) as a function of the problem size NN. Graph densities range from 0.100.10 to 0.950.95 (color coding shown in the legend). Black solid line plots the mean parallel quantum annealing GSP as a function of problem size.

This section studies the accuracy of parallel quantum annealing per se. Using the parameters of Section 2.1, we investigate the probability with which a ground state is found as a function of the problem size. Figure 2 shows results for both DW 2000Q (a) and for DW Advantage (b). We observe that, as expected, DW 2000Q finds the ground state probability (GSP) more reliably for smaller problems, though the probability varies quite considerably with the graph density. Denser problem graphs are typically easier than sparser ones. For cliques of roughly size 4040 onward, DW 2000Q is unable to find the ground state. For DW Advantage we observe a similar picture. The GSP is higher for smaller problems than for larger ones, with an even bigger spread by graph density. As before, denser problem graphs are easier to solve reliably than sparser ones. As expected, DW Advantage is able to solve larger inputs, in particular it finds the ground state with appreciable probability of up to 40%40\% even for inputs of size 4040.

Something we noted in these results is that the parallel quantum annealing method never found all ground state solutions in a single anneal. Instead, different sets of ground state solutions were found with each anneal - and over a sequence of anneals we eventually find the solutions to all of the problems. Interestingly, the rate at which the ground state solutions were found was not the same across all problems - indeed it was heavily biased. This bias is likely due to differences in the embeddings.

Similarly to Figure 2, Figure 3 shows the ensemble TTS measure for parallel quantum annealing as a function of the problem size, and for various graph densities. For DW 2000Q (a), we observe that this TTS measure increases as the problem size increases, which is as expected. The reason for the ”jump” at around problem size 4040 is unknown, but the behavior of D-Wave could be related to an internal hardware issue (during the time frame in which some of this data was taken on DW 2000Q, around problem sizes 38-46, there was an initially undetected temperature increase from the typical 0.0150.015 Kelvin of the hardware, likely leading to a poorer solution quality). In accordance with Figure 2, dense problem inputs seem to be easier for D-Wave, and thus incur a lower TTS value. For DW Advantage (b), we observe a similar picture. Interestingly, for almost all problem sizes that can be solved on DW 2000Q (roughly up to size 4040), the DW 2000Q is actually the better solver with respect to the TTS metric. For sizes above 40, DW Advantage is the better solver with respect to TTS.

\begin{overpic}[width=212.47617pt]{figures/LANL_2000Q/TTS_parallel_QA_0_0_50_0.2.pdf} \put(46.0,-1.0){{(a)}} \end{overpic}\begin{overpic}[width=212.47617pt]{figures/Advantage1.1/TTS_parallel_QA_0_0_50_0.2.pdf} \put(46.0,-1.0){{(b)}} \end{overpic}
Figure 3: Ensemble TTS for parallel quantum annealing as a function of the problem size for DW 2000Q (a) and DW Advantage (b). Graph densities range from 0.100.10 to 0.950.95 (color coding shown in the legend). Black solid line plots the mean parallel quantum annealing ensemble TTS as a function of problem size.

2.4 Comparison of sequential to parallel quantum annealing

\begin{overpic}[width=212.47617pt]{figures/LANL_2000Q/compare_GSP_parallel_non_parallel0_0_50_0.2.pdf} \put(46.0,-1.0){{(a)}} \end{overpic}\begin{overpic}[width=212.47617pt]{figures/Advantage1.1/compare_GSP_parallel_non_parallel0_0_50_0.2.pdf} \put(46.0,-1.0){{(b)}} \end{overpic}
Figure 4: Difference between the ground state probabilities (GSP) of sequential and parallel quantum annealing as a function of the problem size for DW 2000Q (a) and DW Advantage (b). Values above zero indicate that sequential quantum annealing finds higher ground state probabilities than parallel quantum annealing. Averages indicated with a solid black line. Graph densities range from 0.100.10 to 0.950.95 (color coding shown in legend).

To compare the benefits and trade-offs of solving several problems in parallel on the D-Wave quantum annealers, we first investigate the difference between the GSP found by both sequential and parallel quantum annealing. Here, sequential quantum annealing refers to solving the problems separately one after the other on the D-Wave architecture. The GSP in the parallel case is simply defined per problem as the probability of finding its individual ground state. Figure 4 shows results for DW 2000Q (a) and for DW Advantage (b). We observe that, on average across the graph densities considered, solving problems simultaneously causes D-Wave to find a ground state slightly less frequently than when solving them sequentially. For DW Advantage, this phenomenon is even more pronounced.

The reduced GSP we observe for parallel quantum annealing can be explained as follows. First, the problem complexity (i.e., the number of variables and quadratic terms) increases when solving multiple problems as opposed to one, potentially causing difficulties for the D-Wave annealer. Second, multiple embeddings on the chip mean that there is a closer physical proximity between the qubits used per embedding, potentially causing increased leakage and interaction between the involved qubits. However, unlike TTS, the GSP metric does not account for the fact that we are finding solutions to multiple problems.

\begin{overpic}[width=212.47617pt]{figures/LANL_2000Q/compare_TTS_parallel_non_parallel_0_0_50_0.2.pdf} \put(46.0,-1.0){{(a)}} \end{overpic}\begin{overpic}[width=212.47617pt]{figures/Advantage1.1/compare_TTS_parallel_non_parallel_0_0_50_0.2.pdf} \put(46.0,-1.0){{(b)}} \end{overpic}
Figure 5: TTS speedup of using parallel quantum annealing compared to sequential quantum annealing as a function of the problem size for D-Wave 2000Q (a) and D-Wave Advantage 1.1 (b). The TTS speedup is computed as the quotient of the sequential and parallel TTS metrics. Mean values are shown by the solid black line. Graph densities range from 0.100.10 to 0.950.95 (color coding shown in the legend).

Next, Figure 5 compares the sequential and parallel quantum annealing efficiencies with respect to the TTS metric. The TTS speedup is computed as the quotient of the sequential and parallel TTS values. For DW 2000Q (a), we observe that the proposed parallel quantum annealing approach yields substantial speedups (an average of 152152-fold over all problem sizes considered) compared to solving the same problems sequentially on the D-Wave chip. A stratification by graph density is not observable in this case. As expected, the speedup is more pronounced for smaller problem sizes, as in this case more problems can be solved in parallel. For DW Advantage, a considerable speedup (on average around 2020-fold) is observed over all problem sizes considered.

Note that the CPU unembedding time increases when unembedding larger datasets; meaning that for larger backends (and more variables used per anneal) the total unembedding time will increase. Therefore, the differences in TTS between the two backends seen in Figure 5 are not only caused by the differences in GSP seen in Figure 4.

2.5 Comparison with a fast classical solver

\begin{overpic}[width=212.47617pt]{figures/LANL_2000Q/fmc_0_0_50_0.2.pdf} \put(46.0,-1.5){{(a)}} \end{overpic}\begin{overpic}[width=212.47617pt]{figures/Advantage1.1/fmc_0_0_50_0.2.pdf} \put(46.0,-1.5){{(b)}} \end{overpic}
Figure 6: Speedup of using parallel quantum annealing compared with the classical FMC solver as a function of the problem size for DW 2000Q (a) and DW Advantage (b). Log scale on the y-axis. Graph densities range from 0.100.10 to 0.950.95 (color coding shown in the legend).

We compare the performance of the proposed parallel quantum annealing approach with a classical solver for the Maximum Clique problem. We employ the fast maximum clique (FMC) solver [20]. FMC is run in exact mode throughout this experiment.

Figure 6 shows the speedup of parallel quantum annealing over FMC. For the probabilistic parallel QA, we measure ensemble TTS (see Section 2.2), whereas for the classical FMC solver, we measure CPU times (this is justified since TTS reduces to CPU time in the classical case). The speedup is computed as the quotient of the parallel TTS metric and the CPU process time used by FMC. For DW 2000Q (a), we observe that for both small problems (of any density) and for high densities up to problem size 4040, using parallel quantum annealing yields better TTS values than using FMC. In turn, FMC is superior to DW 2000Q for low density graphs. For problems exceeding size 4040, DW 2000Q is not the preferred choice (as seen in previous experiments), making FMC the faster choice for all densities. Interestingly, for DW Advantage (b), we observe that D-Wave is not superior to FMC apart from very high densities. In turn, the DW Advantage architecture does not suffer from a reduced accuracy for large problem sizes (compared to the DW 2000Q).

2.6 Improving a single hard problem’s TTS using parallel quantum annealing

Finally, we consider how parallel quantum annealing can be used to increase the GSP, and therefore TTS, of a single hard QUBO problem. Such problems need to be run sequentially many times on the quantum annealer to get even a single optimal solution. Our idea is to run multiple replicas (identical copies of the QUBO) simultaneously using parallel quantum annealing.

We compare sequential quantum annealing and parallel quantum annealing on DW Advantage for a fixed set of 1616 random graphs with N=40N=40 nodes and p=0.5p=0.5. Each problem is first solved sequentially, and afterwards solved in parallel by embedding it 1616 times onto the hardware. Results are shown in Figure 7, where we see that parallel quantum annealing can reduce TTS due to improved GSP. Over all problems considered, the average TTS speedup (the quotient of sequential QA TTS and parallel QA TTS) is 77, and the average GSP increase (the quotient of parallel QA GSP and sequential QA GSP) is 1010. Note that problem 0 had worse TTS in the parallel quantum annealing case. This is due to an increased cost in the CPU unembedding time in the parallel case.

To compute the TTS measure of the parallel implementation, we use eq. (3), and not eq. (4). This is due to the fact that there are not KK distinct problem being solved in parallel; instead, we are solving the same problem KK times during the same annealing cycle. Therefore, pp is simply the proportion of anneals that found the optimal solution. Note, however, that we can have cases where the ground state solution is found multiple times in the same anneal (we do not distinguish between finding the ground state once or more than once in the embedded problems).

\begin{overpic}[width=212.47617pt]{figures/same_problem/same_problem_TTS.pdf} \put(53.0,-1.5){{(a)}} \end{overpic}\begin{overpic}[width=212.47617pt]{figures/same_problem/same_problem_GSP.pdf} \put(53.0,-1.5){{(b)}} \end{overpic}
Figure 7: TTS (a) and GSP (b) for the 1515 test problems we consider, both for parallel quantum annealing (blue) and sequential quantum annealing (green). Log scale on the y-axis.

3 Discussion

This article investigates a proposed method called parallel quantum annealing, which allows one to solve many independent problems on a quantum annealer during the same annealing cycle. The idea is justified by the fact that independent QUBOs, that is those not sharing any variable, can be added together to yield a combined QUBO that preserves the bitstring solution of each, and has a new ground state probability that is the sum of the individual ground state probabilities.

We evaluate the proposed approach with respect to both accuracy and the TTS metric on both DW 2000Q and DW Advantage. We compare to sequential quantum annealing (i.e., solving the same problems sequentially on D-Wave), and the classical FMC solver. We demonstrate that while the solution accuracy is slightly decreased for simultaneously solved problems, parallel quantum annealing can yield a considerable speedup of up to two orders of magnitude. Interestingly, using the newer DW Advantage does not yield more pronounced speedups in TTS than the previous generation DW 2000Q.

Notably, parallel quantum annealing is different from classical parallelism in that parallel quantum annealing solves multiple problems on a single quantum chip, whereas classical parallel computing solves independent problems at the same time, but using different physical circuits. Therefore, for a sufficiently large quantum annealer (and assuming the solution quality does not suffer), a very large number of problems could be solved in a very short amount of time (determined by the annealing time and the number of anneals).

The number of qubits in new quantum annealers have been steadily increasing, but there has been a concern that higher error rates may prevent such machines from solving very large problems. Parallel quantum annealing may be an alternative to making future larger machines usable.

This work leaves scope for future research avenues: 1. This work only considers solving Maximum Clique problems simultaneously on D-Wave. Investigating other important NP-hard problems remains for future work. In particular, the QUBOs being solved in parallel do not have to stem from the same problems, and they do not need to be of equal size.
2. Optimizing the annealing parameters (such as annealing time, chain strengths, etc.) for DW Advantage might significantly improve its performance.
3. It remains to investigate why parallel quantum annealing unfairly finds the ground states of some problems at higher rates than others; and how this effect can be mitigated. A related question is how fairly parallel quantum annealing samples the degenerate ground states of the embedded problems.

4 Methods

This section discusses some of the elements of the proposed parallel quantum annealing technique, in particular the QUBO formulation of the Maximum Clique problem (Section 4.1), combining multiple small QUBOs into a single one (Section 4.2), the choice of the embedding (Section 4.3), and the specifics of the D-Wave execution (Section 4.4).

4.1 Maximum Clique QUBO

The Maximum Clique QUBO we use is given by [5, 26]

HMC=AvVxv+B(u,v)E¯xuxv,\displaystyle H_{MC}=-A\sum_{v\in V}x_{v}+B\sum_{(u,v)\in\overline{E}}x_{u}x_{v}, (5)

where the constants can be chosen as A=1A=1, B=2B=2 (see [26]).

4.2 Combining QUBOs of independent problems

Consider two QUBOs H1(x1,,xn)H_{1}(x_{1},\ldots,x_{n}) and H2(xn+1,,xm)H_{2}(x_{n+1},\ldots,x_{m}) with respective minimum solutions x1,,xnx_{1}^{\ast},\ldots,x_{n}^{\ast} and xn+1,,xmx_{n+1}^{\ast},\ldots,x_{m}^{\ast}. Since both QUBOs do not share any variable, the ground state of H1+H2H_{1}+H_{2}, which is now a function in x1,,xn,xn+1,,xmx_{1},\ldots,x_{n},x_{n+1},\ldots,x_{m}, is precisely the sum of the ground states of H1H_{1} and H2H_{2}, and the minimum bitstring yielding the ground state will be x1,,xn,xn+1,,xmx_{1}^{\ast},\ldots,x_{n}^{\ast},x_{n+1}^{\ast},\ldots,x_{m}^{\ast}. This idea lays at the heart of parallel quantum annealing. Given we have an embedding that allows us to place H1H_{1} and H2H_{2} simultaneously on the quantum chip, we can therefore embed and solve H1H_{1} and H2H_{2} in one D-Wave call. Naturally, H1H_{1} and H2H_{2} do not need to be QUBOs of the same type of optimization problem, and neither do they need to be of equal size. Generalization to a larger number of QUBOs is straightforward.

The D-Wave hardware has a hardware precision limit [27, 28, 29] for the provided problem coefficients. Therefore, when constructing these independent QUBOs, another point to consider is how the minimum and maximum range of these independent QUBOs compare to each other. If one of these QUBOs’ coefficients dominate all of the other QUBOs, then most of the QUBO coefficients will effectively be lost in noise. Therefore, it might be necessary to normalize the minimum and maximum QUBO coefficients across all of the QUBOs that are being solved at the same time. In the experiments shown in Section 2, all of the problems are Maximum Clique QUBOs, which always have the same minimum and maximum coefficient range - therefore no coefficient normalization is used.

4.3 Choice of the embedding

We aim to create multiple disjoint minor-embeddings of a complete graph of size NN onto the DW 2000Q and the DW Advantage hardware. This allows us to embed arbitrary QUBOs of size NN onto the hardware connectivity graph, meaning that we can solve multiple problems of arbitrary structure (up to size NN) in parallel in a single backend call.

\begin{overpic}[width=212.47617pt]{figures/embedding_sizes.pdf} \put(53.0,-1.5){{(a)}} \end{overpic}\begin{overpic}[width=212.47617pt]{figures/embedding_chain_lengths.pdf} \put(53.0,-1.5){{(b)}} \end{overpic}
Figure 8: (a): Number of embeddings as a function of the embedding size (clique size) NN for DW 2000Q (blue) and DW Advantage (red). (b): Scatter plot of the chain lengths occurring in the embeddings as a function of the embedding size NN, with the average depicted as a line. For DW 2000Q, we consider cliques of sizes N{8,,46}N\in\{8,\ldots,46\}, while N{10,,60}N\in\{10,\ldots,60\} for DW Advantage.

We compute separate embeddings for clique sizes NN in the range of N{8,,46}N\in\{8,\ldots,46\} for DW 2000Q, and N{10,,60}N\in\{10,\ldots,60\} for DW Advantage. As shown in Figure 1, ideally the embeddings should be chosen such that both the number of cliques (complete graphs) embedded on the quantum hardware, as well as the physical distance between the embeddings is maximized. The latter is conjectured to decrease spurious interactions between qubits of different QUBOs on the chip, thus helping to increase the solution quality.

The disjoint embeddings for both DW 2000Q and DW Advantage are computed using the method minorminer [30, 31]. In particular, we supply the graph to be embedded to minorminer as simply the union of the disjoint cliques we wish to embed. This is done for all clique sizes NN for which an embedding can be found by minorminer. Figure 8 shows both the number of embeddings we are able to place on the D-Wave quantum chip, as well as the chain lengths occurring in those embeddings for both DW 2000Q and DW Advantage.

Supplying minorminer with a graph consisting of the disjoint cliques we wish to embed is sub-optimal. In particular, when computing the embeddings, we would ideally be optimizing for the separation between the embeddings on the hardware. This remains for future work, thus close physical proximity of our embeddings may cause leakage between qubits.

As described in Section 1, after annealing, we need to unembed all chains. For this, we tried both the majority vote method and the random weighted method. All results shown in this paper use the random weighted unembedding method.

4.4 D-Wave procedure

For each of the number of disjoint embeddings we fixed in Section 4.3, we generate an Erdős–Rényi random graph G(n,p)G(n,p) of size n=Nn=N, varying edge probability pp from 0.100.10 to 0.950.95 in increments of 0.050.05 such that each random graph is connected, does not have any degree 0 nodes, and is not a full clique itself. Then, we compute the exact Maximum Clique solution using FMC [20] and Networkx [32, 33, 34, 35] (the Networkx solver is used to find any degenerate Maximum Clique solutions). This step is necessary in order to find the CPU time used by FMC, as well as determine the exact solution(s) for the purpose of computing ground state probability (GSP), which is used to compute the quantum annealer Time-to-Solution (TTS) metric.

Next, we arbitrarily assign each of the random graphs to one of the disjoint embeddings. Then, the Maximum Clique QUBO is computed for each of these random graphs, and each QUBO is then embedded onto its assigned embedding. For example, Figure 1 (a) shows that on DW 2000Q for N=20N=20, we generate 1212 random graphs of size 2020 and then embed those 12 Maximum Clique QUBOs onto each of their respective (independent) embeddings. Then, this is repeated for each density.

Once the problems are embedded, we call the D-Wave backend using the parameter setting of Section 2.1. In order to get reliable results for the TTS computation, we call the backend 100100 times for each problem, and each backend call requests 1,0001,000 anneals. This results in 100,000100,000 samples in total per random graph. From these results, we first unembed the samples using the random weighted method, and then we compute how many of the 100,000100,000 samples correctly found the Maximum Clique solution. Using this information, we then compute the ensemble TTS metric using eq. (4).

Next, we apply the sequential quantum annealing method; meaning that we solve each of those QUBOs using separate backend calls. Importantly, each QUBO still uses the same physical hardware embedding as it did in the parallel case, so that differences in the embeddings will not change solution quality. In the example of Figure 1 (a), we would perform the 100100 backend calls for each one of the 1212 distinct problems, without the other 1111 embedded into the chip. This procedure is then repeated for all graph densities as well. Once again, we unembed using the random weighted method, and we also compute GSP for each of the problems. This procedure allows us to directly compare the solution quality for each of these problems when solving them separately and in parallel. Using this data, we can compute the TTS using eq. 3 for each of the random graphs. Thus, the total TTS for a group of graphs that were solved sequentially is the sum of each of their individual TTS values (in contrast to using parallel quantum annealing where we compute the ensemble TTS using eq. 4).

It is important to note that, especially for larger clique sizes (e.g., cliques of size 4040 and larger for DW 2000Q, and clique sizes 5050 onward for DW Advantage), not all problems can be solved to optimality by the quantum annealer. In this case, we encounter missing information in the computation of the TTS metric. We mitigate this problem by attempting to solve another randomly generated problem instead until the TTS metric can be computed. However, in settings where particular graphs need to be solved (as opposed to the setting with random graphs we consider), it is likely that D-Wave fails for large problems solved using a fixed embedding [1]. These failures can be attributed to longer chain lengths leading to broken chains, and therefore worse solution quality. Our results in section 2.6 show a possible solution to this problem; that is using many different embeddings to solve the same problem in parallel might allow D-Wave to more consistently find optimal solutions for large problem sizes.

Acknowledgments

The research presented in this article was supported by the Laboratory Directed Research and Development program of Los Alamos National Laboratory under the project number 20190065DR. The work of Hristo Djidjev has been also partially supported by Grant No. BG05M2OP001-1.001-0003, financed by the Science and Education for Smart Growth Operational Program (2014-2020) and co-financed by the European Union through the European Structural and Investment Funds.

Code and data availability

The code and the data (e.g., the embeddings) are available at

https://github.com/lanl/Parallel-Quantum-Annealing [36]

References

  • [1] Aaron Barbosa, Elijah Pelofske, Georg Hahn and Hristo N. Djidjev “Using Machine Learning for Quantum Annealing Accuracy Prediction” In Algorithms 14.6, 2021, pp. 187 DOI: 10.3390/a14060187
  • [2] Immanuel M. Bomze, Marco Budinich, Panos M. Pardalos and Marcello Pelillo “The Maximum Clique Problem” In Handbook of Combinatorial Optimization: Supplement Volume A Boston, MA: Springer US, 1999, pp. 1–74 DOI: 10.1007/978-1-4757-3023-4˙1
  • [3] Ryan A. Rossi, David F. Gleich, Assefaw H. Gebremedhin and Md. Ali Patwary “Fast Maximum Clique Algorithms for Large Graphs” In Proceedings of the 23rd International Conference on World Wide Web New York, NY, USA: Association for Computing Machinery, 2014, pp. 365–366 DOI: 10.1145/2567948.2577283
  • [4] Elijah Pelofske, Georg Hahn and Hristo Djidjev “Solving Large Maximum Clique Problems on a Quantum Annealer” In Quantum Technology and Optimization Problems Springer International Publishing, 2019, pp. 123–135 DOI: 10.1007/978-3-030-14082-3˙11
  • [5] Guillaume Chapuis, Hristo Djidjev, Georg Hahn and Guillaume Rizk “Finding Maximum Cliques on the D-Wave Quantum Annealer” In J Signal Process Sys 91.3-4, 2019, pp. 363–377 DOI: 10.1007/s11265-018-1357-8
  • [6] Elijah Pelofske, Georg Hahn and Hristo Djidjev “Decomposition Algorithms for Solving NP-hard Problems on a Quantum Annealer” In Journal of Signal Processing Systems 93.4, 2021, pp. 405–420 DOI: 10.1007/s11265-020-01550-1
  • [7] Wenbo Li, Longyin Wen, Mooi Choo Chuah and Siwei Lyu “Category-Blind Human Action Recognition: A Practical Recognition System” In 2015 IEEE International Conference on Computer Vision (ICCV), 2015, pp. 4444–4452 DOI: 10.1109/ICCV.2015.505
  • [8] Steven Maenhout, Bernard De Baets and Geert Haesaert “Graph-Based Data Selection for the Construction of Genomic Prediction Models” In Genetics 185.4, 2010, pp. 1463–1475 DOI: 10.1534/genetics.110.116426
  • [9] Guillaume Chapuis et al. “Parallel Seed-Based Approach to Multiple Protein Structure Similarities Detection” In Sci. Program. 2015, 2015, pp. 279715:1–279715:12 DOI: 10.1155/2015/279715
  • [10] P. Ray, B.K. Chakrabarti and Arunava Chakrabarti “Sherrington-Kirkpatrick model in a transverse field: Absence of replica symmetry breaking due to quantum fluctuations” In Phys. Rev. B 39, 1989, pp. 11828–11832 DOI: 10.1103/PhysRevB.39.11828
  • [11] Tadashi Kadowaki and Hidetoshi Nishimori “Quantum annealing in the transverse Ising model” In Phys. Rev. E 58, 1998, pp. 5355–5363 DOI: 10.1103/PhysRevE.58.5355
  • [12] Edward Farhi et al. “A Quantum Adiabatic Evolution Algorithm Applied to Random Instances of an NP-Complete Problem” In Science 292.5516, 2001, pp. 472–475 DOI: 10.1126/science.1057726
  • [13] Giuseppe E. Santoro, Roman Marton̆ák, Erio Tosatti and Roberto Car “Theory of Quantum Annealing of an Ising Spin Glass” In Science 295.5564, 2002 DOI: 10.1126/science.1068774
  • [14] Arnab Das and Bikas K. Chakrabarti “Colloquium: Quantum annealing and analog quantum computation” In Rev. Mod. Phys. 80, 2008, pp. 1061–1081 DOI: 10.1103/RevModPhys.80.1061
  • [15] Travis S. Humble et al. “Quantum Computers for High-Performance Computing” In IEEE Micro 41.5, 2021, pp. 15–23 DOI: 10.1109/MM.2021.3099140
  • [16] Konrad Jałowiecki, Andrzej Więckowski, Piotr Gawron and Bartłomiej Gardas “Parallel in time dynamics with quantum annealers” In Scientific Reports 10.1 Nature Publishing Group, 2020, pp. 1–7 DOI: 10.1038/s41598-020-70017-x
  • [17] Navid Anjum Aadit et al. “Massively Parallel Probabilistic Computing with Sparse Ising Machines”, 2021 arXiv:2110.02481
  • [18] Elijah Pelofske et al. “Quantum annealing algorithms for Boolean tensor networks” In Scientific Reports 12.1 Springer ScienceBusiness Media LLC, 2022 DOI: 10.1038/s41598-022-12611-9
  • [19] Kelly Boothby, Paul Bunyk, Jack Raymond and Aidan Roy “Next-Generation Topology of D-Wave Quantum Processors” arXiv, 2020 DOI: 10.48550/ARXIV.2003.00133
  • [20] Bharath Pattabiraman et al. “Fast Algorithms for the Maximum Clique Problem on Massive Sparse Graphs” In Algorithms and Models for the Web Graph Springer International Publishing, 2013, pp. 156–169 DOI: 10.1007/978-3-319-03536-9˙13
  • [21] D-Wave Systems Uniform Torque Compensation Github”, 2020
  • [22] James King et al. “Benchmarking a quantum annealing processor with the time-to-target metric” arXiv, 2015 DOI: 10.48550/ARXIV.1508.05087
  • [23] Aaron Barbosa, Elijah Pelofske, Georg Hahn and Hristo N. Djidjev “Optimizing Embedding-Related Quantum Annealing Parameters for Reducing Hardware Bias” In Parallel Architectures, Algorithms and Programming Springer Singapore, 2021, pp. 162–173 DOI: 10.1007/978-981-16-0010-4˙15
  • [24] Michael Ryan Zielewski and Hiroyuki Takizawa “A Method for Reducing Time-to-Solution in Quantum Annealing Through Pausing” In International Conference on High Performance Computing in Asia-Pacific Region, HPCAsia2022 Virtual Event, Japan: Association for Computing Machinery, 2022, pp. 137–145 DOI: 10.1145/3492805.3492815
  • [25] Ryan Hamerly et al. “Experimental investigation of performance differences between coherent Ising machines and a quantum annealer” In Science Advances 5.5, 2019, pp. eaau0823 DOI: 10.1126/sciadv.aau0823
  • [26] Andrew Lucas “Ising formulations of many NP problems” In Frontiers in Physics 2, 2014, pp. 1–5 DOI: 10.3389/fphy.2014.00005
  • [27] Kristen L. Pudenz, Tameem Albash and Daniel A. Lidar “Quantum annealing correction for random Ising problems” In Physical Review A 91.4, 2015 DOI: 10.1103/physreva.91.042302
  • [28] John E. Dorband “Extending the D-Wave with support for Higher Precision Coefficients”, 2018 arXiv:1807.05244
  • [29] D-Wave D-Wave Error Sources for Problem Representation”, 2021
  • [30] D-Wave D-Wave Ocean Software Documentation: Minorminer”, 2021
  • [31] Jun Cai, William G. Macready and Aidan Roy “A practical heuristic for finding graph minors” arXiv, 2014 DOI: 10.48550/ARXIV.1406.2741
  • [32] Networkx Find Cliques method”, 2021
  • [33] Coen Bron and Joep Kerbosch “Algorithm 457: Finding All Cliques of an Undirected Graph” In Commun. ACM 16.9, 1973, pp. 575–577 DOI: 10.1145/362342.362367
  • [34] Etsuji Tomita, Akira Tanaka and Haruhisa Takahashi “The worst-case time complexity for generating all maximal cliques and computational experiments” In Theoretical Computer Science 363.1, 2006, pp. 28–42 DOI: https://doi.org/10.1016/j.tcs.2006.06.015
  • [35] F. Cazals and C. Karande “A note on the problem of reporting maximal cliques” In Theoretical Computer Science 407.1, 2008, pp. 564–568 DOI: https://doi.org/10.1016/j.tcs.2008.05.010
  • [36] Elijah Pelofske, Georg Hahn and Hristo N. Djidjev Parallel-Quantum-Annealing”, 2021 URL: https://github.com/lanl/Parallel-Quantum-Annealing