This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Methods for Accelerating Geospatial Data Processing Using Quantum Computers

Max Henderson, Jarred Gallina, Michael Brett
(April 2020)
Abstract

Quantum computing is a transformative technology with the potential to enhance operations in the space industry through the acceleration of optimization and machine learning processes. Machine learning processes enable automated image classification in geospatial data. New quantum algorithms provide novel approaches for solving these problems and a potential future advantage over current, classical techniques. Universal Quantum Computers, currently under development by Rigetti Computing and other providers, enable fully general quantum algorithms to be executed, with theoretically proven speed-up over classical algorithms in certain cases. This paper describes an approach to satellite image classification using a universal quantum enhancement to convolutional neural networks: the quanvolutional neural network. Using a refined method, we found a performance improvement over previous quantum efforts in this domain and identified potential refinements that could lead to an eventual quantum advantage.

We benchmark these networks using the SAT-4 satellite imagery data set in order to demonstrate the utility of machine learning techniques in the space industry and the potential advantages that quantum machine learning can offer.

1 Introduction

1.1 Quantum Computing for Aerospace Applications

From solving systems of nonlinear equations to processing massive amounts of “big data”, the aerospace field is rife with computational challenges. We explore here the problem of image classification using lower-resolution data. Many applications within aerospace depend on reliable satellite imagery, which has historically been an expensive resource to collect and utilize. However, the entry of small satellite operators to the market has increased the supply of lower cost imagery. Although this data is of lower resolution than that from more traditional providers, powerful post-processing techniques bolster the utility of the data, making its usefulness rival their more expensive counterparts for certain applications. For example, techniques for deriving sub-pixel details from lower resolution imagery have been applied in a variety of use cases like crop mapping [1]. Post-processing techniques that leverage machine learning algorithms garner information from low-resolution data that previous techniques cannot. However, these algorithms can still be challenging, costly and time consuming. One successful but computationally-expensive technique is convolutional neural networks (CNNs) [2]. CNNs are a class of neural networks commonly used to analyze visual imagery. In order to address overfitting, they are more regularised than deep belief networks, and prioritise correctness over complexity. One avenue for improvement of CNNs is to look to the field of Quantum Machine Learning (QML) [3].

1.2 Introduction to Quantum Computing

Quantum computing is a novel technology that has seen an explosion in academic and industrial interest. Quantum computers are devices composed of quantum systems (typically quantum bits, or qubits) that leverage fundamental quantum mechanical phenomena, such as superposition, entanglement, and tunnelling. These quantum phenomena allow for the development of completely new types of “quantum” algorithms, some of which have been shown to be exponentially more efficient than any known classical computing counterpart [4]. Such promising results have spurred the development of improved quantum hardware alongside investigations into quantum computing algorithms for application spaces with highly important and classically-intractable problems. Solving aerospace big data problems using quantum computation has become a matter of focus for NASA [5, 6, 7], with potential applications for exact and approximate optimization, sampling, clustering, anomaly detection, and simulating quantum many-body systems for materials science and chemistry. Quantum computing’s potential has also drawn interest from other governmental institutions such as the National Academy of Sciences [8] and the Executive branch of the U.S. Government[9]. Quantum computing is currently in an era of noisy (error-prone), intermediate-scale (<<100 qubits) quantum devices or ‘NISQ ’[10], and developers have focused on the most promising near-term quantum computing application areas of chemistry / physical simulation, optimization, and machine learning.

1.3 Quantum machine learning

The field of quantum machine learning (QML) seeks to improve upon machine learning algorithms using quantum computing, which may be quantum variants of classical approaches or wholly novel algorithms. For example, researchers have leveraged quantum computing approaches to enhance support vector machines for handwriting recognition [11] and train fully-connected Boltzmann machines for sampling purposes [12, 13]. Another recent QML algorithm is a quantum version of a CNN called the quanvolutional neural network (QNN), which provides a new and potentially powerful modeling technique [14]. While the QNN algorithm approach using ”quantum convolutional” or ”quanvolutional” layers is explained in great detail in [14], we will briefly summarize the approach and argument for a quantum advantage here.

The QNN approach transforms input images, represented mathematically as a NN-by-NN-by-MM tensor, where NN is the pixel width and height of the square image and M is the number of channels. A QNN can be considered simply as a CNN extended by adding one or more new layers composed of quanvolutional filters. A quanvolutional filter acts just as a convolutional filter, processing sub-samples of these input tensors. Each pixel of the input sub-sample is mapped to a different variational parameter in a structured quantum circuit; we can name these pixel values θ\theta. We can represent a single quanvolutional filter ii operating on input θ\theta as:

|Ψi=Ui(θ)|0n|\Psi_{i}\rangle=U_{i}(\theta)\left|0\right\rangle^{\otimes n} (1)

wherein |Ψi|\Psi_{i}\rangle is the output quantum state of quanvolutional filter ii operating on input θ\theta, n is the number of qubits in the circuit, and UiU_{i} is the unitary matrix representing the quantum circuit transformation, which contains gates that have been parameterized by θ\theta (rotational angle parameterization is an update to the original method; see 2.4 for additional details). This output quantum state captures the extracted “feature information”, which is decoded using a user-defined decoding protocol function dd to generate the final output feature oio_{i}:

oi=d(|Ψi)o_{i}=d(|\Psi_{i}\rangle) (2)

wherein each oio_{i} value is a scalar output, just as a single scalar output is produced for each convolutional filter jj running onto the same input sub-samples θ\theta.

As laid out in [14], the argument for a potential quantum advantage is based on the following assumptions:

  1. 1.

    The quantum transformations executed on classical input data by the quantum device are computationally intractable to calculate (or simulate) classically.

  2. 2.

    The quantum transformations provide a benefit over other similar, tractable, classical transformations.

This quanvolutional neural network (QNN) approach was first tested on the MNIST data set [15]. Experiments showed that while the quantum features generated did provide a benefit over a classical convolutional neural network (CNN) without the same quantum transformations, it did not outperform a CNN with an additional layer of classically-tractable non-linear transformations. We conducted further experiments to determine potential quantum advantage [16], implementing a classification task on satellite images of the DeepSat data set [17]. Using a layer of 3-by-3 quanvolutional filters (requiring nine qubits with the basis state encoding protocol specified in [14] on a single color channel), the QNN trained faster but converged on a lower test set accuracy than the comparative classical CNN.

2 Materials and Methods

2.1 Caveats and objectives

Before examining our experimental details, it is worth briefly noting the objectives of both the QNN approach and the DeepSat modeling problem in this work. First, while the eventual goal of QML research is to show a clear quantum advantage over purely classical techniques, the state of technology is still nascent. Considerable effort and research is still required to incrementally better understand algorithmic performance and optimize the various components in the entire stack. This paper endeavours to improve on the shortcomings of the original QNN approach [14] and show improvements in terms of accuracy and processing capability (larger tensors), with the goal of illuminating a path forward to potential quantum advantage. Second, in terms of the machine learning application we selected to study, we deliberately chose a relatively standard image classification problem. While this does not include the data processing particulars that go into a complete space application pipeline, as well as the more specific end-user applications such as accelerating disaster responses [18], or identifying maritime objects from space [19], CNNs are fundamental components of these pipelines. For simplicity and more direct comparison to QNNs, we therefore chose a more straightforward image classification task with a standard CNN implementation, recognizing that if there is a benefit for this type of task, a similar benefit should be realized in a more complex stack.

2.2 Data set

In this experiment we benchmark the QNN using low-resolution satellite imagery to demonstrate the utility of such technology in this field. Specifically, all experiments were trained using the freely available SAT-4 data set [20]. SAT-4 is a collection of 500,000 28-by-28 pixel 4-channel patches labelled as 4 distinct classes – barren, trees, grassland and other.

In order to reduce the size of the available data and demonstrate the utility of both machine learning and quantum machine learning techniques in aerospace, we use only 10,000 of the available patches (9,000 for training, and 1,000 for testing).

2.3 Neural network architecture

We compared a CNN and QNN model, both of which are comprised of an assortment of convolutional (CONV), pooling (POOL), and fully-connected (FC) layers. The QNN model also contains one quanvolutional (QUANV) layer. The structure of the CNN model was CONV1-POOL1-CONV2-POOL2-FC. Both CONV layers use a stride of 1, while the POOL layers have a stride of 2. CONV1 uses five 5×55\times 5 filters, while CONV2 uses twelve 3×33\times 3 filters. POOL1 is an average pooling layer with a 5×55\times 5 filter, and POOL2 is a maxpooling layer with a 2×22\times 2 filter. The QNN model follows a similar architecture. The only modification is the replacement of the layer CONV1 and POOL1 with a quanvolutional layer QUANV1.

2.4 Potential QNN improvements

Our analysis of our earlier experiments and their overall performance identified potential avenues for improvement for the QNN model approach in [14].

  1. 1.

    Preserving more classical information. Our previous method of basis state encoding required that each pixel value is thresholded to 0 or 1; this loses considerable information from the input data. An encoding protocol which preserves more of this information is desirable.

  2. 2.

    Processing larger input tensors. While 3-by-3 pixel filters were sufficient for the MNIST data set, this input size was too small to properly capture input features in the DeepSat data set. This limitation was due to both the encoding and decoding protocols of [14]. The encoding basis state protocol required the same number of qubits as pixels, while the decoding protocol required the full probability distribution of all possible basis states. This limited us to 3-by-3 filters as simulating 25 qubit systems (which would be required for a 5-by-5 pixel input) would have taken a prohibitively long time. An alternate approach is needed to implement larger quanvolutional filters efficiently.

  3. 3.

    Using structured rather than random circuits. In our original work [14], we showed that adding more quanvolutional filters did increase accuracy for the QNN networks as expected (see Figure 2). However, adding many more quanvolutional filters did not increase accuracy results dramatically; this is likely due to the construction of the quanvolutional circuits implementing their transformations. The circuits were simply different random quantum circuits, formed by randomly applying 1 and 2 qubit gates to the qubits in the circuit. As the random quantum circuit becomes sufficiently deep, the resulting output distribution approaches that of the Haar measure [21]. This means in practice, although the circuits all had very different random gates choices, they likely were extracting very similar features from the input data. Selecting the quantum circuits more methodically should increase the variety of information the quanvolutional filters extract.

  4. 4.

    Large pre-processing requirements. Running large numbers of quantum circuit simulations is computationally expensive and slow. Rather than do these calculations on the fly, in [14] a large amount of pre-processing was undertaken to calculate all possible output values for each possible basis state encoding for each quanvolutional filter circuit. While this was computationally feasible for 9 qubit circuits, it is grossly inefficient at scale. An alternate approach to pre-processing may improve the scalability.

2.5 Improved QNN approach

Our revised approach here aims to address each of these avenues for potential improvement with a new quanvolutional layer processing structure. The quanvolutional framework still has the same components as shown in Figure 1B of [14], namely (1) encoding, (2) quantum circuit processing, and (3) decoding protocols. We implemented changes in each of these protocols as follows:

  • Encoding protocol: variational parameters. Instead of using basis encoding and discretizing all input pixels, we choose to preserve the pixel information by encoding them as continuous values into rotational angles using 1-qubit gate rotations.

  • Quantum circuit ansatz: non-random, structured circuits. Figure 1 shows the quantum computer used in this work, Rigetti’s Aspen-7-25Q-B quantum processing unit (QPU), which has 25 qubits with 24 programmable two-qubit gate interactions. This fit nicely with the number of parameters in our 5-5-4 tensor blocks of input data (see Figure 2), which when flattened have 100 parameters. By generating graphs with the same topology as the Aspen-7-25Q-B and with random edge weights, we were able to create natively executable quantum approximate optimization algorithm (QAOA) [22] circuits. These circuits assume that our randomly-weighted graphs with Aspen topology were MaxCut instances to solve, and returned solutions. We assigned a variational parameter for each term in the cost Hamiltonian, which maps to each edge in the graph, thus leading to 24 parameters. Additionally, after applying the cost Hamiltonian terms, we have an additional variational parameter for each application of the driver Hamiltonian, leading to 25 total parameters per pp layer in the QAOA design (see Appendix A for additional details on QAOA circuit interactions). If p=4p=4, there is exactly 1 variational parameter in the circuit for each input tensor pixel. This allows a natural way of mapping larger tensors into structured quantum circuits. However, we only ran the p=1p=1 case to attempt to circumvent gate depth issues (see Section 4.1 for more details).

    Refer to caption
    Figure 1: Visualization of the Rigetti Aspen-7-25Q-B QPU topology. Each number represents a programmable qubit, and edges represent the ability to run two-qubit gates between those qubits. By counting, one can verify this device has 25 qubits and 24 two-qubit gates.
  • Running on real QPU hardware. Simulating a single shot of a 25+ qubit system can take several minutes, while on a real QPU device this process is on the order of microseconds. While still limited in terms of available access time due to other users reserving the QPU, we were able to run orders of magnitudes more results than would have been possible within the same block of time using simulations.

  • Decoding protocol. Instead of pre-computing all possible input/output basis state pairs, which is intractable at scale, we developed a dynamic mapping protocol. It works as follows: within a particular compute window, we run as many input tensors as possible through the QPU and generate output for each of these inputs. After the window is closed, we take all the input tensors which were successfully processed and use them to construct a balltree data structure [23]. This then acts as a fast “map”, so that for a new, unseen tensor (i.e. one that was unprocessed by the QPU during the time available) we can quickly “map” this tensor to the most similar tensor that was successfully processed. The QPU output of this tensor is then used as an approximation for the unprocessed tensor. This method allows for a dynamic mapping regardless of the exact number of input tensors that were processed for each quanvolutional filter circuit.

Refer to caption
Figure 2: Example DeepSat-4 image (left) broken into 5-5-4 tensor blocks (right). Each of the complete 5-5-4 tensors blocks (25 per image) are sent as tensors to be run through the QPU circuit.

3 Results

Our experimental results are shown in Figure 3, which take the average test set accuracy results as a function of training iterations across 10 different CNN and QNN models, respectively. As expected, for both the CNN and QNN models, test set accuracy increased as a function of training iterations. While the average CNN model results were more accurate than the QNN results, they converged to similar values towards the end of training with both models getting to around 70% accuracy. While not yet outdoing the classical model, these results indicate that the new QNN protocol described in 2.5 has significantly increased performance. In the previous results exploring the performance of the original QNN algorithm [14] on the same DeepSat data modeling problem [16], the QNN model test set accuracy did not reach 60% accuracy using 25 quanvolutional filters. These results show the improved QNN reaching approximately 70% accuracy with only 5 quanvolutional filters.

Refer to caption
Figure 3: Results comparing two neural networks: the first being a QNN with an initial quanvolutional layer composed of 5 quanvolutional QAOA circuits and the second being a CNN with an additional CONV + POOL with 5 convolutional filters. The plotted results are the average across 10 networks, for each of CNN and QNN models.

4 Discussion

Near-term quantum computing is a constant challenge. Dealing with one limitation almost inevitably unearths some new challenge. While the improvements to the QNN algorithm in this work have shown benefits over the original implementation of [14], there are still inefficiencies that could be further improved. To improve the algorithm, we believe that we need to address the challenges raised by the new method.

4.1 New QNN challenges and potential solutions

  • Determine ideal number of QPU shots per input. While simulators give exact quantum states, real QPU hardware provides information one measurement (‘shot’) at a time, providing at best an approximate quantum state. More shots give a better representation of the true quantum state after applying the quantum circuit in question. In our experiments it is unclear how performance is affected by the number of shots taken in the quantum circuit. This could be explored experimentally in the future to optimize performance vs run-time.

  • Improving heuristic dynamic mapping. While computationally tractable, our decoding protocol essentially does a nearest-neighbor approximation of what the quantum output would be if that input tensor was run through a particular quanvolutional filter. The assumption that this approximation is sufficiently close to the value that would be returned by running the actual input through the QPU needs to be tested, and its impact on performance quantified. In particular, the heuristic dynamic mapping takes longer (1) as the input tensors grow larger and (2) the more quantum runs which are completed (i.e. more possible outputs to map to). To make this overall algorithm as strong as possible, the fastest, most scalable classical mapping to find the nearest processed tensor needs to be investigated. This could also involve some form of GPU parallelization as the nature of the underlying problem (calculating distances between many vectors) is highly parallelizable.

  • Mitigating gate depth limitations. While the speed benefit of running on real QPU hardware is impressive compared to simulation times, noise becomes a major factor when working with real quantum circuits. Error in quantum circuits both (1) corrupts results and consequently (2) limits total gate depth (i.e. number of gate operations while still being under some total error ϵ\epsilon). Error increases exponentially as a function of circuit depth for real quantum circuits in the absence of error-correction, so even near-perfect hardware will become error-dominant when running a sufficiently deep circuit. If we used the original p=4p=4 QAOA mapping, our total circuit depth would be roughly 150, which would result in extremely noisy output. We reduced the gate depth by grouping our pixels into groups of 4 and taking the average across those 4 for each encoding parameter in the p=1p=1 QAOA case, which led to a gate depth of around 40. However, this reduces the potential expressible power of the full p=4p=4 circuit considerably. To get better results and use deeper circuits, there are essentially two lines of improvement. First are changes made at the hardware level: each improvement in noise performance in the QPU hardware should have a positive impact on QNN performance. Second are solutions at the software level: by using various forms of quantum error correction protocols [24], we may be able to improve the overall performance and explore deeper circuit ansatz.

  • Experimenting with different structured quantum circuits. In this study, we saw good modeling using variations of the same structured QAOA circuit ansatz. However, there are other types of ansatz that could be analyzed for particular data sets. This could lead to better understandings for optimal design choice of the number and type of quanvolutional filter circuits.

5 Conclusion

Quantum computing proffers a powerful alternative to purely classical computing methods for machine learning applications. While not yet showing a practical advantage by outperforming a classical state-of-the-art nonlinear layer (CONV + POOL), our new QNN algorithm demonstrates progress by building off our earlier work [14]. This work provides more insight into how the quanvolutional approach might be applied more effectively in terms of accuracy and run-time efficiency. Such an iterative process, wherein each improvement in the overall QNN algorithm should raise the bar and help clarify the next tangible goals to improve, suggests that the road to quantum advantage will be incremental, not sudden. Our follow on work will continue this and try to work in the potential improvements mentioned in Section 4 as we continue to reach towards a quantum machine learning implementation that shows a clear quantum advantage.

6 Acknowledgements

We would like to thank Duncan Fletcher for his time spent reviewing this work and helpful edits.

References

  • [1] Clement Atzberger and Felix Rembold. Mapping the Spatial Distribution of Winter Crops at Sub-Pixel Level Using AVHRR NDVI Time Series and Neural Nets. Remote Sensing, 5:1335–1354, 2013.
  • [2] Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2323, 1998.
  • [3] Jacob Biamonte, Peter Wittek, Nicola Pancotti, Patrick Rebentrost, Nathan Wiebe, and Seth Lloyd. Quantum Machine Learning. Nature, 549(7671):195–202, 2018.
  • [4] Peter W. Shor. Polynomial-time algorithms for prime factorization and discrete logarithms on a quantum computer. SIAM Journal on Computing, 26(5):1484–1509, Oct 1997.
  • [5] Eleanor G. Rieffel, Stuart Hadfield, Tad Hogg, Salvatore Mandrà, Jeffrey Marshall, Gianni Mossi, Bryan O’Gorman, Eugeniu Plamadeala, Norm M. Tubman, Davide Venturelli, Walter Vinci, Zhihui Wang, Max Wilson, Filip Wudarski, and Rupak Biswas. From Ansatze to Z-gates: a NASA View of Quantum Computing. May 2019.
  • [6] Vadim N. Smelyanskiy, Eleanor G. Rieffel, Sergey I. Knysh, Colin P. Williams, Mark W. Johnson, Murray C. Thom, William G. Macready, and Kristen L. Pudenz. A Near-Term Quantum Computing Approach for Hard Computational Problems in Space Exploration. Apr 2012.
  • [7] Alejandro Perdomo-Ortiz, Marcello Benedetti, John Realpe-Gómez, and Rupak Biswas. Opportunities and challenges for quantum-assisted machine learning in near-term quantum computers. Quantum Science and Technology, 3(3):030502, Jul 2017.
  • [8] Engineering National Academies of Sciences and Medicine. Quantum Computing: Progress and Prospects. The National Academies Press, Washington, DC, 2019.
  • [9] National Strategic Overview for Quantum Information Science.
  • [10] John Preskill. Quantum computing in the nisq era and beyond. Quantum, 2:79, Aug 2018.
  • [11] Patrick Rebentrost, Masoud Mohseni, and Seth Lloyd. Quantum Support Vector Machine for Big Data Classification. Phys. Rev. Lett., 113(13):130503, Sep 2014.
  • [12] Marcello Benedetti, John Realpe-Gomez, Rupak Biswas, and Alejandro Perdomo-Ortiz. Quantum-assisted learning of graphical models with arbitrary pairwise connectivity. CoRR, abs/1609.0, 2016.
  • [13] Maxwell Henderson, John Novak, and Tristan Cook. Leveraging Quantum Annealing for Election Forecasting. Journal of the Physical Society of Japan, 88(6):61009, Mar 2019.
  • [14] Maxwell Henderson, Samriddhi Shakya, Shashindra Pradhan, and Tristan Cook. Quanvolutional neural networks: powering image recognition with quantum circuits. Quantum Machine Intelligence, 2(1):1–9, 2020.
  • [15] Yann LeCun and Corinna Cortes. MNIST handwritten digit database. 2010.
  • [16] Michael Brett, Jarred Gallina, and Maxwell Henderson. Methods for Accelerating Geospatial Data Processing Using Quantum Computers. In International Astronautical Congress 2019, 2019.
  • [17] Saikat Basu, Sangram Ganguly, Supratik Mukhopadhyay, Robert DiBiano, Manohar Karki, and Ramakrishna R. Nemani. Deepsat - A learning framework for satellite imagery. CoRR, abs/1509.03602, 2015.
  • [18] Vyron Antoniou and Chryssy Potsiou. A Deep Learning Method to Accelerate the Disaster Response Process. Remote Sensing, 12(3):544, feb 2020.
  • [19] Domenico Velotto, Björn Tings, Carlos Augusto Bentes da Silva, Anja Frost, and James Imber. Machine learning approaches to classify maritime objects from space radar. In 69th International Astronautical Congress (IAC) 2018, oct 2018.
  • [20] Saikat Basu, Sangram Ganguly, Supratik Mukhopadhyay, Robert DiBiano, Manohar Karki, and Ramakrishna Nemani. DeepSat - A Learning framework for Satellite Imagery. Sep 2015.
  • [21] Winton Brown. Random quantum dynamics: From random quantum circuits to quantum chaos. 2011.
  • [22] Edward Farhi, Jeffrey Goldstone, and Sam Gutmann. A Quantum Approximate Optimization Algorithm. Nov 2014.
  • [23] Stephen M Omohundro. Five balltree construction algorithms. International Computer Science Institute Berkeley, 1989.
  • [24] Alexander Erhard, Joel J Wallman, Lukas Postler, Michael Meth, Roman Stricker, Esteban A Martinez, Philipp Schindler, Thomas Monz, Joseph Emerson, and Rainer Blatt. Characterizing large-scale quantum computers via cycle benchmarking. Nature Communications, 10(1):5347, 2019.

Appendix A QAOA algorithm details

All quanvolutional filters in this experimental study were QAOA ansatz, each solving a different randomly weighted graph.

The QAOA algorithm is a universal gate-based quantum computing analogue to running the quantum annealing algorithm. Quantum annealing is used to solve quadratic, unconstrained binary optimization (QUBO) problems, of which many important NP-complete problems (such as MaxCut) can be expressed. In this section, we deeply explore the most trivial two-qubit case to understand the effects of the algorithm.

Quantum annealing seeks to evolve an initial ground state |Ψ0\left|\Psi_{0}\right\rangle of an initial trivial Hamiltonian, to the ground state |Ψ|\Psi\rangle of a problem or cost Hamiltonian HPH_{P}. This can be expressed via the equation:

|Ψ=eiHPt|Ψ0|\Psi\rangle=e^{-iH_{P}t}\left|\Psi_{0}\right\rangle (3)

where t is time. The QAOA algorithm as a quantum circuit emulates this process by essentially breaking it into 3 steps, which we will focus on for 2 qubits:

HI=HHH_{I}=H\otimes H (4)
HP=Cnot(IRz(θ))CnoteiHPH_{P^{\prime}}=Cnot\cdot\left(I\otimes R_{z}(\theta)\right)\cdot Cnot\approx e^{-iH_{P}} (5)
HD=(IHRZ(β)H)(HRZ(β)HI)eiβHDH_{D^{\prime}}=\left(I\otimes HR_{Z}(\beta)H\right)\cdot\left(HR_{Z}(\beta)H\otimes I\right)\approx e^{-i\beta H_{D}} (6)

where HIH_{I} is the Initialization Hamiltonian, HPH_{P^{\prime}} is the approximate Problem or Cost Hamiltonian term, and HDH_{D^{\prime}} is the approximate Driver or Mixer Hamiltonian evolved for a duration proportional to β\beta, where:

H=22[1111]H=\frac{\sqrt{2}}{2}\left[\begin{array}[]{cc}{1}&{1}\\ {1}&{-1}\end{array}\right],  Cnot =[1000010000010010]\quad\text{ $Cnot$ }=\left[\begin{array}[]{cccc}{1}&{0}&{0}&{0}\\ {0}&{1}&{0}&{0}\\ {0}&{0}&{0}&{1}\\ {0}&{0}&{1}&{0}\end{array}\right],

I=[1001]\quad I=\left[\begin{array}[]{cc}{1}&{0}\\ {0}&{1}\end{array}\right], Rz(θ)=[θe200eθ2]\quad R_{z}(\theta)=\left[\begin{array}[]{cc}{\frac{-\theta}{e^{2}}}&{0}\\ {0}&{e^{\frac{\theta}{2}}}\end{array}\right].

The HPH_{P^{\prime}} transformation is simply a decomposition of the ZZ-Ising gate:

ZZ(θ)=[eθ20000eθ20000eθ20000eθ2]ZZ(\theta)=\left[\begin{array}[]{cccc}{e^{\frac{\theta}{2}}}&{0}&{0}&{0}\\ {0}&{e^{\frac{-\theta}{2}}}&{0}&{0}\\ {0}&{0}&{e^{\frac{-\theta}{2}}}&{0}\\ {0}&{0}&{0}&{e^{\frac{\theta}{2}}}\end{array}\right] (7)

except for a phase shift:

Cnot(IRz(θ))Cnot=[eθ20000eθ20000eθ20000eθ2]Cnot\cdot\left(I\otimes R_{z}(\theta)\right)\cdot Cnot=\left[\begin{array}[]{cccc}{e^{\frac{-\theta}{2}}}&{0}&{0}&{0}\\ {0}&{e^{\frac{\theta}{2}}}&{0}&{0}\\ {0}&{0}&{e^{\frac{\theta}{2}}}&{0}\\ {0}&{0}&{0}&{e^{\frac{-\theta}{2}}}\end{array}\right] (8)
Refer to caption
Figure 4: Visualizing the effect of a ZZ gate controlling the coupling strength between two qubits ran on actual quantum hardware (top) vs the exact analytic solution (bottom).

By running our full circuit on an initial ground state of |00,|00\rangle, we get:

HDHPHI|00=|Ψ.H_{D^{\prime}}H_{P^{\prime}}H_{I}|00\rangle=|\Psi\rangle. (9)

Using |Ψ|\Psi\rangle, we can calculate the exact probabilities of measuring if the qubits are in the same state (i.e. both in |00|00\rangle or |11|11\rangle) as:

f(θ,β)=Ψ00|Ψ00+Ψ11|Ψ11=12(1+sin(θ)sin(2β))f(\theta,\beta)=\left\langle\Psi_{00}|\Psi_{00}\right\rangle+\left\langle\Psi_{11}|\Psi_{11}\right\rangle=\frac{1}{2}(1+\sin(\theta)\sin(2\beta)) (10)

with clear maximum when sin(θ)sin(2β)=1 (i.e. θ=π2+2πn and β=π4+nπ)\left.\sin(\theta)\sin(2\beta)=1\text{ (i.e. }\theta=\frac{\pi}{2}+2\pi n\text{ and }\beta=\frac{\pi}{4}+n\pi\right) and minima when sin(θ)sin(2β)=1\sin(\theta)\sin(2\beta)=-1 (i.e. either θ=3π2+2πn\theta=\frac{3\pi}{2}+2\pi n and β=π4+nπ\beta=\frac{\pi}{4}+n\pi or θ=π2+2πn\theta=\frac{\pi}{2}+2\pi n and β=\beta= 3π4+nπ)\left.\frac{3\pi}{4}+n\pi\right). This can be visualized in Figure 4 for three different β\beta angles and a range between 0 and 2π2\pi for θ\theta.