arxiv:1911.11071v1 [cs.lg] 25 nov 2019

10
Learning to Optimize Variational Quantum Circuits to Solve Combinatorial Problems Sami Khairy, 1 Ruslan Shaydulin, 2 Lukasz Cincio, 3 Yuri Alexeev, 4 Prasanna Balaprakash 4 1 Illinois Institute of Technology, 2 Clemson University, 3 Los Alamos National Laboratory, 4 Argonne National Laboratory [email protected], [email protected], [email protected], [email protected], [email protected] Abstract Quantum computing is a computational paradigm with the potential to outperform classical methods for a variety of problems. Proposed recently, the Quantum Approximate Op- timization Algorithm (QAOA) is considered as one of the leading candidates for demonstrating quantum advantage in the near term. QAOA is a variational hybrid quantum- classical algorithm for approximately solving combinatorial optimization problems. The quality of the solution obtained by QAOA for a given problem instance depends on the per- formance of the classical optimizer used to optimize the vari- ational parameters. In this paper, we formulate the problem of finding optimal QAOA parameters as a learning task in which the knowledge gained from solving training instances can be leveraged to find high-quality solutions for unseen test in- stances. To this end, we develop two machine-learning-based approaches. Our first approach adopts a reinforcement learn- ing (RL) framework to learn a policy network to optimize QAOA circuits. Our second approach adopts a kernel den- sity estimation (KDE) technique to learn a generative model of optimal QAOA parameters. In both approaches, the train- ing procedure is performed on small-sized problem instances that can be simulated on a classical computer; yet the learned RL policy and the generative model can be used to efficiently solve larger problems. Extensive simulations using the IBM Qiskit Aer quantum circuit simulator demonstrate that our proposed RL- and KDE-based approaches reduce the opti- mality gap by factors up to 30.15 when compared with other commonly used off-the-shelf optimizers. 1 Introduction Recently, a number of quantum computing devices have be- come available on the cloud. Current devices, which are commonly referred to as Noisy Intermediate-Scale Quan- tum (NISQ) devices, operate on a small number of qubits and have limited error-correction capabilities. Demonstrat- ing quantum advantage on these devices, which is the ability to solve a problem more efficiently by using quantum com- putational methods compared with classical state-of-the-art methods, requires the development of algorithms that can run using a modest quantum circuit depth. Copyright c 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. The Quantum Approximate Optimization Algorithm (QAOA) is one of the leading candidate algorithms for achieving quantum advantage in the near term. QAOA is a hybrid quantum-classical algorithm for approximately solv- ing combinatorial problems (Farhi, Goldstone, and Gutmann 2014). QAOA combines a parameterized quantum state evo- lution that is performed on a NISQ device, with a classical optimizer that is used to find optimal parameters. Conceiv- ably, the quality of the solution produced by QAOA for a given combinatorial instance depends on the quality of the variational parameters found by the classical optimizer. De- signing robust optimization methods for QAOA is therefore a prerequisite for achieving a practical quantum advantage. Optimizing QAOA parameters is known to be a hard problem because the optimization objective is nonconvex with low-quality nondegenerate local optima (Shaydulin, Safro, and Larson 2019; Zhou et al. 2018). Many approaches have been applied to QAOA parameter optimization, includ- ing gradient-based (Romero et al. 2018; Zhou et al. 2018; Crooks 2018) and derivative-free methods (Wecker, Hast- ings, and Troyer 2016; Yang et al. 2017; Shaydulin, Safro, and Larson 2019). Noting that the optimization objective of QAOA is specific to the underlying combinatorial instance, existing works approach the task of finding optimal QAOA parameters for a given instance as an exclusive task, and they devise methods that require quantum circuit evaluations on the order of thousands. To the best of our knowledge, ap- proaching QAOA parameter optimization as a learning task is underexplored. To that end, we propose two machine- learning-based methods for QAOA parameter optimization, in which the knowledge gained from solving training in- stances can be leveraged to efficiently find high-quality so- lutions for unseen test instances with only a couple of hun- dred quantum circuit evaluations. Our novel treatment of the QAOA optimization task has the potential to make QAOA a cost-effective algorithm to run on near-term quantum com- puters. The main contributions of our work are summarized as follows. First, we formulate the task of learning a QAOA parameter optimization policy as a reinforcement learning (RL) task. This approach can learn a policy network that can exploit geometrical regularities in the QAOA optimization arXiv:1911.11071v1 [cs.LG] 25 Nov 2019

Upload: others

Post on 10-Dec-2021

2 views

Category:

Documents


0 download

TRANSCRIPT

Learning to Optimize Variational Quantum Circuitsto Solve Combinatorial Problems

Sami Khairy,1 Ruslan Shaydulin,2 Lukasz Cincio,3 Yuri Alexeev,4 Prasanna Balaprakash4

1Illinois Institute of Technology, 2Clemson University, 3Los Alamos National Laboratory, 4Argonne National [email protected], [email protected], [email protected], [email protected], [email protected]

Abstract

Quantum computing is a computational paradigm with thepotential to outperform classical methods for a variety ofproblems. Proposed recently, the Quantum Approximate Op-timization Algorithm (QAOA) is considered as one of theleading candidates for demonstrating quantum advantagein the near term. QAOA is a variational hybrid quantum-classical algorithm for approximately solving combinatorialoptimization problems. The quality of the solution obtainedby QAOA for a given problem instance depends on the per-formance of the classical optimizer used to optimize the vari-ational parameters. In this paper, we formulate the problem offinding optimal QAOA parameters as a learning task in whichthe knowledge gained from solving training instances can beleveraged to find high-quality solutions for unseen test in-stances. To this end, we develop two machine-learning-basedapproaches. Our first approach adopts a reinforcement learn-ing (RL) framework to learn a policy network to optimizeQAOA circuits. Our second approach adopts a kernel den-sity estimation (KDE) technique to learn a generative modelof optimal QAOA parameters. In both approaches, the train-ing procedure is performed on small-sized problem instancesthat can be simulated on a classical computer; yet the learnedRL policy and the generative model can be used to efficientlysolve larger problems. Extensive simulations using the IBMQiskit Aer quantum circuit simulator demonstrate that ourproposed RL- and KDE-based approaches reduce the opti-mality gap by factors up to 30.15 when compared with othercommonly used off-the-shelf optimizers.

1 IntroductionRecently, a number of quantum computing devices have be-come available on the cloud. Current devices, which arecommonly referred to as Noisy Intermediate-Scale Quan-tum (NISQ) devices, operate on a small number of qubitsand have limited error-correction capabilities. Demonstrat-ing quantum advantage on these devices, which is the abilityto solve a problem more efficiently by using quantum com-putational methods compared with classical state-of-the-artmethods, requires the development of algorithms that canrun using a modest quantum circuit depth.

Copyright c© 2020, Association for the Advancement of ArtificialIntelligence (www.aaai.org). All rights reserved.

The Quantum Approximate Optimization Algorithm(QAOA) is one of the leading candidate algorithms forachieving quantum advantage in the near term. QAOA is ahybrid quantum-classical algorithm for approximately solv-ing combinatorial problems (Farhi, Goldstone, and Gutmann2014). QAOA combines a parameterized quantum state evo-lution that is performed on a NISQ device, with a classicaloptimizer that is used to find optimal parameters. Conceiv-ably, the quality of the solution produced by QAOA for agiven combinatorial instance depends on the quality of thevariational parameters found by the classical optimizer. De-signing robust optimization methods for QAOA is thereforea prerequisite for achieving a practical quantum advantage.

Optimizing QAOA parameters is known to be a hardproblem because the optimization objective is nonconvexwith low-quality nondegenerate local optima (Shaydulin,Safro, and Larson 2019; Zhou et al. 2018). Many approacheshave been applied to QAOA parameter optimization, includ-ing gradient-based (Romero et al. 2018; Zhou et al. 2018;Crooks 2018) and derivative-free methods (Wecker, Hast-ings, and Troyer 2016; Yang et al. 2017; Shaydulin, Safro,and Larson 2019). Noting that the optimization objective ofQAOA is specific to the underlying combinatorial instance,existing works approach the task of finding optimal QAOAparameters for a given instance as an exclusive task, and theydevise methods that require quantum circuit evaluations onthe order of thousands. To the best of our knowledge, ap-proaching QAOA parameter optimization as a learning taskis underexplored. To that end, we propose two machine-learning-based methods for QAOA parameter optimization,in which the knowledge gained from solving training in-stances can be leveraged to efficiently find high-quality so-lutions for unseen test instances with only a couple of hun-dred quantum circuit evaluations. Our novel treatment of theQAOA optimization task has the potential to make QAOA acost-effective algorithm to run on near-term quantum com-puters.

The main contributions of our work are summarized asfollows. First, we formulate the task of learning a QAOAparameter optimization policy as a reinforcement learning(RL) task. This approach can learn a policy network that canexploit geometrical regularities in the QAOA optimization

arX

iv:1

911.

1107

1v1

[cs

.LG

] 2

5 N

ov 2

019

objective of training instances, to efficiently optimize newQAOA circuits of unseen test instances. Second, we proposea sampling-based QAOA optimization strategy based on akernel density estimation (KDE) technique. This approachcan learn a generative model of optimal QAOA parameters,which can be used to generate new parameters and quicklysolve test instances. In both approaches, we choose a train-ing set of small-sized combinatorial instances that can besimulated on a classical computer, yet the test set includeslarger instances. We conduct extensive simulations using theIBM Qiskit Aer quantum circuit simulator to evaluate theperformance of our proposed approaches. We show that thetwo approaches can reduce the optimality gap by factors upto 30.15 when compared with other commonly used off-the-shelf optimizers.

2 The Quantum Approximate OptimizationAlgorithm

The idea of encoding the solution to a combinatorial opti-mization problem in the spectrum of a quantum Hamiltoniangoes back to 1989 (Apolloni, Carvalho, and de Falco 1989).Using this quantum encoding, one can find the optimal so-lution to the original combinatorial problem by preparingthe highest energy eigenstate of the problem Hamiltonian.Multiple approaches inspired by the adiabatic theorem (Kato1950) have been proposed to achieve this. Of particularinterest is the Quantum Approximate Optimization Algo-rithm, introduced by (Farhi, Goldstone, and Gutmann 2014),and its generalization, the Quantum Alternating OperatorAnsatz (Hadfield et al. 2017). QAOA is arguably one of thestrongest candidates for demonstrating quantum advantageover classical approaches in the current NISQ computers era(Streif and Leib 2019).

In QAOA, a classical binary assignment combinatorialproblem is first encoded in a cost Hamiltonian HC by map-ping classical binary decision variables si ∈ {−1, 1} ontothe eigenvalues of the quantum Pauli Z operator σz . For un-constrained combinatorial problems, the initial state is a uni-form superposition quantum state |ψ〉 = |+〉⊗n, prepared byapplying Hadamard gates on all qubits in the system. QAOAprepares a variational quantum state |ψ(β,γ)〉 by apply-ing a series of alternating operators e−iβkHM and e−iγkHC ,∀k ∈ [p],

|ψ(β,γ)〉 = e−iβpHM e−iγpHC · · · e−iβ1HM e−iγ1HC |+〉⊗n ,(1)

where β,γ ∈ [−π, π] are 2p variational parameters, n isthe number of qubits or binary variables, and HM is thetransverse field mixer Hamiltonian HM =

∑i σ

xi . Alter-

native initial states and mixer operators can be used to re-strict |ψ(β,γ)〉 to the subspace of feasible solutions for con-strained optimization problems (Hadfield et al. 2017). In or-der to find the highest energy eigenstate of HC , a classicaloptimizer is used to vary parameters β,γ to maximize theexpected energy of HC ,

f(β,γ) = 〈ψ(β,γ)|HC |ψ(β,γ)〉 . (2)

For p → ∞, ∃β∗,γ∗ = arg maxβ,γ f(β,γ) such thatthe resulting quantum state |ψ(β∗,γ∗)〉 encodes the opti-mal solution to the classical combinatorial problem (Farhi,Goldstone, and Gutmann 2014). In practice, the value p ischosen based on the trade-off between the achieved approx-imation ratio, the complexity of parameter optimization, andthe accumulated errors. Ideally, increasing p monotonicallyimproves the QAOA approximation ratio (Zhou et al. 2018),although for higher p the complexity of QAOA parameteroptimization can limit the benefits (Huang et al. 2019). Onreal near-term quantum hardware, errors become a domi-nating factor. For instance, on a state-of-the-art trapped-ionquantum computer, increasing p beyond 1 does not leadto improvements in the approximation ratio because of er-rors (Pagano et al. 2019). The hardware is rapidly progress-ing, however, and it is expected that QAOA with 1 ≤ p ≤ 5can be run in the foreseeable future, thus motivating us tochoose p ∈ {1, 2, 4} for our benchmark.

QAOA has been applied to a variety of problems, in-cluding graph maximum cut (Crooks 2018; Zhou et al.2018), network community detection (Shaydulin et al. 2019;Shaydulin et al. 2018), and portfolio optimization, amongmany others (Barkoutsos et al. 2019).

2.1 QAOA for Max-CutIn this paper, we explore QAOA applied to the graph max-imum cut problem (Max-Cut). It is among the most com-monly used target problems for QAOA because of its equiv-alence to quadratic unconstrained binary optimization. Con-sider a graph G = (V,E) where V is the set of vertices andE is the set of edges. The goal of Max-Cut is to partition theset of vertices V into two disjoint subsets such that the totalweight of edges connecting the two subsets is maximized.Let binary variables sk denote the partition assignment ofvertex k, ∀k ∈ [n]. Then Max-Cut can be formulated as fol-lows,

maxs

∑i,j∈V

wijsisj + c, sk ∈ {−1, 1},∀k (3)

where wij = 1 if (i, j) ∈ E, and 0 otherwise, and c is aconstant. The objective in (3) can be encoded in a problemHamiltonian by mapping binary variables sk onto the eigen-values of the Pauli Z operator σz: HC =

∑i,j∈V wij σ

zi σ

zj .

The optimal binary assignment s for (3) is therefore encodedin the highest energy eigenstate of HC . Note that in the def-inition of HC , there is an implicit tensor product with theidentity unitary, σI , applied to all qubits except for qubitsi, j.

In (Crooks 2018) the author shows that QAOA forMax-Cut can achieve approximation ratios exceeding thoseachieved by the classical Goemans-Williamson (Goemansand Williamson 1995) algorithm. A number of theoreti-cal results show that QAOA for Max-Cut can improve onbest-known classical approximation algorithms for somegraph classes (Parekh, Ryan-Anderson, and Gharibian 2019;Wang et al. 2018).

2.2 Linear Algebraic Interpretation of QAOAHere, we provide a short linear algebraic description ofQAOA for readers who are not familiar with the quantumcomputational framework. An n-qubit quantum state is a su-perposition (i.e., a linear combination) of computational ba-sis states that form an orthonormal basis set in a complexvector space C2n ,

|φ〉 = α0 |0..000〉+ α1 |0..001〉+ · · ·+ α2n−1 |1..111〉= [α0, α1, · · · , α2n−1]T ,

(4)where |αi|2,∀i ∈ [2n−1] is the probability that the quantumstate is in |i〉,

∑2n−1i=0 |αi|2 = 1. Quantum gates are unitary

linear transformations on quantum states. For instance, PauliZ and Pauli X operators and the identity operator are

σz =

[1 00 −1

], σx =

[0 11 0

], σI =

[1 00 1

]. (5)

Note that the eigenvectors of the Pauli Z operator are com-putational basis states |0〉 and |1〉 with eigenvalues 1 and−1, respectively. Therefore, e−iβkHM and e−iγkHC are2n × 2n linear operators, which for large n cannot be effi-ciently simulated classically. HC is a Hermitian linear oper-ator with eigenvalues λ0, λ1, · · · , λ2n−1. Based on the min-max variational theorem, the minimum eigenvalue λmin =minx {RHC (x) : x 6= 0}, and the maximum eigenvalueλmax = maxx {RHC (x) : x 6= 0}, where RHC (x) is theRayleigh-Ritz quotient, RHC (x) = x†HCx

x†x, x ∈ C2n\{0}.

Note that for any quantum state 〈φ|φ〉 = φ†φ = 1, the high-est energy of HC in (2) is in fact λmax. That said, QAOAconstructs linear operators parameterized by the 2p param-eters, β,γ, whose net effect is to transform the uniform su-perposition state, |ψ〉 = 1√

2n[1, · · · , 1]T , into a unit-length

eigenvector that corresponds to λmax.1

3 Related WorksIn this section, a literature review of related research work ispresented.

3.1 QAOA Parameter OptimizationA number of approaches have been explored for QAOAparameter optimization, including a variety of off-the-shelf gradient-based (Romero et al. 2018; Crooks 2018)and derivative-free methods (Wecker, Hastings, and Troyer2016; Yang et al. 2017; Shaydulin, Safro, and Larson 2019).QAOA parameters have been demonstrated to be hard to op-timize by off-the-shelf methods (Shaydulin, Safro, and Lar-son 2019) because the energy landscape of QAOA is non-convex with low-quality nondegenerate local optima (Zhouet al. 2018). Off-the-shelf optimizers ignore features ofQAOA energy landscapes and nonrandom concentration ofoptimal parameters. Researchers have demonstrated theo-retically that for certain bounded-degree graphs (Brandao

1Eigenvectors of a quantum Hamiltonian HC are referred to asits eigenstates; therefore, “transforming |ψ〉 into an eigenvector ofHC corresponding to λmax” is a different way of saying “preparingan eigenstate of HC corresponding to the energy λmax.”

et al. 2018), QAOA energy landscapes are graph instance-independent given instances that come from a reasonabledistribution. Motivated by this knowledge, we develop twomachine-learning-based methods exploiting the geometricalstructure of QAOA landscapes and the concentration of op-timal parameters, so that the cost of QAOA parameter opti-mization can be amortized.

3.2 Hyperparameter and Optimizer Learning

Hyperparameter optimization, which involves the optimiza-tion of hyperparameters used to train a machine learn-ing model, is an active field of research (Hutter, Kotthoff,and Vanschoren 2018). Each hyperparameter choice cor-responds to one optimization task, and thus tuning of hy-perparamters can be regarded as a search over differentoptimization instances, which is analogous to our prob-lem formulation. Recent methods devise sequential model-based Bayesian optimization techniques (Feurer, Springen-berg, and Hutter 2015) or asynchronous parallel model-based search (Balaprakash et al. 2018), while older meth-ods devise random-sampling-based strategies (Balaprakash,Birattari, and Stutzle 2007).

On the other hand, learning an optimizer to train machinelearning models has recently attracted considerable researchinterests. The motivation is to design optimization algo-rithms that can exploit structure within a class of problems,which is otherwise unexploited by hand-engineered off-the-shelf optimizers. In existing works, the learned optimizer isimplemented by long short-term memory (Andrychowicz etal. 2016; Verdon et al. 2019) or a policy network of an RLagent (Li and Malik 2016). Our RL-based approach to op-timizer learning differs from that of (Li and Malik 2016)mainly in the choice of reward function and the policy searchmechanism. In our work, a Markovian reward function ischosen to improve the learning process of a QAOA opti-mization policy.

4 Learning Optimal QAOA Parameters

Since evaluating a quantum circuit is an expensive task, theability to find good variational parameters using a smallnumber of calls to the quantum computer is crucial for thesuccess of QAOA. To amortize the cost of QAOA param-eter optimization across graph instances, we formulate theproblem of finding optimal QAOA parameters as a learn-ing task. We choose a set of graph instances, GTrain, whichinclude graph instances that are representative of their re-spective populations, as training instances for our proposedmachine learning methods. The learned models can then beused to find high-quality solutions for unseen instances froma test set GTest. In Section 4.1, we propose an RL-based ap-proach to learn a QAOA parameter optimization policy. InSection 4.2, we propose a KDE-based approach to learn agenerative model of optimal QAOA parameters. Section 4.3elaborates on graph instances in GTrain and GTest.

4.1 Learning to Optimize QAOA Parameterswith Deep Reinforcement Learning

Our first approach aims to learn a QAOA parameter opti-mization policy, which can exploit structure and geometri-cal regularities in QAOA energy landscapes to find high-quality solutions within a small number of quantum circuitevaluations. This approach can potentially outperform hand-engineered that are designed to function in a general setting.We cast the problem of learning a QAOA optimizer as anRL task, where the learned RL policy is used to produceiterative parameter updates, in a way that is analogous tohand-engineered iterative optimizers.

In the RL framework, an autonomous agent learns how tomap its state in a state space, s ∈ S, to an action from itsaction space, a ∈ A, by repeated interaction with an envi-ronment. The environment provides the agent with a rewardsignal, r ∈ R, in response to its action. Based on the re-ward signal, the agent either reinforces the action or avoidsit at future encounters, in an attempt to maximize the ex-pected total discounted rewards received over time (Suttonand Barto 2018).

Mathematically, the RL problem is formalized as aMarkov decision process (MDP), defined by a tuple(S,A,P,P0,R, ζ), where the environment dynamics, P :S × A × S → [0, 1], that is, the model’s state-action-statetransition probabilities, P(st+1|st, at), and P0 : S → [0, 1],which is the initial distribution over the states, are unknown;R : S × A × S → R is a reward function that guides theagent through the learning process; and ζ is a discount fac-tor to bound the cumulative rewards and trade off how far-sighted or shortsighted the agent is in its decision making.A solution to the RL problem is a stationary Markov policythat maps the agent’s states to actions, π(a|s), such that theexpected total discounted rewards is maximized,

π∗ = arg maxπ

E[

T∑t=0

ζtR(st, at, st+1)]. (6)

In our setting, we seek a Markov policy that can be usedto produce iterative QAOA parameter updates. This policymust exploit QAOA structure and geometrical regularitiessuch that high-quality solutions can be achieved despite thechallenges pertaining to QAOA parameter optimization. Todo so, we formulate QAOA optimizer learning as an MDPas follows.

1. ∀st ∈ S, st = {∆ftl,∆βtl,∆γtl}l=t−1,...,t−L; that is,the state space is the set of finite differences in the QAOAobjective and the variational parameters between the cur-rent iteration and L history iterations, S ⊂ R(2p+1)L.

2. ∀at ∈ A, at = {∆βtl,∆γtl}l=t−1; that is, the actionspace is the set of step vectors used to update the varia-tional parameters, A ⊂ R2p.

3. R(st, at, st+1) = f(βt + ∆βt,tl,γt + ∆γtl) −f(βt,γt), l = t − 1; that is, the reward is the changein the QAOA objective between the next iteration and thecurrent iteration.

The motivation behind our state space formulation comesfrom the fact that parameter updates at xt = (βt,γt) shouldbe in the direction of the gradient at xt and the step sizeshould be proportional to the Hessian at xt, both of whichcan be numerically approximated by using the method offinite differences. The RL agent’s learning task is there-fore to find the optimal way of producing a step vectorat = {∆βtl,∆γtl}l=t−1, given some collection of his-torical differences in the objective and parameters space,{∆ftl,∆βtl,∆γtl}l=t−1,...,t−L, such that the expected to-tal discounted rewards are maximized. Note that (6) is max-imized when R(st, at, st+1) ≥ 0, which means the QAOAobjective has been increased between any two consecu-tive iterates. The choice of reward function adheres to theMarkovian assumption and encourages the agent to takestrides in the landscape of the parameter space that yield ahigher increase in the QAOA objective (2), if possible, whilemaintaining conditional independence on historical statesand actions.

Training Procedure Learning a generalizable RL policythat can perform well on a wide range of test instances re-quires the development of a proper training procedure andreward normalization scheme. We use the following strat-egy to train the RL agent on instances in GTrain. A trainingepisode is defined to be a trajectory of length T = 64, whichis sampled from a depth-p QAOA objective (2) correspond-ing to one of the training instances in GTrain. At the end ofeach episode, the trajectory is cut off and restarted from arandom point in the domain of (2). Training instances are cir-culated in a round-robin fashion, with each episode mitigat-ing destructive policy updates and overfitting. Rewards fora given training instance are normalized by the mean depth-p QAOA objective corresponding to that instance, which isestimated a priori by uniformly sampling the 2p variationalparameters. Training is performed for 750 epochs of 128episodes each, and policy updates are performed at the endof each epoch.

Deep RL Implementation We train our proposed deepRL framework using the actor-critic Proximal Policy Opti-mization (PPO) algorithm (Schulman et al. 2017). In PPO, aclipped surrogate advantage objective is used as the trainingobjective,

Lclip(θ) = Et[min( πθ(at|st)πθold (at|st)

At, clip( πθ(at|st)πθold (at|st)

, 1 + ε, 1− ε)At)](7)

The surrogate advantage objective, Et[ πθ(at|st)πθold (at|st)

At], is ameasure of how the new policy performs relative to the oldpolicy. Maximizing the surrogate advantage objective with-out constraints could lead to large policy updates that maycause policy collapse. To mitigate this issue, PPO maxi-mizes the minimum of an unclipped and a clipped versionof the surrogate advantage, where the latter removes the in-centive for moving πθ(at|st)

πθold (at|st)outside of [1− ε, 1 + ε]. Clip-

ping therefore acts as a regularizer that controls how muchthe new policy can go away from the old one while stillimproving the training objective. In order to further ensurereasonable policy updates, a simple early stopping method

is adopted, which terminates gradient optimization on (7)when the mean KL-divergence between the new and old pol-icy hits a predefined threshold.

Fully connected multilayer perceptron networks with twohidden layers for both the actor (policy) and critic (value)networks are used. Each hidden layer has 64 neurons. Tanhactivation units are used in all neurons. The range of out-put neurons is scaled to [−0.1, 0.1]. The discount factor andnumber of history iterations in the state formulation are setto ζ = 0.99 and L = 4, respectively. A Gaussian policywith a constant noise variance of e−6 is adopted throughoutthe training in order to maintain constant exploration andavoid getting trapped in a locally optimal policy. At testing,the trained policy network corresponding to the mean of thelearned Gaussian policy is used, without noise.

4.2 Learning to Sample Optimal QAOAParameters with Kernel Density Estimation

In our second approach, we aim to learn the distribution ofoptimal QAOA parameters and use this distribution to sam-ple QAOA parameters that can provide high-quality solu-tions for test instances. Although previous theoretical re-sults show that optimal QAOA parameters concentrate forgraph instances necessarily coming from the same distri-bution (Brandao et al. 2018), we learn a meta-distributionof optimal QAOA parameters for graph instances in GTrain,which come from diverse classes and distributions. This ap-proach can potentially eliminate the need for graph featureengineering and similarity learning and can dramaticallysimplify the solution methodology.

To learn a meta-distribution of optimal QAOA parame-ters, we adopt a KDE technique. Suppose we have accessto a set of N optimal QAOA parameters Sp∗ = {xi∗ =

(βi∗,γi∗)}N−1i=0 for graph instances in GTrain and a given

QAOA circuit depth, p. A natural local estimate for the den-sity fX at x is fX(x) =

#Xi∗∈B(x)Nω , where B(x) is a neigh-

borhood of width ω around x based on some distance met-ric. This estimate, however, is not smooth. Instead, one com-monly adopts a Parzen-Rosenblatt approach with a Gaussiankernel to obtain a smoother estimate,

fX(x) =1

N

N−1∑i=0

Kω(x,xi∗) (8)

where Kω(x,xi∗) = 1(2πω2)N/2

exp(−(x−xi∗)

T (x−xi∗)2ω2 ).

In order to generate new QAOA parameters z using (8),a data point xi∗ from Sp∗ is chosen uniformly randomlywith probability 1/N , and it is added to a sample X′

∼ N (0, ω2I), that is, a sample drawn from a multivariateGaussian distribution with zero mean and diagonal covari-ance matrix ω2I. Noting that fX′(x) = Kω(x,0), one caneasily confirm that this sampling strategy yields data pointsz distributed according to (8),

P(Z ≤ z) = FZ(z) = P(X + X′ ≤ z)

=

N−1∑i=0

P(X + X′ ≤ z|X = xi∗)P(X = xi∗)

=1

N

N−1∑i=0

P(X′ ≤ z− xi∗)

=1

N

N−1∑i=0

FX′(z− xi∗)

(9)and hence, fZ(z) = 1

N

∑N−1i=0 Kω(z − xi∗,0) =

1N

∑N−1i=0 Kω(z,xi∗).

This methodology assumes that Sp∗ is readily available. Inorder to construct Sp∗ in practice, a derivative-free off-the-shelf optimizer is started from a large number (∼ 10, 000as in (Zhou et al. 2018)) of random points in the domain of(2). Because (2) is known to be a nonconvex function in theparameter space, a portion of these extensive searches mayconverge to low-quality local optima. To tackle this issue,we admit into Sp∗ only parameters that achieve an optimalityratio of 99% or higher for a given training instance and agiven QAOA circuit depth.

4.3 Graph Max-Cut InstancesIn this subsection, we describe graph instances in GTrainand GTest. Four classes of graphs are considered: (1) Erdos-Renyi random graphs GR(nR, ep), where nR is the num-ber of vertices and ep is the edge generation probability, (2)ladder graphs GL(nL), where nL is the length of the lad-der, (3) barbell graphs GB(nB), formed by connecting twocomplete graphs KnB by an edge, and (4) Caveman graphsGC(nC , nk), where nC is the number of cliques and nk isthe size of each clique. Figure 1 shows sample graph in-stances drawn from each graph class.

Random Ladder

Barbell Caveman

Figure 1: Sample graph instances: random GR(nR =8, ep = 0.5), ladder GL(nL = 4), barbell GB(nB = 4),and caveman GC(nC = 2, nk = 4).

To construct GTrain, we choose one representative graphinstance of 8 vertices from each class and distribution,amounting to |GTrain| = 7 training instances. GTest contains94 instances with varying number of vertices, as shown in

00

f(0,

0)(a) GR(nR = 16, ep = 0.5)

0 0

f(0,

0)

(b) GL(nL = 8)

00

f(0,

0)

(c) GB(nB = 8)

00

f(0,

0)

(d) GC(nC = 2, nk = 8)

Figure 2: QAOA energy landscapes for p = 1 (a) random graph, (b) ladder graph, (c) barbell graph, and (d) Ccveman graph.

Table 1: Train and test graph Max-Cut instances.Graph Class GTrain GTest

GR(nR, ep) nR = 8, nR ∈ {8, 12, 16, 20}, seed = {1, 2, 3, 4}ep ∈ {0.5, 0.6, 0.7, 0.8} ep ∈ {0.5, 0.6, 0.7, 0.8}

GL(nL) nL = 4 nL ∈ {2, 3, 5, 6, 7, 8, 9, 10, 11}GB(nB) nB = 4 nB ∈ {3, 5, 6, 7, 8, 9, 10, 11}GC(nC , nk) (nC , nK) = (2, 4) {(nC , 4) : nC ∈ {3, 4, 5}},

{(nC , 3) : nC ∈ {3, 5, 7}},{(2, nK) : nK ∈ {3, 5, 6, 7, 8, 9, 10}}

|GTrain| = 7 |GTest| = 94

Table 1. We choose |GTrain| < |GTest| to demonstrate thatcombining our proposed machine learning approaches withQAOA can be a powerful tool for amortizing the QAOAoptimization cost across graph instances. In addition, wechoose to train on graph instances that are smaller than testinstances, to demonstrate that our proposed approaches areindependent of instance size or complexity. We note thatGTrain ∩GTest = {}.

5 Results and DiscussionIn this section, we present the results of our work. We useQiskit Aer to perform noiseless simulations of QAOA cir-cuits. In Figure 2 we show the expected energy landscapeof the cost Hamiltonian for a depth (p = 1) QAOA cir-cuit with two variational parameters β0 and γ0 for somegraph instances. We can see that the expected energy of thecost Hamiltonian is nonconvex in the parameter space and isnoisy because it is statistically estimated based on quantumcircuit measurements. These features tend to be more severeas the depth of the QAOA circuit increases, thus posing se-rious challenges for commonly used derivative-free off-the-shelf optimizers.

In Figure 3, we present training results associated with ourproposed deep RL- and KDE-based approaches for a depth(p = 1) QAOA circuit. The RL learning curve during thetraining procedure is shown in Figure 3(a). We can see thatthe expected total discounted rewards of the RL agent startsaround 1 at the beginning of training, which means the RLagent is performing as well as random sampling. As train-ing progresses, the performance of the learned optimizationpolicy on the training set improves. This can also be seen inFigure 3(b), which shows the best objective value for one ofthe training instances versus time steps of an episode at dif-ferent stages of training. The optimization policy learned atthe end of training, namely, epoch 750, produces a trajectorythat rises quickly to a higher value compared with the trajec-

tories produced by the optimization policies learned at themiddle and beginning of training. On the other hand, Figure3(c) shows contour lines for the learned bivariate probabilitydensity based on S1

∗ using our proposed KDE technique. Wecan see that optimal QAOA parameters for a depth (p = 1)QAOA circuits of training instances in GTrain concentrate insome regions in the domain of (2).

Next, we benchmark the performance of our trained RL-based QAOA optimization policy and the sampling-KDE-based QAOA optimization strategy by comparing their per-formance with common derivative-free off-the-shelf opti-mizers implemented in the NLopt nonlinear optimizationpackage (Johnson 2019), namely, BOBYQA, COBYLA,and Nelder-Mead, as well as a purely random sampling-based strategy. Starting from 10 randomly chosen variationalparameters in the domain of (2), each optimizer is given 10attempts with a budget of B = 192 quantum circuit eval-uations to solve QAOA energy landscapes correspondingto graph instances in GTest. In each of the 10 attempts, therandom sampling B variational parameters are sampled uni-formly (random), and in the KDE-based approach they arebased on the learned density (KDE); and the parameters withthe highest objective value are chosen as the solution. Sinceour primary focus is to devise methods that find high-qualityparameters with a few quantum circuit evaluations, we usethe learned optimization policy by RL to generate trajecto-ries of length B/2, and we resume the trajectory from thebest parameters found by using Nelder-Mead for the restof B/2 evaluations. This approach is motivated by our ob-servation that the RL-based approach reaches regions withhigh-quality solutions rather quickly, yet subsequent quan-tum circuit evaluations are spent without further improve-ments on the objective value (Khairy et al. 2019). The visu-alizations of the RL agent as it navigates the QAOA land-scape for p = 1 and the results of the pure RL-based methodare reported in (Khairy et al. 2019).

We compare our results with gradient-free methods forthe following reasons. First, analytical gradients are in gen-eral not available, and evaluating gradients on a quan-tum computer is computationally expensive (Crooks 2018;Guerreschi and Smelyanskiy 2017). Under reasonable as-sumptions about the hardware (Guerreschi and Matsuura2019), one evaluation of the objective takes 1 second. There-fore, minimizing the number of objective evaluations neededfor optimization is of utmost importance. Estimating gra-dient requires at least two evaluations for each parameter,

0 200 400 600Training Epoch

0.98

1.00

1.02

1.04

1.06

Expe

cted

Disc

ount

ed R

ewar

ds

(a) Learning curve of RL agent. At thebeginning of training, RL performance ontraining instances is as good as a random-based sampling strategy. As training pro-gresses, the learned RL policy collectshigher expected discounted rewards.

0 10 20 30 40 50 60Evaluation

0.80

0.85

0.90

0.95

1.00

1.05

Best

Obj

ectiv

e Va

lue

Learnt trajectory after Epoch 1Learnt trajectory after Epoch 360Learnt trajectory after Epoch 740

(b) Performance of RL agent on one traininginstance during training. The policy networkat the end of training (epoch 740) could pro-duce a trajectory that reaches a high valuequickly.

2 0 20

2

1

0

1

2

0

0.0

0.3

0.6

0.9

1.2

1.5

(c) Optimal QAOA parameters densitylearned via KDE. Concentration of optimalparameters in some regions of the parame-ters space is evident.

Figure 3: RL and KDE training-related results (p = 1)

making gradient-based methods noncompetitive within thebudget of 192 objective evaluations (≈ 200 sec projectedrunning time on a real quantum computer) that we chose forour benchmark. Second, gradients are sensitive to noise (Zhuet al. 2019), which includes stochastic noise originatingfrom the random nature of quantum computation, and thenoise caused by hardware errors. This is typically addressedby increasing the number of samples used to estimate theobjective. On the other hand, optimizers employed for hy-perparameter optimization, such as Bayesian optimization,require a few hundred evaluations simply to initialize thesurrogate model, ruling out these optimizers from our bench-mark where the evaluation budget is 192.

To report the performance, we group graph instances inGTest in three subgroups, (1) Random graphs, which con-tains all graphs of the form GR(nR, ep), 2) Communitygraphs, which contains graphs of the form GC(nC , nk) andGB(nB), and 3) Ladder graphs, which contains graphs ofthe form GL(nL). In Figure 4, we report a boxplot of theexpected optimality ratio, E[τG] = E[f/fopt], where the ex-pectation is with respect to the highest objective value at-tained by a given optimizer in each of its 10 attempts. Theoptimal solution to a graph instance in GTest is the largestknown f value found by any optimizer in any of its 10 at-tempts for a given depth p. One can see that the medianoptimality ratio achieved by our proposed approaches out-performs that of other commonly used optimizers. Whilethe performance of derivative-free optimizers, random sam-pling, and the RL-based approach degrades as the dimen-sion of parameters increase from p = 1 to p = 4, theKDE-based approach maintains high optimality ratios andis the superior approach in all cases and across differentgraph classes. The RL-based approach and random samplingrank second and third, respectively, in the majority of cases.Random sampling turns out to be a competitive candidatefor sampling of variational parameters, especially for low -epth p = 1 circuits, and for random graph instances. Wenote that this finding hasd not been identified prior to ourwork because random search was not included in the ex-

Table 2: Median optimality gap reduction factor w.r.tNelder-Mead.

Graph Class Proposed Optimizer p = 1 p = 2 p = 4

Random RL 3.07 1.81 1.37KDE 7.06 8.07 6.52

Community RL 2.66 1.47 1.05KDE 26.20 14.15 9.31

Ladder RL 5.59 1.87 1.12KDE 17.68 30.15 8.15

perimental comparison (Shaydulin, Safro, and Larson 2019;Nakanishi, Fujii, and Todo 2019; LaRose et al. 2019; Ver-don et al. 2019). In Table 2, we summarize the median op-timality gap reduction factor with respect to Nelder-Mead,attained by our proposed RL- and KDE-based approaches,namely, ( 1−ENM[τG]

1−ERL[τG] ) and ( 1−ENM[τG]1−EKDE[τG] ). As the table shows,

our proposed approaches reduce the optimality gap by fac-tors up to 30.15 compared with Nelder-Mead, and the gapreduction factor is consistently larger than 1.

In Figure 5, we show a boxplot of the expected approx-imation ratio performance, E[ηG] = E[f/Copt], of QAOAwith respect to the classical optimal Copt found using bruteforce methods across different graph instances in GTest. Wecan see that increasing the depth of QAOA circuit improvesthe attained approximation ratio, especially for structuredgraphs (i.e., community and ladder graph instances).

6 ConclusionIn this paper, we formulated the problem of finding optimalQAOA parameters for approximately solving combinatorialproblems as a learning task. Two machine-learning-basedapproaches have been proposed: an RL-based approach,which can learn a policy network that can efficiently opti-mize new QAOA circuits by exploiting geometrical regulari-ties in QAOA objectives, and a KDE-based approach, whichcan learn a generative model of optimal QAOA parametersthat can be used to sample parameters for new QAOA cir-cuits. Our proposed approaches have been trained on a small

Random Community Ladder

0.85

0.90

0.95

1.00

[G]

COBYLABOBYQANELDERMEADRANDOMRLKDE

(a) p = 1

Random Community Ladder

0.80

0.85

0.90

0.95

1.00

[G]

COBYLABOBYQANELDERMEADRANDOMRLKDE

(b) p = 2

Random Community Ladder

0.80

0.85

0.90

0.95

1.00

[G]

COBYLABOBYQANELDERMEADRANDOMRLKDE

(c) p = 4

Figure 4: Expected optimality ratio performance of different optimizers on graph instances in GTest for a given QAOA circuitdepth p ∈ {1, 2, 4}. Test graph instances are grouped in three subgroups: random graphs, community graphs, and laddergraphs. Our proposed machine-learning-based methods, RL and KDE, outperform commonly used derivative-free off-the-shelfoptimizers such as COBYLA, BOBYQA, and Nelder-Mead.

RANDOM COMMUNITY LADDER

0.7

0.8

0.9

1.0

[G]

p = 1p = 2p = 4

Figure 5: Approximation ratio performance of QAOA withrespect to classical optimal on GTest. With higher-depthQAOA circuit, the attained approximation ratio increases.

set of small-sized training instances, yet they are capable ofefficiently solving larger problem instances. When coupledwith QAOA, our proposed approaches can be powerful toolsfor amortizing the QAOA optimization cost across combina-torial instances.

In our future work, we will investigate machine-learning-based methods for QAOA applied to constrained combina-torial optimization problems such as maximum independentset and max κ colorable subgraphs, which have importantapplications in many disciplines.

AcknowledgmentsThis material is based upon work supported by the U.S. De-partment of Energy, Office of Science, Office of AdvancedScientific Computing Research, under Contract DE-AC02-06CH11357. This research was funded in part by and usedresources of the Argonne Leadership Computing Facility,which is a DOE Office of Science User Facility supportedunder Contract DE-AC02-06CH11357. We gratefully ac-knowledge the computing resources provided on Bebop, ahigh-performance computing cluster operated by the Lab-oratory Computing Resource Center at Argonne NationalLaboratory. LC acknowledges support from LANLs Labora-tory Directed Research and Development (LDRD) program.

LC was also supported by the U.S. Department of Energy,Office of Science, Office of Advanced Scientific Comput-ing Research, under the Quantum Computing ApplicationTeams program.

References[Andrychowicz et al. 2016] Andrychowicz, M.; Denil, M.;Gomez, S.; Hoffman, M. W.; Pfau, D.; Schaul, T.; Shilling-ford, B.; and De Freitas, N. 2016. Learning to learn bygradient descent by gradient descent. In Advances in NeuralInformation Processing Systems, 3981–3989.

[Apolloni, Carvalho, and de Falco 1989] Apolloni, B.; Car-valho, C.; and de Falco, D. 1989. Quantum stochasticoptimization. Stochastic Processes and Their Applications33(2):233–244.

[Balaprakash et al. 2018] Balaprakash, P.; Salim, M.; Uram,T.; Vishwanath, V.; and Wild, S. 2018. DeepHyper: Asyn-chronous hyperparameter search for deep neural networks.In 2018 IEEE 25th International Conference on High Per-formance Computing (HiPC), 42–51.

[Balaprakash, Birattari, and Stutzle 2007] Balaprakash, P.;Birattari, M.; and Stutzle, T. 2007. Improvement strategiesfor the F-Race algorithm: Sampling design and iterativerefinement. In et al, B.-B., ed., Hybrid Metaheuristics,108–122. Berlin, Heidelberg: Springer Berlin Heidelberg.

[Barkoutsos et al. 2019] Barkoutsos, P. K.; Nannicini, G.;Robert, A.; Tavernelli, I.; and Woerner, S. 2019. Improv-ing variational quantum optimization using CVaR. arXivpreprint arXiv:1907.04769.

[Brandao et al. 2018] Brandao, F. G.; Broughton, M.; Farhi,E.; Gutmann, S.; and Neven, H. 2018. For fixed control pa-rameters the quantum approximate optimization algorithm’sobjective function value concentrates for typical instances.arXiv:1812.04170.

[Crooks 2018] Crooks, G. E. 2018. Performance of the quan-tum approximate optimization algorithm on the maximumcut problem. arXiv:1811.08419.

[Farhi, Goldstone, and Gutmann 2014] Farhi, E.; Goldstone,

J.; and Gutmann, S. 2014. A quantum approximate opti-mization algorithm. arXiv:1411.4028.

[Feurer, Springenberg, and Hutter 2015] Feurer, M.; Sprin-genberg, J. T.; and Hutter, F. 2015. Initializing Bayesianhyperparameter optimization via meta-learning. In Twenty-Ninth AAAI Conference on Artificial Intelligence.

[Goemans and Williamson 1995] Goemans, M. X., andWilliamson, D. P. 1995. Improved approximation al-gorithms for maximum cut and satisfiability problemsusing semidefinite programming. Journal of the ACM42(6):1115–1145.

[Guerreschi and Matsuura 2019] Guerreschi, G. G., andMatsuura, A. Y. 2019. QAOA for max-cut requires hundredsof qubits for quantum speed-up. Scientific Reports 9(1).

[Guerreschi and Smelyanskiy 2017] Guerreschi, G. G., andSmelyanskiy, M. 2017. Practical optimization for hybridquantum-classical algorithms. arXiv:1701.01450.

[Hadfield et al. 2017] Hadfield, S.; Wang, Z.; O’Gorman,B.; Rieffel, E. G.; Venturelli, D.; and Biswas, R. 2017.From the quantum approximate optimization algorithm toa quantum alternating operator ansatz. arXiv preprintarXiv:1709.03489.

[Huang et al. 2019] Huang, C.; Szegedy, M.; Zhang, F.; Gao,X.; Chen, J.; and Shi, Y. 2019. Alibaba cloud quantumdevelopment platform: Applications to quantum algorithmdesign. arXiv preprint arXiv:1909.02559.

[Hutter, Kotthoff, and Vanschoren 2018] Hutter, F.; Kot-thoff, L.; and Vanschoren, J., eds. 2018. Automated Ma-chine Learning: Methods, Systems, Challenges. Springer.available at http://automl.org/book.

[Johnson 2019] Johnson, S. G. 2019. The NLopt nonlinear-optimization package.

[Kato 1950] Kato, T. 1950. On the adiabatic theorem ofquantum mechanics. Journal of the Physical Society ofJapan 5(6):435–439.

[Khairy et al. 2019] Khairy, S.; Shaydulin, R.; Cincio, L.;Alexeev, Y.; and Balaprakash, P. 2019. Reinforcement-learning-based variational quantum circuits optimization forcombinatorial problems. arXiv preprint arXiv:1911.04574.

[LaRose et al. 2019] LaRose, R.; Tikku, A.; ONeel-Judy, E.;Cincio, L.; and Coles, P. J. 2019. Variational quantum statediagonalization. npj Quantum Information 5(1):8.

[Li and Malik 2016] Li, K., and Malik, J. 2016. Learning tooptimize. arXiv preprint arXiv:1606.01885.

[Nakanishi, Fujii, and Todo 2019] Nakanishi, K. M.; Fujii,K.; and Todo, S. 2019. Sequential minimal optimization forquantum-classical hybrid algorithms. arXiv:1903.12166.

[Pagano et al. 2019] Pagano, G.; Bapat, A.; Becker, P.;Collins, K.; De, A.; Hess, P.; Kaplan, H.; Kyprianidis, A.;Tan, W.; Baldwin, C.; et al. 2019. Quantum approximateoptimization with a trapped-ion quantum simulator. arXivpreprint arXiv:1906.02700.

[Parekh, Ryan-Anderson, and Gharibian 2019] Parekh,O. D.; Ryan-Anderson, C.; and Gharibian, S. 2019. Quan-

tum optimization and approximation algorithms. Technicalreport.

[Romero et al. 2018] Romero, J.; Babbush, R.; McClean, J.;Hempel, C.; Love, P.; and Aspuru-Guzik, A. 2018. Strate-gies for quantum computing molecular energies using theunitary coupled cluster ansatz. Quantum Science and Tech-nology.

[Schulman et al. 2017] Schulman, J.; Wolski, F.; Dhariwal,P.; Radford, A.; and Klimov, O. 2017. Proximal policy op-timization algorithms. arXiv preprint arXiv:1707.06347.

[Shaydulin et al. 2018] Shaydulin, R.; Ushijima-Mwesigwa,H.; Safro, I.; Mniszewski, S.; and Alexeev, Y. 2018. Com-munity detection across emerging quantum architectures.Proceedings of the 3rd International Workshop on PostMoore’s Era Supercomputing.

[Shaydulin et al. 2019] Shaydulin, R.; Ushijima-Mwesigwa,H.; Safro, I.; Mniszewski, S.; and Alexeev, Y. 2019. Net-work community detection on small quantum computers.Advanced Quantum Technologies 1900029.

[Shaydulin, Safro, and Larson 2019] Shaydulin, R.; Safro,I.; and Larson, J. 2019. Multistart methods for quantumapproximate optimization. 2019 IEEE High PerformanceExtreme Computing Conference (HPEC).

[Streif and Leib 2019] Streif, M., and Leib, M. 2019. Com-parison of QAOA with quantum and simulated annealing.arXiv preprint arXiv:1901.01903.

[Sutton and Barto 2018] Sutton, R. S., and Barto, A. G.2018. Reinforcement learning: An introduction. MIT Press.

[Verdon et al. 2019] Verdon, G.; Broughton, M.; McClean,J. R.; Sung, K. J.; Babbush, R.; Jiang, Z.; Neven, H.; andMohseni, M. 2019. Learning to learn with quantum neu-ral networks via classical neural networks. arXiv preprintarXiv:1907.05415.

[Wang et al. 2018] Wang, Z.; Hadfield, S.; Jiang, Z.; and Ri-effel, E. G. 2018. Quantum approximate optimization al-gorithm for MaxCut: a fermionic view. Physical Review A97:022304.

[Wecker, Hastings, and Troyer 2016] Wecker, D.; Hastings,M. B.; and Troyer, M. 2016. Training a quantum optimizer.Physical Review A 94(2):022309.

[Yang et al. 2017] Yang, Z.-C.; Rahmani, A.; Shabani, A.;Neven, H.; and Chamon, C. 2017. Optimizing variationalquantum algorithms using Pontryagin’s minimum principle.Physical Review X 7(2):021027.

[Zhou et al. 2018] Zhou, L.; Wang, S.-T.; Choi, S.; Pichler,H.; and Lukin, M. D. 2018. Quantum approximate optimiza-tion algorithm: Performance, mechanism, and implementa-tion on near-term devices. arXiv:1812.01041.

[Zhu et al. 2019] Zhu, D.; Linke, N. M.; Benedetti, M.;Landsman, K. A.; Nguyen, N. H.; Alderete, C. H.; Perdomo-Ortiz, A.; Korda, N.; Garfoot, A.; Brecque, C.; et al. 2019.Training of quantum circuits on a hybrid quantum computer.Science advances 5(10):eaaw9918.

The submitted manuscript has been created byUChicago Argonne, LLC, Operator of Argonne Na-tional Laboratory (“Argonne”). Argonne, a U.S. De-partment of Energy Office of Science laboratory, isoperated under Contract No. DE-AC02-06CH11357.The U.S. Government retains for itself, and others act-ing on its behalf, a paid-up nonexclusive, irrevocableworldwide license in said article to reproduce, preparederivative works, distribute copies to the public, andperform publicly and display publicly, by or on be-half of the Government. The Department of Energywill provide public access to these results of federallysponsored research in accordance with the DOE PublicAccess Plan. http://energy.gov/downloads/doe-public-access-plan