Complexity Classification of Product State Problems for Local Hamiltonians
Abstract
Product states, unentangled tensor products of single qubits, are a ubiquitous ansatz in quantum computation, including for state-of-the-art Hamiltonian approximation algorithms. A natural question is whether we should expect to efficiently solve product state problems on any interesting families of Hamiltonians.
We completely classify the complexity of finding minimum-energy product states for Hamiltonians defined by any fixed set of allowed 2-qubit interactions. Our results follow a line of work classifying the complexity of solving Hamiltonian problems and classical constraint satisfaction problems based on the allowed constraints. We prove that estimating the minimum energy of a product state is in if and only if all allowed interactions are 1-local, and -complete otherwise. Equivalently, any family of non-trivial two-body interactions generates Hamiltonians with -complete product-state problems. Our hardness constructions only require coupling strengths of constant magnitude.
A crucial component of our proofs is a collection of hardness results for a new variant of the Vector Max-Cut problem, which should be of independent interest. Our definition involves sums of distances rather than squared distances and allows linear stretches.
We similarly give a proof that the original Vector Max-Cut problem is -complete in 3 dimensions. This implies hardness of optimizing product states for Quantum Max-Cut (the quantum Heisenberg model) is -complete, even when every term is guaranteed to have positive unit weight.
Keywords and phrases:
quantum complexity, quantum algorithms, local hamiltoniansCopyright and License:
![[Uncaptioned image]](x1.png)
2012 ACM Subject Classification:
Theory of computation Quantum complexity theoryAcknowledgements:
Sandia National Laboratories is a multimission laboratory managed and operated by National Technology and Engineering Solutions of Sandia, LLC., a wholly owned subsidiary of Honeywell International, Inc., for the U.S. Department of Energy’s National Nuclear Security Administration under contract DE-NA-0003525. This written work is authored by an employee of NTESS. The employee, not NTESS, owns the right, title and interest in and to the written work and is responsible for its contents. Any subjective views or opinions that might be expressed in the written work do not necessarily represent the views of the U.S. Government. The publisher acknowledges that the U.S. Government retains a non-exclusive, paid-up, irrevocable, world-wide license to publish or reproduce the published form of this written work or allow others to do so, for U.S. Government purposes. The DOE will provide public access to results of federally sponsored research in accordance with the DOE Public Access Plan.Funding:
This work was supported by the U.S. Department of Energy, Office of Science, Office of Advanced Scientific Computing Research, Accelerated Research in Quantum Computing, Fundamental Algorithmic Research for Quantum Computing. O.P. was also supported by U.S. Department of Energy, Office of Science, National Quantum Information Science Research Centers, Quantum Systems Accelerator. JY was partially supported by Scott Aaronson’s Simons Investigator Award.Editors:
Raghu MekaSeries and Publisher:

1 Introduction
Product states, unentangled tensor products of single-qubit states, have served as an effective focus for better understanding quantum phenomena. Because general quantum states cannot be described efficiently, approximation algorithms must be restricted to output some subset of states, an ansatz. Mean-field approaches are common as first steps in statistical mechanics, and recent approximation algorithms for extremal energy states of local Hamiltonians have relied on proving that product states provide good approximations in particular regimes [17, 4, 5, 18, 31, 32]. In fact, because in some natural regimes the ground states are rigorously well-approximated by product states [4], optimal approximation algorithms for local Hamiltonians on arbitrary interaction graphs must be capable of finding good product states. Understanding product state optimization is essential for understanding the complexity of Hamiltonian approximation generally.
Product states are a natural intermediate between classical and quantum states, allowing for superposition but not entanglement. Unlike general quantum states, they have succinct classical descriptions: a single-qubit pure state can be specified by two complex numbers, and an -qubit pure product state can be specified by complex numbers. One could consider “more quantum” intermediates, in the form of reduced states of two or more qubits. However, verifying the consistency of a set of quantum marginals is a -complete problem, even for 2-qubit reduced states [27, 10]. Therefore, product states are uniquely useful when optimizing directly over state vectors.
We study the following question: for a family of Hamiltonians defined by a given set of allowed interactions, what is the complexity of computing the extremal energy over product states? Additionally, how does the complexity of optimizing over product states relate to that of optimizing over general states? For example, for a -hard local Hamiltonian, must finding the optimal product state in turn be -hard?111The product state problem is always in since product states have succinct classical descriptions with which we can compute the expected energy contribution from each local Hamiltonian term in time polynomial in the size of the term.
This question follows a long line of work classifying the complexity of constraint satisfaction problems (CSPs) based on the sets of allowed constraints, clauses, or interactions between variables. In particular, the dichotomy theorem of Schaefer [35] showed that for any set of allowed Boolean constraints, the family of CSPs is either efficiently decidable or is -complete. In the context of quantum problems, Cubitt and Montanaro [14] introduced a similar classification of ground state energy problems for 2-local Hamiltonians, showing that for any fixed set of allowed 2-qubit interactions, the -LH problem is either in or -, -, or -complete (the case relies on the concurrent work of Bravyi and Hastings [6]). We briefly survey some of this line of classical and quantum work in Related Work below.
While the complexity of finding the extremal states (e.g. the ground state) of 2-local Hamiltonians is well understood, the complexity of finding optimal product state solutions has been only sparsely studied [22]. The only -hardness results for such problems are based on mapping a classical problem to a diagonal 0-1 valued Hamiltonian [39].
An additional motivation for our study is the hope of developing new methods for identifying families of local Hamiltonians for which problems involving general ground states are not hard. While a complete complexity classification for the general problem is known, more refined attempts at classification which take into account restrictions on the sign of the weights or geometry of the system are currently incomplete [33]. Developing algorithms for product states is a “mostly classical” problem that is easier to analyze, and progress involving product states may inform our expectations regarding general states.
In this work, we completely classify the complexity of finding optimal product states for families of 2-local Hamiltonians. In fact, we find the complexity of the product state problem is fully determined by the complexity of the general local Hamiltonian problem: if the general problem is -hard, the product state problem is -complete, and otherwise it is in . To arrive at our results, we study a variant of the Vector Max-Cut which should be of independent interest especially to the optimization community. As a corollary to our classification theorem, we give the first published proof that estimating optimal product state energies in the Quantum Max-Cut model is -complete, and we show hardness holds even for unweighted Hamiltonians.222 A more complex unpublished proof based on large graph cycles was known earlier by Wright [40].
1.1 Our Contributions
Formal definitions are given in Section 2. A -local Hamiltonian is a sum of Hamiltonian terms each of which only acts non-trivially on at most qubits, analogous to -variable Boolean clauses. denotes the problem of estimating the ground state energy (the minimum eigenvalue across all states) of a -local Hamiltonian to inverse-polynomial additive precision. Given a set of local terms , -LH is restricted to Hamiltonians such that every term belongs to . Finally, prodLH and are the restrictions of these problems to product states, i.e. to minimize where is the Hamiltonian and ranges over tensor products of single-qubit states.333 An earlier version of this work considered exact versions of product state and graph problems. We have now improved our hardness results to hold up to an inverse-polynomial additive gap.
2-local
The classification of the general ground state energy problem by Cubitt and Montanaro [14] completely classifies -LH for any fixed set of 2-qubit terms, showing it is either in or it is one of -, -, or -complete.
In the same vein, we give a complete classification of product state complexity for families of 2-local Hamiltonians as a function of the set of allowed 2-qubit interactions. For any given set of 2-qubit terms, we prove the problem prodLH is either in or is -complete. To the best of our knowledge, ours is the first systematic inquiry into the complexity of product state problems.
Theorem 1.
For any fixed set of 2-qubit Hamiltonian terms , if every matrix in is 1-local then is in , and otherwise is -complete.
Additionally, our hardness constructions only require coupling strengths (weights) of at most constant magnitude. This is preferable in practice and contrasts with most known -hardness constructions.
The sets for which [14] shows -LH is in are the same for which we show prodLH is in , i.e. those containing only 1-local terms. This immediately implies that a family of 2-local Hamiltonians has efficiently-computable minimum product state energy if and only if it has efficiently-computable ground state energy.
Corollary 2.
For any fixed set of 2-qubit Hamiltonian terms , the problem is in if and only if is in .
Corollary 3.
For any fixed set of 2-qubit Hamiltonian terms , the problem is -hard if and only if is -complete.
Our results imply that hardness of product state approximations is not restricted to Hamiltonians for which product states well approximate ground states: for any -hard family of terms, our result implies that (assuming ) we can construct a family of local Hamiltonians that are -hard to product state approximate and for which the product states do not well approximate the ground state by constructing Hamiltonians on two systems, each with one of these properties, and taking a disjoint union thereof. This implies that algorithms using product states to approximate the ground states of -hard Hamiltonians face a “double penalty”: hardness of approximating product states which themselves imperfectly approximate ground states.
The Stretched Linear Vector Max-Cut problem
Our hardness constructions for product state problems embed an objective function which we prove is -complete. This objective function generalizes the classical Max-Cut problem. Given work on other variations of Max-Cut, we expect this problem and our reductions should be of independent interest, especially to the optimization and approximation communities.
In Max-Cut, one is given a graph and asked to assign to each vertex a label so as to achieve the maximum number of oppositely labeled adjacent vertices:
(1) |
A problem referred to as Vector Max-Cut, Rank--Max-Cut, or () has been studied [8, 7, 22] which generalizes Max-Cut to assigning -dimensional unit vectors so as to maximize the angles between adjacent vertex labels, or equivalently to maximize the squared distances between adjacent vertex labels:
(2) |
Our new problem can be seen as a stretched and linear version of . The goal in () is to assign unit vectors so as to maximize the distance between adjacent labels:
(3) |
where is a fixed diagonal matrix. Comparing and , our problem sums over un-squared distances and incorporates a linear stretch given by . We consider the decision version of this problem, in which the objective is to test whether the optimal solution is at least or no more than , for . Note that, unlike , this is an unweighted problem – one could naturally define a weighted version but our hardness results will not require this.
Geometrically, corresponds to embedding a graph into the surface of a unit sphere with the objective of maximizing the sum of the squared lengths of every edge. Likewise, our problem corresponds to embedding a graph into the surface of a -dimensional ellipsoid, with radii defined by the entries of , with the objective of maximizing the sum of the (non-squared) edge lengths.
Despite being generalizations of the -complete Max-Cut problem, hardness of neither nor is trivial. The Goemans-Williamson approximation algorithm for Max-Cut on an -vertex graph begins with efficiently computing the solution to via an SDP. In fact, deciding is known to be in for any , [28, Theorem 8.4] or [3, (2.2)]. And while it has been conjectured by Lovász that is -complete for all constants , in [29, p. 236] and earlier, no proof has been given for any .
Our main theorem concerning , which is used to prove Theorem 1, is the following.
Theorem 4.
For any fixed non-negative with at least one of nonzero, is -complete.
Quantum Max-Cut Product States and
As a corollary of our classification theorem, we give the first published proof of the fact that product state optimization in the Quantum Max-Cut (QMC) model is -hard. This model, also known as the anti-ferromagnetic Heisenberg model, is equivalent to -LH with . We note that a sketch of a different proof for this specific problem was previously known but unpublished [40]. That proof was based on large graph cycles, and our gadgets are simpler to analyze.
However, the proof of Theorem 1 utilizes Hamiltonian gadgets involving negative weights (unlike the aforementioned proof of [40]). This leaves open whether prodQMC remains -hard on unweighted graphs. In Section 5, we give a direct proof of hardness using the fact that the unweighted product-state version of QMC is equivalent to (Equation 2). Our work then is also the first published proof that is -complete for some (in our case, ), partially resolving a conjecture of Lovász [29, p. 236]. Note that as with , we consider the decision version in which the goal is to determine whether the value is above or below , for and with inverse-poly separation.
Theorem 5.
is -complete.
Corollary 6.
Quantum Max-Cut restricted to product states, prodQMC, is -complete, even when all terms are restricted to have positive unit weight.
1.2 Proof Overview
2-local
As product states have classical descriptions and their energies can be calculated in polynomial time, is automatically in , so we focus on how we show hardness. Our approach is in two parts. We show how to reduce to prodLH, and later we show that -complete (Theorem 4). More precisely, the first part of our approach is to show that for any containing a strictly 2-local term, there exists a corresponding weight matrix meeting the conditions of Theorem 4 so that is -hard. For any instance of with this fixed , we show how to construct a Hamiltonian from such that the minimum product state energy encodes the value, yielding Theorem 1.
We interpret problems over product states as optimization problems on the collection of Bloch vectors for each single-qubit state. For example, consider with the specific set (the QMC model). In this case, by writing each qubit in the Bloch vector representation
the energy contributed by an interaction between qubits and is
(4) |
So, given a Hamiltonian which is the sum of XYZ interactions between pairs of qubits, the problem of estimating the extremal product state energies is equivalent to optimizing the objective function
over 3-dimensional unit vectors, where each edge corresponds to the Hamiltonian’s (weighted) interaction graph. Up to constant shifts and scaling, this is equivalent to , introduced in Equation 2.
More generally, because the Pauli matrices are a basis for all Hermitian matrices, any 2-qubit interaction can be written as , where the are the Pauli matrices, for some , , . Then, the energy of a given product state is calculated similarly to Equation 4. However, the resulting expression potentially contains many terms.
Our approach is to take an arbitrary 2-qubit term and insert it into gadgets which simplify the energy calculations. First, in the proof of Theorem 1, we borrow a trick of [14] and symmetrize the terms. For any 2-qubit interaction on qubits , the combined interaction is symmetric, invariant under swapping the qubits. A similar trick handles the case of anti-symmetric terms.
Second, in the proofs of Lemmas 14 and 15, we show how to use a symmetric or anti-symmetric term to embed the value into the minimum energy of a gadget. We begin by removing the 1-local terms, such as or , again taking inspiration from gadgets used by [14]. For two qubits corresponding to two vertices in a instance, the gadget adds two ancilla qubits and weights each interaction within the gadget to effectively cancel out the 1-local terms. When the two ancilla qubits vary freely, we find the minimum energy contributed by the entire four-qubit gadget is determined by the distance between the states of and . Although each individual edge contributes energy proportional to the squared distance between their states, the overall gadget contributes energy proportional to just the distance of the two “vertex qubits”, . With some massaging, we can treat as a non-negative diagonal matrix which meets the conditions of Theorem 4.
Therefore, as desired, we have that the value for an -complete instance of can be embedded into the minimum product state energy of an prodLH instance.
Stretched Linear Vector Max-Cut
To prove that is -hard for any fixed diagonal non-negative with at least one nonzero entry, we divide into three cases, in Lemmas 23, 21, and 22. Our first proof is a reduction from the standard Max-Cut problem, while the other two are reductions from 3-Coloring.
When there is a unique largest entry of , we reduce from Max-Cut by taking any input graph and forming by adding large star gadgets around each of the vertices of , each using many ancilla vertices. Because the ancilla vertices in each gadget have just one neighbor (the original vertex at the center of the gadget), their optimal vector labels given any choice of labels for the center vertices are the negation of the center vertex labels. This means they heavily penalize assigning the center vertices any labels that are not along the highest-weight axis. Therefore, when the maximum entry of is unique, the optimal assignment to will have its vector labels almost entirely along the highest-weight axis. The assignment can trivially earn the maximum possible value on the star gadgets, and the amount additional amount it can earn on the original edges of corresponds to the Max-Cut value of .
When all of the entries of are equal, we reduce from 3-Coloring. Given a graph , we construct by replacing each edge with a -clique gadget, made by adding one ancilla vertex per gadget, along with a single ancilla vertex shared by every gadget. We show that is 3-colorable iff has a sufficiently large value. Specifically, we show this holds iff there is a vector assignment that simultaneously achieves (nearly) the maximum value on all of the clique gadgets. Achieving the maximum objective value on a clique corresponds to maximizing the total distance between each pair of vectors, and this enforces a predictable arrangement.
When weights are equal, assigning these vectors can be viewed as inscribing vectors in the unit sphere, and it is known that maximal perimeter polyhedra inscribed in the sphere must be regular. So for a 4-clique, the vector labels must form a regular tetrahedron. We carefully argue that for regular tetrahedra, fixing two of the vertices (approximately) fixes the other two vertices up to swapping (several of these geometric facts are proved in Appendix A). This means that the clique gadget corresponding to an edge in shares two vertices with any gadget corresponding to an edge incident to the first edge: and the “global” ancilla shared by all gadgets. This means that, once we fix the vector assigned to the global ancilla, the choice of a vector for restricts the labels of both and to be chosen from a set of two vectors. So simultaneously optimizing every clique gadget is possible iff, for each connected component of , we can 3-color that component with three vectors (corresponding to, for any in the component, the vector assigned to , the vector assigned to the global ancilla vertex, and the two vectors that can share a maximal tetrahedron with those two).
Finally, when the two largest entries of are equal but distinct from the third, we combine the two previous approaches. Inserting star gadgets effectively reduces the problem from three dimensions to two, by penalizing vector assignments not in the 2d space corresponding to the two largest entries of . We then add -clique gadgets, with one ancilla for each edge in , and optimizing these over two dimensions corresponds to inscribing maximal perimeter triangles in the unit circle. Now, assigning a vector to one vertex fixes the optimal vectors assigned to the other two vertices (again up to swapping them) and so there is again a one-to-one correspondence between vector assignments simultaneously optimizing every clique gadget and 3-colorings of the connected components of .
Three-Dimensional Vector Max-Cut
Our proof that with positive unit weights is -complete runs on very similar lines to our 3-Coloring reduction for with all weights equal. However, there is one additional complication: that proof depended on the fact that the sum of the side lengths of the tetrahedron is uniquely maximized by choosing it to be regular. This is not the case when we instead consider the squared side lengths, for instance assigning half the vectors to one pole of the sphere and half to the other would achieve the same bound. So we use a different gadget: instead of replacing every edge with a 4-clique, we replace every edge with a 4-clique that in turn has its edges replaced with triangles. It turns out that this gadget does have a unique optimal assignment, which in particular assigns the vectors in the 4-clique to a regular tetrahedron, allowing us to proceed along the same lines as the aforementioned proof.
1.3 Related Work
Brandão and Harrow [4] give simple conditions under which -local Hamiltonians have product states achieving near-optimal energy, such as systems with high-degree interaction graphs. This suggests that unless , such Hamiltonians cannot be -complete. Since then, there has been a line of work on the relationship between product states and general ground states in other Hamiltonians, in the more general case when the two problems are not equivalent. See e.g. [5, 17, 20].
Product states have especially been studied in the context of the Quantum Max-Cut problem, introduced by Gharibian and Parekh [18]. Briët, de Oliveira Filho, and Vallentin [8] give hardness results conditional on the unique games conjecture for approximating the optimal product state in the QMC model. Related work by Hwang, Neeman, Parekh, Thompson, and Wright [22] also gives tight hardness results for the QMC problem under a plausible conjecture. Parekh and Thompson [32] give an optimal approximation algorithm for QMC when using product states.
See [22] for an exposition of the relationship between QMC restricted to product state solutions and the Vector Max-Cut problem. Studying vector solutions to Max-Cut has a long history [29], including the seminal Goemans-Williamson algorithm [19]. This study is usually with the goal of a solution to the original Max-Cut problem, which relates to approximation ratios of integer and semidefinite programs. Bounding these ratios has been referred to in terms of Grothendieck problems and inequalities: see [7, 2] for further context on this nomenclature. A tight NP-hardness result is known for the non-commutative Grothendieck problem [9] which also generalizes the “little” Grothendieck problem over orthogonal groups [2]. Iterative algorithms (heuristics) for solving are also well-studied in the literature (see [11] and citing references), since in practice solving is often faster than solving the corresponding SDP relaxation.
Cubitt and Montanaro [14] classified -LH for sets of 2-qubit terms. This work relies in turn on the work of Bravyi and Hastings [6] to classify the case. [14] also examined a variant where is assumed to always contain arbitrary single-qubit terms. Follow-up work by Piddock and Montanaro [33] began investigating classifying the complexity of -LH under the additional restrictions of positive weights (anti-ferromagnetic model) and/or interactions restricted to a spatial geometry such as a 2D grid. [15, 34] continued along these lines, and introduced Hamiltonian simulations rather than computational reductions.
In classical computer science, Schaefer [35] gave a dichotomy theorem showing that given a fixed set of allowed constraints, the family of CSPs is either decidable in or is -complete; but see Section 2 or [14] for some important assumptions that are made in the quantum versus classical models. In fact, Schaefer’s classification offers a more fine-grained classification with classes within . Later, [12, 25] gave a similar dichotomy theorem for the complexity of Max-SAT and related optimization problems, where the question is not just whether all clauses are simultaneously satisfiable but how many are simultaneously satisfiable. Applying weights to constraints becomes relevant with Max-SAT, and is covered in their work. This work is especially relevant given -LH is more analogous to Max-SAT than to SAT. Continuing, Jonsson [23] classified these problems when both positive and negative weights are allowed. While arbitrary weights seem natural in the quantum setting, previous classical work simply assumed all weights were non-negative. The book [13] offers an excellent survey of this area. The more recent results of [24, 36] extend classical classification results to problems with non-binary variables (analogous to qudits).
1.4 Open Questions
-
1.
We have shown a relationship between when product state problems and general Hamiltonian problems are hard. This points toward an important question: can some “interesting” class of local Hamiltonians or a Hamiltonian problem for which we do not know an explicit efficient algorithm be proven not hard, e.g. neither - nor -hard, by showing the corresponding product state problem is in ?
-
2.
Is a more refined classification of the complexity of product state problems, taking into account allowable weights or spatial geometry in the vein of [33], or imposing other promises, possible?
-
3.
While little progress has been made classifying the general problem for higher-locality families, can we classify prodLH for families of -local Hamiltonians with ?
-
4.
Can we relate approximability instead of just complexity? For example, does the ability to -approximate the product state problem imply the ability to -approximate the general ground state problem on families defined by some sets of allowed interactions but not others?
-
5.
As mentioned above, we make the first progress towards a conjecture of Lovász [29] that is -hard for any . We only focus on here because of our interest in prodLH. Can our proof be generalized to other values of ?
2 Preliminaries
We assume familiarity with the conventions of quantum computation [38] and complexity theory [1, 26]. See also [16, 14] for surveys of Hamiltonian complexity.
2.1 Notation
denotes the identity operator. and denote the minimum and maximum eigenvalues of an operator . In the same manner as with asymptotic notation, we use to denote a term that can be bounded by some fixed polynomial in .
For an operator , we use superscripts such as to indicate acts on individual qubits and . Unless is symmetric, the order matters and is different than . If no superscripts are used, then the action is implicit in the ordering of the terms (left versus right).
When clear from context, we will denote the tensor product of two operators simply by . All terms implicitly are tensor the identity on any systems not specified.
will denote the 2-qubit operator exchanging and while and unchanged. We call a 2-qubit term symmetric if , meaning the ordering of the qubits does not matter. Alternatively, is antisymmetric if .
The single-qubit Pauli matrices are denoted or . Recall that is a basis for Hermitian matrices. The Pauli decomposition of a 2-qubit Hermitian matrix is written in the Pauli basis,
(5) |
with all coefficients real and the matrix referred to as the correlation matrix. Generally, Equation 5 should include a term , but we will work with traceless terms such that .
Unless otherwise stated, all graphs are undirected and simple, meaning there are no self-loops and no multi-edges. We assume all graphs are connected, as it is straightforward to extend any of our constructions to disconnected graphs. When summing over edges, , we do not double-count and . Finally, denotes the unit sphere in -dimensional space.
2.2 Definitions and Assumptions
A -local Hamiltonian on qubits is a Hermitian matrix that can be written as such that each is Hermitian and acts non-trivially on at most qubits. More precisely, each acts on some subset of at most qubits and each term in the sum is , but we generally leave this implicit. We usually consider constant values of , so each term is of constant size independent of . The -qubit terms are often referred to as interactions between qubits. We may refer to eigenvalues and expectation values as the energy of the state in the system described by . In particular, the ground state energy and ground state refer to the minimum eigenvalue and an associated eigenvector.
Estimating the minimum eigenvalue of a Hamiltonian is a natural quantum generalization of estimating the maximum number of satisfiable clauses in a Boolean formula.
Definition 7 (-LH).
Given a -local Hamiltonian acting on qubits with , the entries of each specified by at most bits, and the norms polynomially bounded in , and two real parameters such that , decide whether is at most (YES) or at least (NO), promised that one is the case.
In this work, we are interested in restricted to families of local Hamiltonians, where the families are determined by sets of allowed interactions. In particular, we will be interested in sets of 2-qubit interactions.
Definition 8 (-LH).
For any fixed set of Hamiltonian terms, define -LH as the problem with the additional promise that any input is of the form where each is an element of assigned to act on some subset of qubits and the weights have magnitude polynomially-bounded in .
Remark 9.
There are several standard assumptions implicit in our definition of -LH. Some are not physically realistic in the context of the condensed-matter literature but allow us to precisely characterize the complexity of these problems. First, although classical CSPs generally allow a constraint to take as input multiple copies of the same variable, this makes less sense in the quantum setting and we do not allow it. Second, the definition of -local only restricts the dimension of each term, it does not imply any spatial locality or geometry. Therefore, any term in may be applied to any subset of qubits with the qubits arranged in any order. In particular this means that, if contains a directed term , then the family of Hamiltonians allowed as input to is equivalent to the family allowed to for . Third, for the purpose of classifying the complexity of -LH, we may assume , since adding or removing a term is equivalent to simply shifting the input parameters by . For containing 2-qubit terms, this fact also implies we may assume all terms in are traceless. Fourth, except when noted, we allow both positive and negative weights.
Classifying the complexity of systems under additional, more physically natural restrictions appears to be a significantly more difficult problem [14, 33].
Given this setup, our interest will be in the problems and restricted to product states.
Definition 10 (Product state).
A state where each is a single-qubit state.
Definition 11 (prodLH).
Given a -local Hamiltonian on qubits with , the entries of each specified by at most bits, and the norms polynomially bounded in , and two real parameters , decide whether there exists a product state with (YES) or all product states satisfy (NO), promised that one is the case.
The problem is defined analogously. In both definitions, the fact that product states have concise classical descriptions allows us to naturally consider any choice of parameters, even an exact decision problem with , in contrast to -LH.
By convexity, a product state achieves an extreme value of if and only if there exists a pure product state which achieves that value. Similarly, mixtures of product states, known as separable states, reduce to considering pure product states.
Remark 12.
In the context of or given some fixed set , we will describe as “containing” a Hamiltonian term , or that we “have access to” , even when formally . As previously referenced in Remark 9, given a set , the family of Hamiltonians allowed as input to may be equivalent to the family allowed given some other set . For example, may include for and any permutation of the qubits . Similarly, adding constant multiples of the terms in or any linear combinations of terms from does not change the family of allowed Hamiltonians. So, when discussing , we may implicitly refer to elements of the largest set such that and each have the same family of allowed inputs.
Additionally, we note that -LH is reducible to for any which can be used to implement all elements of – whether because formally or through other means. In the opposite direction, if the terms in a set can be used to construct some term and we wish to show hardness of , then it is sufficient to show hardness of .
Finally, for a 2-local Hamiltonian, we may refer to the interaction graph, with vertices associated with each qubit such that vertex is adjacent to vertex whenever a nonzero interaction exists on qubits and . When all interactions are symmetric, then the graph is undirected. Notably, when is defined with a singleton, then an input is fully specified by its weighted interaction graph.
3 Classification of prodLH
In this section, we prove a dichotomy theorem classifying the complexity of estimating the minimum expectation of product states for given families of 2-local Hamiltonians. In particular, we show that for any set of 2-qubit terms such that at least one term is not 1-local,444We would prefer a more concise name for 2-qubit terms that are not 1-local, but are unaware of any. One option is 2-qubit terms with Pauli degree 2. Alternatively, these are 2-qubit terms which have nonzero Pauli rank, referring to the rank of the correlation matrix in the Pauli decomposition (Equation 5). the problem is -complete. These are precisely those 2-local families such that (as shown in [14]) is -, -, or -complete. Conversely, if all terms are 1-local, then both and are in .
See 1 Our -hardness results hold with coupling strengths of at most constant magnitude.
For comparison with our classification, we recall the tetrachotomy theorem of Cubitt and Montanaro [14] classifying -LH for families of 2-local Hamiltonians. They proved that for every set of 2-qubit Hamiltonian terms , the problem is either in or -, -, or -complete, and described properties of the set which determine the problem’s complexity. We note that both Theorem 13 and our Theorem 1 classify the complexity of all sets of 2-qubit terms.
Theorem 13 (Theorem 7 of [14]).
For any fixed set of 2-qubit Hamiltonian terms:
-
If every matrix in is 1-local, then -LH is in .
-
Otherwise, if there exists a single-qubit unitary such that locally diagonalizes all elements of (i.e. is diagonal for each 2-qubit ), then -LH is -complete.
-
Otherwise, if there exists a single-qubit unitary such that for each 2-qubit ,
for some and 1-local Hermitian matrices , then -LH is -complete.
-
Otherwise, -LH is -complete.
Combining our classification of prodLH with the classification of -LH gives us Corollaries 2 and 3.
To prove Theorem 1, showing containment in and are straightforward, and our effort is to prove -hardness. In the proof, we will use a simple symmetrization trick that allows us to consider only antisymmetric or symmetric Hamiltonian terms. We then prove two lemmas, one for each case.
Lemma 14.
If contains a 2-qubit antisymmetric term that is not 1-local, then with is polynomial-time reducible to .
Lemma 15.
If contains a 2-qubit symmetric term that is not 1-local, then there exists a fixed non-negative with at least one of nonzero such that is polynomial-time reducible to .
We state some helpful facts in Section 3.1 below and then prove the two lemmas in Section 3.2. We will now use these lemmas to prove our main theorem.
Proof of Theorem 1.
First consider the case where only contains 1-local terms. Then we can write , where acts only on the qubit. If is the single-qubit state minimizing , then minimizes , and so can be solved by finding the ground state of single-qubit Hamiltonians, which takes time.
Now suppose is a set of of 2-qubit Hamiltonian terms such that at least one element of is not 1-local. Let be any such element. As previously mentioned, for any fixed , as product states have concise classical descriptions which can be used to efficiently calculate expectation values for a given local Hamiltonian. If is antisymmetric, then with fixed is reducible to prodLH by Lemma 14, and with such a is -hard by Theorem 4, so prodLH is -complete. If is symmetric, then by Lemma 15 there exists a non-negative nonzero matrix such that is reducible to prodLH, it is -hard by Theorem 4, and so prodLH is -complete. If is neither of these, then we use our freedom to permute the direction is applied to any pair of qubits to apply both and , which is equivalent to implementing the symmetric term . So, we can say effectively contains , or formally that hardness of implies hardness of , and again referring to Lemma 15 concludes the proof.
3.1 Closure Properties of
Before proving the two lemmas required in the proof of Theorem 1, we review several more facts regarding 2-qubit Hamiltonian terms and operations under which the complexities of and are unaffected. This section mostly reviews observations made in [14].
First, for a single-qubit unitary and an operator , define simultaneous conjugation by to mean . When discussing sets of -qubit terms, we define simultaneous conjugation to mean .
Fact 16.
For any single-qubit unitary , the complexities of -LH and prodLH are equal to the complexities of -LH and prodLH, respectively, where is simultaneously conjugated by .
Observe that . Simultaneous conjugation by gives a bijection between Hamiltonians allowed in -LH and -LH as well as prodLH and prodLH. The above fact follows from observing that this bijection preserves expectation values, and that is a product state iff is.
Fact 17.
For any choice of permutation on and any choice of two of , there exists a single-qubit unitary and corresponding third coefficient s.t. simultaneous conjugation by maps the Pauli matrices to . So, writing every element of in the Pauli basis, relabeling all with in the decompositions of each element of does not change the complexity of or , where and two of the coefficients can be chosen arbitrarily.
To justify the above fact, consider simultaneously rotating the three axes of the Bloch sphere. Next, we quote the following, more involved, fact without proof.
Fact 18 ([14, 21]).
Let be a 2-qubit Hamiltonian term with Pauli decomposition . For any single-qubit unitary ,
(6) |
for some . Likewise, for any , there exists a single-qubit such that the Pauli decomposition of matches Equation 6.
A further straightforward observation from [14] is that in the Pauli decomposition (Equation 5), if is symmetric then the correlation matrix is symmetric, and if is antisymmetric then is skewsymmetric, meaning .
Finally, the below observation combines some of the above facts to establish a “normal form” for symmetric and antisymmetric terms.
Fact 19.
If a 2-qubit Hamiltonian term is symmetric, and so the associated correlation matrix is symmetric, there exists which diagonalizes . Combining Facts 18 and 16, for any 2-qubit symmetric term , there exists a symmetric term of the form such that the complexities of and are respectively the same as and .
3.2 Proofs of Antisymmetric and Symmetric Lemmas
We now prove the two lemmas required in the proof of the main theorem, respectively handling the cases that contains an antisymmetric term and that contains a symmetric term. In both cases, it is sufficient for to contain just a single term. Interestingly, our construction in Lemma 14 for antisymmetric terms is unweighted, meaning all weights are . In this case, the final Hamiltonian is fully determined by the specification of a single 2-qubit term and the interaction graph. Our construction in Lemma 15 uses positive and negative unit weights, . Intuitively, antisymmetric terms inherently allow negativity by simply permuting the qubits they act on, while for symmetric terms we must use negative weights.
Proof of Lemma 14.
Consider an arbitrary instance of the problem with . For a given graph , the objective function is
Given a parameter , we must decide whether is at least or at most . To reduce to prodLH, we first construct a gadget using the promised antisymmetric term. Then, we apply this gadget according to the graph such that the minimum energy of a product state in our final Hamiltonian will equal .
Denote the assumed 2-qubit antisymmetric term that is not 1-local in by . By Fact 18, our antisymmetric term may be mapped to a term of the form
where all coefficients are real, and we have since the term is not 1-local. As explained in Remark 12 and Section 3.1, the complexity of is equivalent to that of for derived using a variety of operations, including permutations and linear combinations. If is negative, then we redefine the direction the term acts in, versus , so that is positive. Finally, we scale555If we want the terms to have unit weights, we could forgo scaling the term and reduce to instead. As this problem has the same complexity as . the term so that and define a single-qubit Hermitian matrix . Given the complexity is unchanged using or , we simply redefine the original term, so that
Next, we use a symmetrization gadget to remove the 1-local terms . For four qubits , define
Note that here the direction of the interaction matters, since the terms are asymmetric. Then
Now we consider how interacts with product states on four qubits. For , define
with the Pauli operators and the Bloch vector. Then, writing any product state on qubits as , the expectation value on is . After eliminating terms, we find this equals
It is helpful to note that .
Now, consider the graph given as input to . Associate a qubit with each vertex and call these the “vertex qubits”. For each edge , construct a copy of such that it acts on qubits and two ancilla qubits. The vertex qubits may be shared among several gadgets, while the ancilla qubits are part of only one gadget. In particular, we choose to associate the vertex qubits with qubits and in each copy of , letting and be the ancilla. We will refer to the copy of which acts on vertex qubits and as . Our Hamiltonian is then
Before analyzing the full Hamiltonian , consider the minimum expectation of a single gadget if the two vertex qubits are fixed, i.e. . The minimum is achieved when and , which yields an expectation of
Therefore, given an arbitrary graph , applying our gadget to every edge constructs a Hamiltonian such that the minimum expectation of any product state is equal to
For any graph , the objective is equal for and . Alternatively, we may use our freedom to relabel Paulis to redefine the 2-qubit term such that the final weight matrix is .
Finally, multiplying the full Hamiltonian by gives us that the minimum expectation of any product state equals
We conclude deciding reduces to deciding prodLH on . Since is entirely constructed from the antisymmetric term , this completes the desired reduction of with to prodLH.
Next we prove the lemma dealing with containing a symmetric term. The construction is nearly the same as the in the previous proof, but requires negative weights to implement the symmetrization gadget removing 1-local terms.
Proof of Lemma 15.
Given fixed , we will show there exists some fixed such that reduces to -LH. But, before describing , we must analyze .
Denote the assumed 2-qubit symmetric term that is not 1-local in by . As in the previous proof, without changing the complexity of we may conjugate and scale as necessary so that
where all coefficients are real and at least one of is nonzero since is nonzero. The superscripts in the above equations are to differentiate the coefficients of from the entries of , which must be non-negative.
We again use a symmetrization gadget to remove the 1-local terms, but now require negative weights. Given four qubits , define . This is a rectangle with two positive edges and two negative edges. Then
Now we see how interacts with product states on four qubits. For , we again define with and . Then, writing any product state on as , the expectation value on is , which equals
which is in turn equal to
If we fix the state of qubits and and minimize the expectation on , the minimum is achieved when and . This minimum expectation is . Observe this expectation value is equivalent to for where .
Now we are prepared to set up our reduction. For , which is non-negative and nonzero, consider an arbitrary instance of . For a given graph , the objective function is again , and given a parameter , we must decide whether is at least or at most .
Associate a “vertex qubit” with each vertex and construct a copy of the gadget for each edge , such that it acts on and two ancilla qubits, and denote it . The vertex qubits may be shared among several gadgets, while the ancilla qubits are part of only one gadget. In particular, we choose and in each gadget to be the vertex qubits.
Substituting our gadget for every edge constructs a Hamiltonian such that the minimum expectation of any product state is equal to
Simply multiplying by makes this equal to .
We conclude that given contains a 2-qubit symmetric term that is not 1-local, there exists some non-negative with at least one nonzero entry such that reduces to . Since was constructed using only the symmetric term , this is also a reduction to an instance of prodLH, as desired.
4 The Stretched Linear Max-Cut Problem
We study a generalization of the classical Max-Cut problem which arises naturally from our study of product states and which is likely of independent interest. Both Max-Cut and its generalization
were introduced in Section 1. As the above equation emphasizes, maximizing the distance between vectors is equivalent to maximizing the angle, optimally being anti-parallel.
Our new problem is defined with two significant changes. First, the sum is over distances rather than squared distances. Second, the distances are allowed to incorporate a linear stretch.
Definition 20 ( ()).
For a fixed diagonal matrix , given an -vertex graph and thresholds with , decide whether
is at least or at most .
A comparison of the geometric interpretations of and was given in Section 1. A further interpretation comes from treating the edges of the graph as springs or rubber bands. As explored in [29], the potential energy of a spring is quadratic in its length, so the value represents the total potential energy of the system given a particular embedding. On the other hand, the force or tension of each spring is linear in its length. So, gives the total force, tension, or pressure such an arrangement of springs would apply to the surface of the sphere (or ellipsoid, more generally).
In both problems, the objective is a linear sum, of either the distances or the inner products. Both problems generalize the traditional Max-Cut problem, since when restricted to labels, distances are directly proportional to squared distances. Previous work was likely motivated to focus on squared distances because approximation algorithms like SDPs naturally apply to inner products but not to square roots of inner products.
Our main theorem concerning is the below.
See 4
Our hardness proofs do not require any edge weights (unlike our Hamiltonian constructions in the previous section).
Containment in is immediate, and we break the proof of -hardness into three cases based on the entries of . The three cases depend on how many entries of are equal, requiring different approaches for dealing with degenerate solutions. We assume throughout that ; as we show in the final proof of Theorem 4, this suffices by scaling and symmetry. Lemma 21 considers the case when all three entries are equal. Lemma 22 considers the case when the largest entry is unique. Lemma 23 finally considers the case when the two largest entries are equal and distinct from the third, combining techniques from the previous two proofs. Note that these cases are not entirely disjoint.
When , we prove hardness by reducing from the -complete 3-Coloring problem. We replace every edge in the graph with a 4-clique, or tetrahedron. To deal with the symmetry created by equally weighted axes, all of the gadgets are connected to a new sink vertex, removing a degree of freedom. We then argue that there is an assignment to the new graph that simultaneously (nearly) maximizes all of these cliques iff the original graph is 3-colorable.
Lemma 21.
For , is -hard.
Proof.
We will reduce from 3-Coloring. Consider an arbitrary graph on vertices and edges. We construct as follows: Start with . For each edge , add vertices and connect to form a 4-clique. Then add a sink vertex and an edge for each . therefore consists of edge-disjoint 4-cliques, each containing one edge from , and additional edges from vertices to .
We claim that if is 3-colorable, then . Conversely, we claim that if , for an we will choose later, then is 3-colorable.
First, suppose is 3-colorable. We will show how to derive a vector assignment to attaining from a 3-coloring of . Define
Let be vectors corresponding to a regular tetrahedron inscribed in the unit sphere, known to achieve the maximum perimeter of any inscribed tetrahedron at [30]. We 3-color (and therefore the vertices of other than , , and ) with , assigning each vertex the vector matching its color. Then for each , we assign the vector in not assigned to or , and to . Finally, we assign to . By construction, this assignment will yield a value of on each 4-clique gadget, and the edges will each contribute exactly 1. Thus , as desired.
Now we will show the converse. Suppose there exists an assignment of vectors achieving greater than on . We construct a -coloring of each connected component of as follows: Choose an edge in the corresponding component of (note that there is a 1-to-1 correspondence between components of and ). Let be the vectors assigned to the vertices of the corresponding clique in . Our coloring will use these four vectors as colors, which we denote as the set . We assign each vertex the color corresponding to the element of that is closest to the vector assigned to in the assignment. We will show that this is a proper coloring, and that it assigns the same color to every , and therefore gives a 3-coloring of .
To show that this is a proper coloring, we need to show that every pair of adjacent vertices in are assigned different colors, i.e. that the closest elements of to the vectors assigned to them in the assignment are distinct. As every edge in is contained in some 4-clique corresponding to some edge of , it will suffice to show the following: for every edge in the component of containing , if is the smallest number of edges in a path in starting with and ending with , each of , , and is within of a different element of .
First we note that, as consists of edge-disjoint 4-cliques and other edges, and the maximum any assignment can earn on an edge is , the lower bound on the total amount the assignment earns implies that every 4-clique earns at least and the other edges earn at least each. So by trigonometry we immediately have that every is within of , and therefore within of each other, and in particular.
For the other vertices, we proceed by induction on . We have the desired result trivially for , as in this case . Now suppose it holds for . Let be the end of a -edge path starting with . Without loss of generality let be the vertex of earlier in the path, and let be the immediately prior vertex in the path, so by the inductive hypothesis , , , and are within of different elements of . Furthermore, as both and are vertices of tetrahedra with perimeters at least , by Lemma 29 the edge lengths of these tetrahedra are all in the interval . So as we have already shown that , , and are all within of each other, then the criteria of Lemma 30 are satisfied with and and so and are each within of (different) elements of . So then the result follows by the triangle inequality.
We now have that every vertex in is assigned a vector within of an element of , and so taking to be a sufficiently small constant times , we can take these distances to be at most . Moreover, for any two adjacent vertices, the choice of color will be different, and so as by Lemma 29 the distances between these four vectors are at least , this implies any two adjacent vertices are assigned different colors, and so we have a proper 4-coloring of . Finally, note that, as every is within of every other one, this implies all the were assigned the same color, and so no vertex in uses this color. So this gives us a proper 3-coloring of .
Next, in Lemma 22 we consider the case and , in which the maximum weight is unique. The approach in the proof of Lemma 21, reducing from 3-Coloring, can in fact be modified to cover this case, but analyzing triangles inscribed in ellipses instead of circles is more technical. Instead, we take a different approach and in the case of a unique maximum (whether equals or not) give a reduction from Max-Cut instead of 3-Coloring.
We insert ancilla vertices so that every vertex in the original graph is the center of a large star. These star gadgets amplify any deviation from the highest-weight axis such that any near-optimal solution must approximate a standard 1-dimensional labeling, as in Max-Cut.
Lemma 22.
For any with , is -hard.
Proof.
We reduce from the standard -complete Max-Cut problem. For any graph with vertices and edges, we construct a graph with and by, for each , adding ancilla vertices , and then adding an edge from each of these vertices to so that is the center of a -star. Now and .
We claim that, for any and for large enough , implies and implies .
First, suppose there is a cut of with value at least . We construct a corresponding assignment of vectors to vertices in . First assign the vector to all vertices in which are labeled in and to those with labels . Then, for every vertex in , which by construction is at the center of a star of ancilla qubits in , assign the vector opposite the one assigned to to each of the ancilla vertices. This assignment of vectors gives an objective value of at least on the edges from the original graph and on the edges of the star gadgets, and so the value of this assignment is .
Now suppose there exists an assignment of vectors achieving value greater than on . We will show that the cut given by for each (i.e. projecting to the -axis and checking whether it is or ) has value at least .
First, for each , let . We will show that this is close to . Because the original graph can contribute at most to the objective, and each star gadget can contribute at most , each star gadget must contribute at least . By Lemma 26, for the -star to achieve at least , the vector assigned to must satisfy
We will use this fact to show that the value earned by the vector assignment on the original graph is close to the value of the cut we defined. We have
which is for large enough .
Using for a second time the fact that the edges of the star gadgets can contribute at most to the , the vector assignment must achieve at least on , and so
implying that the value of our cut is strictly greater than for sufficiently large . Therefore, as it is integer-valued it is at least , concluding the proof.
Finally, we give Lemma 23, in which and . Our proof combines the techniques in the previous two proofs. Similar to the proof of Lemma 21, we show hardness by reducing from 3-Coloring. Our construction begins by replacing every edge in the graph with a 3-clique. Then, as in the proof of Lemma 22, we insert large star gadgets on every vertex, forcing solutions away from the low-weight -axis. With solutions restricted to two dimensions, we are able to argue there is an assignment to the new graph that simultaneously (nearly) maximizes all of the cliques iff the original graph was 3-colorable.
Lemma 23.
For any with , is -hard.
Proof.
Consider an arbitrary instance of 3-Coloring on a graph with vertices and edges. Construct a new graph by taking and, for each edge , adding a vertex and edges and , so that each edge in corresponds to a 3-clique ( in . Note that the cliques constructed this way are edge-disjoint. Next, for each vertex in , add ancilla vertices , each connected to so that is the center of a -star. Call the final graph , for which we have vertices and edges.
We claim that if is 3-colorable, then . Conversely, we claim that if , for an we will choose later, then is 3-colorable. As testing 3-colorability is -hard, this will prove the theorem.
First, suppose is 3-colorable. Let be any set of three vectors in the -plane achieving the maximum value of . Given any 3-coloring of , we assign one of these vectors to each color, and thus assign vectors from to with no two adjacent vertices having the same vector. We extend this assignment to by, for each of our constructed 3-cliques , assigning the vector in that was not assigned to either or to . Now each of these cliques contributes to the objective, and as there are of them and they are edge-disjoint, this contributes to the objective. To extend to and its star gadgets, every edge of a star centered on vertex can trivially contribute to the objective value. Because all vectors are in the -plane, and so the unit circle, this means they all contribute 1. So in total, there exists an assignment with value .
Now suppose there exists a vector assignment achieving greater than on . For a vector , let denote the vector . We assign colors as follows. Choose any of the gadgets in corresponding to an edge in the original graph. Let be the vectors assigned to the vertices, respectively. Assign three colors to the vertices. Then, choose any gadget adjacent to the first. For each vertex in the second clique, consider its assigned vector and round it via , determine which of the rounded vectors it is closest to, and assign the same color. We continue coloring adjacent cliques in this way, comparing rounded vectors to the original set , until the coloring is propagated to the entire graph. This colors all of the vertices in , which are the centers of star gadgets in . We will show no adjacent vertices in were assigned the same color. Since is a subgraph of , this also implies a proper coloring of , as desired.
We first show that just as the assigned vectors must achieve a large objective value on the clique gadgets, the star gadgets force the rounded vectors to achieve a similar score. Because the clique gadgets can each contribute at most to the objective value, the star gadgets in must contribute at least . Similarly, because the edges of each star can contribute at most , each star must achieve at least . For any vertex in in , consider the star gadget centered on it in . As shown in Lemma 26, for the star to achieve , the vector assigned to must satisfy
On the other hand, because the star gadgets can contribute at most to the objective value, the vector assignment must achieve at least on the remaining edges, the ones in . Given the vectors are close, we have that
Therefore, for , the set of rounded vectors must achieve at least on the rest of . Because each of the clique gadgets can contribute at most , the rounded vectors must achieve at least on each individual clique.
The rounded vectors exist in the -plane, which means they are inscribed in the unit circle. Maximizing the sum of the edge lengths on a gadget is equivalent to maximizing the perimeter of a triangle. For a triangle in the unit circle to have nearly maximal perimeter, it must be nearly regular; as shown in Lemma 27, if the perimeter is at least , then each edge length must be in the interval .
Now, consider any two adjacent vertices in , which are at the center of star gadgets in . We must show they were assigned different colors. The two vertices exist in some gadget in , and the coloring procedure, starting from , must have reached them in at most rounds. At each step in the procedure, there is a colored clique and a successive adjacent clique which share a clique. Two nearly regular triangles which share a vertex must nearly share their other vertices; as shown in Lemma 28, the distance between the first clique’s rounded vectors and the second clique’s is at most .
After rounds of the coloring procedure, we conclude and must be assigned colors such that are each at distance at most away from the same-colored rounded vectors in the initial clique. Both the initial clique and the clique containing and must have vectors separated by at least . For sufficiently small ( with a sufficiently small constant suffices), cannot also be within of the same initial vector, so we conclude they must be associated with different colors, and this is a proper coloring. We now conclude with a proof of the main theorem of this section, that is -complete for any fixed diagonal non-negative nonzero .
Proof of Theorem 4.
The containment of for any diagonal is straightforward. With a constant, given a claimed vector assignment, the value can be verified in time linear in the number of edges.
To show hardness, we make two simplifications. First, because for a constant , we can easily reduce to an instance in which we assume the largest entry of equals . Second, although rearranging the entries of diagonal requires changing any vector assignment, it does not change the objective value. So and any other rearrangement of , and we can assume the entries are ordered . With this, the theorem follows by Lemmas 21, 23, and 22.
5 NP-Hardness of Unweighted Quantum Max-Cut
Our lower bounds elsewhere in this paper are for local Hamiltonian problems in which terms can be given positive or negative weight. They apply in the model where all weights are restricted to being of constant weight but require some terms with negative weight to work. In this section we show that this restriction can be removed for one of the best-studied local Hamiltonian problems: the Quantum Max-Cut (or QMC) problem.
The Quantum Max-Cut problem can be defined as -LH with . We will write for .666Other work frequently includes a multiplicative factor, e.g. , in the definition of and/or of QMC.
Definition 24 (Quantum Max-Cut (QMC)).
Given a Hamiltonian acting on qubits where each , all are real, polynomially-bounded, and specified by at most bits, and two real parameters such that , decide whether is at least (YES) or at most (NO).
Note that we have written QMC as a maximum eigenvalue problem (with a flip and shift of the local terms) rather than in terms of the minimum eigenvalue as for -LH in Definition 7. This is to follow the norm in previous work; note that as both the terms and the objective function are flipped, an instance of the problem defined this way will be equivalent to an instance of the corresponding -LH problem with the same weights.
When minimizing (maximizing) the eigenvalue and the weights are restricted to be non-negative (non-positive), it is referred to as the anti-ferromagnetic Heisenberg model. Flipping the restrictions, e.g. minimizing with non-positive weights, is referred to as the ferromagnetic Heisenberg model. The latter case is trivial when viewed as an optimization problem (as earns on the local term, and so the problem is optimized by assigning to every qubit), so we will be interested in hardness results for the former.
It is straightforward to verify that for two qubits with pure states ,
where are the corresponding Bloch vectors. This shows that deciding QMC restricted to product states, which we denote prodQMC, is equivalent to the standard Vector Max-Cut problem in three dimensions:
Definition 25 (()).
Given an -vertex graph and thresholds such that , decide whether
is at least or less than .
Note that this is different from the we studied in Section 4 as it considers squared distances. Furthermore, while Max-Cut is a classic -complete problem, is not expected to be hard for all values of , and in particular is tractable when .
Our main result classifying immediately implies , and therefore also prodQMC and , is -complete. However, the proof of Lemma 15 utilizes Hamiltonian gadgets involving negative weights. This leaves open whether prodQMC and remain -hard on unweighted graphs. We now prove that is -hard even when restricted to positive unit weights. This is the first published proof of this fact, although we note that a sketch of a different proof was previously known for this specific problem [40].
Our approach is similar to the proof of Lemma 21, which demonstrated hardness of for by replacing every edge with a 4-clique (with one vertex connected to a source vertex), and showing that the resulting graph could simultaneously optimize all of these 4-cliques (by assigning vectors corresponding to a regular tetrahedron) if and only if the original graph was 3-colorable.
However, the change in objective function from distances to squared distances causes a problem: while the regular tetrahedron is the unique optimal solution for , in the case of , setting would also be optimal. So we replace each edge of the 4-cliques with triangles, which penalize the degenerate solution and forces the vectors assigned to the vertices of the 4-cliques toward regular tetrahedra.
See 5
Proof of Theorem 5.
Clearly is in . To show hardness, we reduce 3-Coloring to . Given an instance on vertices and edges, we first replace every edge with a copy of . That is, for each edge , we add vertices and add edges to form a 4-clique. Call the new graph . Then in , we replace every edge with a copy of . That is, for all we add a vertex and edges and . Finally, we add a sink vertex and for every edge , in the original graph, add edge . Call the resulting graph . For future reference, let denote a copy of with each edge replaced by a copy of , which we may call a “tetrahedron with adjoined triangles”.
We claim that if has a proper 3-coloring then . Conversely, we claim that if , for an we will choose later, then there is a 3-coloring of . Later, we will show .
First, suppose is 3-colorable. Let consist of the following three unit vectors in :
Along with , these are the four vertices of a regular tetrahedron inscribed in the unit sphere. Assign each one of the vectors of to each color, such that every vertex in is assigned a vector and no adjacent vertices have the same vector. We copy those vectors to the vertices of , and for each vertex we assign the vector in not assigned to or . We assign to each vertex . We copy these vectors to the vertices of . In we assign to . Finally, for each edge and 3-clique , the only uncolored vertex is , and we assign the unique unit vector that is antiparallel to the sum of the vectors assigned to and .
Now we calculate the objective value that this assignment achieves. For vertex , let denote the assigned vector. For any edge in , the vectors assigned to the associated gadget correspond to vertices of a regular tetrahedron, and it can be directly calculated that for any of the six edges , we have . For edges , we have . For edges , given and , the inner product is . In total, for each edge in , the graph has a gadget with six edges in a gadget, twelves edges in gadgets, and one edge incident on . Plugging the inner products we calculated into the definition of gives a total objective value of .
Second, suppose there is a set of unit vectors for such that . We color the vertices of , comprised of the gadgets, as follows. Pick some color and assign it to all of the vertices . Then, choose any gadget in corresponding to an edge in the original graph. Let be the vectors assigned to the vertices, respectively. Assign three colors to the uncolored vertices arbitrarily. Then, choose any gadget adjacent to the first. For each uncolored vertex in the second clique, consider its assigned vector, determine which of it is closet to, and assign the same color. We continue coloring adjacent cliques in this way, comparing vectors to the original set , until the coloring is propagated to the entire graph.
Although we assigned four colors, one of them is used exclusively by the vertices . Since these vertices are not in , if this is a proper 4-coloring of , then it gives a proper 3-coloring of the subgraph . So, in the remainder of the proof we will show that no two adjacent vertices in were assigned the same color.
The graph is comprised of edge-disjoint gadgets as well as edges . In order for the total objective value to be greater than , each gadget must contribute at least and each edge must contribute at least . As shown in Lemma 31, the sum of squared edge lengths of a “tetrahedron with adjoined triangles” is nearly maximized iff the lengths of the edges forming the tetrahedron are nearly regular. Specifically, for a gadget to achieve at least , each side of the tetrahedron must be in the interval .
Now consider any two adjacent vertices in . The two vertices exist in some gadget in , and the coloring procedure must have reached it after at most rounds. At each step in the procedure, there is a colored clique and a successive adjacent clique. There are two copies of in which contain those two 4-cliques from . The gadgets share one vertex. Additionally, each gadget has a vertex adjacent to the sink , whose assigned vectors must be at distance at least from . Since all of these are in the unit sphere, standard trigonometry shows those two assigned vectors must themselves be within of each other. As verified in Lemma 30, considering two nearly regular tetrahedra which share one vertex and have a pair of similar vertices, the other vertices of the tetrahedra must each be within of each other.
After rounds of the coloring procedure, we conclude that are each at distance at most away from the same-colored vectors in the original set . Since all of the edges of the 4-clique gadgets must be at least , for sufficiently small the vectors cannot be near the same original vector. Therefore, they must been colored differently, as desired.
By the correspondence described earlier in the section, we immediately have the desired bound on . See 6
References
- [1] Sanjeev Arora and Boaz Barak. Computational complexity: a modern approach. Cambridge University Press, 2009.
- [2] Afonso S. Bandeira, Christopher Kennedy, and Amit Singer. Approximating the little Grothendieck problem over the orthogonal and unitary groups. Mathematical programming, 160:433–475, 2016. doi:10.1007/s10107-016-0993-7.
- [3] Alexander I. Barvinok. Problems of distance geometry and convex properties of quadratic maps. Discrete & Computational Geometry, 13:189–202, 1995. doi:10.1007/BF02574037.
- [4] Fernando G.S.L. Brandão and Aram W. Harrow. Product-state approximations to quantum states. Commun. Math. Phys., 342:47–80, 2016. doi:10.1007/s00220-016-2575-1.
- [5] Sergey Bravyi, David Gosset, Robert König, and Kristan Temme. Approximation algorithms for quantum many-body problems. Journal of Mathematical Physics, 60(3):032203, 2019. doi:10.1063/1.5085428.
- [6] Sergey Bravyi and Matthew Hastings. On complexity of the quantum Ising model. Communications in Mathematical Physics, 349(1):1–45, 2017. doi:10.1007/s00220-016-2787-4.
- [7] Jop Briët, Harry Buhrman, and Ben Toner. A generalized Grothendieck inequality and nonlocal correlations that require high entanglement. Communications in mathematical physics, 305(3):827–843, 2011. doi:10.1007/s00220-011-1280-3.
- [8] Jop Briët, Fernando Mário de Oliveira Filho, and Frank Vallentin. The positive semidefinite Grothendieck problem with rank constraint. In International Colloquium on Automata, Languages, and Programming, pages 31–42. Springer, 2010. doi:10.1007/978-3-642-14165-2_4.
- [9] Jop Briët, Oded Regev, and Rishi Saket. Tight hardness of the non-commutative Grothendieck problem. Theory of Computing, 13(15):1–24, 2017. doi:10.4086/toc.2017.v013a015.
- [10] Anne Broadbent and Alex Bredariol Grilo. QMA-hardness of consistency of local density matrices with applications to quantum zero-knowledge. SIAM Journal on Computing, 51(4):1400–1450, 2022. doi:10.1137/21M140729X.
- [11] Samuel Burer and Renato D.C. Monteiro. A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization. Mathematical programming, 95(2):329–357, 2003. doi:10.1007/s10107-002-0352-8.
- [12] Nadia Creignou. A dichotomy theorem for maximum generalized satisfiability problems. Journal of Computer and System Sciences, 51(3):511–522, 1995. doi:10.1006/jcss.1995.1087.
- [13] Nadia Creignou, Sanjeev Khanna, and Madhu Sudan. Complexity classifications of Boolean constraint satisfaction problems. SIAM, 2001.
- [14] Toby Cubitt and Ashley Montanaro. Complexity classification of local Hamiltonian problems. SIAM Journal on Computing, 45(2):268–316, 2016. doi:10.1137/140998287.
- [15] Toby S. Cubitt, Ashley Montanaro, and Stephen Piddock. Universal quantum Hamiltonians. Proceedings of the National Academy of Sciences, 115(38):9497–9502, August 2018. doi:10.1073/pnas.1804949115.
- [16] Sevag Gharibian, Yichen Huang, Zeph Landau, Seung Woo Shin, et al. Quantum Hamiltonian complexity. Foundations and Trends in Theoretical Computer Science, 10(3):159–282, 2015. doi:10.1561/0400000066.
- [17] Sevag Gharibian and Julia Kempe. Approximation algorithms for QMA-complete problems. SIAM Journal on Computing, 41(4):1028–1050, 2012. doi:10.1137/110842272.
- [18] Sevag Gharibian and Ojas Parekh. Almost Optimal Classical Approximation Algorithms for a Quantum Generalization of Max-Cut. In Approximation, Randomization, and Combinatorial Optimization (APPROX/RANDOM 2019), volume 145 of Leibniz International Proceedings in Informatics (LIPIcs), pages 31:1–31:17. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPIcs.APPROX-RANDOM.2019.31.
- [19] Michel X. Goemans and David P Williamson. Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming. J. ACM, 42(6):1115–1145, 1995. doi:10.1145/227683.227684.
- [20] Sean Hallgren, Eunou Lee, and Ojas Parekh. An approximation algorithm for the max-2-local Hamiltonian problem. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2020). Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2020. doi:10.4230/LIPIcs.APPROX/RANDOM.2020.59.
- [21] Ryszard Horodecki et al. Information-theoretic aspects of inseparability of mixed states. Phys. Rev. A, 54(3):1838–1843, 1996. doi:10.1103/PhysRevA.54.1838.
- [22] Yeongwoo Hwang, Joe Neeman, Ojas Parekh, Kevin Thompson, and John Wright. Unique games hardness of Quantum Max-Cut, and a conjectured vector-valued Borell’s inequality. In Proceedings of the 2023 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 1319–1384. SIAM, 2023. doi:10.1137/1.9781611977554.ch48.
- [23] Peter Jonsson. Boolean constraint satisfaction: complexity results for optimization problems with arbitrary weights. Theoretical Computer Science, 244(1-2):189–203, 2000. doi:10.1016/S0304-3975(98)00343-0.
- [24] Peter Jonsson, Mikael Klasson, and Andrei Krokhin. The approximability of three-valued Max CSP. SIAM Journal on Computing, 35(6):1329–1349, 2006. doi:10.1137/S009753970444644X.
- [25] Sanjeev Khanna, Madhu Sudan, and David P Williamson. A complete classification of the approximability of maximization problems derived from boolean constraint satisfaction. In Proceedings of the twenty-ninth annual ACM symposium on Theory of computing, pages 11–20, 1997. doi:10.1145/258533.258538.
- [26] Alexei Yu Kitaev, Alexander Shen, and Mikhail N Vyalyi. Classical and quantum computation. Number 47 in Graduate Studies in Mathematics. American Mathematical Soc., 2002.
- [27] Yi-Kai Liu. Consistency of local density matrices is QMA-complete. In Josep Díaz, Klaus Jansen, José D. P. Rolim, and Uri Zwick, editors, Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, pages 438–449, Berlin, Heidelberg, 2006. Springer Berlin Heidelberg. doi:10.1007/11830924_40.
- [28] L. Lovász. Semidefinite Programs and Combinatorial Optimization, pages 137–194. Springer New York, New York, NY, 2003. doi:10.1007/0-387-22444-0_6.
- [29] László Lovász. Graphs and geometry, volume 65. American Mathematical Soc., 2019.
- [30] Hiroshi Maehara. On the total edge-length of a tetrahedron. The American Mathematical Monthly, 108(10):967–969, 2001. URL: http://www.jstor.org/stable/2695418.
- [31] Ojas Parekh and Kevin Thompson. Beating random assignment for approximating quantum 2-local Hamiltonian problems. In 29th Annual European Symposium on Algorithms (ESA 2021), volume 204 of Leibniz International Proceedings in Informatics (LIPIcs), pages 74:1–74:18. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2021. doi:10.4230/LIPIcs.ESA.2021.74.
- [32] Ojas Parekh and Kevin Thompson. An optimal product-state approximation for 2-local quantum Hamiltonians with positive terms, 2022. arXiv:2206.08342v1.
- [33] Stephen Piddock and Ashley Montanaro. The complexity of antiferromagnetic interactions and 2D lattices. Quantum Info. Comput., 17(7-8):636–672, 2017. doi:10.5555/3179553.3179559.
- [34] Stephen Piddock and Ashley Montanaro. Universal qudit Hamiltonians, 2018. arXiv:1802.07130.
- [35] Thomas J. Schaefer. The complexity of satisfiability problems. In Proceedings of the tenth annual ACM symposium on Theory of computing, pages 216–226, 1978. doi:10.1145/800133.804350.
- [36] Johan Thapper and Stanislav Živnỳ. The complexity of finite-valued CSPs. Journal of the ACM, 63(4):1–33, 2016. doi:10.1145/2974019.
- [37] G. Thompson. Normal forms for skew-symmetric matrices and Hamiltonian systems with first integrals linear in momenta. Proc. of the Amer. Math. Soc., 104(3):910–916, 1988. doi:10.2307/2046815.
- [38] John Watrous. The theory of quantum information. Cambridge university press, 2018.
- [39] Pawel Wocjan and Thomas Beth. The 2-local Hamiltonian problem encompasses NP. International Journal of Quantum Information, 1(03):349–357, 2003. doi:10.1142/S021974990300022X.
- [40] John Wright. Personal communication, 2022.
Appendix A Geometry Lemmas
First, we give a simple lemma regarding star graphs. Our goal is to show that inserting star gadgets into a graph forces maximal solutions to to be close to the highest weighted axes.
Lemma 26.
Consider a star graph with center vertex and neighbors. Consider any and any with and .
Let denote the projector onto the axes for which (so the projector onto the -axis or the -plane). Similarly, let be the largest which is not equal to 1 (either or ). Let .
If vectors are assigned to which achieve at least , and is the vector assigned to , then
Proof.
Suppose to the contrary that . We know that
where the inequality holds for . Combining, we have that , which implies . Now we are ready to find that the maximum objective value earned on the star is less than half of
So, the objective value is less than , which is less than , a contradiction.
Next, we study the geometry of triangles. For a triangle , let denote the sum of the edge lengths. It is straightforward to show that for a triangle inscribed in a circle of radius , , which is uniquely achieved by an equilateral triangle [30].
Lemma 27.
Consider a triangle inscribed in the unit circle and any . If , then each edge length is in the interval .
Proof.
For the sake of contradiction, suppose and that there exists an edge length outside the interval .
First suppose that some edge length is greater than . Because the maximum value of is , this implies at least one edge length is less than . So whether an edge length is above or below the interval, some edge length is less than .
Given and some edge length is less than , some edge length must be greater than , so there exists a pair of edge lengths whose difference is greater than .
We relabel the triangle so that and . We can rotate the unit circle as necessary so that is of the form
Define . We will show , implying is less than , a contradiction.
Both and are sums of 3 edge lengths. The edge is shared and we have that , so we are interested in . We may calculate
From here, we may verify
and so
and
To lower bound the left-hand side, we use our lower bound on the difference between the edge lengths and the upper bound that all lengths in the unit circle are less than . We find
Overall, we have found that .
Lemma 28.
Consider triangles and inscribed in the unit circle. If all edges of the triangle have lengths in the interval , then the points are each within distance of (different) points in .
Proof.
Because these are vectors restricted to a unit circle, we can assume . Given is fixed, we can characterize the constraints on the other points as follows. must lie in the intersection of the unit circle with a circular shell bounded by radii and centered at . Given , it is clear this intersection is two disjoint segments of the unit circle.
We wish to upper bound the distance between points in either of these regions. This is the distance between the ends of the regions, along a chord from the point at distance to the point at distance .
The length of the chord can be bound as follows. Given a chord length , the internal angle is . Given an angle , the chord length is . So, we can convert the known distances to angles, take the difference, and convert back to a distance: . This is equivalent to for . Observing this function is increasing in , we can upper bound the value by taking . The Taylor series of gives that this is at most .
Next, we transition to considering tetrahedra. For a tetrahedron , let denote the sum of the edge lengths. It is known that for a tetrahedron inscribed in a sphere of radius , , and the maximum is uniquely achieved by a regular tetrahedron [30].
Lemma 29.
Consider a tetrahedron inscribed in the unit sphere and any . If , then each edge length is in the interval .
Proof.
For the sake of contradiction, suppose and that there exists an edge length outside the given interval.
First consider the case that some edge length is greater than . Because the maximum of in the unit sphere is , at least one of the other edge lengths must be less than . So, whether an edge length is above or below the interval, in both cases some edge length is less than .
If and some edge length is less than , then some edge length must be greater than . So, there exists a pair of edges such that the difference in their lengths is greater than .
If two edge lengths in a tetrahedron differ by more than , then there must exist a pair of adjacent edges which differ by more than . This is because either the two edges are adjacent, or they are both adjacent to a third edge and that length must be closer to or to , i.e. We relabel the tetrahedron so that .
Next, without changing the value of , we can rotate such that are of the form . Let and , We will show , implying must in fact be less than .
Both and are sums of 6 edge lengths. We can ignore . We can directly verify that is
and so is clearly non-negative. We may calculate
and similar expressions for .
Just as in the proof of Lemma 27, we may verify
Similarly, is twice the difference of the quadratic mean and arithmetic mean of , which is non-negative by the generalized mean inequality.
Overall, we have found that .
Lemma 30.
Consider tetrahedra and inscribed in the unit sphere with . If all edges of the tetrahedra have lengths in the interval , then the points are each within distance of (different) points in .
Proof.
Because this is the unit sphere, we can rotate so that are in the -plane and equidistant from the -axis.
Given are fixed, we can characterize the constraints on the other points as follows. Let denote the region bounded by a spherical shell centered at with radii and . Similarly, let denote a region of the same size and shape but centered at . The points must exist in the intersection of , and the unit sphere. The constraints imply the intersection is two disjoint sectors of the unit sphere opposite and , call them . Similarly, the points must exist within of and .
We wish to show that any two points in must be close, and similarly for . Consider any two points in . Fixing and letting , the function is convex in over . Therefore, the extremes of the distance will occur at the four extremal points of , identified by their distances from and .
If is at from and from , then is maximized with at from and . From point , moving towards and then towards is an upper bound on the distance to , so .
The other extreme may occur with and as the distances from to and , respectively, and as the distances from to and , respectively. Recalling that we were able to assume are in the -plane and symmetric about the -axis, it is clear that must be symmetric such that . This implies are coplanar, and so is a cyclic quadrilateral inscribed in a circular cross-section of the unit sphere. We know the lengths of three sides and the diagonals of this quadrilateral and want to know the fourth side, . Ptolmey’s Theorem for cyclic quadrilaterals gives us that
Simplifying, we find
Overall, we conclude that given any two points in , the distance between the points is at most , and similarly for .
Finally, we return to the shapes and . Since the distance must be in , one point must be in and one point in . Similarly, one of must be near and one near . So arbitrarily, are within of and are within of . We can conclude and are at most , as desired.
We will use the name “tetrahedron with adjoined triangles” to refer to a three-dimensional tetrahedron such that for each edge there is an additional point with which the ends of the edge are joined to form a triangle. In total, there are 10 vertices and 12 edges. The shape will be assumed to be inscribed in the unit sphere, so a vector describing any vertex of the tetrahedron or a triangle is a unit vector. For some instance of this shape with assigned vertex locations, let denote the sum of the squared edge lengths. The following lemma says that for to be maximized, the edges of the tetrahedron must approximate a regular tetrahedron.
Lemma 31.
Given a tetrahedron with adjoined triangles , . Furthermore, If any edge of the tetrahedron has a length outside of the interval , then
Proof.
Suppose the vertices of the tetrahedron are , and the additional vertex forming a triangle with edge is . Let denote the length of edge .
Note that the optimal location (which maximizes ) of a vertex is entirely determined by the locations of and , since it has no other neighbors. Given , the point which maximizes the squared distances is proportional to . Moreover, because we are in the unit sphere, the maximum objective achieved by a triangle when are fixed is entirely determined by the length of alone; their exact coordinates do not matter, due to rotational symmetry. Combining these two facts and applying trigonometry, we may calculate that given an edge length of the tetrahedron , the maximum objective value of the triangle is
Now, because is the disjoint union of the six triangles which each have one of the tetrahedron’s edges as a side, we can express
A regular tetrahedron inscribed in the unit sphere has edge lengths equal to . So, if is formed by a regular tetrahedron and the ideal points for the triangles, then .
Now, we suppose the tetrahedron is not regular and some side length is not in the interval . As noted in prior lemmas, it is known that the maximum sum of the (unsquared) edge-lengths of a tetrahedron is [30], so we have . If any edge length is greater than , then the sum of the other five sides must be greater (otherwise, could be trivially increased, meaning our assumption gives a lower bound on the difference), so some side is less than . So whether some side is above or below the interval, there exists a side .
Analyzing , we see it is increasing and concave down on the interval . The maximum occurs at . We can assume all edge lengths are at least and in the concave-down region, as otherwise the total edge length is less than , which is bounded away from the optimum . Therefore, the sum of over the edges of a regular tetrahedron are greater than over a non-regular tetrahedron:
In particular,
After extensive calculation, the difference of the sum over a regular tetrahedron (the LHS) minus the sum over the non-regular tetrahedron (the RHS) is .