Abstract 1 Introduction 2 The Metropolis Process 3 Our results 4 Further Related work 5 Hardness Results for Bipartite Graphs 6 Lower and Upper Bounds for the Time Complexity of Simulated Annealing in Trees 7 Conclusion References

Time Lower Bounds for the Metropolis Process and Simulated Annealing

Zongchen Chen ORCID Georgia Institute of Technology, Atlanta, GA, USA Dan Mikulincer ORCID University of Washington, Seattle, WA, USA Daniel Reichman ORCID Worcester Polytechnic Institute, MA, USA Alexander S. Wein ORCID University of California, Davis, CA, USA
Abstract

The Metropolis process (MP) and Simulated Annealing (SA) are stochastic local search heuristics that are often used in solving combinatorial optimization problems. Despite significant interest, there are very few theoretical results regarding the quality of approximation obtained by MP and SA (with polynomially many iterations) for NP-hard optimization problems.

We provide rigorous lower bounds for MP and SA with respect to the classical maximum independent set problem when the algorithms are initialized from the empty set. We establish the existence of a family of graphs for which both MP and SA fail to find approximate solutions in polynomial time. More specifically, we show that for any ε(0,1) there are n-vertex graphs for which the probability SA (when limited to polynomially many iterations) will approximate the optimal solution within ratio Ω(1n1ε) is exponentially small. Our lower bounds extend to graphs of constant average degree d, illustrating the failure of MP to achieve an approximation ratio of Ω(log(d)d) in polynomial time. In some cases, our lower bounds apply even when the temperature is chosen adaptively. Finally, we prove exponential-time lower bounds when the inputs to these algorithms are bipartite graphs, and even trees, which are known to admit polynomial-time algorithms for the independent set problem.

Keywords and phrases:
Metropolis Process, Simulated Annealing, Independent Set
Category:
RANDOM
Funding:
Dan Mikulincer: DM was partially supported by the Brian and Tiffinie Pang Faculty Fellowship.
Alexander S. Wein: ASW was partially supported by an Alfred P. Sloan Research Fellowship.
Copyright and License:
[Uncaptioned image] © Zongchen Chen, Dan Mikulincer, Daniel Reichman, and Alexander S. Wein; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Complexity classes
Related Version:
Full Version: https://arxiv.org/abs/2312.13554 [11]
Acknowledgements:
We would like to thank Eric Vigoda for his useful feedback regarding this work.
Editors:
Alina Ene and Eshan Chattopadhyay

1 Introduction

Simulated Annealing [42, 60] (SA) is a family of randomized local search heuristics, widely applicable for combinatorial optimization problems. In the maximization version we are given a finite search space 𝒞 of feasible solutions, and a cost f(x), for every solution x𝒞. Additionally, for every solution x𝒞 there is a set N(x)𝒞 of neighboring solutions accessible via a single move of the search algorithm. In contrast to hill-climbing methods that consistently choose an element yN(x) with f(y)f(x), SA may choose a neighboring solution y satisfying f(y)<f(x) with probability eΔT where Δ:=f(x)f(y) and T>0 is a temperature parameter, that governs the behavior of the algorithm. Typically one gradually reduces the temperature over time using a predetermined cooling schedule allowing for a more exploratory algorithm in the early stages. The idea is that since the algorithm is allowed to accept downhill moves it should be able to escape local maxima and find (hopefully in polynomial time) near-optimal solutions. The case where the temperature T is fixed throughout the algorithm has received attention as well [51]: in this case, the algorithm is called the Metropolis process (MP).

Since its inception in the 1980s SA was found empirically to be highly effective for numerous optimization problems in diverse fields such as VLSI design, pattern recognition, and quantum computing. The great popularity of SA is acknowledged in several dedicated books, articles, surveys, and textbooks concerned with algorithm design [43, 14, 1, 4, 39, 40].

Considering the wide applicability of SA for optimization problems, one may wonder what rigorous results can be obtained regarding the algorithm’s performance. It is well-known [26] that for a suitable cooling schedule, SA, if run for sufficiently many iterations, will almost surely converge to a global optimizer. However, there is no guarantee that the running time will be polynomial in the size of the input. This begs the question of what can be said with respect to SA when it is constrained to run for polynomially many steps. This question is explicitly mentioned as challenging in several papers [2, 4, 25, 37]. For example, it is mentioned in [2] that “The polynomial time behavior of simulated annealing is notoriously hard to analyze rigorously”. In the field of approximation algorithms for NP-hard optimization problems not much seems to be known with respect to upper and lower bounds regarding the approximation factor that can be achieved efficiently with SA. The situation for MP is similar: Little is known about the approximation ratio achievable by MP (when run for polynomially many steps) for NP-hard optimization problems. As stated by [37], “Rigorous results…about the performance of the Metropolis algorithm on non-trivial optimization problems are few and far between”. Despite some recent developments [13, 12, 47], the literature on rigorous results for MP and SA remains sparse and experts have noticed the “gap between theory and practice…for Simulated Annealing” [15].

The lack of runtime complexity lower bounds for SA and MP is not a coincidence, since some natural approaches run into difficulties. One direction to prove time lower bounds for MP and SA is to rely on known bounds for the mixing time of the relevant Markov chains. For such bounds, there is a wealth of existing established techniques [46, 52, 20, 9, 16, 6, 54]. In particular, it is well known that for a fixed fugacity parameter λ, the Metropolis process for independent sets converges to a stationary distribution known as the hardcore model. This distribution assigns to each independent set I the value λ|I|/Z where Z is the normalizing constant. There are numerous lower bounds for the mixing time of Markov chains converging to the hardcore model [22, 23, 55, 58, 54]. However, slow mixing does not necessarily imply anything about the efficiency of MP as an approximation algorithm. For example, a simple conductance argument shows that MP has exponential mixing time, with any temperature parameter, when searching for the maximum independent set in the complete bipartite graph Kn,n. On the other hand, it is easily seen that with an appropriate temperature, MP will find an optimal independent set in Kn,n in polynomial time. Furthermore, lower bounds on mixing times imply the existence of a “bad” initial state from which the expected time of the chain to mix is super-polynomial. These kinds of statements do not usually carry information about initialization from specific states, as done in practice. For example, as observed in [35] and further elaborated in [12], conductance lower bounds on the mixing time do not imply comparable lower bounds on the time it takes MP for the independent set problem to find an optimal or even near-optimal solution when using the natural empty state initialization. Finally, as noted in [35], common techniques to prove mixing times lower bounds for homogeneous Markov chains do not generalize in a straightforward way to inhomogeneous chains such as SA.

A brief summary of our results

The overarching aim of this paper is to advance our understanding of the theoretical guarantees afforded by the above-mentioned algorithms and investigate the possible limitations and hard instances. Specifically, we focus on the maximum independent set problem. Recall that given a graph G, an independent set in G is a subset of vertices that spans no edge. The cardinality of a maximum independent set in G is denoted by α(G). Computing α(G) exactly or approximately is a classical NP-hard problem. We establish super polynomial lower bounds on the number of iterations required by MP or SA to obtain a reasonable approximation for α(G) for several different families of graphs: dense graphs, graphs of bounded average degree, bipartite graphs, and trees (for more details please see Section 3). For each such class, we prove corresponding lower bounds, that differ in the obtained approximation ratio and allowed cooling schedule, used in SA. Notably, for dense graphs, which represent the most general category, we present particularly robust results that extend beyond SA, applying to any cooling schedule, even adaptive ones. While the specifics of our results, and their proofs, differ across graph classes, the common thread is this: We establish the existence of graph families where either MP or SA must run for an exponentially large number of steps to approximate the maximum independent set. The only exception is our lower bound for trees where we study the time to find the optimal solution (as we will see in this case that MP does succeed in efficiently finding an approximate solution for tree instances). Finally, we observe that our results imply that SA cannot sample a uniformly random independent set efficiently when the input is a bipartite graph, answering a recent question raised in [32].

Proving lower bounds for MP and SA has proven difficult to achieve even for instances where the independent set problem (or equivalently the clique problem) is believed to be intractable such as sparse random graphs [13] and the planted clique problem [12, 35]. Despite significant effort, it is not known whether MP and SA in their full generality can find efficiently the optimal solution in these instances. Tackling first the easier challenge of proving the limitations of these algorithms for worst-case instances could be instrumental in proving superpolynomial-time lower bounds for MP for finding nearly optimal solutions in sparse random graphs or random graphs containing a planted clique.

It is well known that the independent set problem is NP-hard not only to compute exactly but also to approximate [30, 63, 5]. Nevertheless, the study of unconditional limitations for the quality of approximate solutions to NP-hard problems that can be achieved efficiently by specific algorithmic methods such as linear and semi-definite programming has received significant attention [45, 21, 62]. We believe that proving unconditional limitations for the quality of approximation that can be efficiently achieved by widely used algorithmic methods such as SA and MP is of interest as well. One takeaway of the current work is that rigorous mathematical study of these algorithms can be achieved from first principles. We hope that this will encourage the mathematical study of the performance of SA and MP for additional NP-hard optimization problems.

The exact dynamics of the Metropolis process, Simulated Annealing, and the various variants we consider for the independent set problem are introduced in Section 2, and we refer the reader there for exact definitions. Let us mention that the main differences between the different algorithms lie in how the temperature, also called (inverse) fugacity, changes over the execution of the algorithm. In the classical Metropolis process, the temperature is fixed and does not change, while for Simulated Annealing there is some fixed schedule for decreasing the temperature over time. Some of our results also apply to a more general class of algorithms where the temperature can be chosen adaptively during the algorithm’s execution. In the sequel, for simplicity we shall colloquially refer to all these algorithms as the Metropolis process (MP) and make sure to mention the different temperature schedules when relevant.

2 The Metropolis Process

In this section, we introduce the different variants of the Metropolis process for which we prove lower bounds. Our description differs slightly from the algorithm described in the introduction; we will work on a logarithmic scale for the temperature and parametrize the algorithm by the inverse of the temperature, also called fugacity. Ultimately this is a choice for convenience and the two descriptions are equivalent.

All considered algorithms for finding (random) large independent sets of a given input graph fall under one very general stochastic process, called the Universal Metropolis Process (UMP). Among others, UMP incorporates the Randomized Greedy algorithm, the Metropolis process, and the Simulated Annealing process. Let G=(V,E) be an input graph, and T be the number of steps of the process. For each t1, let λt[1,] be the fugacity, or inverse-temperature, at time t. We refer to the collection {λt}t=1 of all fugacities as the fugacity schedule of the process. With these definitions, the Universal Metropolis Process is described by Algorithms 1 and 2.

Algorithm 1 Universal Metropolis Process (UMP).
Algorithm 2 Update(I,v,ζ,λ).

We observe that:

  • If λt for all t>0, then UMP corresponds to the Randomized Greedy algorithm where in each iteration a vertex is chosen randomly and added to the maintained independent set if it is not a neighbor of a previously added vertex. As the deletion probability is zero, no deletions of added vertices are possible.

  • If λtλ for all t>0 and for some λ[1,), independent of t, then UMP corresponds to the Metropolis process with fugacity λ, which is a Markov chain whose stationary distribution is the Gibbs distribution μ of the hardcore model on G (i.e., μ(I)λ|I| for each independent set I of G).

  • If {λt} forms a predetermined non-decreasing sequence, i.e. λ1λ2, then UMP corresponds to the Simulated Annealing algorithm with fugacity schedule {λt}.

Of course, UMP goes beyond these three well-known cases and allows for, say, non-monotone fugacities schedules (that could depend on the input graph G in complicated ways). In general, if {λt} is some arbitrary deterministic sequence or if it is random, but independent from the randomness of {It}, we shall call it a non-adaptive schedule. On the other hand, if λt can depend on {Is}s=1t1 (and so it is necessarily random), we call the schedule adaptive. As an example for an adaptive schedule, one may set λt=1|It|+1 where It is the currently maintained independent set, resulting with decreased deletion probability as the cardinality of It increases.

3 Our results

Our main result consist of exponential lower bounds on the time complexity of MP when approximating the value of α(G). We construct infinite families of graphs G1,G2,,Gn, (with Gn having n vertices) such that the following holds. There is a function p:(0,1) and a constant η>0, such that if MP runs for fewer than enη iterations the probability it will find an independent set in Gn larger than p(n)α(Gn) is at most enη. In other words, when run for less than an exponential number of steps, MP gives a multiplicative approximation of at most p(n) to α(Gn). As an instructive case, for general graphs, we show that one can take p(n)=1n12ε, while α(Gn)=n1ε, for some ε>0 arbitrarily small. Thus, even though Gn contains an independent set of nearly linear size, MP may struggle to even find an independent set of size nε. All of our results hold when MP is initialized from the empty set. As previously noted, proving lower bounds for these algorithms when starting from a given state has proven challenging.

Results for general graphs

Our first main result is rather general and establishes lower bounds for the classical Metropolis process (with constant temperature) on graphs parametrized by their average degree. In Section 3.1 we outline the key ideas used in the proof and the complete proof can be found in [11].

Theorem 1.

Let {dn}n=1 satisfy dnC for some large enough constant C>0, and dn=o(nlog2(n)). There exists a sequence of graphs {Gn}n0 satisfying:

  • Gn has Θ(n) vertices, the average degree of Gn is Θ(dn), and α(Gn)=Θ(nlog(dn)).

  • If {It}t0 is the process of independent sets maintained by MP with any fixed111By fixed we mean that the temperature does not change during the algorithm. The temperature parameter may depend on n. temperature,

    (maxtexp(nCdn)|It|Clog(dn)dnα(Gn))exp(ndnlog(dn)),

    where C>0 is a universal constant.

Let us unpack Theorem 1 and consider the extreme cases for the average degree. The largest degree we can take and still obtain super-polynomial bounds is dn=npolylog(n). In this case, Theorem 1 guarantees that Gn has an independent set of nearly linear size npolylog(n). However, with any fixed temperature, if MP runs for only a polynomial number of steps, it will fail to find an independent set of even polynomial size and will only result in a set of size polylog(n). To get exponential lower bounds we can slightly lower the average degree and take dn=n1ε for any fixed ε>0. For these slightly sparser graphs, if MP runs for exp(nε) iterations, it will only find a set of size at most O~(nε). To put this result in context, as was mentioned above, it is known that it is NP-hard to approximate the maximum independent set to within an O(1n1ε) factor [30, 63, 41] for any ε(0,1). Thus, the exponential lower bound is predicted by this hardness result and should be seen as an unconditional proof of this prediction for MP.

Theorem 1 also applies when the average degree of the graph is a constant, that does not depend on the number of vertices. For these sparse graphs, there is extensive literature surrounding the question of approximating the maximum independent set [28, 5, 29, 3]. Hence, it is interesting to study the approximation achieved by MP (with polynomial running time) for sparse graphs. Theorem 1 allows to take dnd, for some large enough constant average degree d>0 and obtain a sparse graph. For our sparse graphs, MP will only find an O(log(d)d) approximation of α(G). As an algorithmic counterpart to our lower bound, the randomized greedy algorithm will find an independent set of expected size at least nd+1. Below, in Section 2 we explain how the randomized greedy algorithm can be instantiated as an MP algorithm, which shows that our lower bound is tight up to the log(d) factor.

Simulated Annealing in dense graphs

To go beyond MP, and allow the temperature to change over time, we specialize Theorem 1 to denser graphs. A key appealing feature of our result in this case is that the theorem applies to any sequence of temperatures. In particular, the sequence can be adaptive (see Section 2 for the exact meaning of an adaptive sequence) and may be changed adversarially during the algorithm’s execution.

Theorem 2.

For every ε(0,13), there exists a sequence of graphs {Gn}n0 satisfying:

  • Gn has Θ(n) vertices and α(Gn)=Θ(n12ε).

  • For any temperature schedule, which can be adaptive, if {It}t0 is the process of independent sets maintained by MP, then

    (maxtenη|It|2nε)enη,

    for some constant η>0.

As mentioned above, when the temperature is some predetermined sequence that decreases over time, the MP algorithm is also known as Simulated Annealing. Therefore, by considering this temperature scheduling, Theorem 2 bounds the best approximation ratio SA can achieve. As discussed, this bound precisely matches the best-known results that follow from NP-hardness and again serves as proof of their prediction. The theorem goes beyond SA and, unsurprisingly, shows that there is no way to change the temperature schedule (even if one is allowed to make changes during execution) to go beyond the hurdle suggested by NP-hardness results. Adaptive changes to the temperature in the SA algorithm have been suggested before [33]. We are not aware of previous rigorous results about the benefits or limitations of adaptivity when using these methods to efficiently solve NP-hard optimization problems.

The proof of Theorem 2 appears in the full version of this paper [11].

Results for bipartite graphs

A key feature of our construction of hard instances for bipartite graphs is that they are built from bipartite graphs. These graphs are then augmented by blowing up some vertices into cliques, losing the bipartite structure. Given our construction, it is also interesting to study the performance of MP on general bipartite graphs. The point is that, in this case, there exists a simple linear time algorithm to obtain a 12 approximation of α(G) by finding a bipartition. Furthermore, the standard linear programming relaxation for the independent set problem can recover the exact size of α(G) in polynomial time. Keeping in mind the tractability of the problem for bipartite graphs, it seems natural to expect that there exists some variant of MP that will fare similarly in these instances. On the contrary, our next result shows that in general, MP with any temperature schedule fails to come close to the performance of the mentioned algorithms.

Theorem 3.

Let dnlog(n)100. There exists a sequence of bipartite graphs {Gn}n0 satisfying.

  • Gn has Θ(n) vertices, average degree Θ(dn).

  • For any temperature schedule, which can be adaptive, if {It}t0 is the process of independent sets maintained by MP, then

    (maxtenη|It|(4+o(1))log(dn)dnα(Gn))enη,

    for some constant η>0.

Theorem 3 implies that there exist n-vertex bipartite graphs for which SA cannot efficiently approximate the size of the largest independent set within a ratio better than O(1log(n)). We did not attempt to optimize the hardness ratio as our main point is that MP, with any temperature (thus also covering SA), fails to find an approximate solution even in instances where the independent set problem is tractable. It is possible that stronger inapproximability results hold for SA: It might be that it fails to efficiently approximate the independent set problem in n-vertex bipartite graphs within a ratio larger than 1/nc for some c(0,1). Studying this question is left for future work.

A direct consequence of our time lower bound for SA is that SA (with empty state initialization) cannot (even approximately) sample efficiently uniformly random independent sets in bipartite graphs. It is a central open question in approximate counting whether it is possible to approximately sample in polynomial time a uniformly distributed independent set in a bipartite graph [7]. As noted, [32] asked whether SA can be used for this purpose and we answer this question negatively.

The proof of Theorem 3 and its implications for approximate counting can be found in section Section 5.

Performance of Simulated Annealing on trees

Theorem 3 shows that, even on tractable instances, MP and its variants achieve significantly worse approximation when compared to polynomial-time algorithms. In our final hardness result, we further emphasize this point by considering the, arguably, easiest class of graphs for the independent set problem: trees. Trees are a strict and simpler sub-class of bipartite graphs. A simple greedy algorithm will return the maximum independent set in polynomial time [43]. In our next theorem, we give a complete characterization of the performance of MP, with a non-adaptive temperature schedule (like in SA), on trees. In particular, we show that MP is not competitive with polynomial-time algorithms: with less than an exponential number of iterations, there are trees where it will fail to find the maximum independent set. We complement this hardness result by establishing that MP can return an arbitrarily good approximation to α(G) in polynomial time.

Theorem 4.

The following hold:

  • There exists a sequence of n-vertex trees {Fn}n0 such that for any constant η(0,14), if {It}t0 is the process of independent sets maintained by MP with any non-adaptive sequence of fugacities,

    (maxtexp(nη)|It|=α(Fn))eΩ(n).
  • For any constant δ(0,1) and any n-vertex forest Fn, there exists λ=λ(δ,n) such that, if {It}t0 is the process of independent sets maintained by MP with fixed fugacity λ,

    (maxtpoly(n)|It|(1δ)α(Fn))1o(1).

The proof of Theorem 4 can be found in Section 6. Our time lower bound for finding optimal solutions in trees holds only for non-adaptive schedules. Obtaining an analgous lower bound for adaptive schedules is an interesting question for future research.

Greedy algorithms vs. MP

As a final remark, one may wonder if there are instances where MP (when restricted to polynomially many iterations) is superior to the well-studied greedy [27] and randomized greedy [24] algorithms for the maximum independent set problem. If this was not the case, one could prove lower bounds for MP by constructing hard instances for greedy algorithms. While greedy algorithms were shown to achieve comparable results to MP for certain problems [38, 8] we provide simple examples in [11] where greedy algorithms achieve significantly worse approximation compared to MP in approximating the independent set problem.

3.1 Proof approach

We first explain our proof approach for Theorems 1 and 2 which establish the failure of the Metropolis process for finding large independent sets in graphs of given edge density characterized by the average degree. We prove this by carefully constructing a family of random graphs. Naively, we would hope to use Erdős-Rényi random bipartite graphs which are significantly unbalanced as bad instances. More specifically, suppose that the vertex set is partitioned into V=LR and every edge connects one vertex from L and one from R. Ideally, we want to have |L||R|, and be able to show that the Metropolis process is more likely to pick up vertices in L, and thus reaches independent sets mostly contained in L. However, a moment of thought immediately shows this cannot be the case. Especially, if in each step the Metropolis process picks a vertex uniformly at random, then vertices in R are more likely to be chosen and MP will get independent sets of large overlap with R, which is nearly optimal.

Instead of simply using an Erdős-Rényi random bipartite graph, we further augment it with a blowup construction. More specifically, we replace each vertex uL from the smaller side with a clique of size and connect this clique to all neighbors of u. We pick sufficiently large so that |L||R|. This immediately provides two advantages. First, in every step, MP is more likely to choose vertices from the new L, now a disjoint union of |L| cliques of size , because |L||R|. Second, independent sets of the blowup graph are in one-to-one correspondence to independent sets of the original graph since each clique can have at most one occupied vertex. Furthermore, it is much more difficult to remove an occupied vertex in order to make the corresponding clique unoccupied, since the MP has to pick the correct occupied vertex among all vertices in the clique.

We can then argue that the MP will not be able to find a large independent set for these blowup graphs with polynomially many steps. Suppose that |L|=n and |R|=kn and that k1. Then, after a suitable burn-in phase, we will show that the MP will pick at most a tiny fraction of vertices in R, and at least a constant fraction of vertices in L. Thus, within the burn-in phase, MP only reaches independent sets mostly contained in L. In particular, there is a set L1L of at least n/10 occupied vertices in L, and a set R0R of at least (k1)n unoccupied vertices in R, at the end of the burn-in phase. These vertices induce a smaller Erdős-Rényi random bipartite graph and the MP on it with the initialization L1 will contain vertices mostly from L1 and barely from R0, via simple conductance (i.e., bottleneck) arguments. Thus, within polynomially many steps the MP cannot reach independent sets with too many vertices from R0. In fact, the obtained independent set contains at most n vertices from R0 with high probability and consequently has size at most 3n since |L|=n and |RR0|n. Meanwhile, R is an independent set of size kn3n, exhibiting the failure of MP. We remark that the whole argument works even for constant k, and edge density O(1/n) so that the average degree is constant, though all “” will be replaced with explicit inequalities.

Our construction of the bipartite graphs, appearing in Theorem 3, is based on the t-blowup operation. In this blowup, every vertex is replaced by an independent set (“cloud”) of size t and two clouds that correspond to neighboring vertices (before the blowup) are connected by a complete bipartite graph. The main observation is that once a vertex from a cloud is chosen, MP is much more likely to keep adding vertices to the cloud as opposed to deleting vertices from it. By taking many (identical) duplicates of the t-blowup of an initial bipartite graph, we get that for a large fraction of the duplicates no cloud is deleted (assuming a vertex from it is chosen by the algorithm) in polynomial time, hence resulting in essentially the randomized greedy algorithm where deletions do not occur. To conclude the proof we need to provide a bipartite graph for which randomized greedy does badly: This can serve as the “base graph” on which we perform the blow-up and duplication. We prove that the random balanced bipartite graph of size 2n with edge probability d/n, for a large enough constant222Our lower bounds can be extended to d=O(log(n)). d>0, is with high probability a hard instance for randomized greedy. Our proof follows a martingale argument and may be of independent interest. The limitation of greedy algorithms for coloring (and implicitly independent set) has been observed before [44] for random multipartite graphs m-vertex graphs where the partition includes mΩ(1) parts. We are not aware of a previous hardness result for the randomized greedy for approximating the size of the maximum independent set in bipartite graphs.

For trees, the core ingredient of the hard instance is a “star-shaped” tree composed of a root r connected to k nodes a1,,ak, and each node ai has a single leaf neighbor bi. The unique maximum independent set consists of r together with all the vi’s. In the first m=k1/2ε iterations, MP will add roughly m/2 neighbors of r. Let I denote the set of indices i for these chosen neighbors. The crux of the argument is to track the configuration of the branches I and show that they behave roughly like an i.i.d. collection of random variables supported on three states (free, ai chosen, bi chosen) where the probability ai is chosen is at least 1/4. It follows that with high probability, some ai will be occupied for exponential time, blocking the root r from ever being added. The argument is completed by duplicating many copies of the hard tree, ensuring the probability that the process runs for less than exponentially many iterations is exponentially small. Finally, the forest can be made into a tree by connecting all the roots of the trees to a single additional vertex. The upper bound, showing that MP can efficiently approximate the optimum solution in a tree within a factor of 1δ for arbitrary δ(0,1), is a simple consequence of the rapid mixing of MP for trees [10, 17], taking λ to be a sufficiently large constant to ensure the partition function is concentrated on large independent sets. The complete proofs of upper and lower bounds of running times for trees can be found in Section 6.

4 Further Related work

One of the earliest works studying lower bounds for the Metropolis process is due to Jerrum [35]. In his work, he considered the planted clique problem where one seeks to find a clique of size nβ for β(0,1) planted in the Erdős-Rényi random graph G(n,12). Jerrum proved using a conductance argument the existence of an initial state for which the Metropolis process for cliques fails to find a clique of size (1+ε)logn assuming β<12 so long as it is executed for less than nΩ(log(n)) iterations.

Several open questions were raised in [35] regarding whether one could prove the same lower bound for MP when initialized from the empty set, whether the same lower bound holds for arbitrary β<1 (as opposed to β<12) and whether similar lower bounds could be extended to SA as opposed to MP. The recent paper [12] made the first substantial progress towards answering Jerrum’s question; when the inverse temperature satisfies λO(log(n)), the super-polynomial lower bounds hold with respect to MP initialized from the empty set even when β<1. Under the same assumptions, the result of [12] also applies to SA (initialized from the empty set) for a certain temperature scheduling termed simulated tempering [49]. The lower bounds in [12] do not rule out that MP could solve the planted clique problem (even for β<1/2) when the inverse-temperature is set to Clogn for a suitable constant C. Establishing a lower bound on the running time of MP for every possible temperature is mentioned as an open question in [12].

It is natural to compare the results in [12] to our lower bounds, such as Theorems 1 and 2 which apply without any restriction on the temperature and establish exponential (as opposed to quasi-polynomial) time lower bounds. Moreover in the dense case, which is the analog of G(n,12), our bounds go beyond the restriction of simulated tempering and cover any sequence of temperatures. It should be noted however that [12] focused on resolving Jerrum’s questions, while our work is geared towards proving general lower bounds. Thus, for example, in the planted clique problem a quasi-polynomial lower bound is the best one can hope for, as MP can be shown to solve the problem with high probability in nO(log(n)) iterations. In contrast, to prove our lower bounds we choose carefully crafted instances of random distributions on graphs, which allows for more flexibility in establishing lower bounds.

In [13] exponential lower bounds were proven with respect to the mixing time of MP for independent sets in sparse random graphs G(n,dn) assuming d is a large enough constant. Observe however that lower bounds on the mixing time do not imply lower bounds for the time it takes MP to encounter a good approximation of α(G). It is still open whether MP (with empty set initialization) fails to find an independent set of size (1+ε)log(d)dn in polynomial time in G(n,dn). While the setting and proofs in [13] are different from those in our paper, a common theme in both papers is that a barrier for MP is the need to delete many vertices from a locally optimal solution in order to reach a superior approximation to the optimum.

Few additional lower bounds are known for the time complexity of SA when used to approximately solve combinatorial optimization problems. Sasaki and Hajek [57] proved that, for certain instances, when searching for a maximum matching in a graph, the running time of SA can be exponential in the number of vertices (even when initialized from the empty set). Let us stress the fact that the lower bound in [57] concerns finding an exact solution, rather than approximating the solution. Moreover, their family of hard instances cannot be used to prove SA requires super-polynomial time to approximate maximum matching (and hence the independent set problem) within a factor of α for a fixed α(0,1), as they prove that for every fixed ε(0,1) MP yields a 1ε multiplicative approximation for the maximum matching problem in polynomial time. A different proof showing that MP can find a 1ε approximation for maximum matching in polynomial time, was discovered later in [36].

In [61] it is shown, by providing a family of hard instances, that SA cannot find in polynomial time a 1+o(1) multiplicative approximation for the minimum spanning tree problem. However, this result cannot be extended in a substantial way to show that SA fails to find a 1+ε approximation for fixed ε>1: For the MST problem (with non-negative weights), it was proven in [15] that SA can find a 1+ε approximation for the optimal solution in polynomial time for any fixed ε>0, extending an earlier result of [61]. In another direction, [6] proves exponential lower bounds on the mixing time of SA-based algorithms, which can be seen as a crucial hindrance for finding approximate solutions. Manthey and van Rhijn [48] have studied recently the performance of SA for the TSP problem in certain random instances. They conjuncture that for these random instances SA will require exponential time to find the optimal solution.

In terms of approximation ratios that can be achieved in polynomial time by SA, [39, 40] provide extensive empirical simulations for SA when applied to NP-hard optimization problems such as min bisection and graph coloring. In terms of rigorous results regarding the approximation ratio achieved efficiently by SA in the worst case, [25] provides an algorithm inspired333The authors in [25] mention that “We should remark that our algorithm is somewhat different from a direct interpretation of simulated annealing.” For more details, see [25]. by SA that achieves in polynomial time a 0.41-approximation for unconstrained submodular maximization and a 0.325-approximation for submodular maximization subject to a matroid independence constraint. In [59] it is proven that certain properties associated with the energy landscape of a solution space may lead SA to efficiently find the optimal solution.

Compared to SA, somewhat more is known regarding lower bounds for MP. Sasaki [56] introduced families of n-vertex instances of min-bisection and the traveling salesman problem (TSP) where MP requires exponential time to find the optimal minimum solution. Sasaki’s proof is based on a “density of states” argument. The main step is to show, via a conductance argument, that when the number of optimal solutions is smaller by an exponential multiplicative factor than the number of near-optimal solutions, there exists an initialization where the expected hitting time of the optimal solution is exponential in n. It is explicitly mentioned in [56] that the proof methods do not imply lower bounds for SA. While our hard instance for trees has the property that the number of optimal solutions of value OPT is smaller by an exponential factor than the number of solutions of value OPT1, our proof that SA requires exponential time to find the optimal solution differs from the proofs in [56] (indeed our proof applies for SA whereas the proof in [56] applies only for the fixed temperature case). The paper [50] contains constructions of instances of the traveling salesman problem (TSP) where MP takes exponential time to find the optimal solution but SA with an appropriate cooling schedule finds a tour of minimum cost in polynomial time. Both [56, 50] prove lower bounds for exact computation of the optimum and do not prove lower bounds for algorithms approximating the optimum. An informative survey of these and additional results related to SA and MP can be found in [34].

Recently, MCMC methods were shown useful in algorithmic applications despite slow mixing [47]. For example, despite exponential mixing time of the Glauber dynamics (a lazy version of MP) for the hardcore model in bipartite graphs [16], it can find an independent set of size Ω(logddn) in an n-vertex graph of max degree d in polynomial time.

There is an extensive literature on efficient approximation algorithms for the independent set problem. As mentioned, for general n-vertex graphs, it is NP-hard to approximate α(G) within a factor of 1n1ε for any ε(0,1) [30, 63]. Under a certain complexity-theoretic assumption it was shown in [41] that approximating the size of an independent set in n-vertex graphs within a factor larger than 2(logn)3/4+γ/n (for arbitrary γ>0) is impossible. The current best efficient approximation algorithm for the independent set problem achieves a ratio of Ω(log(n)3nloglog(n)2) [18]. For graphs with average degree d it has long been known that a simple greedy algorithm achieves an approximation of Ω(1d). This bound has been gradually improved [28, 29], and the state of the art [3] is an algorithm based on the Sherali-Adams hierarchy achieving an approximation ratio of Ω~(log(d)2d) for graphs of maximum degree d (the Ω~ sign hides poly(loglogd) multiplicative factors). The running time of this algorithm is polynomial in n and exponential in d. This matches, up to poly(loglogd) factors, the lower bound in [5] showing that obtaining an approximation ratio larger than O(log2(d)d) is NP-hard in graphs of maximum degree d (assuming d is a constant independent of n). In contrast to these lower bounds our lower bounds for the time complexity needed to find approximate solutions apply to specific algorithms (MP and SA). On the other hand, our results are unconditional and also apply to instances (such as bipartite graphs) where polynomial-time algorithms are known to find the optimal solution.

5 Hardness Results for Bipartite Graphs

In this section, we focus on bipartite graphs and prove Theorem 3. The main technical novelty of the proof is a reduction between the Metropolis algorithm and the randomized greedy algorithm on bipartite graphs. The idea is to blow up a hard instance for randomized greedy in a way that makes the added randomness of Metropolis inefficient.

Recall that the randomized greedy algorithm is equivalent to the Metropolis process at 0 temperature, or alternatively λt. In other words, the algorithm chooses random vertices uniformly at random, adds them to a growing independent set whenever possible, and never deletes them. Thus, the algorithm always terminates after each vertex is chosen at least once, which happens in finite time almost surely. For the remainder of this section, for a graph G, we shall denote by IG, the (random) independent set obtained at the termination of the randomized greedy algorithm on G.

Having introduced the randomized greedy algorithm Theorem 3 is now an immediate consequence of the following two results.

Proposition 5.

Suppose that there exists a family of bipartite graphs {Gn}n0 on Θ(n) vertices, and a function r:(0,1), such that,

(|IGn|>r(n)α(Gn))=enη,

for some η>0. Then, for ε>0 small enough, there exists a family of bipartite graphs {G~n}n0 on poly(n) vertices, such if It stands for the metropolis process on G~n with any temperature,

(maxt<enη|It|(r(n)+2n1ε)α(Gn))=enη,

for some ηη.

Proposition 6.

Let dn<log(n)100 be a sequence of numbers. There exists a sequence of bipartite graphs Gn on Θ(n) vertices and average degree Θ(dn), such that

(|IGn|>(4+o(1))log(dn)dnα(Gn))=enη,

for some η>0.

Proposition 5 essentially says that hard instances for randomized greedy can be used to construct hard instances for Metropolis, with any cooling schedule, and Proposition 6 asserts that hard instances for randomized greedy do exist. Theorem 3 immediately follows by coupling these two facts and appropriately choosing dn. Thus, the rest of this section is devoted to the proofs of the two propositions. In Section 5.1 below we prove Proposition 5. The proof of Proposition 6 can be found in the full version [11].

Implications to sampling independent sets in bipartite graphs

Proposition 5 implies the existence of an infinite family of 2n-vertex bipartite graphs Gn with both sides of cardinality n such that the following holds. When SA is ran on Gn for polynomially many iterations, it will return with probability 1o(1) an independent set of size at most n/100. Namely, the maximum cardinality of an independent set the SA algorithm will encounter is at most n/100 ( The constant 1/100 can be improved to a function g(n) tending to zero with n but we choose it for simplicity and as it suffices for our purpose). It follows that SA will fail to provide a (nearly) uniform sample of independent sets in polynomial time in Gn, as it will only encounter independent sets of size at most n/100 whose total number is at most (2nn/100)(600)n/100: a negligible fraction of the total number of independent sets in Gn which is at least 2n+1.

5.1 From randomized greedy to Metropolis

We begin by explaining how to blow-up hard instances for randomized greedy into hard instances for Metropolis. Let G be a graph on n vertices {vi}i=1n, and denote by IG the (random) independent set obtained by running randomized greedy on G. Assume that, for some r(n)>0, and η>0,

(|IG|>r(n)α(G))enη. (1)

Our goal is to show that there exists a graph for which similar estimates hold when we replace randomized greedy with the metropolis process with polynomially many steps.

Let us describe the hard instance. For K,M>10, define the the graph GK,M in the following way:

  1. 1.

    First, let GM be the disjoint union of M copies of G.

  2. 2.

    GK,M is obtained as the K blow-up of GM. That is, every node is replaced by an independent set of size K, and a complete bipartite graph replaces every edge.

As usual, we shall write It for the set maintained by MP on G and enumerate the vertices of GK,M as vi,m,k, the kth element in the blow-up of vi in the mth copy of G. We shall refer to the set of vertices vi,m={vi,m,k}k=1K corresponding to a vertex vi and in the mth copy, as a cloud. Clearly, every cloud has exactly K vertices. The following quantity shall play a central role,

vi,mt:=|It{vi,m,1,,vi,m,K}|,

the load in the cloud vi,m, at time t. The idea is that once a cloud becomes occupied, MP is more likely to add more vertices from the cloud than to remove existing ones. Thus it is very unlikely that a cloud will become empty once occupied. This should be seen as an analogy to randomized greed; as long as no cloud has emptied one can simulate randomized greedy on a given component of the graph. Having many different copies ensures that on most copies no cloud will empty.

In light of this, we first show that once the load becomes positive, and so a cloud becomes occupied, it is very unlikely to drop to 0 again.

Lemma 7.

For any t00, and t0<t<KK1/32 it holds that,

P(vi,mt+t0=0|vi,mt0=1)CK1/3,

for some absolute constant C>0.

Proof.

Let t0>t0+K13 be the first time that K13 where chosen from the cloud vi,m, and observe

(t[t0,t0]:vi,mt=0|vi,mt0=1)(vi,mt0=K13|vi,mt0=1)1K13. (2)

Indeed, this estimate follows from a standard Balls and Bins argument. The probability that after the first K13 choices of vertices in vi,m the algorithm deletes a vertex from It is at most

i=1K13iKK23K1K13,

since by the union bound this upper bounds the probability K13 balls will have at least one collision when distributed randomly across K bins. This estimate gives the right inequality in (2). The left inequality in (2) then follows from the fact that to reach a load K13 every chosen vertex, out of the K13 was added to the cloud, and so no vertex was removed.

Conditional on the event vi,mt0=K13 denote now Yt=vi,mt+t0 and observe that as long as Yt<2K13 we have that, at any t in which the value Yt changes,

P(YtYt1=1)K2K13K,
P(YtYt1=1)2K13K.

Define the stopping time τ=min{t|Yt=2K13 or Yt=0}. Similar to the proof of Theorem 2, since Yt is stochastically dominated by a biased random walk with the above increments and since Y0=K13, the results in [19, (2.8), Chapter XIV.2] imply

(Yτ=0)(2K13K2K13)K13(4K13K)K134K2K133.

The proof concludes by iterating this argument for a polynomial number of steps and applying a union bound.

For fixed time t>0 we now define the number of deloaded clouds as

deload(t):=#{vi,m|t<t′′t such that vi,mt=1 and vi,mt′′=0}.

In words, deload(t) measures the number of clouds that were at some point occupied (had at least one vertex chosen by the algorithm) and later become unoccupied (all vertices in the cloud were deleted at a later point). The main upshot of the previous result is that the number of deloaded vertices remains very small after polynomially many iterations.

Lemma 8.

Suppose that CnMK131. Then, for any t<KK1/32, and ε>0

(deload(t)nεnMK13)=eΩ(nε).
Proof.

Observe that GK,M contains nM clouds. Thus, for t<KK1/32, by Lemma 7, deload(t) is stochastically dominated by B:=binomial(nM,CK13). Since 𝔼[B]=CnMK131, by Chernoff’s inequality,

(BnεnMK13)eΩ(nε).

The proof is complete. We can now prove Proposition 5.

Proof of Proposition 5.

Set M=n and K=n6, so that nMK13=1. By Lemma 8, with probability eΩ(nε) at most nε clouds were de-loaded by time tKK1/32. With no loss of generality, they belong to the first nε copies in GM, {Gm}m=1nε. On the other Mnε copies, {Gm}m=nεM, since no deloading happened we can couple Metropolis on Gm with (a lazy version of) randomized greedy on the base graph G in the following way:

If Metropolis chooses vertex vi,m,k, from cloud vi,m we let randomized greedy choose vi from G. Since all clouds have the same size, the probability of choosing a vertex from cloud vi,m is equal to the probability of choosing the vertex vi. Moreover, since edges exist only between clouds, whenever vi,m,k can be added to It then either vi can also be added, if vi,mt=0, or vi was already added, if vi,mt>0. If metropolis removes vi,m,k then randomized greedy does nothing and maintains its chosen vi in the independent set. This coupling remains valid until the first deloading happens in Gm.

Thus, by the assumption (1) on the base graph G, with probability 1eΩ(nε) for every mnε, (|ItGm|>r(n)Kα(G))enα, where we allow metropolis to fill in all K vertices from every occupied cloud. So, suppose that ε<α, then

(mnε,|ItGm|<r(n)Kα(G))12Menα.

On the other hand, clearly for mnε we have |ItGi|Kn. It follows that

(|It|Knnε+r(n)KMα(G))12Menα.

Finally, by construction GK,M has a maximal independent set of size α(G)Mk, so the approximation ratio is Knnε+r(n)KMα(G)α(G)MK. Let us choose now M=n and K=n2 to get an approximation ratio of nεα(G)+r(n)2n1ε+r(n). Here we have used the fact that if G is a bipartite graph on n vertices then α(G)n2.

6 Lower and Upper Bounds for the Time Complexity of Simulated Annealing in Trees

Here we analyze the performance of MP on trees and forests with an aim to prove Theorem 4. The proof of Theorem 4 is separated into two parts. First, in Section 6.1 we construct a determinstic hard instance for MP. The hard instance is a union of identical trees, each having weak hardness guarantees. It turns out that taking polynomially many copies of the same tree is enough to imply exponential lower bounds, and so proves the first point of Theorem 4. In Section 6.2 we apply recent results about mixing times of MP on graphs with bounded treewidth to prove the second point of Theorem 4.

6.1 Exponential lower bound

We consider MP with an arbitrary non-adaptive fugacity schedule λt, and fix a small constant ε(0,1/2). To construct the hard instance, first consider a “star-shaped” tree Tk that consists of a root r connected to k nodes a1,,ak, and each node ai has a single leaf neighbor bi. That is, Tk consists of a root connected to k edge-disjoint length-2 paths. Let A={a1,,ak} and B={b1,,bk}. The unique optimal independent set in Tk is rB, which has size k+1.

We first describe the first phase of the algorithm and show that it is hard for MP to include the root.

Lemma 9 (Burn-in phase).

The following holds with high (1ok(1)) probability. After m=k1/2ε iterations, MP has at least (1/2ε)m vertices from A (and, as a result, does not include the root r).

Proof.

With high probability, the root r is not selected during the first m iterations. By a standard balls-and-bins argument [53], with high probability the vertices selected during the first m iterations are all distinct, and furthermore at most one vertex from each branch is selected (ai or bi, but not both). Thus, MP adds all the vertices that are selected during the first m iterations and does not delete any. With high probability, at least (1/2ε)m of the selected vertices belong to A.

Now condition on the first m steps of MP and suppose the high-probability event from Lemma 9 holds. Letting I[k] denote the set of branches i for which MP includes ai at the end of m steps, we have |I|(1/2ε)m. We will now focus on only the branches I and show that MP continues to include at least one of the corresponding ai’s for exponential time, blocking the root r from being added.

For the analysis, it will be convenient to consider an auxiliary process MP’, defined as follows. MP’ has the same underlying graph Tk and starts at the same state as MP at timestep m. MP’ has the same update rule as MP except it never adds the root r (even if r is selected and none of its neighbors are present). The random choices of the two processes are coupled so that MP and MP’ share the same state until the first time that MP adds the root.

Fix a time horizon T>m. Our goal is to show that with high probability, at each timestep tT, MP includes at least one of the vertices {ai:iI}. It suffices to show that the same holds for MP’, as the presence of any ai prevents MP from adding the root. Therefore we turn our attention to the analysis of MP’.

Now reveal, and condition on, the choice of which branch is selected by MP’ at each timestep t for m<tT. That is, we reveal the variable σt which is equal to iI if MP’ chooses either ai or bi at step t, and equal to otherwise (i.e., if MP’ selects r or a branch outside I). The random choice between ai and bi is not revealed yet.

For iI, let si denote the number of timesteps t (where m<tT) for which σt=i. For each iI, we define an associated Markov chain X0(i),X1(i),,Xsi(i) on the state space {𝒜,,} with initial state X0(i)=𝒜. The state 𝒜 encodes that MP’ has ai (but not bi), the state encodes that MP’ has bi (but not ai), and the state encodes that MP’ has neither ai nor bi. Every time branch i updates (that is, t such that σt=i), the Markov chain X(i) updates according to the following rules.

  • If the previous state is , the new state is

    {𝒜with probability 1/2,with probability 1/2.
  • If the previous state is 𝒜, the new state is

    {w.p.  1/(2λt),𝒜w.p.  11/(2λt).
  • If the previous state is , the new state is

    {w.p.  1/(2λt),w.p.  11/(2λt).

Note that (conditioned on {σt}) the Markov chains X(1),,X(|I|) are independent (since MP’ never adds the root, by definition).

Lemma 10.

For any fixed iI and any fixed 0si, we have (X(i)=𝒜)1/4.

Proof.

Proceed by induction on , strengthening the induction hypothesis to include both (i) (X(i)=𝒜)1/4 and (ii) (X(i)=𝒜)(X(i)=). The base case =0 is immediate, as X(i) is defined to start at state 𝒜. For the inductive step, we analyze the update rules given above. If X(i) takes values 𝒜,, according to the vector of probabilities (a,b,c), then X+1(i) is distributed according to the vector of probabilities

(a,b,c):=(c2+(112λt)a,c2+(112λt)b,12λt(a+b)).

By induction we have a1/4 and ab. Note that from ab we immediately have ab, which proves (ii). For (i), ab implies b1/2 and so

a=c2+(112λt)ac+a2=1b214,

completing the proof.

Therefore, by independence across branches, we have for any fixed timestep t (with m<tT), the probability that MP’ has none of the vertices {ai:iI} is at most (3/4)|I|. Taking a union bound over t, the probability that MP’ intersects {ai:iI} until time T is at least 1T(3/4)|I|. For Texp(k1/22ε) this is 1ok(1), recalling |I|(1/2ε)m=(1/2ε)k1/2ε. As discussed above, we conclude the same result for the original process MP.

Proposition 11.

Fix any constant η(0,1/2) and consider MP run on Tk with an arbitrary non-adaptive fugacity schedule λt. With probability 1ok(1), MP does not add the root r at any point during the first exp(kη) iterations (and thus does not find a maximum independent set).

The hardness result in Theorem 4 now follows by bootstrapping Proposition 11.

Proof of first item in Theorem 4.

Let us first prove the result for a forest rather than a tree. Consider k disjoint copies of Tk, which has a total of k(2k+1) vertices. Additional isolated vertices can be added to create a forest of any desired size. To find the maximal independent set in the union, MP must add the root of every copy of Tk. Conditional on the exact number of updates per copy, MP evolves independently on each copy. Thus, the independence of the different copies and Proposition 11 yield an exponential bound exp(Ω(k)) on the probability that it adds all the roots by time exp(kα).

We now modify the construction, turning the forest into a tree. Add a new vertex and connect it to the root of each Tk (and also to any isolated vertices that were added to pad the instance size). The maximum independent set still includes the root of every Tk. If it were not for the new vertex, we have from above that with probability 1exp(Ω(k)), there is at least one copy of Tk where the root is never added before time exp(kα). With the new vertex present, this copy of Tk still will never add the root before time exp(kα), as the new vertex can only affect Tk by blocking the root from being added. This completes the proof.

6.2 The Metropolis process can efficiently find approximate solutions on trees

Here we show that if one seeks an approximate solution to the size of the largest independent set in a forest then MP finds such a solution efficiently:

Proof of second item in Theorem 4.

Let N be the maximum cardinality of an independent set in Fn. Since Fn is bipartite we have that Nn/2. Set the fugacity as λ41/δ+log2nδn. We upper bound the contribution of independent sets whose size is smaller than (1δ)N to the partition function:

|I|<(1δ)Nλ|I|2nλ(1δ)NλN/n.

Therefore, for the stationary distribution of MP with this value of λ the probability we get an independent set of size smaller than (1δ)N is at most 1/n. The desired result now follows from the fact [17, 10] that MP on forests mixes in time nO(1) (assuming λ is a fixed constant independent of n) as well as the fact [46] that after tmixlogn iterations of MP the total variation distance between the chain and the stationary distribution is at most 1/n. Therefore, MP finds an independent set of size at least (1δ)N with probability at least 12/n.

 Remark 12.

For graphs of treewidth t it is known that the mixing time of MP is nO(t) [17, 10]. Since any graph with treewidth t is t-degenerate (has a vertex of degree at most t) it has an independent set of size at least n/(t+1). The argument above shows that assuming t is constant (independent of n), MP will find a 1ε approximation to the optimal solution in polynomial time with high probability.

7 Conclusion

We have proved super polynomial lower bounds on the time complexity of the SA and MP when approximating the size of a maximum independent set in several graph families. Several questions remain:

  • Prove that SA cannot approximate in polynomial time α(G) in graphs with constant average degree d within a factor larger than O(logd/d). Currently we only know how to prove this when the temperature is fixed and does not change throughout the algorithm.

  • Analyze the approximation ratio MP achieves in polynomial time for α(G) in planar graphs. A concrete first step would be to analyze how well MP approximates α(G) on the square grid.

  • It would be interesting to compare MP (with polynomially many iterations) to the well-studied min-degree greedy algorithm [31, 27] which is known [27] to achieve a 3/(Δ+2)-approximation on graphs with maximum degree Δ. Whether this approximation ratio can be achieved (or improved upon) by MP in polynomial time is currently open. The new machinery introduced recently in [47] may be relevant for this question.

References

  • [1] Emile Aarts and Jan Korst. Simulated annealing and Boltzmann machines. Wiley-Interscience Series in Discrete Mathematics and Optimization. John Wiley & Sons, Ltd., Chichester, 1989. A stochastic approach to combinatorial optimization and neural computing.
  • [2] David Aldous and Umesh Vazirani. “Go with the winners” algorithms. In Proceedings of the 35th Annual Symposium on Foundations of Computer Science, pages 492–501. IEEE, 1994.
  • [3] Nikhil Bansal, Anupam Gupta, and Guru Guruganesh. On the Lovász theta function for independent sets in sparse graphs. SIAM J. Comput., 47(3):1039–1055, 2018. doi:10.1137/15M1051002.
  • [4] Dimitris Bertsimas and John Tsitsiklis. Simulated annealing. In Probability and algorithms, pages 17–29. Nat. Acad. Press, Washington, DC, 1992.
  • [5] Amey Bhangale and Subhash Khot. UG-hardness to NP-hardness by losing half. Theory Comput., 18:Paper No. 5, 28, 2022. doi:10.4086/toc.2022.v018a005.
  • [6] Nayantara Bhatnagar and Dana Randall. Torpid mixing of simulated tempering on the Potts model. In Proceedings of the Fifteenth Annual ACM-SIAM Symposium on Discrete Algorithms, pages 478–487. ACM, New York, 2004. URL: http://dl.acm.org/citation.cfm?id=982792.982860.
  • [7] Jin-Yi Cai, Andreas Galanis, Leslie Ann Goldberg, Heng Guo, Mark Jerrum, Daniel Štefankovič, and Eric Vigoda. # bis-hardness for 2-spin systems on bipartite bounded degree graphs in the tree non-uniqueness region. Journal of Computer and System Sciences, 82(5):690–711, 2016. doi:10.1016/J.JCSS.2015.11.009.
  • [8] Ted Carson and Russell Impagliazzo. Hill-climbing finds random planted bisections. In Proceedings of the Twelfth Annual ACM-SIAM Symposium on Discrete Algorithms (Washington, DC, 2001), pages 903–909. SIAM, Philadelphia, PA, 2001. URL: http://dl.acm.org/citation.cfm?id=365411.365805.
  • [9] Fang Chen, László Lovász, and Igor Pak. Lifting Markov chains to speed up mixing. In Annual ACM Symposium on Theory of Computing (Atlanta, GA, 1999), pages 275–281. ACM, New York, 1999. doi:10.1145/301250.301315.
  • [10] Zongchen Chen. Combinatorial approach for factorization of variance and entropy in spin systems. arXiv preprint arXiv:2307.08212, 2023. doi:10.48550/arXiv.2307.08212.
  • [11] Zongchen Chen, Dan Mikulincer, Daniel Reichman, and Alexander S Wein. Time lower bounds for the metropolis process and simulated annealing. arXiv preprint arXiv:2312.13554, 2023. doi:10.48550/arXiv.2312.13554.
  • [12] Zongchen Chen, Elchanan Mossel, and Ilias Zadik. Almost-linear planted cliques elude the Metropolis process. In Proceedings of the 2023 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 4504–4539. SIAM, Philadelphia, PA, 2023. doi:10.1137/1.9781611977554.ch171.
  • [13] Amin Coja-Oghlan and Charilaos Efthymiou. On independent sets in random graphs. Random Structures Algorithms, 47(3):436–486, 2015. doi:10.1002/rsa.20550.
  • [14] Sanjoy Dasgupta, Christos H Papadimitriou, and Umesh Virkumar Vazirani. Algorithms. McGraw-Hill Higher Education New York, 2008.
  • [15] Benjamin Doerr, Amirhossein Rajabi, and Carsten Witt. Simulated annealing is a polynomial-time approximation scheme for the minimum spanning tree problem. In GECCO’22—Proceedings of the Genetic and Evolutionary Computation Conference, pages 1381–1389. ACM, New York, 2022. doi:10.1145/3512290.3528812.
  • [16] Martin Dyer, Alan Frieze, and Mark Jerrum. On counting independent sets in sparse graphs. SIAM J. Comput., 31(5):1527–1541, 2002. doi:10.1137/S0097539701383844.
  • [17] David Eppstein and Daniel Frishberg. Rapid mixing of the hardcore glauber dynamics and other markov chains in bounded-treewidth graphs. arXiv preprint arXiv:2111.03898, 2021. arXiv:2111.03898.
  • [18] Uriel Feige. Approximating maximum clique by removing subgraphs. SIAM J. Discrete Math., 18(2):219–225, 2004. doi:10.1137/S089548010240415X.
  • [19] William Feller. An introduction to probability theory and its applications. Vol. I. John Wiley & Sons, Inc., New York-London-Sydney, third edition, 1968.
  • [20] James Allen Fill. Eigenvalue bounds on convergence to stationarity for nonreversible Markov chains, with an application to the exclusion process. Ann. Appl. Probab., 1(1):62–87, 1991. URL: http://links.jstor.org/sici?sici=1050-5164(199102)1:1<62:EBOCTS>2.0.CO;2-M&origin=MSN.
  • [21] Samuel Fiorini, Serge Massar, Sebastian Pokutta, Hans Raj Tiwary, and Ronald De Wolf. Linear vs. semidefinite extended formulations: exponential separation and strong lower bounds. In Proceedings of the forty-fourth annual ACM symposium on Theory of computing, pages 95–106, 2012. doi:10.1145/2213977.2213988.
  • [22] Andreas Galanis, Qi Ge, Daniel Štefankovič, Eric Vigoda, and Linji Yang. Improved inapproximability results for counting independent sets in the hard-core model. Random Structures & Algorithms, 45(1):78–110, 2014. doi:10.1002/RSA.20479.
  • [23] Andreas Galanis, Daniel Štefankovič, and Eric Vigoda. Inapproximability of the partition function for the antiferromagnetic ising and hard-core models. Combinatorics, Probability and Computing, 25(4):500–559, 2016. doi:10.1017/S0963548315000401.
  • [24] David Gamarnik and David A. Goldberg. Randomized greedy algorithms for independent sets and matchings in regular graphs: exact results and finite girth corrections. Combin. Probab. Comput., 19(1):61–85, 2010. doi:10.1017/S0963548309990186.
  • [25] Shayan Oveis Gharan and Jan Vondrák. Submodular maximization by simulated annealing. In Proceedings of the Twenty-Second Annual ACM-SIAM Symposium on Discrete Algorithms, pages 1098–1116. SIAM, Philadelphia, PA, 2011. doi:10.1137/1.9781611973082.83.
  • [26] Bruce Hajek. Cooling schedules for optimal annealing. Math. Oper. Res., 13(2):311–329, 1988. doi:10.1287/moor.13.2.311.
  • [27] Magnús Halldórsson and Jaikumar Radhakrishnan. Greed is good: Approximating independent sets in sparse and bounded-degree graphs. In Proceedings of the twenty-sixth annual ACM symposium on theory of computing, pages 439–448, 1994. doi:10.1145/195058.195221.
  • [28] Magnús M. Halldórsson and Jaikumar Radhakrishnan. Improved approximations of independent sets in bounded-degree graphs. In Algorithm theory – SWAT ’94 (Aarhus, 1994), volume 824 of Lecture Notes in Comput. Sci., pages 195–206. Springer, Berlin, 1994. doi:10.1007/3-540-58218-5_18.
  • [29] Eran Halperin. Improved approximation algorithms for the vertex cover problem in graphs and hypergraphs. SIAM J. Comput., 31(5):1608–1623, 2002. doi:10.1137/S0097539700381097.
  • [30] Johan Håstad. Clique is hard to approximate within n1ϵ. Acta Math., 182(1):105–142, 1999. doi:10.1007/BF02392825.
  • [31] Dorit S Hochbaum. Efficient bounds for the stable set, vertex cover and set packing problems. Discrete Applied Mathematics, 6(3):243–254, 1983. doi:10.1016/0166-218X(83)90080-X.
  • [32] Brice Huang, Sidhanth Mohanty, Amit Rajaraman, and David X Wu. Weak Poincaré inequalities, simulated annealing, and sampling from spherical spin glasses. arXiv preprint arXiv:2411.09075, 2024. doi:10.48550/arXiv.2411.09075.
  • [33] Lester Ingber. Very fast simulated re-annealing. Mathematical and computer modelling, 12(8):967–973, 1989.
  • [34] Thomas Jansen. Simulated annealing. In Theory of randomized search heuristics, volume 1 of Ser. Theor. Comput. Sci., pages 171–195. World Sci. Publ., Hackensack, NJ, 2011. doi:10.1142/9789814282673_0006.
  • [35] Mark Jerrum. Large cliques elude the Metropolis process. Random Structures Algorithms, 3(4):347–359, 1992. doi:10.1002/rsa.3240030402.
  • [36] Mark Jerrum and Alistair Sinclair. Approximating the permanent. SIAM J. Comput., 18(6):1149–1178, 1989. doi:10.1137/0218077.
  • [37] Mark Jerrum and Alistair Sinclair. The markov chain monte carlo method: an approach to approximate counting and integration. Approximation Algorithms for NP-hard problems, PWS Publishing, 1996.
  • [38] Mark Jerrum and Gregory B. Sorkin. The Metropolis algorithm for graph bisection. Discrete Appl. Math., 82(1-3):155–175, 1998. doi:10.1016/S0166-218X(97)00133-9.
  • [39] David S Johnson, Cecilia R Aragon, Lyle A McGeoch, and Catherine Schevon. Optimization by simulated annealing: An experimental evaluation; part i, graph partitioning. Operations research, 37(6):865–892, 1989. doi:10.1287/OPRE.37.6.865.
  • [40] David S Johnson, Cecilia R Aragon, Lyle A McGeoch, and Catherine Schevon. Optimization by simulated annealing: an experimental evaluation; part ii, graph coloring and number partitioning. Operations research, 39(3):378–406, 1991. doi:10.1287/OPRE.39.3.378.
  • [41] Subhash Khot and Ashok Kumar Ponnuswami. Better inapproximability results for MaxClique, chromatic number and Min-3Lin-Deletion. In Automata, languages and programming. Part I, volume 4051 of Lecture Notes in Comput. Sci., pages 226–237. Springer, Berlin, 2006. doi:10.1007/11786986_21.
  • [42] S. Kirkpatrick, C. D. Gelatt, Jr., and M. P. Vecchi. Optimization by simulated annealing. Science, 220(4598):671–680, 1983. doi:10.1126/science.220.4598.671.
  • [43] Jon Kleinberg and Eva Tardos. Algorithm design. Pearson Education India, 2006.
  • [44] Luděk Kučera. The greedy coloring is a bad probabilistic algorithm. Journal of Algorithms, 12(4):674–684, 1991. doi:10.1016/0196-6774(91)90040-6.
  • [45] James R Lee, Prasad Raghavendra, and David Steurer. Lower bounds on the size of semidefinite programming relaxations. In Proceedings of the forty-seventh annual ACM symposium on Theory of computing, pages 567–576, 2015. doi:10.1145/2746539.2746599.
  • [46] David A Levin and Yuval Peres. Markov chains and mixing times, volume 107. American Mathematical Soc., 2017.
  • [47] Kuikui Liu, Sidhanth Mohanty, Prasad Raghavendra, Amit Rajaraman, and David X Wu. Locally stationary distributions: A framework for analyzing slow-mixing markov chains. arXiv preprint arXiv:2405.20849, 2024. doi:10.48550/arXiv.2405.20849.
  • [48] Bodo Manthey and Jesse van Rhijn. Towards a lower bound for the average case runtime of simulated annealing on tsp. arXiv preprint arXiv:2208.11444, 2022. doi:10.48550/arXiv.2208.11444.
  • [49] Enzo Marinari and Giorgio Parisi. Simulated tempering: a new Monte Carlo scheme. Europhysics letters, 19(6):451, 1992.
  • [50] Klaus Meer. Simulated annealing versus Metropolis for a TSP instance. Inform. Process. Lett., 104(6):216–219, 2007. doi:10.1016/j.ipl.2007.06.016.
  • [51] Nicholas Metropolis, Arianna W Rosenbluth, Marshall N Rosenbluth, Augusta H Teller, and Edward Teller. Equation of state calculations by fast computing machines. The journal of chemical physics, 21(6):1087–1092, 1953.
  • [52] Milena Mihail. Conductance and convergence of Markov chains – a combinatorial treatment of expanders. In Proceedings of the 30th Annual Symposium on Foundations of Computer Science, pages 526–531. IEEE, 1989. doi:10.1109/SFCS.1989.63529.
  • [53] Michael Mitzenmacher and Eli Upfal. Probability and computing: Randomization and probabilistic techniques in algorithms and data analysis. Cambridge university press, 2017.
  • [54] Elchanan Mossel, Dror Weitz, and Nicholas Wormald. On the hardness of sampling independent sets beyond the tree threshold. Probab. Theory Related Fields, 143(3-4):401–439, 2009. doi:10.1007/s00440-007-0131-9.
  • [55] Ricardo Restrepo, Jinwoo Shin, Prasad Tetali, Eric Vigoda, and Linji Yang. Improved mixing condition on the grid for counting and sampling independent sets. Probability Theory and Related Fields, 156(1):75–99, 2013.
  • [56] Galen Sasaki. The effect of the density of states on the Metropolis algorithm. Inform. Process. Lett., 37(3):159–163, 1991. doi:10.1016/0020-0190(91)90037-I.
  • [57] Galen H. Sasaki and Bruce Hajek. The time complexity of maximum matching by simulated annealing. J. Assoc. Comput. Mach., 35(2):387–403, 1988. doi:10.1145/42282.46160.
  • [58] Allan Sly. Computational transition at the uniqueness threshold. In 2010 IEEE 51st Annual Symposium on Foundations of Computer Science, pages 287–296. IEEE, 2010. doi:10.1109/FOCS.2010.34.
  • [59] Gregory B. Sorkin. Efficient simulated annealing on fractal energy landscapes. Algorithmica, 6(3):367–418, 1991. doi:10.1007/BF01759051.
  • [60] V. Černý. Thermodynamical approach to the traveling salesman problem: an efficient simulation algorithm. J. Optim. Theory Appl., 45(1):41–51, 1985. doi:10.1007/BF00940812.
  • [61] Ingo Wegener. Simulated annealing beats Metropolis in combinatorial optimization. In Automata, languages and programming, volume 3580 of Lecture Notes in Comput. Sci., pages 589–601. Springer, Berlin, 2005. doi:10.1007/11523468_48.
  • [62] Mihalis Yannakakis. Expressing combinatorial optimization problems by linear programs. In Proceedings of the twentieth annual ACM symposium on Theory of computing, pages 223–228, 1988.
  • [63] David Zuckerman. Linear degree extractors and the inapproximability of max clique and chromatic number. In STOC’06: Proceedings of the 38th Annual ACM Symposium on Theory of Computing, pages 681–690. ACM, New York, 2006. doi:10.1145/1132516.1132612.