Improved Mixing of Critical Hardcore Model
Abstract
The hardcore model is one of the most classic and widely studied examples of undirected graphical models. Given a graph , the hardcore model describes a Gibbs distribution of -weighted independent sets of . In the last two decades, a beautiful computational phase transition has been established at a precise threshold where denotes the maximum degree, where the task of sampling independent sets transitions from polynomial-time solvable to computationally intractable. We study the critical hardcore model where and show that the Glauber dynamics, a simple yet popular Markov chain algorithm, mixes in time on any -vertex graph of maximum degree , significantly improving the previous upper bound by the recent work [3]. The core property we establish in this work is that the critical hardcore model is -spectrally independent, improving the trivial bound of and matching the critical behavior of the Ising model. Our proof approach utilizes an online decision-making framework to study a site percolation model on the infinite -ary tree, which can be interesting by itself.
Keywords and phrases:
Hardcore model, Phase transition, Glauber dynamics, Spectral independence, Online decision making, Site percolationCategory:
RANDOMCopyright and License:
2012 ACM Subject Classification:
Theory of computation Random walks and Markov chains ; Mathematics of computing Markov processesEditors:
Alina Ene and Eshan ChattopadhyaySeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
The hardcore model is one of the most fundamental undirected graphical models that has been extensively studied in statistical physics, social science, probability theory, combinatorics, and computer science.
Given a graph , we let denote the collection of all independent sets of , where we recall that an independent set is a subset of vertices inducing no edges. The Gibbs distribution associated with the hardcore model on is parameterized by a vertex weight called the fugacity. Each independent set receives a probability density given by
where is a normalizing constant call the partition function and is defined as
Perhaps the most amazing property of the hardcore model is the phase transition phenomenon associated with it. In fact, the hardcore model was originally proposed by statistical physicists to study and understand the phase transition in systems of hardcore gas particles. Let denote the maximum degree of the underlying graph. The tree-uniqueness threshold characterizes the uniqueness of the hardcore Gibbs measure on the infinite -regular tree. Furthermore, it also describes the existence of long-range correlations. Let each vertex be associated with a Bernoulli random variable, called the spin, indicating whether the vertex is occupied (i.e., included in the independent set) or unoccupied (i.e., not included in the independent set). Then, for small fugacity the configuration at distance from the root has a vanishing influence on the root as tends to infinity, while for large fugacity the correlation is always bounded away from zero.
In the past two decades, a beautiful computational phase transition has been fully established for the problem of sampling from the hardcore model on graphs of maximum degree , precisely around the uniqueness threshold . For , there exist deterministic approximate counting algorithms for estimating the partition function [28, 2, 21], which in turn gives approximate samplers via standard reduction. Meanwhile, for , no polynomial-time approximate counting and sampling algorithms exist assuming [23, 24, 14].
While all deterministic approximate counting algorithms run in polynomial time, they suffer from a pretty slow runtime. For example, Weitz’s algorithm [28] runs in time where denotes the maximum degree and the slackness of the fugacity (i.e., ). In practice, Markov chain Monte Carlo (MCMC) algorithms provide a simpler and significantly faster method for generating random samples from high-dimensional distributions, including the hardcore model studied in this work. Among them, the Glauber dynamics (also known as the Gibbs sampler) is one of the most important and popular examples. The Glauber dynamics performs a random walk in the space of independent sets and, in each step, either stays the same or moves to an adjacent set whose Hamming distance to the current set is 1. More specifically, from the current independent set , the algorithm picks a vertex uniformly at random and updates its spin: Let ; if then set ; otherwise, set with probability and, mutually exclusively, set with probability .
Let denote the transition matrix of the Glauber dynamics. From basic Markov chain theories it is easy to show that the Glauber dynamics is irreducible, aperiodic, and reversible with respect to the Gibbs distribution , which is the unique stationary distribution (i.e., ). The mixing time of the Glauber dynamics is defined as
where is the initial independent set, is the distribution of the chain after steps when starting from , and denotes the total variation distance.
In the past years, exciting progress has been made in understanding the mixing time of Glauber dynamics for the hardcore model. Anari, Liu, and Oveis Gharan introduced a highly powerful technique known as spectral independence [1], leading to significant advancements in this area, including resolutions to major open problems regarding mixing properties. We refer to [18, 25] for a thorough introduction of this technique. In the subcritical regime (i.e., ), the mixing time of the Glauber dynamics was shown to be nearly linear [1, 9, 5, 7]. Meanwhile, it was long known that in the supercritical regime (i.e., ), the mixing time could be exponentially large as witnessed by random -regular bipartite graphs [19].
In a very recent work [3], the mixing property is further investigated at the critical point (i.e., ). For the upper bound, the mixing time of Glauber dynamics is on any -vertex graph of maximum degree . For the lower bound, there exists an infinite sequence of graphs such that the mixing time is , which is, in particular, super-linear.
In this work, we present an improved mixing time upper bound for the Glauber dynamics on the critical hardcore model.
Theorem 1.
Let be a constant. For any -vertex graph of maximum degree , the Glauber dynamics for the hardcore model on with fugacity satisfies
Our upper bound scales as , significantly improving the mixing time established in [3].
Similar to [3], Theorem 1 is established via the spectral independence framework. Our main contribution is to show that the critical hardcore model satisfies spectral independence of order , improving the trivial bound of used in [3]. We show this new spectral independence result in a novel way by studying an online decision-making problem, which allows us to understand a site percolation model on the infinite tree, from which spectral independence readily follows. We provide an overview of the necessary background and known results on spectral independence, as well as our new contribution and proof approach in Section 2.
2 Proof Overview
2.1 Notations and definitions
Denote the set of non-negative integers by , and the set of positive integers by . For any integers , define by the minimum of and , i.e, .
Let denote the Bernoulli distribution with success probability . Let denote the binomial distribution with number of trails and success probability . Let denote the total variation distance. For any random variables , let denote that and are equal in distribution.
Let be a graph. For any , let denote the set of neighbors of in , i.e., ; and let denote the subgraph induced in by , i.e., the graph with vertex set and edge set consisting of all edges of that have both endpoints in .
Let be a tree rooted at . For every vertex , let denote the subtree of rooted at that consists of all descendants of ; in particular, . For any , let denote the set of children of in .
We end this subsection by defining the (-fold) convolution of distributions on .
Definition 2 ((-fold) Convolution).
Let be two distributions on . Define a new distribution on by
We call the convolution of and . Define where inductively by and for . We call the -fold convolution of with itself.
2.2 Spectral independence via coupling on trees
The core result of this work is to establish the -spectral independence for the critical hardcore model, from which Theorem 1 readily follows by sophisticated spectral independence techniques that have been developed in a recent line of works.
The following notion of influences is needed to define the meaning of spectral independence.
Definition 3 (Influence, [1]).
Let be a distribution over . For any such that and , define the (pairwise) influence from to as
In the setting of the hardcore model, the influences describe the correlation between two vertices, represented as Bernoulli random variables indicating whether the vertices are occupied. Roughly speaking, the influence of one vertex on the other represents the difference of the marginal distribution on the second vertex when flipping the first vertex from occupied to unoccupied.
Theorem 4 (-Spectral independence of critical hardcore model).
Let be a constant. Consider the hardcore model on an -vertex graph of maximum degree with fugacity . Then, for any vertex , we have
where is a constant depending only on .
Theorem 4 states that the hardcore model in the regime of interest satisfies spectral independence with constant (see the full version of the paper [8] and also [13]). An analogous result for the Ising model was previously shown in [3].
Many methods have been introduced to establish the spectral independence property for various families of distributions. Here we adopt the coupling independence approach introduced in [6] and apply it on a related tree, known as the self-avoiding walk tree [28]. The formal definition and construction of this tree are omitted in this paper as we only need its existence, and we refer interested readers to the works [28, 10].
We are interested in coupling two hardcore models on this tree, where in one copy the root is fixed to be occupied while in the other it is fixed to be unoccupied. As we shall see soon in Proposition 5, controlling the number of discrepancies between these two copies under a simple coupling procedure enables us to deduce spectral independence. To formally describe this coupling, we first need a few definitions. Let be a tree rooted at of maximum degree at most . Consider the hardcore model on with fugacity . For each vertex , let denote the probability that is occupied in the hardcore model on the subtree , i.e., , where we recall that is the subtree of rooted at that consists of all descendants of .
We now describe a natural vertex-by-vertex greedy coupling for the hardcore model on when the spin at root is flipped.
-
Initialization: and ;
-
While there exists whose parent has already been revealed in and :
-
–
If , then couple the whole subtree perfectly, i.e., ;
-
–
If and , then set and sample ;
-
–
If and , then sample and set ;
-
–
-
Return .
It is straightforward to check that when the greedy coupling ends, is an independent set distributed as and as .
Proposition 5 (Coupling on trees implies spectral independence, [4, Lemma 39], [6, Proposition 4.3]).
Consider the hardcore model on an -vertex graph of maximum degree with fugacity . For any , there exists a tree rooted at with maximum degree at most , such that if is the greedy coupling on , then it holds
where denotes the symmetric difference of two sets , and recall that .
Hence, to establish Theorem 4, it suffices to bound the expected number of disagreements in the greedy coupling for the hardcore model on trees when the spins at the root are distinct.
2.3 Coupling on trees via site percolation
From the greedy coupling procedure above, we observe a natural site percolation on the tree describing the appearance of disagreements. Every vertex is open with probability independently, representing the occurrence of a disagreement at , that is, ; otherwise, the vertex is closed. The root is always open, i.e., , since . Our goal is to bound the size of the open cluster containing the root, consisting of all open vertices that are connected to the root via a path of open vertices.
We now introduce some notations for the site percolation model. Let be a tree rooted at . For any , let be the probability that is open, and we call the occupation probability of . For simplicity, we assume . Let be the list of occupation probabilities for all vertices, and call it the occupation probability list of the site percolation model. Finally, for the site percolation on with occupation probability list , let denote the random variable representing the size of the open cluster containing the root.
From the construction of the greedy coupling and the site percolation above, we see that , where for each ; see the full version of the paper [8] for a formal statement. Thus, the problem is reduced to studying a site percolation model.
In order to study this site percolation model, we need to know the conditions satisfied by the occupation probabilities . Since these probabilities correspond to the marginal probability of the roots in the respective subtrees, they satisfy a well-known recurrence called the tree recursion (see Fact 21). In this work, we present a new inequality satisfied by these marginal probabilities, which is crucial in obtaining our main spectral independence result. In particular, Equation 1 below provides a stronger and simpler contraction property of the tree recursion, which was not known in the literature as far as we are aware. For simplicity, here we consider only the exact critical fugacity .
Lemma 6 (Special case of Lemma 20; see also [17]).
Let be a tree rooted at with maximum degree at most . Consider the critical hardcore model on with fugacity . For each vertex , let denote the probability that is occupied in the critical hardcore model on the subtree rooted at , i.e., . Then, for every non-root vertex , it holds
| (1) |
where we recall that denotes the set of children of .
Remark 7.
In [17], a potential function approach was applied to study the contraction of the tree recursion for the subcritical hardcore model. Uncovering their result ([17, Lemma 12]), the corresponding condition they established at criticality can be stated as
| (2) |
Our bound Equation 1 is stronger than Equation 2 since . In fact, going through the proof in [17], one is able to recover the stronger inequality Equation 1 though it was not stated explicitly. In this work, we provide a simpler proof of Equation 1 (and hence Equation 2). To establish the -spectral independence at criticality, we do require the stronger inequality Equation 1, while the weaker Equation 2 is not sufficient.
We are now ready to state our main percolation result, from which spectral independence Theorem 4 readily follows.
Theorem 8 (Main result for site percolation).
Consider the site percolation model on the infinite -ary tree rooted at with occupation probability list where . Let be the size of the open cluster containing the root. Suppose the following hold:
-
1.
For every non-root vertex , it holds
(3) -
2.
At the root , it holds
(4) where is a constant.
Then, we have that for any ,
where the constant in big-O depends only on .
At this point, we transfer our original problem of bounding the mixing time and proving critical spectral independence into a problem of site percolation on trees. In [3], the same strategy was applied to the critical Ising model, another canonical example of graphical models. There, the occupation probabilities are much simpler; in fact, they can all be set to , which is a uniform upper bound on the pairwise influence on from its parent. We remark that the main distinction of [3] and our work lies in the application of Lemma 6 (corresponding to the condition Equation 3) which substantially extends the uniform setting, for all , appeared in the critical Ising model.
2.4 Site percolation on trees via online decision making
The main part of the paper aims to prove Theorem 8. We introduce a novel approach to study such a site percolation model on the infinite tree, by considering an online decision-making game.
Our strategy is to upper bound by understanding the worst-case choice of occupation probabilities . We do this by changing the perspective and thinking as an adversary who is allowed to pick the occupation probabilities. Namely, we study an online decision-making problem where a player is allowed to pick the occupation probabilities every time we need to reveal the status (open or closed) of a few vertices. These probabilities can be arbitrary as long as they satisfy the requirements Equation 3, and the goal of the player is to maximize .
To be more specific, let denote the number of active vertices at time , where a vertex is said to be active if (i) it is open; (ii) there is an open path from it to the root; (iii) the status of its children has not been fully revealed. At the beginning, since only the root is open and we have not revealed any other vertex. Then, the player picks the occupation probabilities for both the children and the grandchildren of ; these probabilities are required to satisfy Equation 3. By sampling from the corresponding Bernoulli distributions independently, we reveal the number of grandchildren of , denoted as , that are active (i.e., open and connected to ). With being deactivated, the number of active vertices becomes . The game is then repeated. Whenever there exists an active vertex at round , the player picks occupation probabilities of the children and grandchildren of satisfying Equation 3, and, after the number of open grandchildren connected to is revealed, updates . This process stops when there are no active vertices, i.e., when .
We remark that we consider only active vertices at even levels because Equation 3 imposes requirements on two adjacent levels.
Suppose the game stops after rounds. Then, the number of open vertices connected to the root at even levels is precisely , since every such vertex becomes active at some point and is deactivated at some other time. Therefore, controlling the number of rounds played in the game allows us to bound where is the number of vertices at even levels that are in the open cluster containing the root in the site percolation. (Since , in the actual proof we aim to bound for every for simplicity. And we can bound by the maximum probability that the game lasts for at least rounds.) Finally, combining the upper bound of with Equation 4, it is then not hard to upper bound as wanted.
Therefore, our goal is to determine the optimal strategy of the player when such an online decision-making game is played. We deal with this in Section 3 where we state and prove our main result, Theorem 19. While a natural guess of the optimal strategy is to set every occupation probability to be so that Equation 3 is satisfied, we show that the optimal strategy of the player should set for one child of and set for all other children . We remark that such choices of are not realizable as marginal probabilities in the hardcore model since they do not satisfy the tree recursion and are way too large (in reality, at criticality); however, they are sufficient to provide meaningful upper bounds on as we need.
3 Online Decision-Making Problem
In this section, we introduce an online decision-making problem that serves as a key tool in proving our main result for site percolation (Theorem 8), and show the optimal strategy as our main result for this section.
First of all, we describe the setup of the online decision-making game in Section 3.1. Then, we define a partial order called second-order stochastic dominance, which plays an important role in finding the optimal strategy, and show some basic properties in Section 3.2. After that, we introduce the Poisson binomial distribution in Section 3.3 and some properties of the random walk hitting time in Section 3.4, which are crucial in the proof of the optimal strategy. Finally, in Section 3.5, we state and prove our main result for the online decision-making game and show the optimal strategy of the game.
3.1 Setup of online decision making
We now describe precisely an online decision-making game under a slightly more general setting.
In the online decision-making game, the player maintains some number of tokens (corresponding to active vertices). Let denote the number of tokens the player owns after round . At the beginning, the player has tokens. There is a family of distributions on non-negative integers (corresponding to choices of occupation probabilities). For round , the player spends one token, assuming , and chooses a distribution . Then, a sample is generated independently, and the player receives tokens as a reward. The number of tokens the player owns then becomes . The game ends when the player uses up all the tokens, i.e., the first time . We denote this stopping time by . Note that it is possible the game never stops, in which case . The goal of the player is to survive for rounds for some given integer . That is, the player wins if and only if . We present the process of the online decision-making game in Algorithm 1.
Define a strategy for the player by a mapping . For any , is defined by the distribution the player will choose when they needs to survive for more rounds to win and currently has tokens. For example, if the winning requirement is rounds, then at the beginning of round , the player will choose the distribution assuming .
For , define the winning probability under strategy to be
Namely, is the probability that the game lasts for at least rounds, assuming the player has tokens at the beginning and uses strategy . We further define
The following lemma establishes a simple recursive formula for the optimal winning probabilities and the existence of an optimal strategy.
Lemma 9.
Let be a family of distributions on . Suppose the metric space is compact, where we recall is the total variation distance. Then the following holds:
-
1.
For all ,
-
2.
There exists a strategy such that holds for any .
Proof.
We note that is well-defined if is defined for all , and the result of does not matter. In our proof, when we write , we only guarantee that for all , is defined.
We verify the recurrence and define the strategy inductively on .
For the base case , by definition, , and for all . Then Item 1 immediately follows. Furthermore, we can define so that holds for any , where is an arbitrary distribution.
Now suppose . Suppose Items 1 and 2 hold for . Let . Suppose the player chooses in the first round and obtains tokens; hence, after the first round, the player has tokens. Then to maximize the winning probability, i.e., to maximize the probability that survive for at least rounds when having tokens initally, the player should use strategy (), and then the winning probability is by the induction hypothesis . Therefore, if the player chooses in the first round, the maximum winning probability for the player is . Hence, to obtain the maximum winning probability, we need to choose an optimal which maximizes . By compactness, such optimal exists, therefore Item 1 for holds, and we can define by for all . Then it is straightforward that holds for any .
3.2 Second-order stochastic dominance
Our goal is to find an optimal strategy for the online decision-making game. A first thought that one may consider is to use first-order stochastic dominance (also simply called stochastic dominance). For any two distributions over , is (first-order) stochastically dominated by if and only if for all . An immediate corollary is that is stochastically dominated by implies . If there is a largest distribution under stochastic dominance in , then it can be easily proved that the player can attain the maximum winning probability when always picking the largest distribution. However, the largest distribution under stochastic dominance does not exist in our case, since any two different distributions with the same mean are not comparable under stochastic dominance. In fact, in our application, there are infinitely many distributions attaining the largest mean in . Therefore, there is no largest distribution under stochastic dominance and we need something else.
It turns out that second-order stochastic dominance (see, e.g., [16, 15] for more background and applications) can address the problem of the lack of a largest distribution. However, as a trade-off, it is not always true that the player can achieve the maximum winning probability when always choosing the largest distribution under second-order stochastic dominance. Nonetheless, if the largest distribution satisfies some nice properties (see Theorem 19 for details), this is indeed true.
We now define second-order stochastic dominance.
Definition 10 (Second-order stochastic dominance).
We define a partial order “” called second-order stochastic dominance on the family of distributions on with finite expectations. For two distributions on with finite expectations, if and only if
The following proposition shows some classical equivalent definitions of second-order stochastic dominance.
Proposition 11 (Equivalent definitions of second-order stochastic dominance [20, Theorem 8.1.1], see also [22]).
Let be distributions on with finite expectations. The following definitions are equivalent:
-
1.
;
-
2.
(Increasing concave order) For any non-decreasing concave function , it holds:
-
3.
There exists a coupling of and such that
The following two lemmas offer easy ways to find an “upper bound” of a given distribution in the sense of “”, and are helpful to us in identifying the largest distribution in .
Lemma 12.
Let be a distribution on with expectation . Then .
Proof.
For any , we have
which implies .
Lemma 13.
If and , then , where the operator “*” represents convolution (see Definition 2).
Proof.
Let be independent random variables with distributions respectively. Let be independent random variables with distributions respectively. We also assume are independent and are independent. For , by and Proposition 11, there exists a coupling of , such that
| (5) |
Let . Then is the law of , and is the law of . Let be the joint distribution of assuming . It is clear that is a coupling of and . For any ,
where the last inequality follows from Equation 5. Then Lemma 13 follows from Proposition 11.
3.3 Poisson binomial distribution
As hinted by Item 2 of Proposition 11, in order to apply second-order stochastic dominance, we hope to have some non-decreasing concave functions as the objective/utility function in our decision making. It turns out that when the largest distribution (under “”) is a Poisson binomial distribution (see, e.g., [26] for a thorough introduction) with expectation at least , the maximum winning probability is non-decreasing and concave with respect to .
We first define the Poisson binomial distribution.
Definition 14 (Poisson binomial distributions (random variables)).
We call a random variable a Poisson binomial random variable if it can be expressed as a finite sum of independent Bernoulli random variables, i.e., where , are independent. We call a distribution a Poisson binomial distribution if it is the distribution of a Poisson binomial random variable.
An immediate property is the following.
Fact 15.
If and are two independent Poisson binomial random variables, then is also a Poisson binomial random variable.
The crucial property that guarantees the concavity of the maximum winning probability is that a Poisson binomial random variable is unimodal with a mode near the mean of the random variable. We next define unimodality and state the property.
Definition 16 (Unimodality [12]).
Let be a distribution on . Let be an integer. The distribution is called unimodal about if
and we call a mode of .
Let be a random variable with distribution . The random variable is called unimodal about if is unimodal about .
Proposition 17 (Darroch’s rule for the mode [11]).
Let be a Poisson binomial random variable. Let be the expectation of . Then there exists if , or if , such that is unimodal about . In particular,
3.4 Random walk hitting time
The following proposition of the random walk hitting time will be used in deriving the formula of the maximum winning probability and proving the concavity of the maximum winning probability.
Proposition 18.
Let be a random walk on . Specifically, , where are independent and identically distributed integer-valued random variables satisfying (left-continuous). For any , define the hitting time , with the convention that if for all . Then the following holds:
-
1.
([27]) For every ,
-
2.
For any ,
3.5 Determining optimal strategy
In this subsection, we show our main result for the online decision-making game. The following theorem implies that if the largest distribution (under “”) of exists and is a Poisson binomial distribution with expectation at least , then an optimal strategy for the player is , in other words, the player can achieve the maximum winning probability when always choosing the largest distribution.
Theorem 19 (Main result for online decision making).
Let be a family of distributions on . Suppose the metric space is compact. Suppose there exists a largest distribution in under the partial order “”, denoted by . Furthermore, suppose is a Poisson binomial distribution with expectation at least . Then the following holds:
-
1.
(Recurrence) For any ,
Namely, the player will pick to achieve the maximum winning probability, and the optimal strategy is for all .
-
2.
(Formula) Let be independent and identically distributed random variables with distribution , and let . For any , define , with the convention that if for all . For any , it holds that
-
3.
(Concavity) For any , the function is concave:
We will prove by induction on . The recurrence follows from the concavity of the maximum winning probability and Proposition 11 about second-order stochastic dominance. Given the recurrence, we obtain the formula using Proposition 18 about the random walk hitting time. Finally, we derive concavity using the formula and the fact that is a Poisson binomial distribution and hence satisfies unimodality with mode near the mean.
Proof.
We prove by induction on .
Base case: .
By definition, . It is easy to check that Items 1 and 3 hold for . For Item 2, the first equality holds trivially for , the proof for the second equality for is the same as that for arbitrary , i.e., applying Item 1 of Proposition 18, which we will show in the inductive step below.
Inductive step.
1. Recurrence for
By Lemma 9, it holds that
| (6) |
By definition, it is clear that is a non-decreasing function with respect to . By Concavity for , is a concave function with respect to . By assumption, holds for any . For any , applying the equivalence between Item 1 and Item 2 of Proposition 11 with and , it holds that
| (7) |
Equation 6 and Equation 7 imply Recurrence for .
2. Formula for
We first explain why is true intuitively. Recall that are independent and identically distributed random variables with distribution , representing the player choosing the largest distribution every time. Then represents the net income of tokens after round when the player chooses the largest distribution every time. Then, is the first time that the net income of tokens is , i.e., the time the game stops if the player initially gets tokens. Therefore, is the probability that the game lasts for at least rounds when the player chooses the largest distribution every time. Then follows from the fact that maximum winning probability can be obtained from choosing every time, where the fact follows from Recurrence for .
We next prove formally from the induction hypothesis. For any , we have that
| (Recurrence for ) | ||||
| (Formula for ) | ||||
as desired.
For the the second equality of Formula for , we apply Item 1 of Proposition 18 with and . Then it follows
as desired.
3. Concavity for
For any , it holds that
| (Recurrence for ) | ||||
| (Concavity for ) | ||||
| (Recurrence for ) |
It remains to prove the case , i.e.,
By Formula for ,
It suffices to prove
| (8) |
and
| (9) |
For any , since is a Poisson binomial distribution, by Fact 15, is a Poisson binomial random variable. By Proposition 17, we have
for any , where , which implies Equation 8.
Applying Item 2 of Proposition 18 with yields Equation 9.
4 Site Percolation on Infinite Tree
In this section, we aim to prove the main result for the site percolation model Theorem 8 and show that it can be applied to the critical hardcore model.
In Section 4.1, we prove Lemma 6 about a contraction property of the tree recursion of the hardcore model, which indicates that the main result for the site percolation model Theorem 8 can be applied to the critical hardcore model. In Section 4.2, we prove the main result for the site percolation model Theorem 8 by applying the online decision-making game introduced in Section 3.
4.1 Contraction of tree recursion: Proof of Lemma 6
In this subsection, we prove a general version of Lemma 6. In the general version, we extend the domain of the fugacity from at most the critical fugacity to at most times the critical fugacity. And the original version (Lemma 6) can be obtained simply by letting .
Lemma 20 (General version of Lemma 6).
Let be a tree rooted at with maximum degree at most . Let be a constant. Consider the hardcore model on with fugacity . For each vertex , let denote the probability that is occupied in the hardcore model with fugacity on the subtree rooted at , i.e., . Then, for every non-root vertex , it holds
The following well-known fact gives a natural recurrence of the probability of the root being occupied in the hardcore model of subtrees.
Proof of Lemma 20.
Fix . Let . Then has at most children, i.e., .
By tree recursion Fact 21, we have
where , and the inequality follows from the AM-GM inequality and . Therefore,
Set
Since , we have . We next bound . By standard calculus calculation, we have
Set . Since the first factor of is always non-negative, the sign of depends on the second factor, i.e., . Clearly, is decreasing on , , , by the Intermediate Value Theorem, there exists a unique zero of on , denoted by . Then,
| (by ) | ||||
Therefore,
When , , it holds that
By is decreasing on , we have . Then,
as desired.
We next prove for . Consider as a function with respect to both and . Let . Then, by the result of the case , we have for all . For all ,
where the last inequality follows from the AM-GM inequality. Then, by the Mean Value Theorem, for any , it holds that
where is a real number between and , and the second inequality follows from . Therefore, when , holds for any . Then,
as desired.
4.2 Site percolation on trees: Proof of Theorem 8
In this subsection, we prove a general version of Theorem 8. In the general version, we extend the upper bound of from to . We note the latter upper bound can be derived from Lemma 20. The original version (Theorem 8) can be obtained simply by letting .
Theorem 22 (General version of Theorem 8).
Consider a site percolation model on the infinite -ary tree rooted at with occupation probability list where . Let be the size of the open cluster containing the root. Let be a positive integer. Suppose the following conditions hold:
-
1.
For every non-root vertex , it holds , where is a constant;
-
2.
At the root , it holds
(10) where is a constant.
Then, it holds that , where the constant in big-O depends only on .
We first show an upper bound with respect to , the number of vertices at even levels that are in the open cluster containing the root, and it is straightforward to prove the upper bound with respect to when combining Equation 10.
Lemma 23.
Under the setting of Theorem 22, let be the number of vertices at even levels that are in the open cluster containing the root. Then, we have
where the constant in big-O depends only on .
Proof of Theorem 22.
Let be children of . Recall that is the number of vertices at even levels that are in the open cluster containing the root. For , let be the number of vertices at odd levels that are both in the open cluster containing the root and , where we recall denotes the subtree rooted at . Let Then,
By Lemma 23, . Similarly, for any .
Therefore, there exists a constant depending only on , such that
| (Equation 10) |
This shows as desired.
We next prove Lemma 23 by considering the process of site percolation described in Section 2.4.
Proof of Lemma 23.
Recall that in Section 2.4, we introduced a process of site percolation working in rounds that reveals the status of all children and grandchildren of an active vertex in each round. We note that the process in Section 2.4 is described under an adversary setting. Here, we restate this process more precisely for a fixed site percolation model. For convenience, we call it the site percolation process.
Let be the number of active vertices after round . Let be the set of active vertices after round . We label all the vertices in by . At first, only the root is open and active, and the status of all other vertices is unrevealed. Therefore, , . For any , at the beginning of round , assuming , we choose the active vertex with the least label in the current set of active vertices , denoted by . Then reveal the status of all the children and grandchildren of . To be specific, for each child or grandchild of , denoted by , sample independently, then let be open if ; be closed if . Let be the set of activated grandchildren of . We note that if a grandchild is activated, then it is open; however, the converse does not hold. Let be the number of activated grandchildren of , i.e., . Then , and . The process stops when there are no active vertices, i.e., the first time . If the process stops after rounds, we have , because every vertex in the open cluster containing the root at even levels becomes active at some time and will be deactivated later in the process, and the number of rounds equals the number of deactivated vertices. If the process never stops, we have .
Let be the set of all possible active vertex sets of the site percolation process. For any with , define the winning probability of the size percolation process by the probability that the site percolation process will last for at least more rounds when currently there are active vertices and the set of active vertices is . Then, for any , we have .
Let be the law of . We next find a family of distributions that contains all possible . Let be children of . Let be the number of children of that are activated in round . Then, we have
Then, , where is the law of . Let be a family of distributions defined by
where are distributions on . Then for any , , i.e., is a family of distributions that contains all possible .
The site percolation process defined above can be considered as a strategy of the online decision-making game introduced in Section 3 with . (We note that this strategy is slightly different from the strategy defined in Section 3.1, which we will discuss later.) This inspires us to consider the online decision-making game with .
We next check satisfies the conditions in Theorem 19. It is easy to check that the metric space is compact. Let . We next show is the largest distribution in under “”. First of all, , where we recall that for a distribution , is the -fold convolution of with itself (see Definition 2). For any , we may assume , where are distributions on with expectations at most . For any , since , by Lemma 12, we have . By Lemma 13, we have . Therefore, is the largest distribution in under “”. Furthermore, it is clear that is a Poisson binomial distribution with expectation .
Therefore, by Theorem 19, for the online decision-making game with , there is a optimal strategy , and the player can achieve the maximum wining probability by using this strategy, i.e., by choosing for each round .
As we mentioned above, the site percolation process can be considered as a strategy for a player playing the online decision-making game with . However, the strategy of the player here is distinct from the strategies defined in Section 3.1, in the sense that the player (the site percolation process) maintains extra storage and uses external randomness (i.e., the evolution of the set of active vertices) beyond provided in the game (the target number of rounds to survive and the number of tokens) to make a decision in each round. However, it is not hard to see intuitively that an optimal strategy should not require any other information and should depend only on , the target number of rounds to survive, and , the current number of tokens. Namely, the strategy from the site percolation process is no better than an optimal oblivious strategy as defined in Section 3.1. We formally state this in the following claim, whose proof is similar to Lemma 9.
Claim 24.
For any with , it holds that
where is the maximum winning probability of the online decision-making game with when the player needs to survive more rounds to win and the current number of tokens is .
Claim 24 implies holds for any . We next bound . By Item 2 of Theorem 19, we have for any ,
where is defined in the same way as in Item 2 of Theorem 19. To be specific, let be independent and identically distributed random variables with distribution , and let , then is defined by , with the convention that if for all . Then holds for any .
Let be an occupation probability list satisfying , and for all . It is straightforward to check that , where is the random variable representing the size of the open cluster containing the root in the site percolation on with occupation probability for all non-root vertices. For simplicity of notation, we write as shorthand for . By [3, Lemmas 4.7 and 4.8], we have
Then, for any ,
where are constants depending only on . Therefore,
as desired. To complete the proof of Lemma 23, we prove Claim 24 by induction.
Proof of Claim 24.
We prove by induction on .
For the base case , by definition, it is easy to check
holds for any with .
For , by definition, . We now assume . For any with , consider that at the beginning of some round of the site percolation process, the current set of activated vertices is , and the process needs to last for more rounds to win. Let be the number of vertices activated in this round, and be the set of active vertices after this round. Let be the law of . Then . After this round, the site percolation process needs to last for more rounds to win, and there are active vertices, and the set of active vertices is . Let be the law of . Then,
| (induction hypothesis) | ||||
| () | ||||
| (Lemma 9) |
Therefore, Claim 24 holds for . By the induction principle, Claim 24 holds for all .
References
- [1] Nima Anari, Kuikui Liu, and Shayan Oveis Gharan. Spectral independence in high-dimensional expanders and applications to the hardcore model. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 1319–1330, 2020. doi:10.1109/FOCS46700.2020.00125.
- [2] Alexander Barvinok. Combinatorics and complexity of partition functions, volume 30 of Algorithms and Combinatorics. Springer, Cham, 2016. doi:10.1007/978-3-319-51829-9.
- [3] Xiaoyu Chen, Zongchen Chen, Yitong Yin, and Xinyuan Zhang. Rapid mixing at the uniqueness threshold. arXiv preprint arXiv:2411.03413, 2024. doi:10.48550/arXiv.2411.03413.
- [4] Xiaoyu Chen and Weiming Feng. Rapid mixing via coupling independence for spin systems with unbounded degree. arXiv preprint arXiv:2407.04672, 2024. doi:10.48550/arXiv.2407.04672.
- [5] Xiaoyu Chen, Weiming Feng, Yitong Yin, and Xinyuan Zhang. Optimal mixing for two-state anti-ferromagnetic spin systems. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 588–599, 2022. doi:10.1109/FOCS54457.2022.00062.
- [6] Xiaoyu Chen and Xinyuan Zhang. A near-linear time sampler for the ising model with external field. In Proceedings of the Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 4478–4503, 2023. doi:10.1137/1.9781611977554.CH170.
- [7] Yuansi Chen and Ronen Eldan. Localization schemes: A framework for proving mixing bounds for Markov chains. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 110–122, 2022.
- [8] Zongchen Chen and Tianhui Jiang. Improved mixing of critical hardcore model. Full version, 2025. arXiv:2505.07515.
- [9] Zongchen Chen, Kuikui Liu, and Eric Vigoda. Optimal mixing of Glauber dynamics: entropy factorization via high-dimensional expansion. In Proceedings of the Annual ACM Symposium on Theory of Computing (STOC), pages 1537–1550, 2021. doi:10.1145/3406325.3451035.
- [10] Zongchen Chen, Kuikui Liu, and Eric Vigoda. Rapid mixing of Glauber dynamics up to uniqueness via contraction. SIAM J. Comput., 52(1):196–237, 2023. doi:10.1137/20M136685X.
- [11] John Newton Darroch. On the distribution of the number of successes in independent trials. Ann. Math. Stat., 35(3):1317–1321, 1964.
- [12] Sudhakar Dharmadhikari and Kumar Joag-dev. Unimodality, Convexity, and Applications. Probability and Mathematical Statistics. Academic Press, San Diego, 1988.
- [13] Weiming Feng, Heng Guo, Yitong Yin, and Chihao Zhang. Rapid mixing from spectral independence beyond the boolean domain. ACM Trans. Algorithms, 18(3):1–32, 2022. doi:10.1145/3531008.
- [14] Andreas Galanis, Daniel Štefankovič, and Eric Vigoda. Inapproximability of the partition function for the antiferromagnetic Ising and hard-core models. Combin. Probab. Comput., 25(4):500–559, 2016. doi:10.1017/S0963548315000401.
- [15] Yuanying Guan, Muqiao Huang, and Ruodu Wang. A new characterization of second-order stochastic dominance. Insur. Math. Econ., 119:261–267, 2024.
- [16] Haim Levy. Stochastic dominance and expected utility: Survey and analysis. Manag. Sci., 38(4):555–593, 1992.
- [17] Liang Li, Pinyan Lu, and Yitong Yin. Correlation decay up to uniqueness in spin systems. In Proceedings of the Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 67–84, 2013. doi:10.1137/1.9781611973105.5.
- [18] Kuikui Liu. Spectral Independence: A New Tool to Analyze Markov Chains. PhD thesis, University of Washington, 2023.
- [19] Elchanan Mossel, Dror Weitz, and Nicholas Wormald. On the hardness of sampling independent sets beyond the tree threshold. Probab. Theory Related Fields, 143(3-4):401–439, 2009.
- [20] Alfred Müller and Dietrich Stoyan. Comparison of Stochastic Models and Risks, volume 389 of Wiley Series in Probability and Statistics. John Wiley & Sons, 2002.
- [21] Han Peters and Guus Regts. On a conjecture of Sokal concerning roots of the independence polynomial. Michigan Math. J., 68(1):33–55, 2019.
- [22] Michael Rothschild and Joseph E. Stiglitz. Increasing risk: I. a definition. J. Econ. Theory, 2(3):225–243, 1970.
- [23] Allan Sly. Computational transition at the uniqueness threshold. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 287–296, 2010. doi:10.1109/FOCS.2010.34.
- [24] Allan Sly and Nike Sun. The computational hardness of counting in two-spin models on d-regular graphs. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 361–369, 2012. doi:10.1109/FOCS.2012.56.
- [25] Daniel Štefankovič and Eric Vigoda. Lecture notes on spectral independence and bases of a matroid: Local-to-global and trickle-down from a Markov chain perspective. arXiv preprint, 2023. arXiv:2307.13826.
- [26] Wenpin Tang and Fengmin Tang. The poisson binomial distribution – Old & new. Statist. Sci., 38(1):108–119, 2023.
- [27] Remco van der Hofstad and Michael Keane. An elementary proof of the hitting time theorem. Amer. Math. Monthly, 115(8):753–756, 2008. URL: http://www.jstor.org/stable/27642587.
- [28] Dror Weitz. Counting independent sets up to the tree threshold. In Proceedings of the Annual ACM Symposium on Theory of Computing (STOC), pages 140–149, 2006. doi:10.1145/1132516.1132538.
