Abstract 1 Introduction 2 Proof Overview 3 Online Decision-Making Problem 4 Site Percolation on Infinite Tree References

Improved Mixing of Critical Hardcore Model

Zongchen Chen ORCID School of Computer Science, Georgia Institute of Technology, Atlanta, GA, USA Tianhui Jiang ORCID Zhiyuan College, Shanghai Jiao Tong University, China
Abstract

The hardcore model is one of the most classic and widely studied examples of undirected graphical models. Given a graph G, the hardcore model describes a Gibbs distribution of λ-weighted independent sets of G. In the last two decades, a beautiful computational phase transition has been established at a precise threshold λc(Δ) where Δ denotes the maximum degree, where the task of sampling independent sets transitions from polynomial-time solvable to computationally intractable. We study the critical hardcore model where λ=λc(Δ) and show that the Glauber dynamics, a simple yet popular Markov chain algorithm, mixes in O~(n7.44+O(1/Δ)) time on any n-vertex graph of maximum degree Δ3, significantly improving the previous upper bound O~(n12.88+O(1/Δ)) by the recent work [3]. The core property we establish in this work is that the critical hardcore model is O(n)-spectrally independent, improving the trivial bound of n and matching the critical behavior of the Ising model. Our proof approach utilizes an online decision-making framework to study a site percolation model on the infinite (Δ1)-ary tree, which can be interesting by itself.

Keywords and phrases:
Hardcore model, Phase transition, Glauber dynamics, Spectral independence, Online decision making, Site percolation
Category:
RANDOM
Copyright and License:
[Uncaptioned image] © Zongchen Chen and Tianhui Jiang; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Random walks and Markov chains
; Mathematics of computing Markov processes
Related Version:
Full Version: https://arxiv.org/abs/2505.07515 [8]
Editors:
Alina Ene and Eshan Chattopadhyay

1 Introduction

The hardcore model is one of the most fundamental undirected graphical models that has been extensively studied in statistical physics, social science, probability theory, combinatorics, and computer science.

Given a graph G=(V,E), we let (G) denote the collection of all independent sets of G, where we recall that an independent set is a subset of vertices inducing no edges. The Gibbs distribution μG,λ associated with the hardcore model on G is parameterized by a vertex weight λ>0 called the fugacity. Each independent set σ(G) receives a probability density given by

μG,λ(σ)=λ|σ|ZG,λ,

where ZG,λ is a normalizing constant call the partition function and is defined as

ZG,λ=σ(G)λ|σ|.

Perhaps the most amazing property of the hardcore model is the phase transition phenomenon associated with it. In fact, the hardcore model was originally proposed by statistical physicists to study and understand the phase transition in systems of hardcore gas particles. Let Δ3 denote the maximum degree of the underlying graph. The tree-uniqueness threshold λc(Δ):=(Δ1)Δ1(Δ2)Δ characterizes the uniqueness of the hardcore Gibbs measure on the infinite Δ-regular tree. Furthermore, it also describes the existence of long-range correlations. Let each vertex be associated with a Bernoulli random variable, called the spin, indicating whether the vertex is occupied (i.e., included in the independent set) or unoccupied (i.e., not included in the independent set). Then, for small fugacity λλc(Δ) the configuration at distance from the root has a vanishing influence on the root as tends to infinity, while for large fugacity λ>λc(Δ) the correlation is always bounded away from zero.

In the past two decades, a beautiful computational phase transition has been fully established for the problem of sampling from the hardcore model on graphs of maximum degree Δ, precisely around the uniqueness threshold λc(Δ). For λ<λc(Δ), there exist deterministic approximate counting algorithms for estimating the partition function [28, 2, 21], which in turn gives approximate samplers via standard reduction. Meanwhile, for λ>λc(Δ), no polynomial-time approximate counting and sampling algorithms exist assuming RPNP [23, 24, 14].

While all deterministic approximate counting algorithms run in polynomial time, they suffer from a pretty slow runtime. For example, Weitz’s algorithm [28] runs in time nO(1δlogΔ) where Δ denotes the maximum degree and δ(0,1) the slackness of the fugacity (i.e., λ=(1δ)λc(Δ)). In practice, Markov chain Monte Carlo (MCMC) algorithms provide a simpler and significantly faster method for generating random samples from high-dimensional distributions, including the hardcore model studied in this work. Among them, the Glauber dynamics (also known as the Gibbs sampler) is one of the most important and popular examples. The Glauber dynamics performs a random walk in the space (G) of independent sets and, in each step, either stays the same or moves to an adjacent set whose Hamming distance to the current set is 1. More specifically, from the current independent set σt(G), the algorithm picks a vertex vV uniformly at random and updates its spin: Let S=σt{v}; if S{v}(G) then set σt+1=S=σt; otherwise, set σt+1=S{v} with probability λ/(1+λ) and, mutually exclusively, set σt+1=S with probability 1/(1+λ).

Let PGD denote the transition matrix of the Glauber dynamics. From basic Markov chain theories it is easy to show that the Glauber dynamics PGD is irreducible, aperiodic, and reversible with respect to the Gibbs distribution μG,λ, which is the unique stationary distribution (i.e., μG,λPGD=μG,λ). The mixing time of the Glauber dynamics is defined as

Tmix(PGD)=maxσ0(G)mint0{dTV(PGDt(σ0,),μG,λ)14},

where σ0 is the initial independent set, PGDt(σ0,) is the distribution of the chain after t steps when starting from σ0, and dTV(,) denotes the total variation distance.

In the past years, exciting progress has been made in understanding the mixing time of Glauber dynamics for the hardcore model. Anari, Liu, and Oveis Gharan introduced a highly powerful technique known as spectral independence [1], leading to significant advancements in this area, including resolutions to major open problems regarding mixing properties. We refer to [18, 25] for a thorough introduction of this technique. In the subcritical regime (i.e., λ<λc(Δ)), the mixing time of the Glauber dynamics was shown to be nearly linear O(nlogn) [1, 9, 5, 7]. Meanwhile, it was long known that in the supercritical regime (i.e., λ>λc(Δ)), the mixing time could be exponentially large exp(Ω(n)) as witnessed by random Δ-regular bipartite graphs [19].

In a very recent work [3], the mixing property is further investigated at the critical point (i.e., λ=λc(Δ)). For the upper bound, the mixing time of Glauber dynamics is O~(n2+4e+O(1/Δ)) on any n-vertex graph of maximum degree Δ. For the lower bound, there exists an infinite sequence of graphs such that the mixing time is Ω(n4/3), which is, in particular, super-linear.

In this work, we present an improved mixing time upper bound for the Glauber dynamics on the critical hardcore model.

Theorem 1.

Let α0 be a constant. For any n-vertex graph G=(V,E) of maximum degree Δ3, the Glauber dynamics for the hardcore model on G with fugacity λ(1+αn)λc(Δ) satisfies

Tmix(PGD)=Oα(n2+2e+2eΔ2logΔ).

Our upper bound scales as O~(n7.44+O(1/Δ)), significantly improving the O~(n12.88+O(1/Δ)) mixing time established in [3].

Similar to [3], Theorem 1 is established via the spectral independence framework. Our main contribution is to show that the critical hardcore model satisfies spectral independence of order O(n), improving the trivial bound of n used in [3]. We show this new spectral independence result in a novel way by studying an online decision-making problem, which allows us to understand a site percolation model on the infinite tree, from which spectral independence readily follows. We provide an overview of the necessary background and known results on spectral independence, as well as our new contribution and proof approach in Section 2.

2 Proof Overview

2.1 Notations and definitions

Denote the set of non-negative integers by 0, and the set of positive integers by +. For any integers a,b, define ab by the minimum of a and b, i.e, ab:=min{a,b}.

Let Ber(p) denote the Bernoulli distribution with success probability p[0,1]. Let Bin(n,p) denote the binomial distribution with number of trails n+ and success probability p[0,1]. Let dTV(,) denote the total variation distance. For any random variables X,Y, let X=𝑑Y denote that X and Y are equal in distribution.

Let G=(V,E) be a graph. For any SV, let S denote the set of neighbors of S in G, i.e., S={vVSuS,{u,v}E}; and let G[S] denote the subgraph induced in G by S, i.e., the graph with vertex set S and edge set consisting of all edges of G that have both endpoints in S.

Let T=(V,E) be a tree rooted at r. For every vertex vV, let Tv denote the subtree of T rooted at v that consists of all descendants of v; in particular, Tr=T. For any vV, let L(v) denote the set of children of v in T.

We end this subsection by defining the (t-fold) convolution of distributions on .

Definition 2 ((t-fold) Convolution).

Let μ,ν be two distributions on . Define a new distribution μν on by

μν(k)=i=+μ(i)ν(ki),k.

We call μν the convolution of μ and ν. Define μt where t+ inductively by μ1=μ and μt=μ(t1)μ for t2. We call μt the t-fold convolution of μ with itself.

2.2 Spectral independence via coupling on trees

The core result of this work is to establish the O(n)-spectral independence for the critical hardcore model, from which Theorem 1 readily follows by sophisticated spectral independence techniques that have been developed in a recent line of works.

The following notion of influences is needed to define the meaning of spectral independence.

Definition 3 (Influence, [1]).

Let μ be a distribution over {0,1}n. For any i,j[n] such that Prμ[σi=0]>0 and Prμ[σi=1]>0, define the (pairwise) influence from i to j as

Ψμ(i,j):=Prσμ[σj=1σi=1]Prσμ[σj=1σi=0].

In the setting of the hardcore model, the influences describe the correlation between two vertices, represented as Bernoulli random variables indicating whether the vertices are occupied. Roughly speaking, the influence of one vertex on the other represents the difference of the marginal distribution on the second vertex when flipping the first vertex from occupied to unoccupied.

Theorem 4 (O(n)-Spectral independence of critical hardcore model).

Let α0 be a constant. Consider the hardcore model on an n-vertex graph G=(V,E) of maximum degree Δ3 with fugacity λ(1+αn)λc(Δ). Then, for any vertex uV, we have

vV|ΨμG,λ(u,v)|C0n,

where C0=C0(α)>0 is a constant depending only on α.

Theorem 4 states that the hardcore model in the regime of interest satisfies spectral independence with constant O(n) (see the full version of the paper [8] and also [13]). An analogous result for the Ising model was previously shown in [3].

Many methods have been introduced to establish the spectral independence property for various families of distributions. Here we adopt the coupling independence approach introduced in [6] and apply it on a related tree, known as the self-avoiding walk tree [28]. The formal definition and construction of this tree are omitted in this paper as we only need its existence, and we refer interested readers to the works [28, 10].

We are interested in coupling two hardcore models on this tree, where in one copy the root is fixed to be occupied while in the other it is fixed to be unoccupied. As we shall see soon in Proposition 5, controlling the number of discrepancies between these two copies under a simple coupling procedure enables us to deduce spectral independence. To formally describe this coupling, we first need a few definitions. Let T=(V,E) be a tree rooted at r of maximum degree at most Δ. Consider the hardcore model on T with fugacity λ>0. For each vertex v, let pv denote the probability that v is occupied in the hardcore model on the subtree Tv, i.e., pv:=PrμTv,λ[σv=1], where we recall that Tv is the subtree of T rooted at v that consists of all descendants of v.

We now describe a natural vertex-by-vertex greedy coupling for the hardcore model on T when the spin at root r is flipped.

  • Initialization: Xr=1 and Yr=0;

  • While there exists vV whose parent u has already been revealed in X and Y:

    • If Xu=Yu, then couple the whole subtree Tv perfectly, i.e., XTv=YTv;

    • If Xu=1 and Yu=0, then set Xv=0 and sample YvBer(pv);

    • If Xu=0 and Yu=1, then sample XvBer(pv) and set Yv=0;

  • Return (X,Y).

It is straightforward to check that when the greedy coupling ends, X is an independent set distributed as PrμT,λ[σr=1] and Y as PrμT,λ[σr=0].

Proposition 5 (Coupling on trees implies spectral independence, [4, Lemma 39], [6, Proposition 4.3]).

Consider the hardcore model on an n-vertex graph G=(V,E) of maximum degree Δ3 with fugacity λ>0. For any uV, there exists a tree T=TSAW(G,u) rooted at r with maximum degree at most Δ, such that if (X,Y)𝒞 is the greedy coupling on T, then it holds

vV|ΨμG,λ(u,v)|𝔼(X,Y)𝒞[|XY|n],

where XY denotes the symmetric difference of two sets X,Y, and recall that |XY|n:=min{|XY|,n}.

Hence, to establish Theorem 4, it suffices to bound the expected number of disagreements in the greedy coupling for the hardcore model on trees when the spins at the root are distinct.

2.3 Coupling on trees via site percolation

From the greedy coupling procedure above, we observe a natural site percolation on the tree T describing the appearance of disagreements. Every vertex v is open with probability pv independently, representing the occurrence of a disagreement at v, that is, XvYv; otherwise, the vertex v is closed. The root r is always open, i.e., pr=1, since XrYr. Our goal is to bound the size of the open cluster containing the root, consisting of all open vertices that are connected to the root via a path of open vertices.

We now introduce some notations for the site percolation model. Let T=(V,E) be a tree rooted at r. For any vV, let pv be the probability that v is open, and we call pv the occupation probability of v. For simplicity, we assume pr=1. Let P={pv}vV be the list of occupation probabilities for all vertices, and call it the occupation probability list of the site percolation model. Finally, for the site percolation on T with occupation probability list P, let N(T,P) denote the random variable representing the size of the open cluster containing the root.

From the construction of the greedy coupling and the site percolation above, we see that |XY|=𝑑N(T,P), where pv=PrμTv,λ[σv=1] for each vV{r}; see the full version of the paper [8] for a formal statement. Thus, the problem is reduced to studying a site percolation model.

In order to study this site percolation model, we need to know the conditions satisfied by the occupation probabilities {pv}. Since these probabilities correspond to the marginal probability of the roots in the respective subtrees, they satisfy a well-known recurrence called the tree recursion (see Fact 21). In this work, we present a new inequality satisfied by these marginal probabilities, which is crucial in obtaining our main spectral independence result. In particular, Equation 1 below provides a stronger and simpler contraction property of the tree recursion, which was not known in the literature as far as we are aware. For simplicity, here we consider only the exact critical fugacity λ=λc(Δ).

Lemma 6 (Special case of Lemma 20; see also [17]).

Let T=(V,E) be a tree rooted at r with maximum degree at most Δ. Consider the critical hardcore model on T with fugacity λ=λc(Δ). For each vertex vV, let pv denote the probability that v is occupied in the critical hardcore model on the subtree Tv rooted at v, i.e., pv:=PrμTv,λ[σv=1]. Then, for every non-root vertex vV{r}, it holds

pvwL(v)pw1Δ1, (1)

where we recall that L(v) denotes the set of children of v.

 Remark 7.

In [17], a potential function approach was applied to study the contraction of the tree recursion for the subcritical hardcore model. Uncovering their result ([17, Lemma 12]), the corresponding condition they established at criticality can be stated as

pvwL(v)pw1. (2)

Our bound Equation 1 is stronger than Equation 2 since |L(v)|Δ1. In fact, going through the proof in [17], one is able to recover the stronger inequality Equation 1 though it was not stated explicitly. In this work, we provide a simpler proof of Equation 1 (and hence Equation 2). To establish the O(n)-spectral independence at criticality, we do require the stronger inequality Equation 1, while the weaker Equation 2 is not sufficient.

We are now ready to state our main percolation result, from which spectral independence Theorem 4 readily follows.

Theorem 8 (Main result for site percolation).

Consider the site percolation model on the infinite d-ary tree 𝕋dary=(V,E) rooted at r with occupation probability list P={pv}vV where pr=1. Let N=N(𝕋dary,P) be the size of the open cluster containing the root. Suppose the following hold:

  1. 1.

    For every non-root vertex vV{r}, it holds

    pvwL(v)pw1d; (3)
  2. 2.

    At the root r, it holds

    wL(r)pwc, (4)

    where c>0 is a constant.

Then, we have that for any n0,

𝔼[Nn]=O(n),

where the constant in big-O depends only on c.

At this point, we transfer our original problem of bounding the mixing time and proving critical spectral independence into a problem of site percolation on trees. In [3], the same strategy was applied to the critical Ising model, another canonical example of graphical models. There, the occupation probabilities are much simpler; in fact, they can all be set to pv=1/d, which is a uniform upper bound on the pairwise influence on v from its parent. We remark that the main distinction of [3] and our work lies in the application of Lemma 6 (corresponding to the condition Equation 3) which substantially extends the uniform setting, pv=1/d for all v, appeared in the critical Ising model.

2.4 Site percolation on trees via online decision making

The main part of the paper aims to prove Theorem 8. We introduce a novel approach to study such a site percolation model on the infinite tree, by considering an online decision-making game.

Our strategy is to upper bound 𝔼[Nn] by understanding the worst-case choice of occupation probabilities {pv}. We do this by changing the perspective and thinking as an adversary who is allowed to pick the occupation probabilities. Namely, we study an online decision-making problem where a player is allowed to pick the occupation probabilities {pv} every time we need to reveal the status (open or closed) of a few vertices. These probabilities can be arbitrary as long as they satisfy the requirements Equation 3, and the goal of the player is to maximize 𝔼[Nn].

To be more specific, let At denote the number of active vertices at time t, where a vertex is said to be active if (i) it is open; (ii) there is an open path from it to the root; (iii) the status of its children has not been fully revealed. At the beginning, A0=1 since only the root is open and we have not revealed any other vertex. Then, the player picks the occupation probabilities for both the children and the grandchildren of r; these probabilities are required to satisfy Equation 3. By sampling from the corresponding Bernoulli distributions independently, we reveal the number of grandchildren of r, denoted as X1, that are active (i.e., open and connected to r). With r being deactivated, the number of active vertices becomes A1=A01+X1=X1. The game is then repeated. Whenever there exists an active vertex v at round t, the player picks occupation probabilities of the children and grandchildren of v satisfying Equation 3, and, after the number Xt of open grandchildren connected to v is revealed, updates At=At11+Xt. This process stops when there are no active vertices, i.e., when At=0.

We remark that we consider only active vertices at even levels because Equation 3 imposes requirements on two adjacent levels.

Suppose the game stops after k rounds. Then, the number of open vertices connected to the root at even levels is precisely k, since every such vertex becomes active at some point and is deactivated at some other time. Therefore, controlling the number of rounds played in the game allows us to bound 𝔼[Nen] where Ne is the number of vertices at even levels that are in the open cluster containing the root in the site percolation. (Since 𝔼[Nen]=m=1nPr[Nem], in the actual proof we aim to bound Pr[Nem] for every m for simplicity. And we can bound Pr[Nem] by the maximum probability that the game lasts for at least m rounds.) Finally, combining the upper bound of 𝔼[Nen] with Equation 4, it is then not hard to upper bound 𝔼[Nn] as wanted.

Therefore, our goal is to determine the optimal strategy of the player when such an online decision-making game is played. We deal with this in Section 3 where we state and prove our main result, Theorem 19. While a natural guess of the optimal strategy is to set every occupation probability pv to be 1/d so that Equation 3 is satisfied, we show that the optimal strategy of the player should set pvpw=1/d for one child wL(v) of v and set pw=0 for all other children wL(v). We remark that such choices of {pv} are not realizable as marginal probabilities in the hardcore model since they do not satisfy the tree recursion and are way too large (in reality, pv=O(λ)=O(1/d) at criticality); however, they are sufficient to provide meaningful upper bounds on 𝔼[Nn] as we need.

We present the proof of Theorem 8 about site percolation on the infinite tree in Section 4, utilizing Theorem 19. Finally, we deduce spectral independence (Theorem 4) and rapid mixing (Theorem 1) from Theorem 8; see the full version of the paper [8] for details.

3 Online Decision-Making Problem

In this section, we introduce an online decision-making problem that serves as a key tool in proving our main result for site percolation (Theorem 8), and show the optimal strategy as our main result for this section.

First of all, we describe the setup of the online decision-making game in Section 3.1. Then, we define a partial order called second-order stochastic dominance, which plays an important role in finding the optimal strategy, and show some basic properties in Section 3.2. After that, we introduce the Poisson binomial distribution in Section 3.3 and some properties of the random walk hitting time in Section 3.4, which are crucial in the proof of the optimal strategy. Finally, in Section 3.5, we state and prove our main result for the online decision-making game and show the optimal strategy of the game.

3.1 Setup of online decision making

We now describe precisely an online decision-making game under a slightly more general setting.

In the online decision-making game, the player maintains some number of tokens (corresponding to active vertices). Let At denote the number of tokens the player owns after round t. At the beginning, the player has A0=a tokens. There is a family 𝒫 of distributions on non-negative integers 0 (corresponding to choices of occupation probabilities). For round t, the player spends one token, assuming At11, and chooses a distribution πt𝒫. Then, a sample Xtπt is generated independently, and the player receives Xt tokens as a reward. The number of tokens the player owns then becomes At=At11+Xt. The game ends when the player uses up all the tokens, i.e., the first time At=0. We denote this stopping time by τ. Note that it is possible the game never stops, in which case τ=. The goal of the player is to survive for m rounds for some given integer m0. That is, the player wins if and only if τm. We present the process of the online decision-making game in Algorithm 1.

Algorithm 1 Online decision-making game.

Define a strategy SS for the player by a mapping SS:+×+𝒫. For any k,a+, SS(k,a) is defined by the distribution the player will choose when they needs to survive for k more rounds to win and currently has a tokens. For example, if the winning requirement is m rounds, then at the beginning of round t, the player will choose the distribution πt=SS(mt+1,At1) assuming At11.

For m,a0, define the winning probability under strategy SS to be

φmSS(a):=PrSS[τmA0=a].

Namely, φmSS(a) is the probability that the game lasts for at least m rounds, assuming the player has a tokens at the beginning and uses strategy SS. We further define

φm(a):=supSSφmSS(a).

The following lemma establishes a simple recursive formula for the optimal winning probabilities and the existence of an optimal strategy.

Lemma 9.

Let 𝒫 be a family of distributions on 0. Suppose the metric space (𝒫,dTV) is compact, where we recall dTV is the total variation distance. Then the following holds:

  1. 1.

    For all m,a+,

    φm(a)=maxπ𝒫𝔼Xπ[φm1(a1+X)];
  2. 2.

    There exists a strategy SS:+×+𝒫 such that φm(a)=φmSS(a) holds for any m,a+.

Proof.

We note that φmSS(a) is well-defined if SS(k,) is defined for all 1km, and the result of SS(k,)(k>m) does not matter. In our proof, when we write φmSS(a), we only guarantee that for all 1km, SS(k,) is defined.

We verify the recurrence and define the strategy inductively on m.

For the base case m=1, by definition, φ1(a)=𝟙[a1], and φ0(a)=1 for all a0. Then Item 1 immediately follows. Furthermore, we can define SS(1,):=π0 so that φ1(a)=φ1SS(a) holds for any a+, where π0𝒫 is an arbitrary distribution.

Now suppose m2. Suppose Items 1 and 2 hold for m1. Let a+. Suppose the player chooses π𝒫 in the first round and obtains Xπ tokens; hence, after the first round, the player has A1=A01+X=a1+X tokens. Then to maximize the winning probability, i.e., to maximize the probability that survive for at least m1 rounds when having a1+X tokens initally, the player should use strategy SS(k,) (k=1,,m1), and then the winning probability is φm1(a1+X) by the induction hypothesis φm1(a1+X)=φm1SS(a1+X). Therefore, if the player chooses π𝒫 in the first round, the maximum winning probability for the player is 𝔼Xπ[φm1(a1+X)]. Hence, to obtain the maximum winning probability, we need to choose an optimal π which maximizes 𝔼Xπ[φm1(a1+X)]. By compactness, such optimal π exists, therefore Item 1 for m holds, and we can define SS(m,) by SS(m,a):=argmaxπ𝒫𝔼Xπ[φm1(a1+X)] for all a+. Then it is straightforward that φm(a)=φmSS(a) holds for any a+.

3.2 Second-order stochastic dominance

Our goal is to find an optimal strategy for the online decision-making game. A first thought that one may consider is to use first-order stochastic dominance (also simply called stochastic dominance). For any two distributions μ,ν over 0, μ is (first-order) stochastically dominated by ν if and only if Prμ[Xi]Prν[Yi] for all i0. An immediate corollary is that μ is stochastically dominated by ν implies 𝔼μ[X]𝔼ν[Y]. If there is a largest distribution under stochastic dominance in 𝒫, then it can be easily proved that the player can attain the maximum winning probability when always picking the largest distribution. However, the largest distribution under stochastic dominance does not exist in our case, since any two different distributions with the same mean are not comparable under stochastic dominance. In fact, in our application, there are infinitely many distributions attaining the largest mean in 𝒫. Therefore, there is no largest distribution under stochastic dominance and we need something else.

It turns out that second-order stochastic dominance (see, e.g., [16, 15] for more background and applications) can address the problem of the lack of a largest distribution. However, as a trade-off, it is not always true that the player can achieve the maximum winning probability when always choosing the largest distribution under second-order stochastic dominance. Nonetheless, if the largest distribution satisfies some nice properties (see Theorem 19 for details), this is indeed true.

We now define second-order stochastic dominance.

Definition 10 (Second-order stochastic dominance).

We define a partial order “(2)” called second-order stochastic dominance on the family of distributions on 0 with finite expectations. For two distributions μ,ν on 0 with finite expectations, μ(2)ν if and only if

𝔼Xμ[Xi]𝔼Yν[Yi],i+.

The following proposition shows some classical equivalent definitions of second-order stochastic dominance.

Proposition 11 (Equivalent definitions of second-order stochastic dominance [20, Theorem 8.1.1], see also [22]).

Let μ,ν be distributions on 0 with finite expectations. The following definitions are equivalent:

  1. 1.

    μ(2)ν;

  2. 2.

    (Increasing concave order) For any non-decreasing concave function f, it holds:

    𝔼Xμ[f(X)]𝔼Xν[f(X)];
  3. 3.

    There exists a coupling 𝒞 of μ and ν such that

    𝔼(X,Y)𝒞[XYY=i]0,i0.

The following two lemmas offer easy ways to find an “upper bound” of a given distribution in the sense of “(2)”, and are helpful to us in identifying the largest distribution in 𝒫.

Lemma 12.

Let μ be a distribution on 0 with expectation 𝔼μ[X]γ1. Then μ(2)Ber(γ).

Proof.

For any i+, we have

𝔼Xμ[Xi]𝔼Xμ[X]γ=𝔼YBer(γ)[Yi],

which implies μ(2)Ber(γ).

Lemma 13.

If μ1(2)ν1 and μ2(2)ν2, then μ1μ2(2)ν1ν2, where the operator “*” represents convolution (see Definition 2).

Proof.

Let X1,X2 be independent random variables with distributions μ1,μ2 respectively. Let Y1,Y2 be independent random variables with distributions ν1,ν2 respectively. We also assume X1,Y2 are independent and X2,Y1 are independent. For i=1,2, by μi(2)νi and Proposition 11, there exists a coupling 𝒞i of (Xi,Yi), such that

𝔼(Xi,Yi)𝒞i[XiYiYi=j]0,j0. (5)

Let X=X1+X2,Y=Y1+Y2. Then μ1μ2 is the law of X, and ν1ν2 is the law of Y. Let 𝒞 be the joint distribution of (X,Y) assuming (Xi,Yi)𝒞i(i=1,2). It is clear that 𝒞 is a coupling of μ1μ2 and ν1ν2. For any k0,

𝔼(X,Y)𝒞[XY|Y=k] =i=12𝔼(X1,Y1)𝒞1,(X2,Y2)𝒞2[XiYi|Y=k]
=i=12j=0kPr(X1,Y1)𝒞1,(X2,Y2)𝒞2[Yi=j|Y=k]𝔼(Xi,Yi)𝒞i[XiYi|Yi=j]
0,

where the last inequality follows from Equation 5. Then Lemma 13 follows from Proposition 11.

3.3 Poisson binomial distribution

As hinted by Item 2 of Proposition 11, in order to apply second-order stochastic dominance, we hope to have some non-decreasing concave functions as the objective/utility function in our decision making. It turns out that when the largest distribution (under “(2)”) is a Poisson binomial distribution (see, e.g., [26] for a thorough introduction) with expectation at least 1, the maximum winning probability φm(a) is non-decreasing and concave with respect to a.

We first define the Poisson binomial distribution.

Definition 14 (Poisson binomial distributions (random variables)).

We call a random variable X a Poisson binomial random variable if it can be expressed as a finite sum of independent Bernoulli random variables, i.e., X=i=1Xi where +, XiBer(pi) are independent. We call a distribution a Poisson binomial distribution if it is the distribution of a Poisson binomial random variable.

An immediate property is the following.

Fact 15.

If X and Y are two independent Poisson binomial random variables, then X+Y is also a Poisson binomial random variable.

The crucial property that guarantees the concavity of the maximum winning probability is that a Poisson binomial random variable is unimodal with a mode near the mean of the random variable. We next define unimodality and state the property.

Definition 16 (Unimodality [12]).

Let π be a distribution on . Let m be an integer. The distribution π is called unimodal about m if

π(i)π(i1),imandπ(i)π(i+1),im,

and we call m a mode of π.

Let Z be a random variable with distribution π. The random variable Z is called unimodal about m if π is unimodal about m.

Proposition 17 (Darroch’s rule for the mode [11]).

Let Z be a Poisson binomial random variable. Let q be the expectation of Z. Then there exists m{q,q+1} if q, or m=q if q, such that Z is unimodal about m. In particular,

Pr[Z=i]Pr[Z=i1],iqandPr[Z=i]Pr[Z=i+1],iq.

3.4 Random walk hitting time

The following proposition of the random walk hitting time will be used in deriving the formula of the maximum winning probability and proving the concavity of the maximum winning probability.

Proposition 18.

Let {Wt}t=0 be a random walk on . Specifically, Wt=i=1tYi, where Yi are independent and identically distributed integer-valued random variables satisfying Pr[Y11]=1 (left-continuous). For any a+, define the hitting time τa:=min{t0Wt=a}, with the convention that τa:= if Wta for all t0. Then the following holds:

  1. 1.

    ([27]) For every m,a+,

    Pr[τa=m]=amPr[Wm=a];
  2. 2.

    For any a+,

    Pr[τa=]=1(1Pr[τ1=])aaPr[τ1=].

We refer the readers to [27] for the proof of Item 1, and the full version of the paper [8] for the proof of Item 2.

3.5 Determining optimal strategy

In this subsection, we show our main result for the online decision-making game. The following theorem implies that if the largest distribution π (under “(2)”) of 𝒫 exists and is a Poisson binomial distribution with expectation at least 1, then an optimal strategy SS for the player is SSπ, in other words, the player can achieve the maximum winning probability when always choosing the largest distribution.

Theorem 19 (Main result for online decision making).

Let 𝒫 be a family of distributions on 0. Suppose the metric space (𝒫,dTV) is compact. Suppose there exists a largest distribution in 𝒫 under the partial order “(2)”, denoted by π. Furthermore, suppose π is a Poisson binomial distribution with expectation at least 1. Then the following holds:

  1. 1.

    (Recurrence) For any m,a+,

    φm(a)=𝔼Xπ[φm1(a1+X)];

    Namely, the player will pick π to achieve the maximum winning probability, and the optimal strategy is SS(m,a)=π for all m,a+.

  2. 2.

    (Formula) Let Z1,Z2, be independent and identically distributed random variables with distribution π, and let St=i=1t(Zi1). For any a+, define τa:=min{t0St=a}, with the convention that τa:= if Sta for all t0. For any m,a+, it holds that

    φm(a)=Pr[τam]=tmatPr[St=a]+Pr[τa=];
  3. 3.

    (Concavity) For any m+, the function φm is concave:

    2φm(a)φm(a1)+φm(a+1),a+.

We will prove by induction on m. The recurrence follows from the concavity of the maximum winning probability and Proposition 11 about second-order stochastic dominance. Given the recurrence, we obtain the formula using Proposition 18 about the random walk hitting time. Finally, we derive concavity using the formula and the fact that π is a Poisson binomial distribution and hence satisfies unimodality with mode near the mean.

Proof.

We prove by induction on m.

Base case: m=1.

By definition, φ1(a)=𝟙[a1]. It is easy to check that Items 1 and 3 hold for m=1. For Item 2, the first equality holds trivially for m=1, the proof for the second equality for m=1 is the same as that for arbitrary m2, i.e., applying Item 1 of Proposition 18, which we will show in the inductive step below.

Inductive step.

For any m2, suppose Items 1, 2, and 3 hold for m1, and we aim to prove Items 1, 2, and 3 hold for m.

1. Recurrence for m

By Lemma 9, it holds that

φm(a)=maxπ𝒫𝔼Xπ[φm1(a1+X)]. (6)

By definition, it is clear that φm1(a1+X) is a non-decreasing function with respect to X. By Concavity for m1, φm1(a1+X) is a concave function with respect to X. By assumption, π(2)π holds for any π𝒫. For any π𝒫, applying the equivalence between Item 1 and Item 2 of Proposition 11 with μ=π,ν=π and f(X)=φm1(a1+X), it holds that

𝔼Xπ[φm1(a1+X)]𝔼Xπ[φm1(a1+X)]. (7)

Equation 6 and Equation 7 imply Recurrence for m.

2. Formula for m

We first explain why φm(a)=Pr[τam] is true intuitively. Recall that Z1,Z2, are independent and identically distributed random variables with distribution π, representing the player choosing the largest distribution π every time. Then St=i=1t(Zi1) represents the net income of tokens after round t when the player chooses the largest distribution π every time. Then, τa is the first time that the net income of tokens is a, i.e., the time the game stops if the player initially gets a tokens. Therefore, Pr[τam] is the probability that the game lasts for at least m rounds when the player chooses the largest distribution π every time. Then φm(a)=Pr[τam] follows from the fact that maximum winning probability can be obtained from choosing π every time, where the fact follows from Recurrence for 1,2,,m.

We next prove φm(a)=Pr[τam] formally from the induction hypothesis. For any a+, we have that

φm(a) =𝔼Xπ[φm1(a1+X)] (Recurrence for m)
=i=0π(i)Pr[τ(a1+i)m1] (Formula for m1)
=i=0Pr[Z1=i]Pr[τamZ1=i]
=Pr[τam],

as desired.

For the the second equality of Formula for m, we apply Item 1 of Proposition 18 with Yt=Zt1 and Wt=i=1tYi=i=1t(Zi1)=St. Then it follows

Pr[τam] =t=mPr[τa=t]+Pr[τa=]
=t=matPr[St=a]+Pr[τa=],

as desired.

3. Concavity for m

For any a2, it holds that

2φm(a) =𝔼Xπ[2φm1(a1+X)] (Recurrence for m)
𝔼Xπ[φm1(a2+X)+φm1(a+X)] (Concavity for m1)
=𝔼Xπ[φm1(a2+X)]+𝔼Xπ[φm1(a+X)]
=φm(a1)+φm(a+1). (Recurrence for m)

It remains to prove the case a=1, i.e.,

2φm(1)φm(2).

By Formula for m,

φm(1) =t=m1tPr[St=1]+Pr[τ1=]
=t=m1tPr[i=1tZi=t1]+Pr[τ1=],
φm(2)=t=m2tPr[i=1tZi=t2]+Pr[τ2=].

It suffices to prove

Pr[i=1tZi=t1]Pr[i=1tZi=t2],t+, (8)

and

2Pr[τ1=]Pr[τ2=]. (9)

For any t+, since π is a Poisson binomial distribution, by Fact 15, i=1tZi is a Poisson binomial random variable. By Proposition 17, we have

Pr[i=1tZi=j]Pr[i=1tZi=j1]

for any jq, where q=𝔼[i=1tZi]=t𝔼[Z1]t, which implies Equation 8.

Applying Item 2 of Proposition 18 with a=2 yields Equation 9.

4 Site Percolation on Infinite Tree

In this section, we aim to prove the main result for the site percolation model Theorem 8 and show that it can be applied to the critical hardcore model.

In Section 4.1, we prove Lemma 6 about a contraction property of the tree recursion of the hardcore model, which indicates that the main result for the site percolation model Theorem 8 can be applied to the critical hardcore model. In Section 4.2, we prove the main result for the site percolation model Theorem 8 by applying the online decision-making game introduced in Section 3.

4.1 Contraction of tree recursion: Proof of Lemma 6

In this subsection, we prove a general version of Lemma 6. In the general version, we extend the domain of the fugacity λ from at most the critical fugacity to at most (1+ε) times the critical fugacity. And the original version (Lemma 6) can be obtained simply by letting ε=0.

Lemma 20 (General version of Lemma 6).

Let T=(V,E) be a tree rooted at r with maximum degree at most Δ. Let ε0 be a constant. Consider the hardcore model on T with fugacity λ(1+ε)λc(Δ). For each vertex vV, let pv denote the probability that v is occupied in the hardcore model with fugacity λ on the subtree Tv rooted at v, i.e., pv:=PrμTv,λ[σv=1]. Then, for every non-root vertex vV{r}, it holds

pvwL(v)pw1+eεΔ1.

The following well-known fact gives a natural recurrence of the probability of the root being occupied in the hardcore model of subtrees.

Fact 21 (Tree recursion, [28]).

For {pv} defined in Lemma 20, and for any vV, it holds that

pv1pv=λwL(v)(1pw).

Proof of Lemma 20.

Fix vV{r}. Let d=Δ1. Then v has at most d children, i.e., |L(v)|d.

By tree recursion Fact 21, we have

pv1pv=λwL(v)(1pw)λ(11dwL(v)pw)d=λ(1p¯)d,

where p¯=1dwL(v)pw, and the inequality follows from the AM-GM inequality and |L(v)|d. Therefore,

pvwL(v)pwλ(1p¯)d1+λ(1p¯)dwL(v)pw=dp¯λ(1p¯)d1+λ(1p¯)d.

Set

f(x)=xλ(1x)d1+λ(1x)d,x[0,1].

Since p¯[0,1], we have pvwL(v)pwdmaxx[0,1]f(x). We next bound maxx[0,1]f(x). By standard calculus calculation, we have

f(x)=λ(1x)d1(1+λ(1x)d)2(λ(1x)d+1+(1+d)(1x)d).

Set g(x)=λ(1x)d+1+(1+d)(1x)d. Since the first factor of f(x) is always non-negative, the sign of f(x) depends on the second factor, i.e., g(x). Clearly, g(x) is decreasing on [0,1], g(0)>0, g(1)<0, by the Intermediate Value Theorem, there exists a unique zero of g(x) on [0,1], denoted by x^. Then,

maxx[0,1]f(x) =f(x^)=x^(111+λ(1x^)d)
=x^(111+d1x^(1+d)) (by g(x^)=0)
=x^+x^1d.

Therefore,

pvwL(v)pwdmaxx[0,1]f(x)=(d+1)x^1.

When ε=0, λλc(Δ), it holds that

g(1d)λc(Δ)(11d)d+1+(1+d)(11d)d=0.

By g(x) is decreasing on [0,1], we have x^1d. Then,

pvwL(v)pw(d+1)x^11d,

as desired.

We next prove for ε>0. Consider f(x)=xλ(1x)d1+λ(1x)d as a function with respect to both x and λ. Let h(λ,x)=f(x)=xλ(1x)d1+λ(1x)d. Then, by the result of the case ε=0, we have h(λ,x)1d2 for all λλc(Δ),x[0,1]. For all λ>0,x[0,1],

hλ(λ,x) =x(1x)d(1+λ(1x)d)2x(1x)d1d((dx+d(1x))d+1)d+1=1d(dd+1)d+1,

where the last inequality follows from the AM-GM inequality. Then, by the Mean Value Theorem, for any λ(1+ε)λc(Δ),x[0,1], it holds that

h(λ,x) =h(λc(Δ),x)+hλ(λ,x)(λλc(Δ))1d2+1d(dd+1)d+1ελc(Δ)
=1d2(1+ε(1+1d21)d+1)1d2(1+εe1d1)1+eεd2,

where λ is a real number between λc(Δ) and λ, and the second inequality follows from 1+xex. Therefore, when λ(1+ε)λc(Δ), f(x)1+eεd2 holds for any x[0,1]. Then,

pvwL(v)pwdmaxx[0,1]f(x)1+eεd,

as desired.

4.2 Site percolation on trees: Proof of Theorem 8

In this subsection, we prove a general version of Theorem 8. In the general version, we extend the upper bound of pvwL(v)pw from 1d to 1d(1+c1n). We note the latter upper bound can be derived from Lemma 20. The original version (Theorem 8) can be obtained simply by letting c1=0,c2=c.

Theorem 22 (General version of Theorem 8).

Consider a site percolation model on the infinite d-ary tree 𝕋dary=(V,E) rooted at r with occupation probability list P={pv}vV where pr=1. Let N=N(𝕋dary,P) be the size of the open cluster containing the root. Let n+ be a positive integer. Suppose the following conditions hold:

  1. 1.

    For every non-root vertex vV{r}, it holds pvwL(v)pw1d(1+c1n), where c10 is a constant;

  2. 2.

    At the root r, it holds

    wL(r)pwc2, (10)

    where c2>0 is a constant.

Then, it holds that 𝔼[Nn]=O(n), where the constant in big-O depends only on c1,c2.

We first show an upper bound with respect to Ne, the number of vertices at even levels that are in the open cluster containing the root, and it is straightforward to prove the upper bound with respect to N when combining Equation 10.

Lemma 23.

Under the setting of Theorem 22, let Ne be the number of vertices at even levels that are in the open cluster containing the root. Then, we have

𝔼[Nen]=O(n),

where the constant in big-O depends only on c1.

Proof of Theorem 22.

Let r1,,rd be d children of r. Recall that Ne is the number of vertices at even levels that are in the open cluster containing the root. For i=1,2,,d, let Ni be the number of vertices at odd levels that are both in the open cluster containing the root and Tri, where we recall Tri denotes the subtree rooted at ri. Let Bi=𝟙[ri is open](i=1,,d). Then,

𝔼[Nn] =𝔼[(Ne+i=1dNi)n]
𝔼[Nen+i=1d(Nin)]
=𝔼[Nen]+i=1dPr[Bi=1]𝔼[Nin|Bi=1].

By Lemma 23, 𝔼[Nen]=Oc1(n). Similarly, 𝔼[Nin|Bi=1]=Oc1(n) for any 1id.

Therefore, there exists a constant K=K(c1) depending only on c1, such that

𝔼[Nn] 𝔼[Nen]+i=1dPr[Bi=1]𝔼[Nin|Bi=1]
Kn(1+i=1dPr[Bi=1])
Kn(1+c2). (Equation 10)

This shows 𝔼[Nn]=Oc1,c2(n) as desired.

We next prove Lemma 23 by considering the process of site percolation described in Section 2.4.

Proof of Lemma 23.

Recall that in Section 2.4, we introduced a process of site percolation working in rounds that reveals the status of all children and grandchildren of an active vertex in each round. We note that the process in Section 2.4 is described under an adversary setting. Here, we restate this process more precisely for a fixed site percolation model. For convenience, we call it the site percolation process.

Let At be the number of active vertices after round t. Let Ut be the set of active vertices after round t. We label all the vertices in V by 1,2,. At first, only the root is open and active, and the status of all other vertices is unrevealed. Therefore, A0=1, U0={r}. For any t+, at the beginning of round t, assuming At11, we choose the active vertex with the least label in the current set of active vertices Ut1, denoted by vt. Then reveal the status of all the children and grandchildren of vt. To be specific, for each child or grandchild of vt, denoted by u, sample BuBer(pu) independently, then let u be open if Bu=1; be closed if Bu=0. Let Qt be the set of activated grandchildren of vt. We note that if a grandchild is activated, then it is open; however, the converse does not hold. Let Xt be the number of activated grandchildren of vt, i.e., Xt=|Qt|. Then Ut=Ut1{vt}Qt, and At=At11+Xt. The process stops when there are no active vertices, i.e., the first time At=0. If the process stops after k rounds, we have Ne=k, because every vertex in the open cluster containing the root at even levels becomes active at some time and will be deactivated later in the process, and the number of rounds equals the number of deactivated vertices. If the process never stops, we have Ne=.

Let 𝒱 be the set of all possible active vertex sets of the site percolation process. For any m,a0,S𝒱 with a=|S|, define the winning probability ψm(a,S) of the size percolation process by the probability that the site percolation process will last for at least m more rounds when currently there are a active vertices and the set of active vertices is S. Then, for any m+, we have Pr[Nem]=ψm(1,{r}).

Let πt be the law of Xt. We next find a family of distributions that contains all possible πt. Let u1,,ud be d children of vt. Let Zi be the number of children of ui that are activated in round t. Then, we have

Xt=i=1dZiand𝔼[Zi]=puiwL(ui)pw1d(1+c1n).

Then, πt=ζ1ζ2ζd, where ζi is the law of Zi. Let 𝒟 be a family of distributions defined by

𝒟={μ1μd|𝔼μi[X]1d(1+c1n),i=1,d},

where μ1,,μd are distributions on 0. Then for any t+, πt𝒟, i.e., 𝒟 is a family of distributions that contains all possible πt.

The site percolation process defined above can be considered as a strategy of the online decision-making game introduced in Section 3 with 𝒫=𝒟. (We note that this strategy is slightly different from the strategy defined in Section 3.1, which we will discuss later.) This inspires us to consider the online decision-making game with 𝒫=𝒟.

We next check 𝒟 satisfies the conditions in Theorem 19. It is easy to check that the metric space (𝒟,dTV) is compact. Let γ=1d(1+c1n). We next show Bin(d,γ) is the largest distribution in 𝒟 under “(2)”. First of all, Bin(d,γ)=(Ber(γ))d𝒟, where we recall that for a distribution μ, μt is the t-fold convolution of μ with itself (see Definition 2). For any μ𝒟, we may assume μ=μ1μd, where μ1,,μd are distributions on 0 with expectations at most γ. For any 1id, since 𝔼μi[X]γ, by Lemma 12, we have μi(2)Ber(γ). By Lemma 13, we have μ=μ1μd(2)Ber(γ)Ber(γ)=(Ber(γ))d=Bin(d,γ). Therefore, Bin(d,γ) is the largest distribution in 𝒟 under “(2)”. Furthermore, it is clear that Bin(d,γ) is a Poisson binomial distribution with expectation dγ=1+c1n1.

Therefore, by Theorem 19, for the online decision-making game with 𝒫=𝒟, there is a optimal strategy SS(m,a)π:=Bin(d,γ), and the player can achieve the maximum wining probability by using this strategy, i.e., by choosing πt=π:=Bin(d,γ) for each round t+.

As we mentioned above, the site percolation process can be considered as a strategy for a player playing the online decision-making game with 𝒫=𝒟. However, the strategy of the player here is distinct from the strategies defined in Section 3.1, in the sense that the player (the site percolation process) maintains extra storage and uses external randomness (i.e., the evolution of the set of active vertices) beyond provided in the game (the target number of rounds to survive and the number of tokens) to make a decision in each round. However, it is not hard to see intuitively that an optimal strategy should not require any other information and should depend only on m, the target number of rounds to survive, and a, the current number of tokens. Namely, the strategy from the site percolation process is no better than an optimal oblivious strategy as defined in Section 3.1. We formally state this in the following claim, whose proof is similar to Lemma 9.

Claim 24.

For any m+,a0,S𝒱 with a=|S|, it holds that

ψm(a,S)φm(a),

where φm(a) is the maximum winning probability of the online decision-making game with 𝒫=𝒟 when the player needs to survive m more rounds to win and the current number of tokens is a.

Claim 24 implies Pr[Nem]=ψm(1,{r})φm(1) holds for any m+. We next bound φm(1). By Item 2 of Theorem 19, we have for any m,a+,

φm(a)=Pr[τam],

where τa is defined in the same way as in Item 2 of Theorem 19. To be specific, let Z1,Z2, be independent and identically distributed random variables with distribution π=Bin(d,γ), and let St=i=1t(Zi1), then τa is defined by τa:=min{t0St=a}, with the convention that τa:= if Sta for all t0. Then Pr[Nem]φm(1)=Pr[τ1m] holds for any m+.

Let P={pv}vV be an occupation probability list satisfying pr=1, and pv=γ for all vV{r}. It is straightforward to check that τ1=𝑑N(𝕋dary,P), where N(𝕋dary,P) is the random variable representing the size of the open cluster containing the root in the site percolation on 𝕋dary with occupation probability γ for all non-root vertices. For simplicity of notation, we write N as shorthand for N(𝕋dary,P). By [3, Lemmas 4.7 and 4.8], we have

Pr[N=]=Oc1(n12)andPr[N=]=O(32).

Then, for any m+,

Pr[Nem] Pr[τ1m]=Pr[Nm]
==mPr[N=]+Pr[N=]
K1(=m32+n12)K2(m12+n12),

where K1=K1(c1),K2=K2(c1) are constants depending only on c1. Therefore,

𝔼[Nen]=m=1nPr[Nem]m=1nK2(m12+n12)=Oc1(n),

as desired. To complete the proof of Lemma 23, we prove Claim 24 by induction.

Proof of Claim 24.

We prove by induction on m.

For the base case m=1, by definition, it is easy to check

ψ1(a,S)=φ1(a)=𝟙[a1]

holds for any a0,S𝒱 with a=|S|.

Now suppose m2 and Claim 24 holds for m1. We aim to prove Claim 24 holds for m.

For a=0, by definition, ψm(0,)=φm(0)=0. We now assume a1. For any S𝒱 with |S|=a, consider that at the beginning of some round of the site percolation process, the current set of activated vertices is S, and the process needs to last for m more rounds to win. Let X be the number of vertices activated in this round, and S be the set of active vertices after this round. Let π^ be the law of X. Then π^𝒟. After this round, the site percolation process needs to last for m1 more rounds to win, and there are a1+X active vertices, and the set of active vertices is S. Let be the law of (X,S). Then,

ψm(a,S) =𝔼(X,S)[ψm1(a1+X,S)]
𝔼(X,S)[φm1(a1+X)] (induction hypothesis)
=𝔼Xπ^[φm1(a1+X)]
supπ𝒟𝔼Xπ[φm1(a1+X)] (π^𝒟)
=φm(a). (Lemma 9)

Therefore, Claim 24 holds for m. By the induction principle, Claim 24 holds for all m+.

References

  • [1] Nima Anari, Kuikui Liu, and Shayan Oveis Gharan. Spectral independence in high-dimensional expanders and applications to the hardcore model. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 1319–1330, 2020. doi:10.1109/FOCS46700.2020.00125.
  • [2] Alexander Barvinok. Combinatorics and complexity of partition functions, volume 30 of Algorithms and Combinatorics. Springer, Cham, 2016. doi:10.1007/978-3-319-51829-9.
  • [3] Xiaoyu Chen, Zongchen Chen, Yitong Yin, and Xinyuan Zhang. Rapid mixing at the uniqueness threshold. arXiv preprint arXiv:2411.03413, 2024. doi:10.48550/arXiv.2411.03413.
  • [4] Xiaoyu Chen and Weiming Feng. Rapid mixing via coupling independence for spin systems with unbounded degree. arXiv preprint arXiv:2407.04672, 2024. doi:10.48550/arXiv.2407.04672.
  • [5] Xiaoyu Chen, Weiming Feng, Yitong Yin, and Xinyuan Zhang. Optimal mixing for two-state anti-ferromagnetic spin systems. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 588–599, 2022. doi:10.1109/FOCS54457.2022.00062.
  • [6] Xiaoyu Chen and Xinyuan Zhang. A near-linear time sampler for the ising model with external field. In Proceedings of the Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 4478–4503, 2023. doi:10.1137/1.9781611977554.CH170.
  • [7] Yuansi Chen and Ronen Eldan. Localization schemes: A framework for proving mixing bounds for Markov chains. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 110–122, 2022.
  • [8] Zongchen Chen and Tianhui Jiang. Improved mixing of critical hardcore model. Full version, 2025. arXiv:2505.07515.
  • [9] Zongchen Chen, Kuikui Liu, and Eric Vigoda. Optimal mixing of Glauber dynamics: entropy factorization via high-dimensional expansion. In Proceedings of the Annual ACM Symposium on Theory of Computing (STOC), pages 1537–1550, 2021. doi:10.1145/3406325.3451035.
  • [10] Zongchen Chen, Kuikui Liu, and Eric Vigoda. Rapid mixing of Glauber dynamics up to uniqueness via contraction. SIAM J. Comput., 52(1):196–237, 2023. doi:10.1137/20M136685X.
  • [11] John Newton Darroch. On the distribution of the number of successes in independent trials. Ann. Math. Stat., 35(3):1317–1321, 1964.
  • [12] Sudhakar Dharmadhikari and Kumar Joag-dev. Unimodality, Convexity, and Applications. Probability and Mathematical Statistics. Academic Press, San Diego, 1988.
  • [13] Weiming Feng, Heng Guo, Yitong Yin, and Chihao Zhang. Rapid mixing from spectral independence beyond the boolean domain. ACM Trans. Algorithms, 18(3):1–32, 2022. doi:10.1145/3531008.
  • [14] Andreas Galanis, Daniel Štefankovič, and Eric Vigoda. Inapproximability of the partition function for the antiferromagnetic Ising and hard-core models. Combin. Probab. Comput., 25(4):500–559, 2016. doi:10.1017/S0963548315000401.
  • [15] Yuanying Guan, Muqiao Huang, and Ruodu Wang. A new characterization of second-order stochastic dominance. Insur. Math. Econ., 119:261–267, 2024.
  • [16] Haim Levy. Stochastic dominance and expected utility: Survey and analysis. Manag. Sci., 38(4):555–593, 1992.
  • [17] Liang Li, Pinyan Lu, and Yitong Yin. Correlation decay up to uniqueness in spin systems. In Proceedings of the Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 67–84, 2013. doi:10.1137/1.9781611973105.5.
  • [18] Kuikui Liu. Spectral Independence: A New Tool to Analyze Markov Chains. PhD thesis, University of Washington, 2023.
  • [19] Elchanan Mossel, Dror Weitz, and Nicholas Wormald. On the hardness of sampling independent sets beyond the tree threshold. Probab. Theory Related Fields, 143(3-4):401–439, 2009.
  • [20] Alfred Müller and Dietrich Stoyan. Comparison of Stochastic Models and Risks, volume 389 of Wiley Series in Probability and Statistics. John Wiley & Sons, 2002.
  • [21] Han Peters and Guus Regts. On a conjecture of Sokal concerning roots of the independence polynomial. Michigan Math. J., 68(1):33–55, 2019.
  • [22] Michael Rothschild and Joseph E. Stiglitz. Increasing risk: I. a definition. J. Econ. Theory, 2(3):225–243, 1970.
  • [23] Allan Sly. Computational transition at the uniqueness threshold. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 287–296, 2010. doi:10.1109/FOCS.2010.34.
  • [24] Allan Sly and Nike Sun. The computational hardness of counting in two-spin models on d-regular graphs. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 361–369, 2012. doi:10.1109/FOCS.2012.56.
  • [25] Daniel Štefankovič and Eric Vigoda. Lecture notes on spectral independence and bases of a matroid: Local-to-global and trickle-down from a Markov chain perspective. arXiv preprint, 2023. arXiv:2307.13826.
  • [26] Wenpin Tang and Fengmin Tang. The poisson binomial distribution – Old & new. Statist. Sci., 38(1):108–119, 2023.
  • [27] Remco van der Hofstad and Michael Keane. An elementary proof of the hitting time theorem. Amer. Math. Monthly, 115(8):753–756, 2008. URL: http://www.jstor.org/stable/27642587.
  • [28] Dror Weitz. Counting independent sets up to the tree threshold. In Proceedings of the Annual ACM Symposium on Theory of Computing (STOC), pages 140–149, 2006. doi:10.1145/1132516.1132538.