Abstract 1 Introduction 2 Preliminaries 3 Overview of our analysis and technical challenges 4 Analysis References Appendix A Missing proofs Appendix B Tools

On the 𝒉-Majority Dynamics with Many Opinions

Francesco d’Amore ORCID Gran Sasso Science Institute, L’Aquila, Italy Niccolò D’Archivio ORCID INRIA, COATI, Université Côte d’Azur, Nice, France George Giakkoupis ORCID INRIA Rennes, France Emanuele Natale ORCID CNRS, I3S, COATI, Université Côte d’Azur, Nice, France
Abstract

We present the first upper bound on the convergence time to consensus of the well-known h-majority dynamics with k opinions, in the synchronous setting, for h and k that are both non-constant values. We suppose that, at the beginning of the process, there is some initial additive bias towards some plurality opinion, that is, there is an opinion that is supported by x nodes while any other opinion is supported by strictly fewer nodes. We prove that, with high probability, if the bias is ω(x) and the initial plurality opinion is supported by at least x=ω(logn) nodes, then the process converges to plurality consensus in O(logn) rounds whenever h=ω(nlogn/x). A main corollary is the following: if k=o(n/logn) and the process starts from an almost-balanced configuration with an initial bias of magnitude ω(n/k) towards the initial plurality opinion, then any function h=ω(klogn) suffices to guarantee convergence to consensus in O(logn) rounds, with high probability. Our upper bound shows that the lower bound of Ω(k/h2) rounds to reach consensus given by Becchetti et al. (2017) cannot be pushed further than Ω~(k/h). Moreover, the bias we require is asymptotically smaller than the Ω(nlogn) bias that guarantees plurality consensus in the 3-majority dynamics: in our case, the required bias is at most any (arbitrarily small) function in ω(x) for any value of k2.

Keywords and phrases:
Distributed Algorithms, Randomized Algorithms, Markov Chains, Consensus Problem, Opinion dynamics, Plurality Consensus
Copyright and License:
[Uncaptioned image] © Francesco d’Amore, Niccolò D’Archivio, George Giakkoupis, and Emanuele Natale; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Distributed algorithms
; Mathematics of computing Probabilistic algorithms ; Mathematics of computing Markov processes ; Applied computing Systems biology ; Networks Network protocols
Related Version:
Full Version: https://arxiv.org/abs/2506.20218 [17]
Acknowledgements:
The authors are grateful to Frédéric Giroire for helpful discussions regarding this project.
Funding:
This work was partially supported by the MUR (Italy) Department of Excellence 2023 - 2027 for GSSI, the AID INRIA-DGA project n°2023000872 “BioSwarm”, and the 3IA Côte d’Azur Investments in the Future project ANR-19P3IA-0002.
Editor:
Dariusz R. Kowalski

1 Introduction

In this paper, we study the convergence time of the well-known h-majority dynamics. In brief, the setting is the following: We are given a network of n agents that interact with each other in synchronous rounds. At round 0, each agent supports one out of k possible opinions in the set [k]: At round i>0, each agent v samples uniformly at random h neighbors u1,,uh (with repetition) which, in turn, send their opinions x1,,xh to v. Then, v adopts as its own opinion the mode of {x1,,xh}, breaking ties uniformly at random. Let 𝐜t(i) count the number of agents supporting opinion i after round number t. The h-majority dynamics gives rise to a random process on the system graph, which we call the h-majority process. The question is to determine the number of communication rounds that the process needs to reach consensus, i.e., a configuration where all agents agree on some opinion i[k]. Furthermore, we are interested in plurality consensus: Suppose that at the beginning of the process there is a plurality opinion, that is, an opinion i such that 𝐜0(i)>𝐜0(j) for all ji. Plurality consensus is consensus on the opinion i, and one wishes to quantify the size of the initial bias that guarantees to obtain plurality consensus with high probability.111The expression with high probability (in short, w.h.p.) refers to an event that holds with probability at least 1nc, where c>0 is any constant that is independent of the size n of the distributed system.

1.1 Our contribution

We give the first non-trivial upper bound to the consensus time of the h-majority process with k opinions for non-constant h and k. Let us denote the difference maximinji{𝐜t(i)𝐜t(j)} by Bt and refer to it as the bias of the configuration 𝐜t at time t. In a nutshell, our main result is the following.

Theorem 1.

Consider the h-majority dynamics with k opinions in the complete graph of n nodes with self-loops 222Assuming self-loops is standard in the literature, as it simplifies the analysis without affecting the possible outcomes.. Let λ1,λ2,λ3 be large enough positive constants. Assume that opinion 1 is the initial plurality opinion with B0λ1𝐜0(1) and 𝐜0(1)λ2logn. If hλ3nlogn/𝐜0(1), then, with high probability, the h-majority process reaches plurality consensus within O(logn) rounds.

Let us discuss Theorem 1 a bit more in details. First, notice that it always holds 𝐜0(1)n/k since kn/𝐜0(1). In the following, we say that a configuration is almost-balanced if there exists a subset I[k], with |I|=O(1), such that 𝐜0(i)=Θ(𝐜0(1)) for all i[k]I, assuming that opinion 1 is the plurality opinion. Notice that, in this case, it must hold that 𝐜0(1)=Θ(n/k) and 𝐜0(i)=Θ(n/k) for all i[k]I.

Case 1: 𝒌=𝑶(𝒏/𝐥𝐨𝐠𝒏).

Here we have two sub-cases depending on how we fix 𝐜0(1). In the first sub-case, we fix 𝐜0(1)=Θ(n/k) such that 𝐜0(1)λ2logn (which is always possible). Here, we must select some large enough h=Θ(nlogn/c0(1))=Θ(klogn) in order to apply Theorem 1, with the required bias that is minimum, that is, B0=Θ(n/k): observe that the configuration is necessarily almost-balanced. If, instead, 𝐜0(1)=ω(n/k), it is guaranteed that 𝐜0(1)=ω(logn). In this case, the configuration cannot be almost-balanced because the required bias is B0=Θ(𝐜0(1))=ω(n/k). Here, we can pick h=Θ(nlogn/𝐜0(1))=o(klogn).

Case 2: 𝒌=𝝎(𝒏/𝐥𝐨𝐠𝒏).

Since k=ω(n/logn) and 𝐜0(1)=Ω(logn) by hypothesis, we cannot have almost balanced configurations: we can always choose some h=o(klogn) and we will have that the minimum bias is B0=ω(n/k). Note that in this setting, h may be larger than n, which we do not exclude as the samples can contain repetitions.

Observations

Despite the assumption hλ3nlogn/𝐜0(1) is undoubtedly a strong one (that is, h needs to be at least roughly klogn for almost-balanced configurations), in Section 3 we show how the analysis for such regime of h is already quite involved and we discuss what the main difficulties to reduce such assumption are. Also, in Section 3 we will highlight why we believe that hn/𝐜0(1) is a “threshold” value, since the two settings hn/𝐜0(1) and hn/𝐜0(1) seem to need complementary approaches, which is also why our analysis cannot be adapted to the case hn/𝐜0(1). However, since h is a “static” parameter of the h-majority dynamics, in order to approach the case h<λ3nlogn/𝐜0(1), one has to show only that there is a time t>0 such that 𝐜t(1)λ3nlogn/h and that opinion 1 is still the plurality opinion with Bt that meets the hypothesis of Theorem 1, which implies that our analysis applies.

Prior to this work, the only result for the h-majority dynamics with non-constant h and k was a lower bound of Ω(k/h2) rounds to reach consensus given by [5] that holds with high probability whenever maxi[k]{𝐜0(i)}3n/(2k). Hence, Theorem 1 shows that the lower bound cannot be pushed to Ω~(k/h) (where the notation Ω~ hides polylog(n) factors), at least as long as maxi[k]{𝐜0(i)}3n/(2k) is satisfied, e.g., when (nB0)/(𝐜0(1)B0)k3n/(2𝐜0(1)), which is always realized when max{(n+(k1)B0)/k,λ2logn}𝐜0(1)3n/(2k) and k3n/(2λ2logn) for any value of B0 meeting the hypothesis of Theorem 1.

Furthermore, as we will discuss in the related works, in the context of opinion dynamics, the minimum required bias in any other opinion dynamics is always asymptotically larger than our minimum required bias. Interestingly, the h-majority dynamics can amplify a very small bias even when h=O(klogn) (this is easy to prove for very large h as we argue in Section 3, but it is not at all trivial for our parameters).

A final remark that we stress is about the message complexity. According to Theorem 1, the message complexity required to reach consensus is O(hnlogn), which can be as high as O(knlog2n) in almost-balanced configurations. By recent work on the 3-majority dynamics (see Section 1.2), the message complexity of the 3-majority dynamics is O(knlogn). We believe that the gap between these two quantities is due to the fact that the analysis of the h-majority dynamics is not yet tight.

1.2 Previous results

There is a vast body of research that investigated the h-majority dynamics when either h or k are constant values, both in the synchronous and in the asynchronous settings [5, 9, 10, 27, 13, 34]. Before exploring the results in the two settings, let us discuss the work [10]. The authors compared the power of the h-majority dynamics and that of the (h+1)-majority dynamics with k=2 opinions. They proved that, starting from the same configuration, the consensus time Th+1 of the (h+1)-majority dynamics is stochastically dominated by the consensus time Th of the h-majority dynamics in the following sense: for each t>0, Pr(Th+1t)Pr(Tht) in both the synchronous and the asynchronous settings. However, to date no such extension to the case of k>2 opinions is known.

Synchronous setting

Let us specify that all the results we mention in this subsection and the next two ones hold in the complete graph of n nodes with self-loops. The work [6] proved an upper bound of O((k2logn+klogn)(k+logn)) rounds to reach consensus that holds w.h.p., provided that knα for a suitable positive constant α<1.

In [5], the authors showed that the synchronous 3-majority dynamics with k opinions converges in O(min{k,(n/logn)13}logn) with high probability, provided that the bias of the initial configuration is at least cmin{2k,(n/logn)13}nlogn for some constant c>0. Moreover, the authors provided a lower bound of Ω(klogn) on the convergence time to consensus, w.h.p., whenever the initial configuration is sufficiently balanced, that is, maxi[k]{𝐜0(i)}n/k+(n/k)1ε for some ε>0 and k(n/logn)1/4.

The work [9] compared the synchronous 3-majority dynamics with the synchronous 2-choices dynamics. The 2-choices dynamics works as follows: At each round, each agent picks two neighbors (with repetition) u.a.r. If the two sampled neighbors support the same opinion, the agent adopts that opinion. Otherwise, the agent keeps its own opinion. The authors first proved a generic lower bound of Ω(min{k,n/logn}) rounds to reach consensus starting from the initial perfectly balanced configuration, w.h.p. Furthermore, they proved that the 3-majority dynamics works better in symmetric configurations (i.e., with no initial bias) when, e.g., maxi[k]{𝐜0(i)}=O(logn). In particular, w.h.p., the 3-majority takes time at most O(n3/4log7/8n) to reach consensus w.h.p., regardless of any other hypothesis on the initial configuration, while the 2-choices needs time Ω(n/logn) whenever maxi[k]{𝐜0(i)}=O(logn). This was the first work to notice that, for a large number of opinions, the 3-majority dynamics is polynomially (in k) faster than the 2-choices dynamics.

The work [27] improved upon [5] and showed that, for the 2-choices with k=O(n/logn) and for the 3-majority with k=O(n1/3/logn), the convergence time to consensus is O(klogn), with high probability. Notice that this upper bound is tight according to the lower bound by [5], at least as long as k(n/logn)1/4. Furthermore, the authors showed that the convergence time on the 3-majority dynamics is O(n2/3log3/2n) with high probability, regardless of the number of opinions.

Very recently, [34] settled almost tightly the complexity of both the 3-majority and the 2-choices dynamics. The authors proved that, w.h.p., the 3-majority dynamics reaches consensus in O(klogn) rounds whenever k=o(n/logn), while it takes time O(nlog2n) for other values of k. Furthermore, [34] proved that plurality consensus is ensured w.h.p. as long as the initial bias is B0=ω(nlogn). As for the 2-choices, they showed that, w.h.p., the dynamics reaches consensus in O(klogn) rounds whenever k=o(n/log2n), while it takes time O(nlog3n) otherwise. In this case, plurality consensus is ensured w.h.p. as long as the initial bias is B0=ω(𝐜0(1)logn).333This is the only example of initial bias that gets “close enough” to what we require in Theorem 1, but with an exceeding logn factor. These results almost match the generic lower bound given by [9], up to logarithmic factors.

Asynchronous setting

The only works that analyzed the 3-majority dynamics in the asynchronous setting are [13, 10]. In [10], the authors consider the binary opinion case and show that the convergence time of the asynchronous 3-majority dynamics is O(nlogn) rounds, w.h.p., and that a bias of Θ(nlogn) is sufficient to ensure plurality consensus, w.h.p. The authors of [13] showed that the convergence time is O(min{knlog2n,n3/2log3/2n}), w.h.p., no matter the number of initial opinions. They also provided a generic lower bound of Ω(min{kn,n3/2/logn}) rounds to reach consensus (starting from balanced configurations), w.h.p. The work [13] (which came before [34]) was the first to establish exactly how the linear-in-k dependence in the consensus time of the 3-majority dynamics breaks when the number of opinions grows over n, in which case the consensus time is sublinear in the number of opinions. The reader may observe that the asynchronous setting has exactly the same convergence time as the synchronous model, but for a multiplicative factor n. This is to be expected: In the asynchronous setting, in a round only one agent (sampled u.a.r.) updates its state, and hence we need roughly n rounds to activate all agents at least once (up to polylogarithmic factors). For processes with small variance (smaller w.r.t. the voter model), convergence times from the synchronous to the asynchronous setting usually scale with such a factor. Note that this is not true for processes with large variance [7].

On the 𝒉-majority dynamics

As previously discussed, [10] established the “hierarchy” of the h-majority dynamics when the number of opinions is k=2, by showing that the (h+1)-majority is “faster” than the h-majority (both in the synchronous and asynchronou settings), and [5] exhibited the lower bound of Ω(k/h2) rounds to reach consensus (that holds w.h.p.). Apart for the two aforementioned works, there is no theoretical result on the h-majority dynamics for h1, which makes it one of the open problems in consensus dynamics. [34] put the h-majority dynamics in the “Open Question” section of their recent work, and wondered whether the techniques developed in [34] applied to the h-majority dynamics as well. We do not know the answer to this question: As we will discuss in Section 3, in the h-majority the main difficulty is to compute the expectation of 𝐜t+1(i) conditional on the whole configuration 𝐜t at the previous rounds, while this is straightforward when h=3 and is at the core of every paper analyzing the 3-majority dynamics [6, 5, 27, 34, 13]. More in general, there is no non-trivial upper bound to the consensus time of the h-majority dynamics with k1 opinions. In this paper, we take a first step into the analysis of the h-majority dynamics. As we will show, our main effort basically is providing a lower bound to 𝔼[𝐜t+1(1)𝐜t+1(2)𝐜t], supposing that opinion 1 is the plurality opinion at round t, and opinion 2 is the opinion supported by the second-largest community.

Other related works

The 3-majority, the h-majority, and other majority-based dynamics have been investigated also in other settings. For example, in presence of communication noise (which corrupts the exchanged messages) or in presence of stubborn agents [18, 19], or when some opinion is preferred among the others, that is, there is some probability that an agent, instead of running the protocol, spontaneously switches to the preferred opinion with some fixed probability [14, 32].

Other opinion dynamics for the (plurality) consensus problem have been investigated both in the synchronous and asynchronous settings. One of the most prominent is the undecided-state dynamics. In the undecided-state dynamics there is an extra opinion, the undecided opinion. The update rule works as follows: A node samples one neighbor u.a.r. and pulls its opinion. If the received opinion is different from the supported one, the node becomes undecided. If the node is undecided, it just copies whatever opinion receives. The undecided-state dynamics has been studied both in the synchronous setting and in the population protocol model (asynchronous setting) [2, 12, 4, 15, 16, 1, 8, 3] and performs roughly the same as the 3-majority dynamics for a small number of opinions (while its analysis for the unconstrained general case is still open). More in general, work on opinion dynamics fits into the recent trend in the distributed computing community of drawing inspiration from social and biological systems. Other then the consensus problem, researchers have analyzed distributed search problems but also more generic problems (MIS, leader election, etc.) in extremely weak computational models (such as the beeping model or the stone-age model) [31, 23, 11, 29, 28, 20, 26, 35, 24, 22].

A special mention goes to [25]. Here, the authors considered noisy distributed systems, i.e., systems in which exchanged messages can be corrupted and changed with some positive probability, a setting that takes inspiration by biological societies, where environmental factors can disturb communication. They characterized the type of noise for which the tasks of information spreading and plurality consensus can be solved, when the number of opinions is constant.444Information spreading is defined as follows: In the distributed system there is only one node that is informed about one out k opinions, while all other nodes are not informed. The task is to design a protocol that informs all nodes as fast as possible. The protocol designed by [25] to solve plurality consensus relies on some majority-based rule, and proves a technical result that is at the heart of our analysis. We will discuss this in more details in Section 3. We nevertheless stress the fact that [25] only considers the case where the number of opinions k is a constant.

2 Preliminaries

We work on a complete graph of n nodes with self-loops. Initially, each node supports one out of k opinions in the set [k]. Time is synchronous and controlled by some global clock. At each round, in parallel, every node v samples h neighbors u.a.r. with repetition and pulls their opinions. Then, v adopts the unique majority opinion, if any, breaking ties u.a.r. among the opinions that are in the majority. For each opinion i[k], we denote by 𝐜t(i) the number of nodes supporting opinion i at time after performing the update-rule at time t. Let 𝐜t=(𝐜t(1),,𝐜t(k)) denote the configuration of the system at time t. Notice that the random variable 𝐜t defines a Markov chain: for any sequence of configurations y1,,yt[n]k, it holds that Pr(𝐜t=yti=1t1{𝐜i=yi})=Pr(𝐜t=yt𝐜t1=yt1) .

Without loss of generality, we assume that at the beginning of the process we have 𝐜0(1)𝐜0(2)𝐜0(k). At any time t, we call plurality opinion at time t the opinion i such that 𝐜t(i)𝐜t(j) for all j[k]. At any time t, the bias of a configuration 𝐜t is the quantity Bt=maxi[k]minji{𝐜t(i)𝐜t(j)}, that is, of how many nodes the community supporting the plurality opinion exceeds the community supporting the second largest one. Note that Bt=0 if there are multiple plurality opinions, while Bt>0 the plurality opinion is unique.

We work from the perspective of a single node. When a node v samples h neighbors u.a.r., we are drawing from a multinomial distribution. In particular, given a configuration 𝐜t at time t and a node v, let 𝐗(t)(v)=(X1(t)(v),,Xk(t)(v)) be a multinomial random variable where Xi(t)(v)Bin(n,pi(t)), with pi(t)=𝐜t(i)/n, counts the number of neighbors supporting opinion i that v samples at time t: it must hold that i=1kXi(t)(v)=n, so the variables are negatively correlated. Let us denote the event that v supports opinion i at the end of round t by 𝒲i(t)(v).

For the rest of the paper, we will use only p1(t),,pk(t), since these are the probabilities defining the multinomial distribution of interest. We define the normalized bias to be δt=maxi[k]minji{pi(t)pj(t)}, which is equal to Bt/n. When the round we are referring to is clear from the context, we only write p1,,pk, δ, 𝐗(v)=(X1(v),,Xk(v)), and 𝒲i(v), omitting the dependency on t, to refer to p1(t),,pk(t), δt, 𝐗(t)(v)=(X1(t)(v),,Xk(t)(v)), and 𝒲i(t)(v). Also, since in any given round t the random variables {𝐗(v)}vV are i.i.d., and the events {𝒲i(v)}vV are mutually independent, when analyzing the distribution of some 𝐗(v) or the probabilities of some 𝒲i(v), we omit the dependency on v and only write 𝐗=(X1,,Xk) and 𝒲i.

From now on, we will refer to the normalized bias δt simply as the bias of the configuration, as we will not make use of Bt anymore. With this notation, our main theorem can be rewritten as follows:

Theorem 2.

Let λ1,λ2,λ3 be large enough positive constants. Assume that opinion 1 is the initial plurality opinion with δ0λ1p1/n and p1λ2logn/n. If hp1λ3logn, then, with high probability, the h-majority process reaches plurality consensus within O(logn) rounds.

3 Overview of our analysis and technical challenges

The whole analysis is an effort to estimate the quantity Pr(𝒲1)Pr(𝒲2) from below (assuming that p1p2pk), and use it to get a lower bound on the expect round-by-round growth of the bias δ. Notice that, for any given configuration 𝐜t at time t, it holds that

𝔼[δt+1𝐜t] 𝔼[maxi[k]minji{Pr(𝒲i(t+1))Pr(𝒲j(t+1))}|𝐜t]
𝔼[minj1{Pr(𝒲1(t+1))Pr(𝒲j(t+1))}|𝐜t]
𝔼[Pr(𝒲1(t+1))Pr(𝒲2(t+1))|𝐜t]. (1)

Suppose that p1(t)=p2(t)+δt>p2(t)pk(t). This implies that Pr(𝒲1(t+1)𝐜t)Pr(𝒲i(t+1)𝐜t)Pr(𝒲1(t+1)𝐜t)Pr(𝒲2(t+1)𝐜t) for all i3 since p2(t)pi(t) for i3. If we proved that Pr(𝒲1(t+1)𝐜t)Pr(𝒲2(t+1)𝐜t)δt(1+x) for some constant x>0, then we would get 𝔼[δt+1𝐜t]δt(1+x), which allows us to concentrate and establish that δt+1δt(1+x/2) w.h.p. (conditional on 𝐜t), for a suitable choice of x>0 and for a large enough δt, thanks to the Hoeffding bound (Lemma 23).

Notice that, in general, the probability of 𝒲i is the probability that Xi is either the unique maximum or, if it is a maximum and is not unique, it must win the tie broken u.a.r. In formulas, the probability is

Pr(𝒲i)=m=1k1m1i1<<imk:ijij[m]Pr(j[m]{Xi=Xij}(j{i,i1,,im}{Xi>Xj})). (2)

Using directly Equation 2 to bound Pr(𝒲1)Pr(𝒲2) is difficult, and we did not find general estimations in the literature.

Trivial case: very large 𝒉

If, e.g., h is very large compared to δ, in one round all nodes would adopt the plurality opinion with high probability. The reason follows.

Suppose p1=p2+δ>p2pk. Then, by the multiplicative Chernoff bound (Lemma 22 in Appendix B) we get that

Pr(X1hp1(1x))1exp(Θ(x2hp1)).

Similarly,

Pr(Xihp2(1+x))1exp(Θ(x2hp2)).

Suppose p1 and p2 are comparable, that is, p1=Θ(p2) (anyway the bias we require is always an o(p1)). Then, X1hp1(1x) and Xihp2(1+x) for all i2 hold w.h.p. as long as hClogn/(p1x2) for a large enough constant C>0. If hp1(1x)>hp2(1+x), then 𝒲1 holds w.h.p., that is, a node would adopt opinion 1 in just 1 round w.h.p. Playing with constants, by the union bound, one would obtain plurality consensus w.h.p. in just 1 round. The condition is satisfied whenever δ>x(p1+p2), that is, for x<δ/(p1+p2). Since hClogn/(p1x2) and x<δ/(p1+p2) must hold at the same time, it is sufficient to choose hCp1logn/δ2 for a large enough constant C>C. Notice that, since our bias can be such that δp1/n, we must have that hp1>(C/λ12)p1nlognlogn for any choice of p1=Ω(logn/n). The minimum that h can be with this approach is Θ(log2n/p1) and can be as large as h=Ω(n2/3/p1) if p11/n1/3. Our Theorem 2 guarantees us that we can always set h=Θ(logn/p1) as long as p1λ2logn/n for a large enough constant λ2>0. For the minimum bias δ=λ1p1/n, we have that p2=p1o(p1). Hence, 𝔼[X2]=hp2=hp1o(hp1)=𝔼[X1]o(logn). Unfortunately, we cannot capture such a deviation from the average of X1 using concentration bounds, so we need to adopt a different approach.

3.1 Our approach

At the heart of our analysis, there is a technical lemma that was proved in [25, Lemma 9].

The lemma can be reformulated as follows. Suppose we are performing the h-majority when in the system there are only k=2 opinions. Then, [25, Lemma 9] provides a tight lower bound on the expected bias after one round, i.e., on Pr(𝒲1)Pr(𝒲2). Notice that, when k=2, Pr(𝒲1)Pr(𝒲2)=Pr(X1>X2)+Pr(X1=X2)/2Pr(X2>X1)Pr(X1=X2)/2, which is equal to Pr(X1>X2)Pr(X2>X1).

Lemma 3 (Lemma 9 in [25]).

For any integer h>0, let X1Binomial(h,p) and X2=hX1Binomial(h,1p), with p>1/2. Let δ=2p1 be the bias. Then, we have that

Pr(X1>X2)Pr(X2>X1)2hπg(δ,h),

where

g(δ,h)={δ(1δ2)h12if δ<1h,1h(11h)h12if δ1h.

We show how to adapt the proof of [25, Lemma 9] to get Lemma 3 in the full version of the paper [17]. Basically, Lemma 3 makes explicit the minimum sample size required by the expectation of the bias in the next round to increase by a constant multiplicative factor relative to the current bias, to which we refer here as δ. More specifically, it is sufficient that the number of samples h satisfies 2hπ(1δ2)h122hπec for some positive constant c>0 as long as δ<1h. If, instead, δ1h, the lemma states that the new expected value of the bias is more than a constant, which allows us to show that the bias increases just by standard concentration arguments around the averages of X1 and X2 (through Berry-Esseen’s inequality or Chernoff bounds depending on how large δ is w.r.t. h).

We would like to exploit something that is similar, in spirit, to Lemma 3 when k1, in order to get that the bias increases each round. However, X1Binomial(h,p1) and X2Binomial(h,p2) but p1+p2<1 and X2hX1. In order to overcome this issue, let us define few events.

Definition 4 (Winning events).

Let i[k]. We define 𝒲i,ties to be the event in which opinion “i” is the most sampled opinion, including possible ties. Formally,

𝒲i,ties=j[k]{i}{XiXj}. (3)

Similarly, we define 𝒲i,strict to be the event in which opinion “i” is the most sampled opinion, without ties. Formally,

𝒲i,strict=j[k]{i}{Xi>Xj}. (4)

Finally, for ij[k], we define 𝒲i,j=𝒲i𝒲j, 𝒲i,j,strict=𝒲i,strict𝒲j,strict, and 𝒲i,j,ties=𝒲i,ties𝒲j,ties.

Clearly, 𝒲i,strict𝒲i𝒲i,ties and 𝒲i,j,strict𝒲i,j𝒲i,j,ties. The idea is to perform a conditional analysis: if we condition on 𝒲1,2, then we already know that either X1 or X2 win at the next round. However, we do not quite get that the conditional random variable (X1𝒲1,2) is a binomial one. At the same time, it is not true that (X2𝒲1,2) counts the number of failures of (X1𝒲1,2). The whole Section 4.1 presents the analytical effort to demonstrate that, in practice, we can get some result that is similar to Lemma 3. The section is quite technical and involved, with ad-hoc results that are needed for this adaptation. We stress that we actually use, as conditional event, the event 𝒲1,2,strict. The reason being that (as we show in Lemma 7)

Pr(𝒲1|𝒲1,2,strict)Pr(𝒲2|𝒲1,2,strict)
= Pr(X1>X2|𝒲1,2,strict)Pr(X2>X1|𝒲1,2,strict), (5)

while the equality is not true if instead of 𝒲1,2,strict we use 𝒲1,2, because we need to take ties into account. Observe that Equation 5 gives a formula that resembles the one that is estimated in Lemma 3.

In Section 4.2, we estimate the probability of 𝒲1,2,strict with respect to 𝒲1,2 and 𝒲1,2,ties. In order to do that, we classify the k opinions in two classes: the strong and the rare opinions. The strong ones are opinions whose probabilities are comparable to p1, say, pi>p1/2. The rare ones are all the others. Through concentration arguments, we can show that rare opinions disappear in one round w.h.p., and it is easy to see, by symmetry arguments, that 𝒲1=Ω(p1). The main result that we obtain in Section 4.2 is that, basically, under the hypothesis that hp1Clogn for some large enough constant C>0, p1/8Pr(𝒲1,ties)/8Pr(𝒲1,strict)Pr(𝒲1)Pr(𝒲1,ties), that is, ties do not matter that much. Notice that the latter fact is, again, trivial if hp1=Ω(max{p12n,log2n}O(logn), while we require a smaller (and, possibly, much smaller) h. Also, observe that Pr(𝒲1,strict)p1/8 implies that Pr(𝒲1,2,strict)=Ω(p1+p2). Computing directly the probability of ties in the multinomial is hard, but we resolve this issue mapping injectively events in which opinion 1 wins with a tie to an event in which opinion 1 wins without a tie. We show that this mapping preserves probabilities up to small constants, therefore we can conclude that events with ties (where opinion 1 wins) have probability weights that are comparable to events without ties (where opinion 1 wins). See Lemma 13 for more details (and note that no initial bias is required for the result of Lemma 13 to hold).

Finally, in Section 4.3 we show how to combine all the ingredients that we have proved in Sections 4.1 and 4.2 to obtain the round-by-round expected growth of the bias and the final statement on plurality consensus. As for the round-by-round expected growth of the bias, we just use the law of total probability which implies that

Pr(𝒲1)Pr(𝒲2) (6)
Pr(𝒲1,2,strict)[Pr(𝒲1|𝒲1,2,strict)Pr(𝒲2|𝒲1,2,strict)]
= Pr(𝒲1,2,strict)[Pr(X1>X2|𝒲1,2,strict)Pr(X2>X1|𝒲1,2,strict)], (7)

and then use Equation 1. Given the expected growth of the bias, we make use of the Bernstein’s inequality (Lemma 24 in Appendix B) to show its round-by-round increase in high probability, and we need to make sure that all other surrounding conditions of the growth are still satisfied in order to iterate in high probability each round.

3.2 Final discussion and open questions

In this paper we present an analysis of the h-majority dynamics when hp1Clogn for a large enough constant C. Interestingly, Theorem 2 answers (partially) a long-standing open question on mathoverflow [30] which basically asks how to estimate Pr(𝒲1)Pr(𝒲2) when h=o(1/δ2), that is, when δ=o(1/h). With Lemma 16 we answer to this question under the hypothesis that hp1Clogn, and we also show that the bias increases w.h.p. each round as long as Cp1/nδ=o(p1/logn) for some large enough constant C>0.

We also remark that δp1/n is the smallest bias that was ever required in the literature on opinion dynamics (see Section 1.2). Indeed, in the proof of Theorem 17, we show that the standard deviation of the bias at the next round is no smaller than some function Θ(p1/n), while its expected growth is at least a multiplicative factor logn. It remains open to understand whether the bias can be reduced for hklogn, which would imply that the growth of the bias is higher than a multiplicative factor logn: however, we conjecture that this growth is optimal in the regime hklogn.

The main open question here is whether the lower bound Ω(k/h2) given by [5] is tight or not. To understand this, one has to improve the assumption on hp1 and analyze especially the case where hp11. We argue more in the next subsection.

We remark that our work does not deal with perfectly-balanced configurations: one simply has to show that in short time the system reaches a configuration with the bias specified in the hypotheses of Theorem 2. We leave the analysis of the symmetry-breaking phase for future research.

On the multinomial distribution

Our analysis offers a corollary for the multinomial distribution that is interesting per-se, and is just a reformulation of Lemma 13.

Corollary 5.

Consider a multinomial random variable (X1,,Xk) where each XiBinomial(h,pi), with p1pk. If hp1Clogn for a large enough constant C>0, then Pr(i=2k{X1>Xi})cp1 for some small enough constant c>0.

We do not know how the statement of Corollary 5 changes when hp1 decreases, which remains an open question. This is linked to the generalized birthday paradox, which asks as follows: If p1==pk, what is the probability that there exists i[k] such that Xi>1? In fact, if hp12=Θ(1), we are in the standard birthday paradox: with constant probability, we have that maxi[k]{Xi}=1, which means that whenever opinion 1 is the maximum, the number of ties is h with constant probability. In contrast, Corollary 5 suggests that when hp1=Ω(logn) the number of ties that involve opinion 1 when it is the maximum is small.

Since the number of ties among the maxima increases when hp1 gets smaller, we conjecture that, for hp11, Pr(𝒲1,strict)=o(p1), implying that our approach, that entirely builds around the fact that Pr(𝒲1,strict)Pr(𝒲1), cannot apply, and new ideas are required.

Another open question is to quantify the number of ties among the maxima, in general, and especially when opinion 1 is in the maxima for a generic configuration where p1pk. This would help bridging the two cases hp11 and hp11.

4 Analysis

In this section we present the proof of Theorem 2, following the overview presented in Section 3. Most of the technical proofs either are deferred to Appendix A or can be found in the full version [17].

4.1 Conditional analysis of the expected behavior of the process

Due to space limitations, the interested reader can find the proofs in the full version [17].

Lemma 3 shows a good lower bound of Pr(𝒲1)Pr(𝒲2) for the h-majority when the number of opinion is 2. In order to use this result for k opinions, our plan is to condition on the event that either opinion 1 or 2 wins, without ties. Let x=maxi3{Xi} the mode of the sample among the opinions {3,,k}. This event can be written as max{X1,X2}>x. In the following lemma, we adapt the result of Lemma 3 to work under the additional conditioning event that opinion 1 or 2 is sampled more than x times.

Lemma 6.

For any integer m>0, let Y1Binomial(m,q) and Y2=mY1Binomial(m,1q), with q>1/2. We have, for any m/2<xm,

Pr(Y1>Y2max(Y1,Y2)>x)Pr(Y2>Y1max(Y1,Y2)>x)C1min(m(2q1),1),

for some constant C1>0.

The next lemma shows that the expected bias at the next round, i.e. Pr(𝒲1)Pr(𝒲2), is equal to Pr(X1>X2)Pr(X2>X1), conditioning on the event either opinion 1 or 2 wins, without ties. The last quantity is easier to estimate, because we don’t have to handle possible ties. This is to get a formula that resembles that of Lemma 3.

Lemma 7.

Recall the Definition 4 of the event 𝒲1,2,strict. We have

Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)
= Pr(X1>X2𝒲1,2,strict)Pr(X2>X1𝒲1,2,strict)

We remark that the statement of Lemma 7 does not hold if we condition on the event 𝒲1,2, which includes possible ties. In fact, the event {𝒲1,𝒲1,2} contains sub-events for which opinion 1 ties with many other opinions, and we could not find a simple close expression for the probability of these sub-events.

To use Lemma 6 after conditioning on the fact either opinion 1 or 2 wins, we need to estimate the number of total samples of opinion 1 and 2. Since, at the end, we assume hp1logn, we can apply the Chernoff bound (Lemma 22 in Appendix B) to show that X1+X2(p1+p2)/2 w.h.p., but we must ensure that this remains true under the aforementioned conditioning. The following lemma shows this is indeed the case. Despite the intuitive nature of the statement (the fact either opinion 1 or 2 wins is positively correlated with the absolute value of X1 and X2) we need to use an involved coupling argument to formally prove it.

Lemma 8.

It holds that

Pr(X1+X2h(p1+p2)/2𝒲1,2,strict)Pr(X1+X2h(p1+p2)/2).

4.2 Estimating the probability of the conditional event

In this section our goal is to estimate a lower bound for Pr(𝒲1,2,strict). More specifically, we show that Pr(𝒲1,2,strict)C(p1+p2) for some constant C>0. As we already argued after Lemma 7, it is important to exclude ties in the event we condition on.

The next definitions introduce terms to refer to opinions that will disappear w.h.p. at the next round, the weak opinions, and their counter part, the strong opinions.

Definition 9 (Rare opinions).

Denote with x:={i[k]:pixp1} be the set of the x-rare opinions, for 0<x<1.

Definition 10 (Strong opinions).

Denote with 𝒮:={i[k]:pi>12p1} be the set of the strong opinions, for 0<x<1.

Note that the set of the strong opinions and the set of 1/2-rare opinions are complementary.

In the next lemma we show that w.h.p. all weak opinions will be sampled less than the opinion 1, and therefore, they will disappear at the next round. The proof is just an application of Chernoff bounds (Lemma 22 in Appendix B) and we defer it to Section A.1.

Lemma 11.

Let C2,C3 be two constants s.t. 0<C2<1 and C3>2. If hp1>C4logn, with C4=(3C311C2)2, then

Pr(iC2{X1>Xi})11nC32.

To bound Pr(𝒲1,strict), we plan to first show that there exists a constant C s.t. 𝒲1,tiesCp1, and, therefore, show that Pr(𝒲1,strict)CPr(𝒲1,ties), for some other constant C. Hence, in the next lemma we show that Pr(𝒲1)Cp1 which, in turn, implies that Pr(𝒲1,ties)Cp1. The idea is to use Lemma 11 to show that only strong opinion compete for the win and in the worst case they have all same probability to win without ties. The thesis follows after showing that the number of strong opinions is at most 2/p1, and that i𝒮Pr(𝒲i)1. Again, the proof is deferred to Section A.1.

Lemma 12.

Let C4=(18)2. If p1hC4logn, then

Pr(𝒲1)13p1.

The next Lemma plays a key role in avoiding the computation of the probability of ties. Assuming that hp1=O(logn), we can map each realization of the multinomial in which opinion 1 wins with ties to a realization in which opinion 1 wins without ties. We will show that these two realizations have the same probability up to a constant. Therefore, we can conclude that tie events have a comparable weight to events without ties. We present its proof in Section A.1 as we believe this is an interesting result per-se, which also proves Corollary 5 in Section 3.2.

Lemma 13.

Let hp1>C4logn for C4=182. It holds that

Pr(𝒲1,strict)16Pr(𝒲1,ties)118p1.

The goal of this section was to show that Pr(𝒲1,2,strict)C(p1+p2) for some constant C>0. This can be easily obtained by applying Lemma 13, because p1+p22p1 and 𝒲1,strict𝒲1,2,strict.

Corollary 14.

Let hp1>C4logn for C4=182. It holds that

Pr(𝒲1,2,strict)136(p1+p2).

4.3 Putting everything together

In this section we can finally put together all the results to give a lower bound to Pr(𝒲1)Pr(𝒲j) and to use it to prove Theorem 2. Next lemma is the statement that is “equivalent in spirit” to Lemma 3 but for many opinions, conditional on 𝒲1,2,strict. Its proof involves a mix of the previous results and is quite intricate.

Lemma 15.

Let p1hC4logn, with C4=182. We have

Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)Cmin{(p1p2)h2(p1+p2),1},

for some constant C>0.

Proof of Lemma 15.

Let

M={m:Pr(X1+X2=m𝒲1,2,strict)>0}{mh(p1+p2)/2}

and

Lm={x:Pr(maxj3Xj=xX1+X2=m,𝒲1,2,strict)>0}.

We have

Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)
=(i) Pr(X1>X2𝒲1,2,strict)Pr(X2>X1𝒲1,2,strict)
(ii) mMPr(X1+X2=m𝒲1,2,strict)[Pr(X1>X2𝒲1,2,strict,X1+X2=m)
Pr(X2>X1𝒲1,2,strict,X1+X2=m)]
=(iii) mMPr(X1+X2=m𝒲1,2,strict)
xLmPr(maxj3Xj=xX1+X2=m,𝒲1,2,strict)
[Pr(X1>X2𝒲1,2,strict,X1+X2=m,maxj3Xj=x)
Pr(X2>X1𝒲1,2,strict,X1+X2=m,maxj3Xj=x)]
=(iv) mMPr(X1+X2=m𝒲1,2,strict)
xLmPr(maxj3Xj=xX1+X2=m,𝒲1,2,strict)
[Pr(X1>X2max{X1,X2}>x,X1+X2=m,maxj3Xj=x)
Pr(X2>X1max{X1,X2}>x,X1+X2=m,maxj3Xj=x)], (8)

where in (i) we used Lemma 7, in (ii),(iii) we used the law of total probabilities, and in (iv) we used that, by Definition 4,

(𝒲1,2,strict,X1+X2=m,maxj3Xj=x)
= (max{X1,X2}>x,X1+X2=m,maxj3Xj=x).

Conditioning on {X1+X2=m}, makes X1,X2 independent from X3,,Xk, and distributed as Y1,Y2, where Y1Binomial(m,q1) with q1=p1p1+p2 and Y2=mY1. We obtain, by Section 4.3

Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)
mMPr(X1+X2=m𝒲1,2,strict)
xLmPr(maxj3Xj=xX1+X2=m,𝒲1,2,strict)
[Pr(Y1>Y2max{Y1,Y2}x)Pr(Y2>Y1max{Y1,Y2}x)]
(i) mMPr(X1+X2=m𝒲1,2,strict)
xLmPr(maxj3Xj=xX1+X2=m,𝒲1,2,strict)
C1min(m(δ2p1+p2),1)
=(ii) mMPr(X1+X2=m𝒲1,2,strict)C1min(m(δ2p1+p2),1)
(iii) Pr(X1+X2h(p1+p2)/2𝒲1,2,strict)C1min((δ2h2(p1+p2)),1)
(iv) C1(1exp(18h(p1+p2)))min((δ2h2(p1+p2)),1)
(v) C1(11n2)min((δ2h2(p1+p2)),1)
Cmin((δ2h2(p1+p2)),1),

where in (i) we used Lemma 6, in (ii) we used the law of total probabilities, in (iii) we used that by definition of M, mh(p1+p2)/2, in (iv) we used Lemma 8 and the multiplicative Chernoff bound (Lemma 22 in Appendix B), and in (v) we used that hp116logn. This concludes the proof of Lemma 15.

In the next lemma, we finally remove the conditional event 𝒲1,2,strict and bound Pr(𝒲1)Pr(𝒲2) from below. We need to use Lemmas 13 and 15, but its proof consists in just mixing in the “right way” previous results, hence we defer it to Section A.2. From now on, we denote the difference p1pj by δ(j).

Lemma 16.

Let p1hC4logn, with C4=182. If δ(j)2(p1+p2)h, then there exists a constant C5>0 s.t.

Pr(𝒲1)Pr(𝒲j)C5(δ(j)hp1+p2)Pr(𝒲1).

Instead, if δ(j)2(p1+p2)h, there exists a constant 0<C6<1 s.t.

Pr(𝒲1)11C6Pr(𝒲j).

Given a configuration (p1,,pk), we denote by pj the probability associated to opinion j at the next round. Fix an opinion j1. Let δ(j) be the bias between opinion 1 and opinion j at the next round, i.e. p1pj.

Before proceeding in the analysis presentation, let us comment Lemma 16. Note that Pr(𝒲j) is the expected value of pj. In the case the bias between opinion 1 and opinion j is more than 2(p1+p2)h, at the next round we have p1(1+C)pj in expectation, for some constant C>0. In the case the bias is smaller than 2(p1+p2)h, by the fact h=Ω(p1logn) and Pr(𝒲1)=Ω(p1+p2), Lemma 16 implies that at the next round the expected bias grows of a Ω(logn) factor. Thanks to standard concentration bounds (Chernoff Bound and Bernstein Inequality) we show that w.h.p. the new bias is sufficiently close to its expectation.

Now we all have all ingredients to prove Theorem 2. Next theorem is just a reformulation of Theorem 2.

Theorem 17.

Consider an initial configuration (p1,,pk) s.t. p1C7lognn, and p1pjC8p1n for all j2. Then the h-majority dynamics with hp1>C4logn converges in time O(logn) to opinion 1, w.h.p., for some constants C4,C7,C8>0.

The proof relies on first quantifying the increase in the bias between the first opinion and other opinions after one round. Then it iterates the same reasoning until consensus (for the complete proof, see the full version [17]). For the first point, we use concentration bounds on the results obtained in Lemma 16. In particular, we distinguish three sub-cases based on the size of the bias between two opinions. For the second point, we need to show that the initial hypotheses are still satisfied after one round to ensure that we can iterate.

In the next lemma, we show that if p1(1+C)pj for some constant C>0, opinion j disappears at the next round, by simply applying the Chernoff bound.

Lemma 18 (Opinion j-disappear at the next round).

Let hp1C4logn. If δ(j)(111+C6)p1 for the constant C6 defined in Lemma 16, with probability at least 11n4 opinion j will disappear at the next round

The next Lemma shows that if the bias is Ω(p1/h), then the new bias will fall within the hypothesis of Lemma 18 in one step. Therefore, we conclude that, in two steps, opinion j will disappear if the bias is large enough. The proof relies on the results obtained in Lemma 16, combined with the Chernoff bound.

Lemma 19 (Opinion j-disappear in two rounds).

Let hp1C4logn. If p1C7lognn and 2(p1+p2)hδ(j)(111+C6)p1 for the constant C6>0 defined in Lemma 16 and for a constant C7>0, we have that

Pr(δ(j)(111+C6)p1)12n4.

The next Lemma addresses the minimum possible bias. We show that, until we fall back into the hypotheses of Lemmas 18 and 19, the bias grows exponentially. The proof is more complex than the previous two lemmas because it requires bounding the second moment of the bias to apply the Bernstein inequality.

Lemma 20 (Bias exponential growth).

Let hp1C4logn. If p1C7lognn and C8p1nδ(j)2(p1+p2)h for the constant C6>0 defined in Lemma 16 and for some constants C7,C8>0, we have that

Pr(δ(j)eδ(j))11n4

The next Lemma ensures that the hypothesis of Theorem 2 on the bias is always satisfied so that we can iterate Lemma 20.

Lemma 21 (The new bias satisfies the initial hypothesis).

Let δ(j)p1n, and a constant C8>0. For n large enough we have

Pr(δ(j)Cp1n)12n4.

References

  • [1] Talley Amir, James Aspnes, Petra Berenbrink, Felix Biermeier, Christopher Hahn, Dominik Kaaser, and John Lazarsfeld. Fast convergence of k-opinion undecided state dynamics in the population protocol model. In Rotem Oshman, Alexandre Nolin, Magnús M. Halldórsson, and Alkida Balliu, editors, Proceedings of the 2023 ACM Symposium on Principles of Distributed Computing, PODC 2023, Orlando, FL, USA, June 19-23, 2023, pages 13–23. ACM, 2023. doi:10.1145/3583668.3594589.
  • [2] Dana Angluin, James Aspnes, and David Eisenstat. A simple population protocol for fast robust approximate majority. Distributed Comput., 21(2):87–102, 2008. doi:10.1007/S00446-008-0059-Z.
  • [3] Gregor Bankhamer, Petra Berenbrink, Felix Biermeier, Robert Elsässer, Hamed Hosseinpour, Dominik Kaaser, and Peter Kling. Fast consensus via the unconstrained undecided state dynamics. In Joseph (Seffi) Naor and Niv Buchbinder, editors, Proceedings of the 2022 ACM-SIAM Symposium on Discrete Algorithms, SODA 2022, Virtual Conference / Alexandria, VA, USA, January 9 - 12, 2022, pages 3417–3429. SIAM, 2022. doi:10.1137/1.9781611977073.135.
  • [4] Luca Becchetti, Andrea Clementi, Emanuele Natale, Francesco Pasquale, and Riccardo Silvestri. Plurality consensus in the gossip model. In Piotr Indyk, editor, Proceedings of the Twenty-Sixth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2015, San Diego, CA, USA, January 4-6, 2015, pages 371–390. SIAM, 2015. doi:10.1137/1.9781611973730.27.
  • [5] Luca Becchetti, Andrea Clementi, Emanuele Natale, Francesco Pasquale, Riccardo Silvestri, and Luca Trevisan. Simple dynamics for plurality consensus. Distributed Computing, 30(4):293–306, August 2017. doi:10.1007/s00446-016-0289-4.
  • [6] Luca Becchetti, Andrea Clementi, Emanuele Natale, Francesco Pasquale, and Luca Trevisan. Stabilizing consensus with many opinions. In Robert Krauthgamer, editor, Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2016, Arlington, VA, USA, January 10-12, 2016, pages 620–635. SIAM, 2016. doi:10.1137/1.9781611974331.CH46.
  • [7] Luca Becchetti, Andrea Clementi, Francesco Pasquale, Luca Trevisan, Robin Vacus, and Isabella Ziccardi. The minority dynamics and the power of synchronicity. In David P. Woodruff, editor, Proceedings of the 2024 ACM-SIAM Symposium on Discrete Algorithms, SODA 2024, Alexandria, VA, USA, January 7-10, 2024, pages 4155–4176. SIAM, 2024. doi:10.1137/1.9781611977912.144.
  • [8] Petra Berenbrink, Felix Biermeier, and Christopher Hahn. Undecided state dynamics with stubborn agents. CoRR, abs/2406.07335, 2024. doi:10.48550/arXiv.2406.07335.
  • [9] Petra Berenbrink, Andrea Clementi, Robert Elsässer, Peter Kling, Frederik Mallmann-Trenn, and Emanuele Natale. Ignore or comply?: On breaking symmetry in consensus. In Elad Michael Schiller and Alexander A. Schwarzmann, editors, Proceedings of the ACM Symposium on Principles of Distributed Computing, PODC 2017, Washington, DC, USA, July 25-27, 2017, pages 335–344. ACM, 2017. doi:10.1145/3087801.3087817.
  • [10] Petra Berenbrink, Amin Coja-Oghlan, Oliver Gebhard, Max Hahn-Klimroth, Dominik Kaaser, and Malin Rau. On the hierarchy of distributed majority protocols. In Eshcar Hillel, Roberto Palmieri, and Etienne Rivière, editors, 26th International Conference on Principles of Distributed Systems, OPODIS 2022, December 13-15, 2022, Brussels, Belgium, volume 253 of LIPIcs, pages 23:1–23:19. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2022. doi:10.4230/LIPICS.OPODIS.2022.23.
  • [11] Andrea Clementi, Francesco D’Amore, George Giakkoupis, and Emanuele Natale. Search via Parallel Lévy Walks on Z2. In Avery Miller, Keren Censor-Hillel, and Janne H. Korhonen, editors, PODC ’21: ACM Symposium on Principles of Distributed Computing, Virtual Event, Italy, July 26-30, 2021, pages 81–91. ACM, 2021. doi:10.1145/3465084.3467921.
  • [12] Andrea Clementi, Mohsen Ghaffari, Luciano Gualà, Emanuele Natale, Francesco Pasquale, and Giacomo Scornavacca. A tight analysis of the parallel undecided-state dynamics with two colors. In Igor Potapov, Paul G. Spirakis, and James Worrell, editors, 43rd International Symposium on Mathematical Foundations of Computer Science, MFCS 2018, August 27-31, 2018, Liverpool, UK, volume 117 of LIPIcs, pages 28:1–28:15. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2018. doi:10.4230/LIPICS.MFCS.2018.28.
  • [13] Colin Cooper, Frederik Mallmann-Trenn, Tomasz Radzik, Nobutaka Shimizu, and Takeharu Shiraga. Asynchronous 3-majority dynamics with many opinions. In Yossi Azar and Debmalya Panigrahi, editors, Proceedings of the 2025 Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2025, New Orleans, LA, USA, January 12-15, 2025, pages 4095–4131. SIAM, 2025. doi:10.1137/1.9781611978322.140.
  • [14] Emilio Cruciani, Hlafo Alfie Mimun, Matteo Quattropani, and Sara Rizzo. Phase transition of the k-majority dynamics in biased communication models. Distributed Comput., 36(2):107–135, 2023. doi:10.1007/S00446-023-00444-2.
  • [15] Francesco D’Amore, Andrea Clementi, and Emanuele Natale. Phase transition of a non-linear opinion dynamics with noisy interactions - (extended abstract). In Andréa Werneck Richa and Christian Scheideler, editors, Structural Information and Communication Complexity - 27th International Colloquium, SIROCCO 2020, Paderborn, Germany, June 29 - July 1, 2020, Proceedings, volume 12156 of Lecture Notes in Computer Science, pages 255–272. Springer, 2020. doi:10.1007/978-3-030-54921-3_15.
  • [16] Francesco D’Amore, Andrea Clementi, and Emanuele Natale. Phase transition of a nonlinear opinion dynamics with noisy interactions. Swarm Intell., 16(4):261–304, 2022. doi:10.1007/S11721-022-00217-W.
  • [17] Francesco d’Amore, Niccolò D’Archivio, George Giakkoupis, and Emanuele Natale. On the h-majority dynamics with many opinions, 2025. doi:10.48550/arXiv.2506.20218.
  • [18] Francesco D’Amore and Isabella Ziccardi. Phase transition of the 3-majority dynamics with uniform communication noise. In Merav Parter, editor, Structural Information and Communication Complexity - 29th International Colloquium, SIROCCO 2022, Paderborn, Germany, June 27-29, 2022, Proceedings, volume 13298 of Lecture Notes in Computer Science, pages 98–115. Springer, 2022. doi:10.1007/978-3-031-09993-9_6.
  • [19] Francesco D’Amore and Isabella Ziccardi. Phase transition of the 3-majority opinion dynamics with noisy interactions. Theor. Comput. Sci., 1028:115030, 2025. doi:10.1016/J.TCS.2024.115030.
  • [20] Niccolò D’Archivio and Robin Vacus. On the limits of information spread by memory-less agents. In Dan Alistarh, editor, 38th International Symposium on Distributed Computing, DISC 2024, October 28 to November 1, 2024, Madrid, Spain, volume 319 of LIPIcs, pages 18:1–18:21. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024. doi:10.4230/LIPICS.DISC.2024.18.
  • [21] Devdatt P. Dubhashi and Alessandro Panconesi. Concentration of Measure for the Analysis of Randomized Algorithms. Cambridge University Press, 2009. URL: http://www.cambridge.org/gb/knowledge/isbn/item2327542/.
  • [22] Ofer Feinerman, Bernhard Haeupler, and Amos Korman. Breathe before speaking: efficient information dissemination despite noisy, limited and anonymous communication. In Magnús M. Halldórsson and Shlomi Dolev, editors, ACM Symposium on Principles of Distributed Computing, PODC ’14, Paris, France, July 15-18, 2014, pages 114–123. ACM, 2014. doi:10.1145/2611462.2611469.
  • [23] Ofer Feinerman and Amos Korman. The ANTS problem. Distributed Comput., 30(3):149–168, 2017. doi:10.1007/S00446-016-0285-8.
  • [24] Pierre Fraigniaud, Amos Korman, and Yoav Rodeh. Parallel exhaustive search without coordination. In Daniel Wichs and Yishay Mansour, editors, Proceedings of the 48th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2016, Cambridge, MA, USA, June 18-21, 2016, pages 312–323. ACM, 2016. doi:10.1145/2897518.2897541.
  • [25] Pierre Fraigniaud and Emanuele Natale. Noisy rumor spreading and plurality consensus. Distributed Computing, 32(4):257–276, 2019. doi:10.1007/S00446-018-0335-5.
  • [26] Matthias Függer, Thomas Nowak, and Joel Rybicki. Majority consensus thresholds in competitive lotka-volterra populations. In Ran Gelles, Dennis Olivetti, and Petr Kuznetsov, editors, Proceedings of the 43rd ACM Symposium on Principles of Distributed Computing, PODC 2024, Nantes, France, June 17-21, 2024, pages 76–86. ACM, 2024. doi:10.1145/3662158.3662823.
  • [27] Mohsen Ghaffari and Johannes Lengler. Nearly-tight analysis for 2-choice and 3-majority consensus dynamics. In Calvin Newport and Idit Keidar, editors, Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing, PODC 2018, Egham, United Kingdom, July 23-27, 2018, pages 305–313. ACM, 2018. URL: https://dl.acm.org/citation.cfm?id=3212738.
  • [28] George Giakkoupis, Volker Turau, and Isabella Ziccardi. Self-stabilizing MIS computation in the beeping model. In Dan Alistarh, editor, 38th International Symposium on Distributed Computing, DISC 2024, October 28 to November 1, 2024, Madrid, Spain, volume 319 of LIPIcs, pages 28:1–28:21. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024. doi:10.4230/LIPICS.DISC.2024.28.
  • [29] George Giakkoupis and Isabella Ziccardi. Distributed self-stabilizing MIS with few states and weak communication. In Rotem Oshman, Alexandre Nolin, Magnús M. Halldórsson, and Alkida Balliu, editors, Proceedings of the 2023 ACM Symposium on Principles of Distributed Computing, PODC 2023, Orlando, FL, USA, June 19-23, 2023, pages 310–320. ACM, 2023. doi:10.1145/3583668.3594581.
  • [30] Immanuel Weihnachten (https://mathoverflow.net/users/23351/immanuel weihnachten). Lower bound on the probability of guessing the mode in a small multinomial sample. MathOverflow. URL:https://mathoverflow.net/q/210018 (version: 2015-06-24). arXiv:https://mathoverflow.net/q/210018.
  • [31] Amos Korman and Robin Vacus. Distributed alignment processes with samples of group average. IEEE Trans. Control. Netw. Syst., 10(2):960–971, 2023. doi:10.1109/TCNS.2022.3212640.
  • [32] Hicham Lesfari, Frédéric Giroire, and Stéphane Pérennes. Biased majority opinion dynamics: Exploiting graph k-domination. In Luc De Raedt, editor, Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23-29 July 2022, pages 377–383. ijcai.org, 2022. doi:10.24963/IJCAI.2022/54.
  • [33] Michael Mitzenmacher and Eli Upfal. Probability and Computing: Randomized Algorithms and Probabilistic Analysis. Cambridge University Press, 2005. doi:10.1017/CBO9780511813603.
  • [34] Nobutaka Shimizu and Takeharu Shiraga. 3-majority and 2-choices with many opinions. CoRR, abs/2503.02426, 2025. To appear at PODC 2025. doi:10.48550/arXiv.2503.02426.
  • [35] Robin Vacus and Isabella Ziccardi. Minimalist Leader Election Under Weak Communication. In Proceedings of the ACM Symposium on Principles of Distributed Computing, PODC ’25, pages 406–416, New York, NY, USA, 2025. Association for Computing Machinery. doi:10.1145/3732772.3733559.

Appendix A Missing proofs

A.1 Omitted proofs from Section 4.2

Proof of Lemma 11.

Define the event 𝒞j={|Xjhpj|<3C3hpjlogn} and 𝒞=j=1k𝒞j. The event 𝒞j can be rewritten as

𝒞j={Xj(hpj3C3hpjlogn,hpj+3C3hpjlogn)}. (9)

For i, we have

hp13C3hp1logn(hpi+3C3hpilogn)
(1C2)hp13C3hp1logn (ihpihC2p1)
> 0,

by taking hp1>C4logn, with C4=(3C311C2)2. This fact, together with Equation 9, implies that 𝒞{iC2{X1>Xi}}, and therefore Pr(iC2{X1>Xi})Pr(𝒞). By the multiplicative Chernoff bound we have

Pr(Cj) =Pr(|Xjhpj|<3C3hpjlogn)
=1Pr(|Xjhpj|3C3hpjlogn)
12exp(3C3logn3)
=12nC3,

and by the union bound and by the fact the number of opinion kn, we obtain

Pr(i{X1>Xi})Pr(𝒞)12knC311nC32,

concluding the proof of Lemma 11.

Proof of Lemma 12.

Recall 𝒮:={i[k]:pi>12p1} be the set of the strong opinions. By definition of 𝒮 we have

1j𝒮pj>|𝒮|p12,

and therefore we have

|𝒮|2p1. (10)

Recall that 1/2:={i[k]:pi12p1} is the set of the 1/2-rare opinions. If i1/2, we have that

Pr(𝒲i1/2{X1>X})=0.

By this fact, and by noticing that an opinion is either strong or 1/2-rare,

1=Pr(j𝒮𝒲j1/2{X1>X})j𝒮Pr(𝒲j1/2{X1>X}). (11)

Since XjBinomial(pj,h) and p1pj, for all j[k], we have that X1 stochastically dominates Xj for all j2. This fact remains true also conditioning on the event {1/2{X1>X}}. Therefore, we have

Pr(𝒲11/2{X1>X})Pr(𝒲j1/2{X1>X}),

for all j𝒮,j2. This fact, together with Equation 11, implies

|𝒮|Pr(𝒲11/2{X1>X})j𝒮Pr(𝒲j1/2{X1>X})1,

and, by using Equation 10, we obtain

Pr(𝒲11/2{X1>X})1|𝒮|p12.

We can conclude the proof of Lemma 12 noticing that, by Lemma 11 with C3=3 and C2=1/2 (which implies C4=(18)2, we have

Pr(𝒲1)Pr(1/2{X1>X})Pr(𝒲11/2{X1>X})(11n)p12p13,

for n large enough.

Proof of Lemma 13.

Define the event 𝒞j={|Xjhpj|<9hpjlogn} and 𝒞=j1𝒞j. By the multiplicative Chernoff bound (Lemma 22 in Appendix B), we obtain

Pr(𝒞j) =Pr(|Xjhpj|<9hpjlogn)
=1Pr(|Xjhpj|9hpjlogn)
12exp(9logn3)
=12n3.

Moreover, by the union bound and by the fact kn, we have, for n large enough,

Pr(𝒞)12kn311n. (12)

Recall 𝒮:={i[k]:pi>12p1} be the set of the strong opinions. We first note that

𝒞 j𝒮{Xj>hpj9hpjlogn}
j𝒮{Xj>12hp19hp1logn} (pi>12p1,pip1, for all i𝒮)
j𝒮{Xj>0}. (hp1>36logn) (13)

Let us denote a particular realization of a multinomial in the following way.

𝒳(x1,,xk)={X1=x1,,Xk=xk}.

We define the event 𝒳(x1,,xk) a 1-tie if x1xi for all i1 and there exists a j1 s.t. xj=x1. In other words, opinion 1 is the most sampled, together with at least another opinion. We denote the set of 1-tie as 𝒯1.

We map a 1-tie to an event in which the opinion 1 gets one extra sample which is stolen from the least scoring strong opinion. Formally, if 𝒳(x1,,xk)𝒯1 and j=max{i:Xi=minr𝒮{Xr}}, we define:

f:𝒳(x1,,xk)𝒳(x1+1,,xj1,xj1,xj+1,,xk).

The map is well-defined if and only if xj>0. If 𝒳(x1,,xk)𝒞, by Equation 13, we have 𝒳(x1,,xk)j𝒮{Xj>0}, and, since j𝒮 by definition, we obtain that xj>0. Therefore, f is well-defined on 𝒞.

Moreover, the map is injective. Using the probability mass function of the multinomial distribution, we compute the following:

Pr(f(𝒳(x1,,xk)))Pr(𝒳(x1,,xk))=xjx1+1p1pjxjx1+1,

where we used that p1pj. Therefore we obtain

𝒳𝒯1,𝒳𝒞Pr(f(𝒳(x1,,xk)))
𝒳𝒯1,𝒳𝒞xjx1+1Pr((𝒳(x1,,xk)))
hpj9hpjlogn1+hp1+9hp1logn𝒳𝒯1,𝒳𝒞Pr(𝒳(x1,,xk)) (by definition of 𝒞)
12hp19hp1logn1+hp1+9hp1logn𝒳𝒯1,𝒳𝒞Pr(𝒳(x1,,xk)) (pi>12p1,pip1, for i𝒮)
= 12hp19logn1hp19logn+2𝒳𝒯1,𝒳𝒞Pr(𝒳(x1,,xk))
14𝒳𝒯1,𝒳𝒞Pr(𝒳(x1,,xk)), (14)

where the last inequality holds because x12x1x+2 is increasing and hp1182logn. Recall Definition 4. Since 𝒯1=𝒲1,ties𝒲1,strict and f(𝒯1𝒞)𝒲1, we can write

Pr(𝒲1,ties,𝒞)
= Pr(𝒲1,strict,𝒞)+Pr(𝒲1,ties𝒲1,strict,𝒞)
𝒳𝒯1,𝒳𝒞Pr(f(𝒳(x1,,xk)))+𝒳𝒯1,𝒳𝒞Pr(𝒳(x1,,xk))
(1+14)𝒳𝒯1,𝒳𝒞Pr(𝒳(x1,,xk)) (by Equation 14)
= (1+14)Pr(𝒲1,ties𝒲1,strict,𝒞).

Hence,

Pr(𝒲1,ties𝒲1,strict,𝒞)45Pr(𝒲1,ties,𝒞),

and

Pr(𝒲1,strict)Pr(𝒲1,strict,𝒞)=Pr(𝒲1,ties,𝒞)Pr(𝒲1,ties𝒲1,strict,𝒞)
15Pr(𝒲1,ties,𝒞)
= 15(𝒲1,ties+Pr(𝒞)Pr(𝒲1,ties𝒞))(by the inclusion–exclusion princ.)
15(𝒲1,ties+Pr(𝒞)1)(Pr(𝒲1,ties𝒞)1)
15(𝒲1,ties1n)(by Equation 12)
16𝒲1,ties118p1

for n large enough. This concludes the proof of Lemma 13.

Proof of Corollary 14.

By Lemmas 13 and 12, we obtain

Pr(𝒲1,2,strict)Pr(𝒲1,strict)16Pr(𝒲1,ties)118p1=136(p1+p1)136(p1+p2).

This concludes the proof of Corollary 14.

A.2 Omitted proofs from Section 4.3

Proof of Lemma 16.

Since p2pj for all j2, we have that Pr(𝒲j)Pr(𝒲2), and, therefore,

Pr(𝒲1)Pr(𝒲2)Pr(𝒲1)Pr(𝒲j),

and,

Pr(𝒲1)11C6Pr(𝒲2)Pr(𝒲1)11C6Pr(𝒲j).

Hence, we will focus on the proof statement for j=2.

Consider the case δ(2)2(p1+p2)h, that, in particular, implies that

min((δ(2)h2(p1+p2)),1)=(δ(2)h2(p1+p2)).

We have

Pr(𝒲1)Pr(𝒲2)
Pr(𝒲1,2,strict)[Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)]
Pr(𝒲1,strict)[Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)] (𝒲1,strict𝒲1,2,strict)
16Pr(𝒲1,ties)[Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)] (by Lemma 13)
16Pr(𝒲1)[Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)] (𝒲1𝒲1,ties)
C6Pr(𝒲1)δ(2)hp1+p2 (by Lemma 15)

Setting C5=C6, we conclude the proof of the first statement of Lemma 16. Now consider the case δ2(p1+p2)h, that implies that (δh2(p1+p2))=1.

Pr(𝒲1)Pr(𝒲2)
Pr(𝒲1,2,strict)[Pr(𝒲1𝒲1,2,strict)Pr(𝒲2𝒲1,2,strict)]
Pr(𝒲1,2,strict)C (by Lemma 15)
Pr(𝒲1,strict)C (𝒲1,strict𝒲1,strict)
C6Pr(𝒲1,ties) (by Lemma 13)
C6Pr(𝒲1) (𝒲1𝒲1,ties)
C6Pr(𝒲1),

if we set 0<C6=min(12,C6)<1. This concludes the proof of Lemma 16.

Appendix B Tools

Lemma 22 (Multiplicative forms of Chernoff bounds [21]).

Let X1,,Xn be independent binary random variables. Let X=i=1nXi and μ=𝔼[X]. Then:

  1. 1.

    For any δ(0,1) and any μμ+n, it holds that

    Pr(X(1+δ)μ+)exp(δ2μ+/3).
  2. 2.

    For any δ(0,1) and any 0μμ, it holds that

    Pr(X(1δ)μ)exp(δ2μ/2).
Lemma 23 (Hoeffding bounds [33]).

Let a<b) be two constants, and X1,,Xn be independent binary random variables such that Pr(aXib)=1 for all i[n]. Let X=i=1nXi and μ=𝔼[X]. Then:

  1. 1.

    For any t>0 and any μμ+, it holds that

    Pr(Xμ++t)exp(2t2n(ba)2).
  2. 2.

    For any t>0 and any 0μμ, it holds that

    Pr(Xμt)exp(2t2n(ba)2).
Lemma 24 (Bernstein Inequality [21]).

Let X1,X2,,Xn be i.i.d. random variables such that |Xi𝔼[Xi]|C and |𝔼[Xi]|C for some C>0. Define the sum of centered variables Sn=i=1n(Xi𝔼[Xi]). Then, for all t>0,

Pr(|Sn|t)2exp(t22nVar(X1)+2Ct3).