Abstract 1 Introduction 2 Preliminaries 3 Proof of Main Theorem References Appendix A Appendix

Gadgetless Lifting Beats Round Elimination: Improved Lower Bounds for Pointer Chasing

Xinyu Mao ORCID Thomas Lord Department of Computer Science, University of Southern California, Los Angeles, CA, USA Guangxu Yang ORCID Thomas Lord Department of Computer Science, University of Southern California, Los Angeles, CA, USA Jiapeng Zhang ORCID Thomas Lord Department of Computer Science, University of Southern California, Los Angeles, CA, USA
Abstract

We prove an Ω(n/k+k) communication lower bound on (k1)-round distributional complexity of the k-step pointer chasing problem under uniform input distribution, improving the Ω(n/kklogn) lower bound due to Yehudayoff (Combinatorics Probability and Computing, 2020). Our lower bound almost matches the upper bound of O~(n/k+k) communication by Nisan and Wigderson (STOC 91).

As part of our approach, we put forth gadgetless lifting, a new framework that lifts lower bounds for a family of restricted protocols into lower bounds for general protocols. A key step in gadgetless lifting is choosing the appropriate definition of restricted protocols. In this paper, our definition of restricted protocols is inspired by the structure-vs-pseudorandomness decomposition by Göös, Pitassi, and Watson (FOCS 17) and Yang and Zhang (STOC 24).

Previously, round-communication trade-offs were mainly obtained by round elimination and information complexity. Both methods have some barriers in some situations, and we believe gadgetless lifting could potentially address these barriers.

Keywords and phrases:
communication complexity, lifting theorems, pointer chasing
Funding:
Xinyu Mao: Supported by NSF CAREER award 2141536.
Guangxu Yang: Supported by NSF CAREER award 2141536.
Jiapeng Zhang: Supported by NSF CAREER award 2141536.
Copyright and License:
[Uncaptioned image] © Xinyu Mao, Guangxu Yang, and Jiapeng Zhang; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Communication complexity
Related Version:
Full Version: https://arxiv.org/abs/2411.10996
Acknowledgements:
We thank Sepehr Assadi, Yuval Filmus and anonymous reviewers for their helpful comments.
Editor:
Raghu Meka

1 Introduction

Pointer chasing is a well-known problem [26] that demonstrates the power of interaction in communication and has broad applications in different areas. It was used for proving monotone constant-depth hierarchy theorem [23, 20], lower bounds on the time complexity of distributed computation [22], lower bounds on the space complexity of streaming algorithms [10, 15, 1], adaptivity hierarchy theorem for property testing [5], exponential separations in local differential privacy [16], memory bounds for continual learning [7] and limitations of the transformer architecture [24]. It is a two-party function defined below.

Definition 1 (k-step pointer chasing function).

For k1, the k-step pointer chasing function PCk:[n]n×[n]n{0,1} is defined as follows. Given input fA,fB[n]n, for r=0,1,,k we recursively define pointers via

ptr(fA,fB)=𝖽𝖾𝖿{1if r=0;fA(ptr1(fA,fB))if r>0 is odd;fB(ptr1(fA,fB))if r>0 is even.

The output of PCk is the parity of the last pointer, namely, PCk(fA,fB)=𝖽𝖾𝖿ptk(fA,fB)mod2.

Upper bounds

If Alice and Bob could communicate for k rounds, a simple protocol is the following: Alice and Bob alternatively send fA(ptr1(fA,fB)) or fA(ptr1(fA,fB)). The total communication cost for this simple protocol is O(klogn). However, if Alice and Bob can only communicate (k1) rounds, the upper bound then becomes non-trivial. Nisan and Wigderson [23] proposed a randomized (k1)-round protocol with O((n/k+k)logn) communication bits.

  • In the beginning, Alice and Bob use public randomness to pick a set of coordinates I[n] of size 10n/k, and then send fA(I) and fB(I) to the other party.

  • On the other hand, Alice and Bob also simulate (r rounds) deterministic protocol but skip one round if one party finds that the pointer is located in I.

  • If the skip round never happens, Alice and Bob simply abort at the last round. A simple calculation shows the probability of this event is low.

This randomized protocol is indeed very simple. Alice and Bob only share coordinate-wise information. In fact, this is a structured rectangle in our setting.

Lower Bounds

Consider (k1) round protocols where Alice speaks first. For deterministic protocols, Nisan and Wigderson [23] proved an Ω(nklogn) communication lower bound. In the same paper, they also proved an Ω(n/k2klogn) communication lower bound for protocols that achieve 2/3 accuracy under uniform input distribution.

Since then, lower bounds for pointer chasing and its close variants have been substantially studied by a good amount of papers [9, 8, 25, 18, 19, 10, 15, 1]. Finally, Yehudayoff [29] proved an Ω(n/kklogn) lower bound for protocols achieving constant advantage under uniform input distribution.

Now the main gap between the upper bound [23] and the lower bound [29] is the extra klogn term. This gap becomes significant if kn. In this paper, we further improve the lower bound and close the gap.

1.1 Our results

We prove that any protocol that achieves constant advantage under uniform input distribution must communicate Ω(n/k+k) bits.

Theorem 2.

Let Π be a (k1)-round deterministic protocol for PCk where Alice speaks first such that

𝐏𝐫fA,fB[n]n[Π(fA,fB)=PCk(fA,fB)]2/3.

Then the communication complexity of Π is Ω(n/k+k).

By Yao’s minimax principle, it implies a lower bound for the (k1) round randomized communication complexity.

Corollary 3.

Every (k1)-round randomized protocol for PCk with error at most 1/3 (where Alice speaks first) has communication complexity Ω(n/k+k).

We observe there is still a (logn) gap between our lower bound and the protocol by [23]. We conjecture that our lower bound is tight and there is a chance to remove the logn factor in the upper bound side. A simple deterministic protocol with (k1) rounds and O(n) communication bits could be the following: Alice and Bob send the parity of fA(x) and fB(x) for all x[n] in the beginning. Hence they can skip the last round as they already know the parity. This simple protocol shows that [23]’s protocol is not tight when k=o(logn). We believe similar ideas could be extended for large k.

Applications

Given the connections between PCk and diverse applications [10, 22, 5, 16, 7, 24], our improved lower bounds automatically lead to several applications. We list two applications below.

Corollary 4 (Direct sum extension of pointer chasing).

The (k1)-round randomized communication complexity of PCk with d pairs of functions is Ω(dn/k2+d)

This corollary improves the previous Ω(dn/k3dklogn2d) lower bound presented in [10], which has applications in BFS trees streaming lower bound.

Corollary 5 (Exponential separations in local differential privacy).

Let A be a (k1)-round sequentially interactive ε-locally private protocol solving PCk with error probability γ1/3. Then the sample complexity of A is Ω(1eε(n/k+k)) and there is a k round protocol with sample complexity O~(klognε2).

This corollary improves the previous Ω(neεk2) lower bound for k<n/logn given by [16].

1.2 Gadgetless Lifting: A New Framework to Prove Communication Lower Bounds

The following two-step approach for proving communication lower bounds often appears in previous works (e.g., [11, 27]):

  1. 1.

    Identify a family of structured protocols.

  2. 2.

    Simulate general protocols by structured protocols and prove communication lower bounds for structured protocols.

This approach culminates in query-to-communication lifting theorems [12, 13, 6, 21].

Query-to-communication lifting theorems

Let f:Zn{0,1} be a function, and let g:X×YZ be a two-party gadget function. The goal is to prove communication lower bounds for the function fgn:Xn×Yn{0,1}. Indeed, all functions for which lower bounds are proven using the above approach can be written as fgn for appropriate f and g. For such functions, a communication protocol can always simulate a decision tree that computes f – such protocols consist of a natural family of structured protocols. Communication complexity for such protocols is essentially the query complexity of f, for which lower bounds are often easy to prove. Hence, the primary job is to show how to simulate general protocols by structured ones.

Though query-to-communication lifting is a beautiful framework, it requires a gadget function g since f is a one-party function. As a consequence, this framework only applies to lifted functions, namely, functions that can be written as fgn. Many important problems, such as pointer chasing, do not fall into this category; hence, lifting theorems do not apply in those cases.

To address this limitation, we propose a new framework called gadgetless lifting. We take a step back to the original approach, reconsidering the choice of structured protocols. In some cases, although the function is not a lifted function, there are simple and natural protocols. The crux of gadgetless lifting is how to decide the structured protocols. In this paper, we capture it as those protocols that “all shared useful information are local information”. For example, the protocol by [23] only share local information such as fA(x) or fB(x) for some x[n]. In lemma 11, we show that any protocol for PCk can be simulated by such protocols. Our proof is inspired by the structure-vs-pseudorandomness decomposition by Göös, Pitassi, and Watson [13] and Yang and Zhang [28], which is a powerful tool that emerged in the study of query-to-communication lifting theorems. Therefore, we call our method “gadgetless lifting”.

In the study of lifted functions, it has been shown that query-to-communication lifting theorems bypassed some fundamental barriers from previous methods. Similarly, gadgetless lifting can also bypass obstacles from existing methods. We discuss two of them below.

Avoiding the loss in round elimination method

Previously, the only method to prove round-communication trade-offs is the round elimination method [23]. In [23] and [29], the authors studied the pointer chasing problem via the round elimination method. Denote by 𝑴1,,𝑴t the messages sent in the first t rounds, and let 𝒁i be the pointer in the i-th round, i.e., , Zi=𝗉𝗍i(X,Y) where X,Y are uniformly chosen from [n]n. As is standard the round elimination method, [23, 29] analyzed the random variables

𝑹t=(𝑴1,,𝑴t,𝒁1,,𝒁t1) for tk.

They proved that 𝐇(𝑹k)Ω(n/k). Together with the fact that 𝐇(𝒁1,,𝒁k)=klogn, it implies that 𝐇(𝑴)Ω(n/kklogn). The (klogn) loss (or something similar) appears in many previous works that adopt round elimination-based [23, 18, 19, 14, 10, 29]. In this paper, we avoid the klogn loss via the gadgetless lifting.

Breaking square-root loss barrier in information complexity

Another popular method in proving communication lower bounds is by way of information complexity. However, as mentioned by Yahudayoff [29], entropy-based analyses are likely to induce a square-root loss barrier. This barrier usually comes from applying Pinsker’s inequality (or its variant) to bound statistical distance from a small entropy gap. As a consequence, many results such as [23] can only prove an Ω(n/k2klogn) lower bound.

As mentioned in [29], the square-root loss also appears in many works when using the entropy-based method to prove lower bounds. For example, it appears in the parallel repetition theorem and is related to the “strong parallel repetition” conjecture which is motivated by Khot’s unique games conjecture [17]. This loss also appears in direct-sum theorems [2] and direct-product theorems [4] in communication complexity.

[29] overcomes this square-root loss barrier by using a non-standard measurement called triangular discrimination. By contrast, our approach overcomes the barrier more naturally without using entropy.

Potential applications

We noticed that our method can also be naturally extended to multiparty settings such as the numbers in hand model. Moreover, some important open problems, such as round-communication tradeoff of bipartite matching problem [3] and set pointer chasing problem [10, 15], are difficult to solve using the round elimination method due to its inherent limitations. Our method offers the potential to solve these challenging problems.

2 Preliminaries

Notations

We use capital letters X to denote a set and use bold symbols like 𝑹 to denote random variables. Particularly, for a set X, we use 𝑿 to denote the random variable uniformly distributed over the set X. We use to denote sampling from a distribution or choosing an element from a set uniformly at random.

2.1 Density-Restoring Partition

Min-entropy and dense distribution

For a random variable 𝑿, we use supp(𝑿) to denote the support of 𝑿.

Definition 6 (Min-entropy and deficiency).

The min-entropy of a random variable 𝐗 is defined by

𝐇(𝑿):=minxsupp(𝑿)log(1𝐏𝐫[𝑿=x]).

Suppose that 𝐗 is supported on [n]J. We define the deficiency of 𝐗 as

𝐃(𝑿):=|J|logn𝐇(𝑿).

For IJ, x[n]J, let x(I)=𝖽𝖾𝖿(x(i))iI[n]I be the projection of x on coordinates in I.

Definition 7 (Dense distribution).

Let γ(0,1). A random variable 𝐗 supported on [n]J is said to be γ-dense if for all nonempty IJ, 𝐇(x(I))γ|I|logn.

The following lemma is the crux of the structure-vs-pseudorandomness method by [13]. It essentially says that a flat random variable could be decomposed into a convex combination of flat random variables with disjoint support and dense properties.

Lemma 8 (Density-restoring partition).

Let γ(0,1). Let X be a subset of [n]M and J[M]. Suppose that there exists an β[n]J¯ such that xX,x(J¯)=β. Then, there exists a partition X=X1X2Xr and every Xi is associated with a set IiJ and a value αi[n]Ii that satisfy the following properties.

  1. 1.

    xXi,x(Ii)=αi;

  2. 2.

    𝑿i(JIi) is γ-dense;

  3. 3.

    𝐃(𝑿i(JIi))𝐃(𝑿(J))(1γ)logn|Ii|+δi, where δi=𝖽𝖾𝖿log(|X|/|jiXj|).

The proof of this lemma, simple and elegant, is included in the appendix for completeness.

2.2 Communication Protocols

We recall basic definitions and facts about communication protocols.

Protocol Tree

Let X and Y be the input space of Alice and Bob respectively. A deterministic communication protocol Π is specified by a rooted binary tree. For every internal vertex v,

  • it has 2 children, denoted by Π(v,0) and Π(v,1);

  • v is owned by either Alice or Bob – we denote the owner by 𝗈𝗐𝗇𝖾𝗋(v);

  • every leaf node specifies an output.

Starting from the root, the owner of the current node 𝖼𝗎𝗋 partitions its input space into two parts X0 and X1, and sets the current node to Π(𝖼𝗎𝗋,b) if its input belongs to Xb.

Fact 9.

The set of all inputs that leads to an internal vertex v is a rectangle, denoted by Πv=Xv×YvX×Y.

The communication complexity of Π, denoted by CC(Π), is the depth of the tree. The round complexity of Π, is the minimum number k such that in every path from the root to some leaf, the owner switches at most (k1) times. Clearly, if a protocol has k round, then its communication complexity is at least k. We can safely make the following assumptions for any protocol Π:

  • Π has k rounds on every input; and

  • Π communicates CC(Π) bits on every input.

Indeed, for any protocol, we can add empty messages and rounds in the end, which boosts the communication complexity by a factor of 2.

3 Proof of Main Theorem

Theorem 10 (Main theorem, Theorem 2 restated).

Let Π be a (k1)-round deterministic protocol for PCk where Alice speaks first such that

𝐏𝐫fA,fB[n]n[Π(fA,fB)=PCk(fA,fB)]2/3.

Then the communication complexity of Π is Ω(n/k+k).

We use a decomposition and sampling process 𝖣𝖲, as shown in Algorithm  1, in our analysis. 𝖣𝖲 takes as input a protocol Π, and samples a rectangle R that is contained in Πv for some leaf node v. Our proof proceeds in three steps:

  1. 1.

    First, Section 3.1 analyzes crucial invariants during the running of 𝖣𝖲.

  2. 2.

    Next, Section 3.2 shows that the accuracy of Π is captured by a quantity called average fixed size, which is a natural quantity that arises in the running of 𝖣𝖲.

  3. 3.

    Finally, Section 3.3 proves that the average fixed size can be bounded from above by O(CC(Π)). Consequently, if Π enjoys high accuracy, we get a lower bound of CC(Π).

3.1 The Decomposition and Sampling Process

During the sampling process, we maintain a useful structure of R mainly by a partitioning-then-sampling mechanism: At the beginning, R is set to be the set of all inputs. Walking down the protocol tree, we decompose the rectangle into structured sub-rectangles; then we sample a decomposed rectangle with respect to its size. In the end, we arrive at a leaf node v and a subrectangle of Πv.

Algorithm 1 The decomposition and sampling process 𝖣𝖲.
Lemma 11 (Loop invariant).

Set γ=𝖽𝖾𝖿10.1logn. Then in the running of 𝖣𝖲(Π), we have the following loop invariants: After each iteration,

()

X×YΠv;

()

𝑿(JA),𝒀(JB) are γ-dense;

()

there exists some αA[n]JA¯,αB[n]JB¯ such that x(JA¯)=αA,y(JB¯)=αB for all xX,yY;

()

there exists some zr[n] such that ptr(fA,fB)=zr for all (fA,fB)X×Y.

Proof.

Item () is true because every time v is updated, X×Y is updated accordingly to a sub-rectangle of Πv and updating X×Y into its sub-rectangles does not violate this condition.

Since we applied density restoring partition at the end of each iteration, Item () and () is guaranteed by Lemma 8 and the way that X,Y,JA,JB are updated.

We prove the last item () by induction. Assume that the statement holds after the first (t1) iterations. WLOG, assume that at the beginning of the t-th iteration, v is owned by Alice. Consider the following two cases.

  • Case 1. Not a new round: Line 13 is not executed in the t-th iteration. Since r remains unchanged and we only update R to be a sub-rectangle of itself, the statement still holds.

  • Case 2. A new round begins: Line 13 is executed and r is increased by 1. Let ρ denote the value of r before Line 13, then after this iteration, we have r=ρ+1. The induction hypothesis guarantees that there exists some zρ1[n] such that

    𝗉𝗍ρ1(fA,fB)=zρ1 for all(fA,fB)X×Y.

    Due to the partition and the update in Line 11 and Line 12, |supp(𝑿(zρ1))|n/2. Hence, 𝑿(zρ1) cannot be γ-dense as we set γ=10.1logn. Observe that after the update in Line 17, 𝑿(JA) is γ-dense. Consequently, we must have zρ1JA¯, and by item (), there exists some zρ[n] such that fA(zρ1)=zρfAX. By definition, for all (fA,fB)X×Y,

    𝗉𝗍ρ(fA,fB)=fA(𝗉𝗍ρ1(fA,fB))=fA(zρ1)=zρ.

    This is exactly the same statement after the t-th iteration (as we have r=ρ+1).

The restricted rectangles in this loop invariant are inspired by the protocols of Nisan and Wigderson [23]. This lemma aims to capture the fact that Alice and Bob cannot get any additional useful information other than coordinate-wise information during their communication.

3.2 Relating Accuracy and Average Fixed Size

From Lemma 11 we know that the coordinates in JA¯ and JB¯ are fixed if we only look at the inputs in X×Y. Intuitively, the advantage of the protocol comes from such fixed coordinates, since the “alive” coordinates JA,JB are dense in the sense that 𝑿(JA),𝒀(JB) is γ-dense. This intuition is formalized in the following lemma.

Lemma 12 (Relating accuracy and avarage fixed size).

Let Π be a (k1)-round deterministic protocol where Alice speaks first. Then

𝐏𝐫fA,fB[n]n[Π(fA,fB)=PCk(fA,fB)]n1γ2+nγ(k1)𝐄(R,JA,JB)𝖣𝖲(Π)[|JA¯|+|JB¯|].

The proof of the lemma is by the following two claims. The first claim readily says that conditioned on the flag 𝖻𝖺𝖽 is not raised, Π has little advantage in the rectangle R output by 𝖣𝖲(Π). The second claim shows the probability that the flag is raised is bounded in terms of the average fixed size.

Claim 13.

If 𝖣𝖲(Π) outputs (R=X×Y,JA,JB) and 𝖻𝖺𝖽=False in the end, then

𝐏𝐫(fA,fB)R[Π(fA,fB)=PCk(fA,fB)]n1γ2.
Claim 14.

𝐏𝐫𝖣𝖲(Π)[𝖻𝖺𝖽=True]nγ(k1)𝐄(R,JA,JB)𝖣𝖲(Π)[|JA¯|+|JB¯|].

Next, we first prove Lemma 12 using the above two claims, and the proof of the claims is followed.

Proof of Lemma 12.

Note that in the running of 𝖣𝖲(Π), we always update R to a randomly chosen rectangle and the probability of each rectangle being chosen is proportional to its size. Consequently,

𝐏𝐫fA,fB[n]n[Π(fA,fB)=PCk(fA,fB)]
=𝐏𝐫(R,JA,JB)𝖣𝖲(Π),(fA,fB)R[Π(fA,fB)=PCk(fA,fB)]
𝐏𝐫𝖣𝖲(Π)[𝖻𝖺𝖽=True]+𝐏𝐫(R,JA,JB)𝖣𝖲(Π),(fA,fB)R[Π(fA,fB)=PCk(fA,fB)𝖻𝖺𝖽=False]
n1γ2+nγ(k1)𝐄(R,JA,JB)𝖣𝖲(Π)[|JA¯|+|JB¯|].

where the last step is by Claim 13 and Claim 14.

It remains to prove the two claims.

Proof of Claim 13.

WLOG, assume k1 is odd and the protocol always has k round. Let zk1 be the pointer guaranteed by the loop invariant (Lemma 11), i.e., ptk1(fA,fB)=zk1 for all (fA,fB)R. Since 𝖻𝖺𝖽=False, we have zk1JA. Again by the loop invariant, 𝐇(𝑿(zk1))γ. Moreover, since R is contained in some leaf node of Π, Π output the same answer in R, say b{0,1}. Consequently,

𝐏𝐫(fA,fB)R[Π(fA,fB)=PCk(fA,fB)] =𝐏𝐫fAX[fA(zk1)mod2=b]
σ[n]:σmod2=b𝐏𝐫fAX[fA(zk1)=σ]
n2nγ.

Proof of Claim 14.

Let denote the event that the flag 𝖻𝖺𝖽 is raised when r=+1 (i.e., when the -th round ends) for the first time. Clearly, 𝐏𝐫[𝖻𝖺𝖽=True]==1k1𝐏𝐫[]. It suffices to bound each 𝐏𝐫[].

Assume is odd, meaning that Alice speaks in the -th round. Let 𝖼𝗈𝗂𝗇 denote the randomness used for the first (1) rounds. Let X(1), JA(1),JB(1) be the sets X,JA,JB when executing 𝖣𝖲(Π) using 𝖼𝗈𝗂𝗇 until the -th round begins. Let z1 be the pointer promised by the invariant. For to happen, we must have 𝖻𝖺𝖽=False until the -th round begins, meaning that z1JA(1).

Note that the random variable 𝒛 exactly has the same distribution as 𝑿(1)(z1). This is because, in the -th round (i.e., until r steps to +1), we decompose X(1) into finer sets and update X to be one of them with probability proportional to their size. Therefore,

𝐏𝐫𝖼𝗈𝗂𝗇[t]=𝐏𝐫𝖼𝗈𝗂𝗇[𝒛JB(1)] =𝐏𝐫fA𝑿(1)[fA(𝒛1)JB(1)]
=σJB(1)¯𝐏𝐫fA𝑿(1)[fA(𝒛1)=σ]
|JB(1)¯|nγ,

where we fix 𝖼𝗈𝗂𝗇 and the probability runs over 𝖼𝗈𝗂𝗇, the randomness used afterward; the last inequality holds because z1JA(1) and 𝑿(1)(JA(1)) is γ-dense (by Item () in Lemma 11). Averaging over 𝖼𝗈𝗂𝗇, we get

𝐏𝐫𝖣𝖲(Π)[]𝐄𝖼𝗈𝗂𝗇[|JB(1)¯|]nγ𝐄(R,JA,JB)𝖣𝖲(Π)[|JB¯|]nγ,

where the second inequality holds because JB becomes smaller and smaller during the execution.

For even ’s, we analogously have 𝐏𝐫[]𝐄[|JA¯|]nγ, and hence the claim follows from union bound.

3.3 Average Fixed Size is Bounded by Communication

Now that the accuracy of a protocol Π is bounded from above by the average fixed size (i.e., 𝐄(R,JA,JB)𝖣𝖲(Π)[|JA¯|+|JB¯|]), in what follows we show that the average fixed size is at most O(CC(Π)). Formally, we prove that

Lemma 15.

Let Π be a (k1)-round deterministic protocol where Alice speaks first. Then

𝐄(R,JA,JB)𝖣𝖲(Π)[|JA¯|+|JB¯|]3CC(Π)(1γ)logn.
 Remark 16.

We shall set γ:=10.1logn and hence the right-handed side equals 30CC(Π).

Proof.

We shall prove this lemma by density increment argument. That is, we study the change of the density function

𝐃(R)=𝖽𝖾𝖿𝐃(𝑿(JA))+𝐃(𝒀(JB)). (1)

in each iteration. Let ϕ𝒕 be the value of 𝐃(R) at the end of the t-th iteration. Assume without loss of generality Alice speaks (i.e., 𝗈𝗐𝗇𝖾𝗋(v)=𝖠𝗅𝗂𝖼𝖾) in the t-th iteration.

We fix the random coins used for the first (t1) iterations and consider the updates in the current iteration.

  1. 1.

    First, X is partitioned into X=X0X1 according to Π. Then, X is updated to Xb with probability |Xb||X|. Consequently, 𝐃(𝑿(JA)) will increase as |X| shrinks, and in expectation (over the random choice of 𝒃) the increment is

    b{0,1}|Xb||X|log(|X||Xb|)1. (2)
  2. 2.

    Next, suppose that updating v leads to the switch of the owner, i.e., Line 13 is triggered. Since we also partition X into two parts and update X with probability proportional to the size of each part, the same argument applies. That is, taking expectation over the random choice of 𝒃, 𝐃(𝑿(JA)) increases by at most 1 in expectation.

  3. 3.

    Finally, we further partition X according to Lemma 8. Say X is partitioned into X=X1Xm and let I1,,Im be the index sets promised by Lemma 8; and for all j[m] we have

    𝐃(𝑿j(JAIj))𝐃(𝑿(JA))(1γ)logn|Ij|+δj,

    where δj=log(|X|/vjXv). With probability pj=𝖽𝖾𝖿|Xj|/|X|, we update X:=Xj and JA:=JAIj. Therefore, taking expectation over the random choice of 𝒋, the density function will decrease by

    𝐃(𝑿(JA))𝐄j𝒋[𝐃(𝑿j(JAIj))]𝐄j𝒋[(1γ)logn|Ij|δj]. (3)

    Note that δj=𝖽𝖾𝖿log1vjpv and thus

    𝐄j𝒋[δj]=j=1mpjlog1vjpj01log11xdx1. (4)

Let t1 be the σ-algebra generated by the random coins used for the first (t1) iterations. Let 𝜷t be the increment of |JA¯| and |JB¯| in the t-th iteration. Observe that 𝜷t=|I𝒋| by definition. By Equation 3 and Equation 4, taking expectation over random choice of 𝒋, 𝐃(𝑿(JA)) decrease by at least (1γ)logn𝐄[𝜷t|t1]1 due to the density restoring partition. Then

𝐄[ϕtϕt1]=𝐄[𝐄[ϕtϕt1|t1]]𝐄[1+𝜼𝒕((1γ)logn𝜷t1)], (5)

where 𝜼𝒕=𝖽𝖾𝖿𝟙[owner switches in the t-th iteration].

Write c=𝖽𝖾𝖿CC(Π) and assume we always have c iterations. 111Namely, Π communicates c bits on all inputs. In the beginning, ϕ0=𝐃([n]n×[n]n)=0. Since the density function is always non-negative by definition, we have ϕc0 and thus 𝐄[ϕcϕ0]0. On the other hand, by telescoping,

𝐄[ϕcϕ0]=t=1c𝐄[ϕtϕt1]2c+t=1c𝐄[𝜼t(1γ)logn𝜷𝒕],

where the inequality follows from Equation 5. Observe that t=1c𝜼t is at most k and t=1c𝜷t=|𝑱A¯|+|𝑱B¯| by definition. We conclude that

𝐄[|𝑱A¯|+|𝑱B¯|]=𝐄[t=1c𝜷t]2c+k(1γ)logn3c(1γ)logn,

as desired.

Proving the main theorem

Now our main theorem easily follows from the two lemmas.

Proof of Theorem 2.

Set γ=𝖽𝖾𝖿10.1logn. By Lemma 15 and Lemma 12, we get

𝙰𝚌𝚌𝚞𝚛𝚊𝚌𝚢(Π) =𝖽𝖾𝖿𝐏𝐫fA,fB[n]n[Π(fA,fB)=PCk(fA,fB)]
n1γ2+nγ(k1)3CC(Π)(1γ)logn
0.54+1.08(k1)n30CC(Π),

where we use n1γ20.54,nγ1.08n. Since we assumed 𝙰𝚌𝚌𝚞𝚛𝚊𝚌𝚢(Π)2/3, we conclude that

CC(Π)2/30.541.0830nk1>0.0039nk1=Ω(n/k).

We also trivially have CC(Π)k1 as Π has (k1) rounds; putting it together we conclude that CC(Π)=Ω(n/k+k).

References

  • [1] Sepehr Assadi, Yu Chen, and Sanjeev Khanna. Polynomial pass lower bounds for graph streaming algorithms. In Proceedings of the 51st Annual ACM SIGACT Symposium on theory of computing, pages 265–276, 2019. doi:10.1145/3313276.3316361.
  • [2] Boaz Barak, Mark Braverman, Xi Chen, and Anup Rao. How to compress interactive communication. In Proceedings of the forty-second ACM symposium on Theory of computing, pages 67–76, 2010. doi:10.1145/1806689.1806701.
  • [3] Joakim Blikstad, Jan Van Den Brand, Yuval Efron, Sagnik Mukhopadhyay, and Danupon Nanongkai. Nearly optimal communication and query complexity of bipartite matching. In 2022 IEEE 63rd Annual Symposium on Foundations of Computer Science (FOCS), pages 1174–1185. IEEE, 2022. doi:10.1109/FOCS54457.2022.00113.
  • [4] Mark Braverman, Anup Rao, Omri Weinstein, and Amir Yehudayoff. Direct products in communication complexity. In 2013 IEEE 54th Annual Symposium on Foundations of Computer Science, pages 746–755. IEEE, 2013. doi:10.1109/FOCS.2013.85.
  • [5] Clément L Canonne and Tom Gur. An adaptivity hierarchy theorem for property testing. computational complexity, 27:671–716, 2018. doi:10.1007/S00037-018-0168-4.
  • [6] Arkadev Chattopadhyay, Yuval Filmus, Sajin Koroth, Or Meir, and Toniann Pitassi. Query-to-communication lifting for bpp using inner product. In 46th International Colloquium on Automata, Languages, and Programming (ICALP 2019). Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPIcs.ICALP.2019.35.
  • [7] Xi Chen, Christos Papadimitriou, and Binghui Peng. Memory bounds for continual learning. In 2022 IEEE 63rd Annual Symposium on Foundations of Computer Science (FOCS), pages 519–530. IEEE, 2022. doi:10.1109/FOCS54457.2022.00056.
  • [8] Carsten Damm, Stasys Jukna, and Jiří Sgall. Some bounds on multiparty communication complexity of pointer jumping. Computational Complexity, 7:109–127, 1998. doi:10.1007/PL00001595.
  • [9] Pavol Duris, Zvi Galil, and Georg Schnitger. Lower bounds on communication complexity. In Proceedings of the sixteenth annual ACM symposium on Theory of computing, pages 81–91, 1984. doi:10.1145/800057.808668.
  • [10] Joan Feigenbaum, Sampath Kannan, Andrew McGregor, Siddharth Suri, and Jian Zhang. Graph distances in the data-stream model. SIAM Journal on Computing, 38(5):1709–1727, 2009. doi:10.1137/070683155.
  • [11] Mikael Goldmann and Johan Håstad. A simple lower bound for monotone clique using a communication game. Information Processing Letters, 41(4):221–226, 1992. doi:10.1016/0020-0190(92)90184-W.
  • [12] Mika Göös, Toniann Pitassi, and Thomas Watson. Deterministic communication vs. partition number. In 2015 IEEE 56th Annual Symposium on Foundations of Computer Science, pages 1077–1088. IEEE, 2015. doi:10.1109/FOCS.2015.70.
  • [13] Mika Göös, Toniann Pitassi, and Thomas Watson. Query-to-communication lifting for bpp. In 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS), pages 132–143, 2017. doi:10.1109/FOCS.2017.21.
  • [14] Sudipto Guha and Andrew McGregor. Lower bounds for quantile estimation in random-order and multi-pass streaming. In Automata, Languages and Programming: 34th International Colloquium, ICALP 2007, Wrocław, Poland, July 9-13, 2007. Proceedings 34, pages 704–715. Springer, 2007. doi:10.1007/978-3-540-73420-8_61.
  • [15] Venkatesan Guruswami and Krzysztof Onak. Superlinear lower bounds for multipass graph processing. Algorithmica, 76:654–683, 2016. doi:10.1007/S00453-016-0138-7.
  • [16] Matthew Joseph, Jieming Mao, and Aaron Roth. Exponential separations in local differential privacy. In Proceedings of the Fourteenth Annual ACM-SIAM Symposium on Discrete Algorithms, pages 515–527. SIAM, 2020. doi:10.1137/1.9781611975994.31.
  • [17] Subhash Khot. On the power of unique 2-prover 1-round games. In Proceedings of the thiry-fourth annual ACM symposium on Theory of computing, pages 767–775, 2002. doi:10.1145/509907.510017.
  • [18] Hartmut Klauck. On quantum and probabilistic communication: Las vegas and one-way protocols. In Proceedings of the thirty-second annual ACM symposium on Theory of computing, pages 644–651, 2000. doi:10.1145/335305.335396.
  • [19] Hartmut Klauck, Ashwin Nayak, Amnon Ta-Shma, and David Zuckerman. Interaction in quantum communication. IEEE Transactions on Information Theory, 53(6):1970–1982, 2007. doi:10.1109/TIT.2007.896888.
  • [20] Maria Klawe, Wolfgang J Paul, Nicholas Pippenger, and Mihalis Yannakakis. On monotone formulae with restricted depth. In Proceedings of the sixteenth annual ACM symposium on Theory of computing, pages 480–487, 1984.
  • [21] Shachar Lovett, Raghu Meka, Ian Mertz, Toniann Pitassi, and Jiapeng Zhang. Lifting with sunflowers. Leibniz international proceedings in informatics, 215, 2022. doi:10.4230/LIPICS.ITCS.2022.104.
  • [22] Danupon Nanongkai, Atish Das Sarma, and Gopal Pandurangan. A tight unconditional lower bound on distributed randomwalk computation. In Proceedings of the 30th annual ACM SIGACT-SIGOPS symposium on Principles of distributed computing, pages 257–266, 2011. doi:10.1145/1993806.1993853.
  • [23] Noam Nisan and Avi Widgerson. Rounds in communication complexity revisited. In Proceedings of the twenty-third annual ACM symposium on Theory of computing, pages 419–429, 1991. doi:10.1145/103418.103463.
  • [24] Binghui Peng, Srini Narayanan, and Christos Papadimitriou. On limitations of the transformer architecture. arXiv preprint, 2024. doi:10.48550/arXiv.2402.08164.
  • [25] Stephen J Ponzio, Jaikumar Radhakrishnan, and Srinivasan Venkatesh. The communication complexity of pointer chasing. Journal of Computer and System Sciences, 62(2):323–355, 2001. doi:10.1006/JCSS.2000.1731.
  • [26] Anup Rao and Amir Yehudayoff. Communication Complexity: and Applications. Cambridge University Press, 2020.
  • [27] Ran Raz and Pierre McKenzie. Separation of the monotone nc hierarchy. In Proceedings 38th Annual Symposium on Foundations of Computer Science, pages 234–243. IEEE, 1997. doi:10.1109/SFCS.1997.646112.
  • [28] Guangxu Yang and Jiapeng Zhang. Communication lower bounds for collision problems via density increment arguments. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, pages 630–639, 2024. doi:10.1145/3618260.3649607.
  • [29] Amir Yehudayoff. Pointer chasing via triangular discrimination. Combinatorics, Probability and Computing, 29(4):485–494, 2020. doi:10.1017/S0963548320000085.

Appendix A Appendix

The following lemma and proof are from Lemma 5 in [13].

Lemma 17 (Lemma 8 restated).

Let γ(0,1). Let X be a subset of [n]M and J[M]. Suppose that there exists an β[n]J¯ such that xX,x(J¯)=β. Then, there exists a partition X=X1X2Xr and every Xi is associated with a set IiJ and a value αi{0,1}Ii that satisfy the following properties.

  1. 1.

    xXi,x(Ii)=αi;

  2. 2.

    𝑿i(JIi) is γ-dense;

  3. 3.

    𝐃(𝑿i(JIi))𝐃(𝑿(J))(1γ)logn|Ii|+δi, where δi=𝖽𝖾𝖿log(|X|/|jiXj|).

Proof.

We prove it by a greedy algorithm as follows.

Algorithm 2 Greedy Algorithm.

Item 1 is guaranteed by the construction of Xi and Ii.

We prove Item 2 by contradiction. Assume towards contradiction that 𝑿i(JIi) is not γ-dense for some i. By definition, there is a nonempty set KJIi and β[n]K violating the min-entropy condition, namely, 𝐏𝐫[𝑿(K)=β]>nγ|K|. Write Xi=𝖽𝖾𝖿jiXi. Then

𝐏𝐫[𝑿i(IiK)=(αi,β)]=𝐏𝐫[𝑿i(Ii)=αi]𝐏𝐫[𝑿i(K)=β]>nγ|Ii|nγ|K|=nγ|IiK|,

where the first equality holds as (𝑿i|𝑿i(Ii)=αi)=𝑿i. However, this means at moment that Ii is chosen, the set IiKJ also violates the min-entropy condition (witnessed by (αi,β)), contradicting the maximality of Ii.

Finally, Item 3 is proved by straightforward calculation:

𝐃(𝑿i(JIi)) =|JIi|lognlog|Xi|
(|J|logn|Ii|logn)log(|Xi|nγ|Ii|)
=(|J|lognlog|X|)(1γ)|Ii|logn+log(|X||Xi|)
=𝐃(𝑿(J))(1γ)|Ii|logn+δi.