Abstract 1 Introduction 2 Preliminaries 3 The Lower Bound 4 Applications to Streaming References

Optimal Communication Complexity of Chained Index

Janani Sundaresan ORCID Cheriton School of Computer Science, University of Waterloo, Canada
Abstract

We study the chain communication problem introduced by Cormode et al. [ICALP 2019]. For k1, in the chainn,k problem, there are k string and index pairs (Xi,σi) for i[k] such that the value at position σi in string Xi is the same bit for all k pairs. The input is shared between k+1 players as follows. Player 1 has the first string X1{0,1}n, player 2 has the first index σ1[n] and the second string X2{0,1}n, player 3 has the second index σ2[n] along with the third string X3{0,1}n, and so on. Player k+1 has the last index σk[n]. The communication is one way from each player to the next, starting from player 1 to player 2, then from player 2 to player 3 and so on. Player k+1, after receiving the message from player k, has to output a single bit which is the value at position σi in Xi for any i[k]. It is a generalization of the well-studied index problem, which is equivalent to chainn,2.

Cormode et al. proved that the chainn,k problem requires Ω(n/k2) communication, and they used it to prove streaming lower bounds for the approximation of maximum independent sets. Subsequently, Feldman et al. [STOC 2020] used it to prove lower bounds for streaming submodular maximization. However, it is not known whether the Ω(n/k2) lower bound used in these works is optimal for the problem, and in fact, it was conjectured by Cormode et al. that Ω(n) bits are necessary.

We prove the optimal lower bound of Ω(n) for chainn,k when k=o(n/logn) as our main result. This settles the open conjecture of Cormode et al., barring the range of k=Ω(n/logn). The main technique is a reduction to a non-standard index problem where the input to the players is such that the answer is biased away from uniform. This biased version of index is analyzed using tools from information theory. As a corollary, we get an improved lower bound for approximation of maximum independent set in vertex arrival streams via a reduction from chain directly.

Keywords and phrases:
communication complexity, index communciation problem
Funding:
Janani Sundaresan: Supported in part by Sepehr Assadi’s Sloan Research Fellowship and startup grant from University of Waterloo.
Copyright and License:
[Uncaptioned image] © Janani Sundaresan; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Communication complexity
Related Version:
Full Version: https://arxiv.org/abs/2404.07026
Acknowledgements:
The author is thankful to Sepehr Assadi for introducing them to the problem and for insightful discussions on the proof. The author would also like to thank Parth Mittal for useful comments, Christian Konrad for introducing them to the Augmented Chain problem, and the anonymous reviewers of ITCS 2025 for helpful comments and suggestions. The author is very grateful to Mi-Ying Huang, Xinyu Mao, Guangxu Yang and Jiapeng Zhang for an illuminating discussion about the problem. They pointed out an important flaw in an earlier version of this work, and the discussion was instrumental for the new proofs in the current version.
Editors:
Raghu Meka

1 Introduction

The index problem is one of the foundational problems in communication complexity. For n1, in the indexn problem, there are two players Alice and Bob. Alice has a string X{0,1}n and Bob has an index σ[n], and Bob has to output the value of X at position σ. If the communication is one-way from Alice to Bob, it is easy to show that Alice needs to send Ω(n) bits to get any constant advantage [1, 30]. This problem has been well-studied in multiple settings, and we know tight trade-offs in the two-party communication model for communication complexity [34], information complexity [24], and quantum communication complexity [8, 24].

Among the numerous applications of communication complexity, one that is of interest to us is proving lower bounds for streaming algorithms. index and its variants, in particular, have been quite useful in this context, for example, in [23, 18, 20, 21, 16, 10]. This is by no means an exhaustive list.

In this paper, we study a natural generalization of index, called chained index (chainn,k for n,k1) introduced by [12]. There are k different instances of indexn, correlated so that they have the same answer. They are “chained” together, where each player holds the index to the previous instance, and also the string for the next instance.

Definition 1 (Informal).

In chainn,k, there are k instances of indexn, all with the same answer. Players 1 and 2 take on the role of Alice and Bob respectively in the first instance, players 2 and 3 take on the role of Alice and Bob respectively for the second instance, and so on, all the way till players k and k+1 for the last instance.

Communication is one-way from each player to the next in ascending order. The last player has to output the answer. The communication cost is the total number of bits in all the messages sent by the players. See Figure 1 for an illustration.

Figure 1: An illustration of the chainn,k problem with k correlated sub-instances of indexn from Definition 1. The arrows illustrate that the message is from Pi to Pi+1 for i[k].

In [12], a reduction from chain was employed to get a lower bound for approximation of maximum independent sets in vertex arrival streams. Before its introduction, [28] used the problem implicitly to get a lower bound of (11/e) in the approximation factor for maximum matching in O~(n) space in vertex arrival streams.

The problem has been used by the breakthrough result of [19] to study the multi-party communication complexity of submodular maximization. They proved that any randomized p-party protocol which maximizes a monotone submodular function f:{0,1}N, subject to a cardinality constraint of at most p and an approximation factor of at least (1/2+ϵ), uses Ω(Nϵ/p3) communication. This also gave a lower bound for streaming submodular maximization. The chain problem was used by [17] also for similar purposes, but subject to stronger matroid constraints. [7] used a reduction from chain to prove lower bounds for interval independent set selection in streams of split intervals.

We do not know tight bounds for the communication complexity of the chain problem, despite finding varied applications of it. There is a trivial protocol of O(n) bits, where any player can send the entire string to the next player who holds the index. Another simple protocol is for each player to send O(n/k) bits randomly sampled from their strings using public randomness, and with constant probability, in at least one of the k instances, we send the special bit to the player holding the index. However, this still takes Ω(n) total bits of communication.

In [12], they prove a lower bound of Ω(n/k2) for any k1 through a reduction from conservative multi-party pointer jumping problem, introduced by [14]. They state without proof that a stronger lower bound of Ω(n/k) can be obtained for a restricted range of kO((n/logn)1/4). They posed the following conjecture on the optimal communication lower bound.

Conjecture 2 (​​[12]).

Any protocol that solves chainn,k requires Ω(n) bits of communication.

[19] made some progress on 2 by showing that among all the messages sent by the players, there is at least one message with Ω(n/k2) bits for every k1. But the original conjecture is still open, and this is the focus of our work.

1.1 Our Results

We settle 2 almost fully by proving the optimal lower bound of Ω(n) barring the corner case of when k is too large. As far as we know, this corner case is not a focus for existing reductions from chainn,k.

Theorem 3.

For any n,k1, any protocol for chainn,k with probability of success at least 2/3, requires Ω(nklogn) total bits of communication.

Therefore, as long as k=o(n/logn), we get the optimal Ω(n) lower bound from Theorem 3.

The proof of Theorem 3 can be found in Section 3. We prove the lower bound in the more general blackboard model of communication, instead of private messages between players (see Section 2.1 for details). The main idea is to analyze indexn where Alice and Bob already have some prior advantage in guessing the answer. We elaborate on our techniques in Section 1.2.

As a direct corollary of Theorem 3, we get improvements in streaming lower bounds in [12, 19] through reductions from chain immediately. In particular, we get that any algorithm which α-approximates the size of a maximum independent set in vertex arrival streams requires Ω(n2/α5logn) space, while the previous bound was Ω(n2/α7) in [12]. We present the implications of our result in Section 4.

A further generalization of chain, called Augmented Chain was defined in [15]. Here, instances of Augmented Index are chained together instead. Our lower bound of Ω(nklogn) can be extended to Augmented Chain also, and the details are covered in Section 3.4.

1.2 Our Techniques

In this subsection, we give an overview of the challenges in proving the lower bound and a summary of our techniques. We start by going over the prior techniques.

Prior Techniques

We will briefly talk about the technique used in [19] to prove that there is at least one player who sends Ω(n/k2) bits for any k1. We will argue that these techniques can be extended to proving a lower bound of Ω(n/k) for the total number of bits, but not all the way to Ω(n) bits.

The first step in proving a lower bound for chainn,k in [19] is a decorrelation step – the k instances of indexn have the same answer, and they remove this correlation with a hybrid argument. These arguments have been used extensively in the literature (see e.g., [26, 3, 27, 6]). Intuitively, any protocol that tries to solve chainn,k may attempt to solve “many” of the instances of indexn, albeit each with a “small” advantage over 1/2, in the hope that the “small” advantages may accrue to get a constant probability of success overall (taking advantage of the fact that all the instances of indexn have the same answer). The hybrid argument is a way to reduce proving a lower bound on the overall problem, to proving a lower bound for k different indexn problems against these low advantages.

Let us assume, for simplicity, that the protocol tries to get an advantage of Ω(1/k) in each instance of indexn, to get a constant total advantage. We can prove that any protocol that gets an advantage of Ω(1/k) for indexn uses Ω(n/k2) bits of communication using basic tools from information theory [9] (this is quite standard, see e.g., [2] for a direct proof), and this is known to be tight. Therefore, for k instances, we get a lower bound of Ω(n/k) bits in total. Now, we will argue why this is not the optimal lower bound.

On one hand, for indexn, it is known that for any δ(0,1/2), there is a protocol that uses O(nδ2) bits of communication to get a probability of success 1/2+δ. This means that indexn can be solved with advantage Ω(1/k) in O(n/k2) bits; in other words, each “hybrid step” of the previous lower bound argument is optimal. So, then, why can we not get a good protocol for chainn,k by running the protocol for indexn with δ=1/k on all k instances? This protocol would have O(n/k) bits of communication in total. The reason is that the 1/k small advantages do not add up as we would like them to. To illustrate this, we will briefly talk about the protocol that gets δ advantage in O(nδ2) bits for indexn.

Protocol for index𝒏

First, we will sketch a protocol for δ=1/n that uses O(1) bits of communication. Let us imagine that the input X is chosen uniformly at random from {0,1}n and the index σ is chosen uniformly at random from [n]. Then, Alice finds the majority bit from her string X and sends it to Bob. Bob just outputs the bit sent by Alice. We know, from simple anti-concentration bounds on the binomial distribution, that the number of indices with the majority bit in X is at least n/2+cn with constant probability for some appropriate constant c. The protocol succeeds if Bob holds the index σ to a majority bit, and this happens with probability at least 12+cn.

The assumption that the input X is chosen uniformly at random from {0,1}n can be removed using public randomness. Alice and Bob collectively sample a random string A{0,1}n, and Alice changes her input to XA so that each bit is 0 or 1 with equal probability. Similarly, the assumption that the index σ is chosen uniformly at random from [n] can be removed by Alice and Bob sampling a random permutation of [n]. Alice permutes string X according to this permutation, and Bob changes his input to the index that the permutation maps σ to. This is termed as the self-reducibility property of index.

We can also extend the protocol to any δ by partitioning the string X into nδ2 blocks at random using public randomness and sending the majority bit in each block. Bob knows the block that his input index σ belongs to as public randomness is used.

Challenges for chain𝒏,𝒌

If we use the protocol we described for chainn,k, we are left with k bits from each instance of indexn, which may be the correct answer to the problem with probability 1/2+Θ(1/k), but, the variance of each of these bits is 1/4Θ(1/k2). We have a protocol for chainn,k, where, out of the k instances of indexn, it finds the right answer for k/2+Θ(1) instances in expectation. However, the standard deviation of the number of right answers is k/2, which is enough to mask the Θ(1) improvement over k/2 we get in expectation. If we use a hybrid argument over the total variation distance, each of the smaller “hybrid steps” is optimal, whereas the overall lower bound is not optimal. We cannot achieve a lower bound stronger than Ω(n/k).

Our Solution

Instead of keeping track of progress in terms of advantage gained in guessing the answer, we directly keep track of the “information” the message reveals about the answer. Formally, this translates to the change in entropy of the answer, after each successive message. Initially, the players have no information about the answer, and it is uniform over {0,1} (the entropy is 1). Any protocol with a large enough probability of success must reduce the entropy of the answer by a large factor (see Fano’s inequality in Proposition 6). We prove that after each message, the entropy is reduced only by an additive factor linear in the length of the message, by a reduction to the index problem. This will give us a lower bound on the total length of the messages.

For indexn, in any protocol with probability of success at least ε, the entropy of the answer conditioned on the message is at most H2(ε) where H2(x)=xlog(1/x)+(1x)log(1/(1x)) is the binary entropy function. In the standard version where the initial entropy is 1, the reduction in entropy is 1H2(ε), and it is known that the protocol requires Ω(n(1H2(ε))) communication [24]. This is not sufficient for our application due to the following reason: after the message of 𝒫1, when 𝒫2 and 𝒫3 attempt to solve indexn, they already have some prior advantage that the message of 𝒫1 gives them. Therefore, we need to analyze indexn when the answer is not uniform over {0,1}.

Biased Index

We define the biased index problem, parametrized by θ[1/2,1/2]. Alice and Bob receive input Y{0,1}n and ρ[n] respectively, such that the value at position ρ in Y (denoted by Y(ρ)) is 1 with probability 1/2+θ and 0 otherwise. Alice sends a message M to Bob and Bob has to output Y(ρ). The initial entropy of the answer is H2(1/2+θ). We prove that entropy of Y(ρ) conditioned on M is smaller by at most O((|M|+logn)/n) compared to the initial H2(1/2+θ), which is our main contribution (see Lemma 12).

We can show that the entropy of input to Alice, i.e. the random string Y, in such a distribution is at least Ω(nH2(1/2+θ)). Hence, after a message of length s from Alice, the entropy of string Y reduces to Ω(nH2(1/2+θ)s). For a randomly chosen position in Y after conditioning on the message, the entropy is at least H2(1/2+θ)s/n, which gives a lower bound on s.

In the distribution given to Alice and Bob, however, ρ is not chosen uniformly at random, and in fact, is correlated with the distribution of Y. Such versions of index where the distributions of Alice and Bob are correlated have been studied before (see Sparse Indexing in Appendix A of [5] and Section 3.3 of [36]). This correlation is the main issue in analyzing biased index with information theoretic tools.

Adapted from the approach in Appendix A of [5], we restrict the randomness in Y to a fixed set of indices of a carefully chosen size (based on θ), and break this correlation. The loss in entropy of Y is not significant enough to hinder us, and the restriction then allows us to use standard information theoretic tools to analyze biased index (see Section 3.3 for more details).

Independent and Concurrent Work

Independently and concurrently of this work, [33] made progress on 2. They showed a lower bound of Ω(n/k+n) for oblivious protocols (where the length of the message sent by each player does not depend on the input), and a lower bound of Ω(n/kk) for general protocols. We show a lower bound of Ω(nklogn) for all protocols.111The authors of [33] pointed out an important flaw in the arguments of an earlier version of this work which was posted at around the same time as [33]. This flaw was subsequently fixed in the current version using a global change to the original argument, which now recovers optimal result for k=o(n/logn).

Quantitatively, our lower bounds are a factor of almost k stronger, and are optimal for k=o(n/logn); moreover, our lower bound also holds for the Augmented Chain problem of [15]. In terms of techniques, however, the two works are entirely disjoint: their proof is based on a new method of analysis through min-entropy and we use information theoretic approaches.

2 Preliminaries

In this section, we will present the required notation and definitions for our proof.

Notation

For any tuple A=(A1,A2,,Am) of m items, we use A<i to denote the tuple (A1,A2,,Ai1) for all i[m]. We use sans-serif font to denote random variables. For any random variable 𝖠, we use A𝖠 to denote any A sampled from the distribution of the random variable 𝖠.

For any string X{0,1}n, we use X(σ) to denote the bit at position σ in X for σ[n]. We use X(<σ) to denote the string of σ1 bits preceding X(σ) in X. We use X(S) for any S[n] to denote the bits at positions in set S.

For any x[0,1], we use H2:[0,1][0,1] to denote the binary entropy function.

H2(x)=xlogx(1x)log(1x).

We need the following standard approximation of binomial coefficients (see Lemma 7 of Chapter 10 in [32]).

Fact 3 (c.f. [32]).

For any p1 and any q[p1], we have,

2pH2(q/p)n8πq(pq)(pq)2pH2(q/p)n2πq(pq)

2.1 Communication Complexity Model

We use the standard number-in-hand multi-party model of communication. Only the basic definitions are given in this subsection. More details can be found in textbooks on communication complexity [35, 31].

For any k1, let f be a function from 𝒜1×𝒜2××𝒜k to {0,1}. There are k players 𝒫1,𝒫2,,𝒫k where 𝒫i gets an input ai𝒜i for i[k]. There is a shared blackboard visible to all the players. The players have access to a shared tape of random bits, along with their own private randomness. In any protocol π for f, the players send a message to the blackboard in increasing order (𝒫1 sends a message followed by 𝒫2, and so on till 𝒫k). The last player 𝒫k, after all the messages are sent, outputs a single bit denoted by π(a1,a2,,ak). Protocol π is said to solve f with probability of success at least 1δ if, for all i[k], for any choice of ai𝒜i, we have,

Pr[π(a1,a2,,ak)f(a1,a2,,ak)]δ.

The communication cost of a protocol π is defined as the worst case total communication of all the players on the blackboard at the end of the protocol.

Definition 4.

The randomized communication complexity of f, with probability of error δ, is defined as the minimum communication cost of any protocol which solves f with probability of success at least 1δ.

2.2 Information Theoretic Tools

Our proof relies on tools from information theory, and we state the basic definitions and the inequalities we need in this section. Proofs of the statements and more details can be found in Chapter 2 of a textbook on information theory by Cover and Thomas [13].

Definition 5 (Shannon Entropy).

For any random variable 𝖷 over support 𝒜, the Shannon entropy of 𝖷, denoted by (𝖷) is defined as,

(𝖷)=A𝒜Pr[𝖷=A]log(1/Pr[𝖷=A]).

For any event we define (𝖷) in the same way, as the entropy of distribution of 𝖷 conditioned on the event . For any two random variables 𝖷 and 𝖸, the entropy of 𝖷 conditioned on 𝖸, denoted by (𝖷𝖸) is defined as,

(𝖷𝖸)=𝔼Y𝖸(𝖷𝖸=Y).
Fact 5.

We know the following about entropy and mutual information:

  1. 1.

    For any random variable 𝖷, the entropy obeys the bound: 0(𝖷)log2(|𝒳|) where 𝒳 is the support of 𝖷.

  2. 2.

    For any two random variables 𝖷,𝖸, (𝖷𝖸)(𝖷) with equality holding iff 𝖷𝖸.

  3. 3.

    Chain Rule of Entropy: For m1 and any tuple of random variables 𝖷=(𝖷1,𝖷2,,𝖷m), (𝖷)=i[m](𝖷i𝖷<i).

  4. 4.

    Subadditivity of entropy: For m1 and any tuple of random variables 𝖷=(𝖷1,𝖷2,,𝖷m), (𝖷)i[m](𝖷i).

We also need the following proposition, which relates entropy to the probability of correctness while estimating a random variable.

Proposition 6 (Fano’s inequality).

Given a binary random variable 𝖷 and an estimator random variable Y and a function g such that g(Y)=X with probability at least 1δ for δ<1/2,

(𝖷𝖸)H2(δ).

This concludes our preliminaries section.

3 The Lower Bound

In this section, we will prove our lower bound of Ω(n) on the communication complexity of the chainn,k problem for k=o(n/logn). Let us formally define the chainn,k communication problem first.

Definition 7.

The chainn,k communication problem is defined as follows. Given k+1 players 𝒫i for i[k+1] where,

  • 𝒫i has a string Xi{0,1}n for each i[k], and,

  • 𝒫i for 1<ik+1 has an index σi1[n],

such that,

Xi(σi)=z,

for some bit z{0,1}. The players have a blackboard visible to all the parties. For i[k] in ascending order, 𝒫i sends a single message Mi, after which the index σi is revealed to the blackboard at no cost. 𝒫k+1 has to output whether z is 0 or 1. Refer to Figure 2 for an illustration.

Figure 2: An illustration of the chainn,k problem from Definition 7. The solid arrows illustrate that player 𝒫i writes a message Mi to the board. The dashed arrows indicate that 𝒫i can read the contents of the board. It also shows the order in which the messages are sent by the players and indices are released.

Let us recall the statement of our main result.

Theorem 3 (restated).

For any n,k1, any protocol for chainn,k with probability of success at least 2/3, requires Ω(nklogn) total bits of communication.

We give our hard distribution for chainn,k in Section 3.1 and give the proof of Theorem 3 in Section 3.2 except for the analysis of biased index, which is given in Section 3.3. Lastly, we extend the arguments to Augmented Chain in Section 3.4.

3.1 Setting Up the Problem

In this subsection, we start by defining the notation for our proof, and we describe the input distributions to chainn,k.

The input hard distribution is as follows. Let {0,1}n be the subset of strings where the number of ones is exactly equal to n/2.

Distribution 𝒟 for chainn,k:

  1. 1.

    Pick a bit z uniformly at random from {0,1}.

  2. 2.

    For each i[k], sample (Xi,σi) uniformly at random from ×[n] and independently conditioned on Xi(σi)=z.

Notation

We use 𝖷i to denote the random variable corresponding to string Xi and 𝝈i to denote the random variable corresponding to index σi for i[k]. To denote the random variable corresponding to the first i1 strings and indices, we use 𝖷<i and 𝝈<i respectively.

We use 𝖬i to denote the random variable corresponding to the message Mi sent by 𝒫i to the blackboard. We use M=(M1,M2,,Mk) to denote the tuple containing the messages of all the players, and 𝖬 to denote the random variable corresponding to M.

Let π be a deterministic protocol for chainn,k with probability of success at least 2/3 when the input is distributed according to 𝒟. We use Γ to denote the random variable corresponding to the contents of the blackboard (referred to as a transcript), and we use γ to also denote transcripts sampled from Γ. We use Γi to denote the random variable of the tuple (Mi,σi) for i[k]. We use π(γ<i,Xi) to denote the output of 𝒫i when the contents of the blackboard are γ<i and the input is Xi for i[k].

Let s be the total length of all the messages sent by the players in γ. We assume that the total length of the messages is exactly s by padding. For any random variable 𝖠, we use 𝒟(𝖠) to denote the distribution of the random variable 𝖠, as the input is distributed according to 𝒟. We replace 𝒟(𝖠𝖡=b) with 𝒟(𝖠b) for ease of readability whenever it is clear from context.

Let 𝖹 denote the random variable corresponding to bit z. The bit z corresponds to the answer to chainn,k. We show the lower bound of Ω(nklogn) for distinguishing between the case when z=0 and z=1 based on the contents of the blackboard.

We need one important observation about the distribution 𝒟.

Observation 8.

For any i[k], random variable 𝖷i,𝛔i is independent of Γ<i conditioned on 𝖹.

Proof.

For any fixed value of 𝖹, 𝖷i,𝝈i are chosen uniformly at random from ×[n] such that 𝖷i(𝝈i)=𝖹. This choice is independent of any 𝖷j,𝝈j with ji, and thus independent of Γ<i.

We are ready to proceed with the proof of our main theorem.

3.2 Proof of Lower Bound

We start by showing that in any successful protocol, the entropy of the distribution of 𝖹 conditioned on the message must be small.

Claim 9 (The transcript reveals information about 𝖹.).
(𝖹Γ)24/25.
Proof.

We know that 𝒫k+1 successfully finds the value of z with probability of success at least 2/3 using the transcript. Thus, using Fano’s inequality in Proposition 6, we have,

(𝖹Γ)H2(2/3)24/25.

The main part of the proof is that we show a lower bound the entropy of 𝖹 conditioned on the transcripts using the entropy of the message.

Lemma 10.

For any protocol π,

112n((𝖬)+klogn)(𝖹Γ).

Before we prove Lemma 10, we can easily show that it implies Theorem 3.

Proof of Theorem 3.

Combining Lemma 10 with Claim 9, we get,

112n((𝖬)+klogn)(𝖹Γ)2425.

This gives that,

(𝖬)n2512klogn.

We know from Section 2.2-(1) that (𝖬)log(2s)=s, which proves that the total number of bits s=Ω(nklogn) for any deterministic protocol. By Yao’s minimax principle, we get a lower bound of Ω(nklogn) on the randomized communication complexity of chainn,k.

The proof of Lemma 10 employs a reduction to the two player indexn. However, in these instances of indexn, Alice and Bob already have some partial information about the answer. We call this problem the biased index problem, and it is defined based on parameter θ[1/2,1/2], which is the initial bias known about the answer.

Definition 11.

The biased index distributional communication problem, denoted by bias-ind(θ) for θ[1/2,1/2], is defined as follows.

Sample W{0,1} such that W=1 with probability 1/2+θ, and W=0 otherwise. Sample (Y,ρ) uniformly at random from ×[n] conditioned on Y(ρ)=W. Give string Y to Alice, and index ρ to Bob. Bob has to output Y(ρ) after a single message Mindex from Alice.

Let πindex be a deterministic protocol for bias-ind(θ). Let 𝖶,𝖸,𝝆,𝖬index denote the random variables corresponding to W,Y, ρ and Mindex respectively. Let 𝒟θ denote the joint distribution of 𝖶,𝖸,𝝆 and 𝖬index in bias-ind(θ).

We prove the following lemma about bias-ind(θ) in Section 3.3.

Lemma 12 (Biased Index).

For any protocol πindex for bias-ind(θ),

(𝖶𝖬index,𝝆)H2(1/2+θ)2n((𝖬index)+logn).

We can prove Lemma 10 using Lemma 12, but the proof is deferred to the full version.

3.3 Biased Index

In this subsection, we will prove Lemma 12. Let us first recall the input distribution to bias-ind(θ).

Distribution 𝒟θ: Sample W=1 with probability 1/2+θ and set W=0 otherwise. Sample (Y,ρ)×[n] conditioned on Y(ρ)=W.

In 𝒟θ, the distribution of 𝖸 and 𝝆 are highly correlated. We give an alternate way of sampling Y,ρ so that this correlation is removed partially.

Distribution 𝒟θ:

For θ0:

  1. 1.

    Sample set T[n] of size b=n/(1+2θ) uniformly at random.

  2. 2.

    Sample a set S of n/2 indices from T uniformly at random and set them to 1, and set [n]S to 0 to get Y.

  3. 3.

    Sample ρ by sampling an index uniformly at random from T.

For θ<0:

  1. 1.

    Sample set T[n] of size b=n/(12θ) uniformly at random.

  2. 2.

    Sample a set S of n/2 indices from T uniformly at random and set them to 0, and set [n]S to 1 to get Y.

  3. 3.

    Sample ρ by sampling an index uniformly at random from T.

In this section, we assume that θ0. For the case when θ<0, the proof follows in the same vein, and is not presented. Let 𝖳,𝖲 denote the random variables corresponding to set T and set S respectively. We show that distributions 𝒟θ and 𝒟θ are in fact identical, and the proofs can be found in the full version.

Claim 13.

In distribution 𝒟θ, for any (Y,ρ)×[n],

Pr(𝖸=Y,𝝆=ρ)={1+2θn(nn/2)when Y(ρ)=1,12θn(nn/2)when Y(ρ)=0.
Claim 14.

Distribution 𝒟θ is the same as 𝒟θ.

Using this alternate way of sampling, it is easy to see that random variables 𝖸 and 𝝆 are independent of each other conditioned on 𝖳. It can also be extended to include random variable 𝖬index, as it is only a function of 𝖸.

Observation 15.

In distribution 𝒟θ, conditioned on 𝖳=T, for any iT, distribution of random variables 𝖸,𝖬index is independent of event 𝛒=i.

Proof.

Conditioned on 𝖳=T, string Y is chosen by picking an n/2 size set S uniformly at random from T, and setting these indices to 1, and this choice also fixes 𝖬index as the protocol is deterministic. And index ρ is chosen uniformly at random from T, independently of the choice of S, by definition of 𝒟θ. Thus, choice of 𝝆=i is independent of random variables 𝖸,𝖬index.

Next, we will show that even conditioned on 𝖳, the entropy of 𝖸 remains large.

Claim 16.
(𝖸𝖳)n(1+2θ)H2(1/2+θ)2logn.
Proof.

We assume that θ<1/2, as otherwise, the statement is vacuously true. H2(1)=0 by definition, and entropy is always non-negative by Section 2.2-(1).

Conditioned on 𝖳=T, we know that 𝖸 is fixed by choosing set 𝖲 uniformly at random. Thus,

(𝖸𝖳) =log((bn/2)) (by Section 2.2-(1))
log(2bH2(n/2b)b8(n/2)(bn/2)) (by Section 2, and n/2<b, as θ<1/2)
=bH2(n/2b)+12log(b8(n/2)(bn/2))
=n(1+2θ)H2(1/2+θ)+12log(14n(1/2θ))
n(1+2θ)H2(1/2+θ)+12log(1/4n) (as 1/2θ1)
n(1+2θ)H2(1/2+θ)2logn.  

We are ready to prove Lemma 12.

Proof of Lemma 12.

We can lower bound the entropy of 𝖶 conditioned on 𝖬index,𝝆 as,

(𝖶𝖬index,𝝆) (𝖶𝖬index,𝝆,𝖳) (as conditioning reduces entropy, Section 2.2-(2))
=𝔼𝖳=T[1bρT(𝖶𝖬index,𝝆=ρ,𝖳=T)] (as 𝝆 is uniform over T)
=𝔼𝖳=T[1bρT(𝖸(ρ)𝖬index,𝝆=ρ,𝖳=T)] (as W=Y(ρ) by definition of 𝒟θ )
=𝔼𝖳=T[1bρT(𝖸(ρ)𝖬index,𝖳=T)] (as 𝖸,𝖬index(𝝆=ρ)𝖳=T, by 15)
𝔼𝖳=T[1b(𝖸(T)𝖬index,𝖳=T)] (by subadditivity of entropy, Section 2.2-(4))
=𝔼𝖳=T[1b(𝖸𝖬index,𝖳=T)] (as Y([n]T) is fixed to be 0)
=1b(𝖸𝖬index,𝖳).

Thus it follows that,

(𝖸𝖬index,𝖳)b(𝖶𝖬index,𝝆). (1)

We also have,

(𝖸𝖳) =(𝖸,𝖬index𝖳) (as M is fixed by Y)
=(𝖬index𝖳)+(𝖸𝖬index,𝖳) (by chain rule of entropy, Section 2.2-(3))
(𝖬index𝖳)+b(𝖶𝖬index,𝝆) (by Eq 1)
(𝖬index)+b(𝖶𝖬index,𝝆). (as conditioning reduces entropy, by Section 2.2-(2))

Combining with Claim 16, we get,

(𝖬index)+b(𝖶𝖬index,𝝆) bH2(1/2+θ)2logn (as b=n/(1+2θ))
b(𝖶𝖬index,𝝆) bH2(1/2+θ)((𝖬index)+2logn). (rearranging the terms)

Dividing both sides by b, we get,

(𝖶𝖬index,𝝆) H2(1/2+θ)1b((𝖬index)+2logn)
H2(1/2+θ)2n((𝖬index)+2logn), (as bn/2)

finishing the proof.

3.4 Extension to Augmented Chain

In this subsection, we will extend our lower bound to the Augmented Chain problem introduced by [15]. We begin by defining Augmented Index.

Augmented Index is a close variant of the index problem. Here, in addition to having the index σ, Bob also has the bits X(<σ). We know that this generalization also requires Ω(n) communication when Alice sends a message to Bob [34]. This problem is particularly useful for proving lower bounds for turnstile streams (see e.g., [11, 25, 16], and references therein). Tight information cost trade-off for this variant in the two-way communication model was proved by [9].

The formal definition of Augmented Chain follows.

Definition 17 (Augmented Chain).

The aug-chainn,k communication problem is defined as follows. Given k+1 players 𝒫i for i[k+1] where,

  • 𝒫i has a string Xi{0,1}n for each i[k],

  • 𝒫i for 1<ik+1 has an index σi1[n], and a string Xi1(<σi1),

such that,

Xi(σi)=z,

for some bit z{0,1}. The players have a blackboard visible to all the parties. For i[k] in ascending order, 𝒫i sends a single message Mi, after which the index σi, and the string Xi1(<σi1) are revealed to the blackboard at no cost. 𝒫k+1 has to output whether z is 0 or 1. Refer to Figure 3 for an illustration.

Figure 3: An illustration of the aug-chainn,k problem from Definition 17. The solid arrows illustrate that player 𝒫i writes a message Mi to the board. The dashed arrows indicate that 𝒫i can read the contents of the board. The order in which the messages are sent and indices, strings are released is also shown.

For the chained version of Augmented Index, the lower bound that at least one player sends Ω(n/k2) bits holds true, and the proof follows with minimal changes. Using this, [15] proved lower bounds for interval independent set selection in turnstile streams with weighted intervals.

We prove the following result about aug-chainn,k.

Theorem 18.

For any n,k1, any protocol for aug-chainn,k with probability of success at least 2/3 requires communication Ω(nklogn).

A proof sketch detailing the changes needed to prove Theorem 18 is given in the full version. Most parts of the proof are similar to the proof of Theorem 3.

4 Applications to Streaming

In this section we give applications of our main result to independent sets in vertex arrival streams and streaming submodular maximization.

4.1 Independent Sets

In edge arrival streams, for any graph G=(V,E), the vertex set V with n vertices is given, and the edges E arrive in any arbitrary order. We are required to process the graph in limited space.

In vertex arrival streams, for any graph G=(V,E), the edges are grouped by their incident vertices. Vertices from V arrive one by one (in any arbitrary order), and when a vertex arrives, all the edges connecting it to any previously arrived vertices are revealed. This makes the vertex arrival stream a strictly easier model than the edge arrival stream, as the order of edges is restricted.

Indeed, for the maximal independent set problem, we know that finding algorithms in vertex arrival streams is easier; the greedy algorithm produces a maximal independent set in O~(n) space, whereas, in edge arrival streams, any algorithm which finds a maximal independent set requires Ω(n2) space [4, 12].

Maximum independent set (MIS), however, is provably hard in both vertex arrival streams and edge arrival streams. It is known that, any algorithm which performs an α-approximation of MIS in edge arrival streams requires Ω(n2/α2) space from [22]. In vertex arrival streams, [12] proved a lower bound of Ω(n2/α7). They also gave the following connection between chain problem and MIS in the proof of Theorem 9 of their paper.

Proposition 19 (Rephrased from [12]).

For any α1, any algorithm that gives an α-approximation of maximum independent sets in vertex arrival streams for n-vertex graphs using space at most s and probability of success at least 2/3 can be used to solve chainn2/64α4,2α with communication at most 2αs bits and success probability at least 2/3.

Our lower bound Theorem 3, along with Proposition 19 directly gives the following corollary.

Corollary 20.

For α1, any α-approximation of maximum independent sets in n-vertex graphs in vertex arrival streams uses Ω(n2/α5logn) space.

This further reduces the gap between the lower bounds for α-approximation of MIS in vertex arrival streams and edge arrival streams by an α2 factor.

4.2 Submodular Maximization

In this subsection, we will summarize our slight improvements to lower bounds for streaming submodular maximization.

A function f over ground set V from f:2V is submodular if and only if, for any two sets ABV and for any element xVB,

f(B{x})f(B)f(A{x})f(A).

This captures the diminishing returns property of any submodular function.

We are interested in maximization of a monotone submodular function subject to a cardinality constraint. That is, for a given , we want to find a subset SV with |S| such that for any other set TV with |T|, f(S)f(T).

We are given oracle access to function f, however, we do not have access to the entirety of the ground set. The elements of the ground set V arrive one by one, and the algorithm has space s to either store the incoming element or to discard it. The algorithm can query the oracle to f with any subset of the elements currently in storage. We want the storage to be roughly the same as the output size, which is .

In this model, [29] gave an algorithm which finds a (1/2ϵ)-approximation in O(/ϵ) space. [19] showed that a better approximation was not possible. They proved that any algorithm which gets a (1/2+ϵ)-approximation uses Ω(ϵ|V|/3). They give the following connection to the chain problem in Theorem 1.3 and Theorem 1.4 of their paper.

Proposition 21 (Rephrased from [19]).

For any ϵ>0, there exists a constant 0 such that for any 0, any randomized streaming algorithm which maximizes a monotone submodular function f:2V subject to cardinality constraint of at most , using space at most s and with approximation factor at least (1/2+ϵ) in expectation can be used to solve chain|V|/, with probability of success at least 2/3 and communication at most sO(/ϵ).

We get an improvement of factor over the current state-of-art lower bound in [19] as a corollary of Theorem 3.

Corollary 22.

For any ϵ>0, there exists a constant 0 such that for any 0, any streaming algorithm that maximizes a monotone submodular function f:2V subject to a cardinality constraint of at most , with an approximation factor at least (1/2+ϵ), requires Ω(|V|ϵ/2εlog(|V|)) space.

References

  • [1] Farid Ablayev. Lower bounds for one-way probabilistic communication complexity and their application to space complexity. Theoretical Computer Science, 157(2):139–159, 1996. doi:10.1016/0304-3975(95)00157-3.
  • [2] S. Assadi. Lecture notes on sublinear algorithms. https://sepehr.assadi.info/courses/cs514-s20/lec8.pdf, 2020.
  • [3] S. Assadi, G. Kol, R. R. Saxena, and H. Yu. Multi-pass graph streaming lower bounds for cycle counting, max-cut, matching size, and other problems. In 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), pages 354–364, Los Alamitos, CA, USA, November 2020. IEEE Computer Society. doi:10.1109/FOCS46700.2020.00041.
  • [4] Sepehr Assadi, Yu Chen, and Sanjeev Khanna. Sublinear algorithms for (Δ + 1) vertex coloring. In Proceedings of the Thirtieth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2019, San Diego, California, USA, January 6-9, 2019, pages 767–786, 2019. doi:10.1137/1.9781611975482.48.
  • [5] Sepehr Assadi, Sanjeev Khanna, and Yang Li. Tight bounds for single-pass streaming complexity of the set cover problem. In Proceedings of the Forty-Eighth Annual ACM Symposium on Theory of Computing, STOC ’16, pages 698–711, New York, NY, USA, 2016. Association for Computing Machinery. doi:10.1145/2897518.2897576.
  • [6] Sepehr Assadi and Janani Sundaresan. (noisy) gap cycle counting strikes back: Random order streaming lower bounds for connected components and beyond. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, pages 183–195, New York, NY, USA, 2023. Association for Computing Machinery. doi:10.1145/3564246.3585192.
  • [7] Sujoy Bhore, Fabian Klute, and Jelle J. Oostveen. On streaming algorithms for geometric independent set and clique. In Parinya Chalermsook and Bundit Laekhanukit, editors, Approximation and Online Algorithms, pages 211–224, Cham, 2022. Springer International Publishing. doi:10.1007/978-3-031-18367-6_11.
  • [8] Harry Buhrman and Ronald Wolf. Communication complexity lower bounds by polynomials. In Proceedings of the Annual IEEE Conference on Computational Complexity, pages 120–130, February 2001. doi:10.1109/CCC.2001.933879.
  • [9] Amit Chakrabarti, Graham Cormode, Ranganath Kondapally, and Andrew McGregor. Information cost tradeoffs for augmented index and streaming language recognition. SIAM Journal on Computing, 42(1):61–83, 2013. doi:10.1137/100816481.
  • [10] Lijie Chen, Gillat Kol, Dmitry Paramonov, Raghuvansh R. Saxena, Zhao Song, and Huacheng Yu. Near-optimal two-pass streaming algorithm for sampling random walks over directed graphs. In International Colloquium on Automata, Languages and Programming, 2021. URL: https://api.semanticscholar.org/CorpusID:232014583.
  • [11] Kenneth L. Clarkson and David P. Woodruff. Numerical linear algebra in the streaming model. In Proceedings of the Forty-First Annual ACM Symposium on Theory of Computing, STOC ’09, pages 205–214, New York, NY, USA, 2009. Association for Computing Machinery. doi:10.1145/1536414.1536445.
  • [12] Graham Cormode, Jacques Dark, and Christian Konrad. Independent Sets in Vertex-Arrival Streams. In Christel Baier, Ioannis Chatzigiannakis, Paola Flocchini, and Stefano Leonardi, editors, 46th International Colloquium on Automata, Languages, and Programming (ICALP 2019), volume 132 of Leibniz International Proceedings in Informatics (LIPIcs), pages 45:1–45:14, Dagstuhl, Germany, 2019. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ICALP.2019.45.
  • [13] Thomas M. Cover and Joy A. Thomas. Elements of information theory (2. ed.). Wiley, 2006.
  • [14] Carsten Damm, Stasys Jukna, and Jirí Sgall. Some bounds on multiparty communication complexity of pointer jumping. In Claude Puech and Rüdiger Reischuk, editors, STACS 96, 13th Annual Symposium on Theoretical Aspects of Computer Science, Grenoble, France, February 22-24, 1996, Proceedings, volume 1046 of Lecture Notes in Computer Science, pages 643–654. Springer, 1996. doi:10.1007/3-540-60922-9_52.
  • [15] Jacques Dark, Adithya Diddapur, and Christian Konrad. Interval Selection in Data Streams: Weighted Intervals and the Insertion-Deletion Setting. In Patricia Bouyer and Srikanth Srinivasan, editors, 43rd IARCS Annual Conference on Foundations of Software Technology and Theoretical Computer Science (FSTTCS 2023), volume 284 of Leibniz International Proceedings in Informatics (LIPIcs), pages 24:1–24:17, Dagstuhl, Germany, 2023. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.FSTTCS.2023.24.
  • [16] Jacques Dark and Christian Konrad. Optimal lower bounds for matching and vertex cover in dynamic graph streams. In Shubhangi Saraf, editor, 35th Computational Complexity Conference, CCC 2020, July 28-31, 2020, Saarbrücken, Germany (Virtual Conference), volume 169 of LIPIcs, pages 30:1–30:14. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2020. doi:10.4230/LIPICS.CCC.2020.30.
  • [17] Ashkan Norouzi Fard, Moran Feldman, Ola Svensson, and Rico Zenklusen. Submodular maximization subject to matroid intersection on the fly. In 30th Annual European Symposium on Algorithms, ESA 2022, September 5-9, 2022, Berlin/Potsdam, Germany, pages 52:1–52:14, 2022. doi:10.4230/LIPICS.ESA.2022.52.
  • [18] Joan Feigenbaum, Sampath Kannan, Andrew McGregor, Siddharth Suri, and Jian Zhang. Graph distances in the data-stream model. SIAM J. Comput., 38(5):1709–1727, 2008. doi:10.1137/070683155.
  • [19] Moran Feldman, Ashkan Norouzi-Fard, Ola Svensson, and Rico Zenklusen. The one-way communication complexity of submodular maximization with applications to streaming and robustness. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, STOC 2020, pages 1363–1374, New York, NY, USA, 2020. Association for Computing Machinery. doi:10.1145/3357713.3384286.
  • [20] Sudipto Guha and Andrew McGregor. Stream order and order statistics: Quantile estimation in random-order streams. SIAM J. Comput., 38(5):2044–2059, 2009. doi:10.1137/07069328X.
  • [21] Venkatesan Guruswami and Krzysztof Onak. Superlinear lower bounds for multipass graph processing. In Proceedings of the 28th Conference on Computational Complexity, CCC 2013, K.lo Alto, California, USA, 5-7 June, 2013, pages 287–298, 2013. doi:10.1109/CCC.2013.37.
  • [22] Magnús M. Halldórsson, Xiaoming Sun, Mario Szegedy, and Chengu Wang. Streaming and communication complexity of clique approximation. In Artur Czumaj, Kurt Mehlhorn, Andrew M. Pitts, and Roger Wattenhofer, editors, Automata, Languages, and Programming - 39th International Colloquium, ICALP 2012, Warwick, UK, July 9-13, 2012, Proceedings, Part I, volume 7391 of Lecture Notes in Computer Science, pages 449–460. Springer, 2012. doi:10.1007/978-3-642-31594-7_38.
  • [23] P. Indyk and D. Woodruff. Tight lower bounds for the distinct elements problem. In 44th Annual IEEE Symposium on Foundations of Computer Science, 2003. Proceedings., pages 283–288, 2003. doi:10.1109/SFCS.2003.1238202.
  • [24] Rahul Jain, Jaikumar Radhakrishnan, and Pranab Sen. A property of quantum relative entropy with an application to privacy in quantum communication. J. ACM, 56(6), September 2009. doi:10.1145/1568318.1568323.
  • [25] Daniel M. Kane, Jelani Nelson, and David P. Woodruff. On the exact space complexity of sketching and streaming small norms. In Proceedings of the Twenty-First Annual ACM-SIAM Symposium on Discrete Algorithms, SODA ’10, pages 1161–1178, USA, 2010. Society for Industrial and Applied Mathematics. doi:10.1137/1.9781611973075.93.
  • [26] Michael Kapralov, Sanjeev Khanna, and Madhu Sudan. Streaming lower bounds for approximating MAX-CUT. In Piotr Indyk, editor, Proceedings of the Twenty-Sixth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2015, San Diego, CA, USA, January 4-6, 2015, pages 1263–1282. SIAM, 2015. doi:10.1137/1.9781611973730.84.
  • [27] Michael Kapralov, Amulya Musipatla, Jakab Tardos, David P. Woodruff, and Samson Zhou. Noisy Boolean Hidden Matching with Applications. In Mark Braverman, editor, 13th Innovations in Theoretical Computer Science Conference (ITCS 2022), volume 215 of Leibniz International Proceedings in Informatics (LIPIcs), pages 91:1–91:19, Dagstuhl, Germany, 2022. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ITCS.2022.91.
  • [28] Mikhail Kapralov. Better bounds for matchings in the streaming model. In ACM-SIAM Symposium on Discrete Algorithms, 2012. URL: https://api.semanticscholar.org/CorpusID:448251.
  • [29] Ehsan Kazemi, Marko Mitrovic, Morteza Zadimoghaddam, Silvio Lattanzi, and Amin Karbasi. Submodular streaming in all its glory: Tight approximation, minimum memory and low adaptive complexity. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 3311–3320. PMLR, 09–15 June 2019. URL: https://proceedings.mlr.press/v97/kazemi19a.html.
  • [30] Ilan Kremer, Noam Nisan, and Dana Ron. On randomized one-round communication complexity. In Proceedings of the Twenty-Seventh Annual ACM Symposium on Theory of Computing, STOC ’95, pages 596–605, New York, NY, USA, 1995. Association for Computing Machinery. doi:10.1145/225058.225277.
  • [31] Eyal Kushilevitz and Noam Nisan. Communication complexity. Cambridge University Press, 1997.
  • [32] F. J. (Florence Jessie) MacWilliams and N. J. A. (Neil James Alexander) Sloane. The theory of error-correcting codes. North-Holland mathematical library ; v. 16. North-Holland Pub. Co., Amsterdam, 1978 - 1977.
  • [33] Guangxu Yang Mi-Ying Huang, Xinyu Mao and Jiapeng Zhang. Breaking square-root loss barriers via min-entropy. Electron. Colloquium Comput. Complex., pages TR24–067, 2024. URL: https://eccc.weizmann.ac.il/report/2024/067/.
  • [34] Peter Bro Miltersen, Noam Nisan, Shmuel Safra, and Avi Wigderson. On data structures and asymmetric communication complexity. Journal of Computer and System Sciences, 57(1):37–49, 1998. doi:10.1006/jcss.1998.1577.
  • [35] Anup Rao and Amir Yehudayoff. Communication Complexity: and Applications. Cambridge University Press, 2020. doi:10.1017/9781108671644.
  • [36] Mert Saglam. Tight bounds for data stream algorithms and communication problems. Master’s thesis, Simon Fraser University, 2011.