A Min-Entropy Approach to Multi-Party Communication Lower Bounds
Abstract
Information complexity is one of the most powerful techniques to prove information-theoretical lower bounds, in which Shannon entropy plays a central role. Though Shannon entropy has some convenient properties, such as the chain rule, it still has inherent limitations. One of the most notable barriers is the square-root loss, which appears in the square-root gap between entropy gaps and statistical distances, e.g., Pinsker’s inequality. To bypass this barrier, we introduce a new method based on min-entropy analysis. Building on this new method, we prove the following results.
-
An randomized communication lower bound of the -party set-intersection problem where the -th party holds a random set of size .
-
A tight randomized lower bound of the -party Tree Pointer Jumping problems, improving an lower bound by Chakrabarti, Cormode, and McGregor (STOC 08).
-
An lower bound of the Chained Index problem, improving an lower bound by Cormode, Dark, and Konrad (ICALP 19).
Since these problems served as hard problems for numerous applications in streaming lower bounds and cryptography, our new lower bounds directly improve these streaming lower bounds and cryptography lower bounds.
On the technical side, min-entropy does not have nice properties such as the chain rule. To address this issue, we enhance the structure-vs-pseudorandomness decomposition used by Göös, Pitassi, and Watson (FOCS 17) and Yang and Zhang (STOC 24); both papers used this decomposition to prove communication lower bounds. In this paper, we give a new breath to this method in the multi-party setting, presenting a new toolkit for proving multi-party communication lower bounds.
Keywords and phrases:
communication complexity, lifting theorems, set intersection, chained indexFunding:
Mi-Ying (Miryam) Huang: Supported by NSF CAREER award 2141536.Copyright and License:
Jiapeng Zhang; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Communication complexityAcknowledgements:
We thank anonymous reviewers for their helpful comments.Editors:
Srikanth SrinivasanSeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
Information complexity is one of the most powerful tools in proving communication complexity lower bounds [17, 5, 6, 23, 40] and streaming lower bounds [5, 16, 2, 31, 3, 13, 37, 12]. The idea of information complexity is to analyze the mutual information between the inputs held by the communication parties and the communication transcript. The definition of information complexity is similar to communication complexity, with information cost replacing communication cost. For a protocol , a popular notion of information cost is defined by , where and are the input distribution of Alice and Bob respectively and I is the mutual information. Intuitively, captures the mutual information of the inputs and the communication transcript, which is a lower bound of the communication cost. Besides this specific definition, there are many different variants that are smartly designed for diverse applications. However, they all share a similar idea: capture the information cost (usually by Shannon entropy) between the input distribution and the transcript.
Despite a vast number of applications successfully given by the information complexity-based approaches, this framework still has some inherent limitations. Indeed, some significant barriers are not only associated with some specific variants of information cost notions, but further deeply caused by the entropy itself. In this direction, one notable limitation is the square-root loss barrier.
Square-root loss barrier
We first use a simple example to illustrate this phenomenon. Let be a random variable that outputs with probability and with probability . This is a biased coin with a statistical distance to the uniform distribution. However, on the other hand, the entropy gap between them has only . This square gap is not significant if is a constant. However, the loss would become very large when it becomes very small. Beyond this simple example, this is indeed a general gap between entropy loss and statistical distance. For example, any result that applies Pinsker’s inequality has a good chance of creating this gap.
Lemma 1 (Pinsker’s inequality).
If and are two distributions, then
Here is the total variation distance of and and is the KL-divergence.
This quadratic gap makes it difficult to get good bounds via entropy-based analysis in many applications. For instance, proofs of multiparty unique-set disjointness [5], set disjointness under product distribution [23, 40], the chained index problem [21], multi-party pointer jumping problem [15], tree pointer jumping problem [16], pointer chasing problem [39], among others, all meet the square-root loss comparing with the upper bounds.
Despite resolving the square-root loss for some specific problems, these efforts are ad-hoc and use non-standard variants of Shannon entropy. Hence, it is hard to extend them for broader applications. A natural question arises: Could we use any measurement other than the Shannon entropy (or its close variants)?
Now, we revisit the example above. For a random variable supported on with entropy , we know the statistical distance between and the uniform distribution is by Pinsker’s inequality. Furthermore, improving Pinsker’s inequality is hard as it is tight in general. However, on the other hand, for a random variable with min-entropy , a simple calculation shows that the statistical distance between and the uniform distribution is . In this paper, min-entropy is a good candidate for avoiding square root loss in general settings.
Analysis of min-entropy via structure-vs-pseudorandomness
Though the min-entropy itself does not meet the square-root loss, there are other challenges in analyzing it. One of the most significant challenges is that, unlike the Shannon entropy, there is no chain rule for min-entropy, where a chain rule is an essential tool in entropy.
In order to overcome this issue, we adopt the structure-vs-pseudorandomness decomposition to serve as the “chain rule” in min-entropy analysis. This approach has been successfully applied in sunflower lemmas [36, 1] and query-to-communication lifting theorems [29, 30, 35, 46, 38]. Though this approach has been successfully applied in several areas, it has not been studied in multi-party settings. In this paper, we extend this approach to the multi-party setting. Beyond the three problems studied in this paper, we believe the min-entropy-based analysis could provide more applications to multi-party problems.
1.1 Our Results
Building on min-entropy analysis, we improve the lower bounds for three communication problems: (1) Set Intersection [7], (2) Tree Pointer Jumping [16], and (3) Chained Index [21].
1.1.1 Set Intersection Problem
To show the advantages of our min-entropy approach, we consider the search version of the set-disjointness problem, which is called the set-intersection problem. There are two versions of the set-intersection problems. The first one requires the players to find the whole intersection; the second one only asks the players to find one element from the intersection. Together, the set-intersection problems have been studied in many papers [34, 42, 11, 14, 41, 7, 45, 27, 26, 32, 9, 40]. In this paper, we focus on the second version. The setting is: each player is assigned a subset of , and the goal changes to finding an element .
We consider the communication complexity under product distribution here, and there are two typical product distributions that have been widely studied before. One is the fixed-size product distribution, where each player receives a uniformly random subset with . The other one is the Bernoulli product distribution, where each player receives a random set sampled as follows: for each element , independently with probability . Babai, Frankl, and Simon [4] first proposed the communication complexity of the set-disjointness problem under fixed-size product distribution where , and gave an lower bound. Their proof could also be adapted to the setting of Bernoulli product distribution with . Recently, this bound was extended to the -party setting by a recent paper by Dershowitz, Oshman, and Roth [23]. They showed that when , the communication complexity under the Bernoulli product distribution, where , is . Both of these decision-version lower bounds gave various applications.
In the context of the set-intersection, lower bounds are less known, though it also provides many applications. Bauer, Farshim, and Mazaheri [7] first gave a lower bound under Bernoulli product distribution with applications to cryptography. To be more specific, they proved:
Theorem 2 ([7]).
For the -party set-intersection problem under Bernoulli product distribution, where , , its communication complexity is
Note that this problem is exactly the search version of the set-disjointness problem considered in [4, 23]. Compared to the set-disjointness problem, set-intersection could be studied in a larger range of parameters, i.e., , where the intersection could be very large, i.e., as large as with high probability.
However, the theorem by [7] is far from tight and does not provide a non-trivial bound when are small. One of the main obstacles here is that the size of the intersections is large, but players only need to find one common element from many valid answers.
We consider the communication problem in [7], and extend it to the -party setting. Concretely, we assume that each player holds a (random) set of size with , and prove the following results for set-intersection.
Theorem 3.
For the -party set-intersection problem under Bernoulli product distribution, where , and 111We assume all the distributions considered in this paper satisfy this constraint.:
-
1.
the communication complexity is to achieve a constant accuracy;
-
2.
there exists a protocol that solves this problem under the distribution mentioned above with a constant accuracy and uses communication cost.
Note that this theorem establishes the first non-trivial lower bound when (we assume here). Actually, it implies that is tight (up to logarithmic factors when ) for the distributional-version of set-intersection considered in [7]. Also, our bound is similar to [23] when , where they proved a lower bound of for Similar to the two-party setting, our result significantly strengthens theirs when the size of the intersection is large.
1.1.2 Tree Pointer Jumping Problem
The Tree Pointer Jumping problem is a communication problem introduced by Chakrabarti, Cormode, and McGregor [16] with applications in streaming lower bounds. For , we consider a complete -level -ary tree rooted at . The -party Tree Pointer Jumping problem, denoted by , takes as an input a function with if is a leaf of , where is the set of nodes of . We use to denote the set of all valid functions here. For each input , we define the functions by,
The output of is defined by In the communication setting, the input is distributed to players. The problem is described as follows:
-
Player receives the labels of the -th level nodes, i.e., the first player receives ,…, and the last player receives the labels of the leaves.
-
In each round, players send messages in reverse order: from the last player to the first player. The cost of this round is the total number of bits sent by all players.
-
Players could communicate rounds, and the first player outputs the answer.
The goal of the players is to compute while minimizing the maximum cost of each round. For any -round protocol , we use to denote the maximum communication cost in all rounds. In this direction, [16] first proved the following lower bound.
Theorem 4 ([16]).
For any -round protocol with , we have that
Chakrabarti, Cormode, and McGregor [16] first used Theorem 4 to improve multi-pass streaming lower bound for median finding. Later on, Chakrabarti and Wirth [18] used this theorem to show a pass/approximation trade-off for the SET-COVER in the semi-streaming setting. In this paper, we improve the lower bound from Theorem 4 based on min-entropy analysis.
Theorem 5.
For any -round protocol with , we have that
1.1.3 Chained Index Problem
The Chained Index problem, introduced by Cormode, Dark, and Konrad [21], is another useful tool with many applications in streaming lower bounds [21, 24, 25, 10, 22]. For this problem, we consider the following communication setting.
-
There are players. Each player receives an input
-
It is promised that Here is the -th coordinate of .
-
Their goal is to compute through a one-way communication from the first player to the last player, where the last player should output the answer.
We say that a one-way protocol solves the Chained Index problem if for every input , the last player always outputs the correct answer with probability . The communication cost of this protocol is the total communication bits of all players. Using tools from information complexity, Cormode, Dark, and Konrad proved the following lower bounds for the Chained Index problem.
Theorem 6 ([21]).
Any one-way communication protocol that solves the Chained Index problem has randomized communication complexity
Since it has been introduced, many streaming lower bounds [21, 24, 25, 10, 22] were built on Theorem 6. Interested readers can find detailed discussions in the papers mentioned above. Theorem 6 was obtained by direct entropy-based analysis. In this paper, we improve this lower bound.
Theorem 7.
Any one-way communication protocol that solves the Chained Index problem has randomized communication complexity .
1.2 Proof Outline
In this section, we give a brief overview to our proof technique and we use the set-intersection problem for illustration. The proofs for chained index problem is similar in spirit.
Instead of considering Bernoulli distributions, we consider the following product distribution to simplify our presentation:
-
Each player independently and uniformly samples elements from (may have duplicates), where equals here.
Thus, each player receives a vector in and gets its set by removing the duplicate elements in the vector. In general, for any and , we consider as a subset of in a similar way. We prove the lower bound under this distribution, and then reductions are established in Section 3.3 to prove our main theorem.
It is well known that a deterministic protocol partitions the input domain into rectangles by step-by-step communication. The crucial idea of our proof is to further partition these leaf rectangles in the protocol tree into many structured rectangles. A structured rectangle satisfies that: for each , is fixed on some coordinates and pseudorandom on the remaining coordinates. The formal definition is given below.
Definition 8 (Structured rectangles).
Assuming , where each is a subset of , is a rectangle. We say is a structured rectangle if there exist subsets of coordinates with satisfying that
-
For each , there exists a such that . Here, is the complement of defined by and is the values of on .
-
For each , has a high block-wise min-entropy (see definitions in Section 2) on the coordinates .
The notion of structured rectangle has also been widely used in query-to-communication lifting theorems [30, 19, 35].
In the decomposition, we recursively (starting from the root to the leaves) decompose all rectangles in the protocol tree, i.e., for a node (which is also a rectangle), we decompose it based on the decomposition of its ancestors. This is the key step compared to existing decomposition (pre-sampling techniques) in cryptography, which may lead to new applications. The formal process of this decomposition is referred to Section 3.
After the decomposition process, each leaf has been partitioned into many structured rectangles. For a structured rectangle associated with and for , we say that:
-
1.
is bad if .
-
2.
is good if . We also call good structured rectangles as pseudorandom rectangles.
Then, our proof consists of the following two parts.
Combining the two parts, we are able to prove the main theorem. We defer the detailed proofs to Section 3.
Comparison with existing methods
Similar questions have been widely studied in several recent papers [7, 32, 23, 40]. All of these papers used standard known techniques in communication complexity such as information complexity.
These papers achieved tight bounds for set-disjointness (decision version), or set-intersection enumeration (finding whole intersections). However, all of their bounds for search problems are sub-optimal whenever the size of the set intersection (the solution space for the search problem) is large. By contrast, our method follows the structure-vs-pseudorandomness approach by Yang and Zhang [46], which was inspired by lifting theorems [30, 19, 35]. In [46], authors first introduced the techniques in proving query-to-communication lifting theorems directly to a communication setting without gadgets and proved the communication complexity lower bound for the collision problem in a two-party setting.
Compared to [46], we further extend their approach in two aspects: 1) we generalize this method into the multi-party setting; 2) we adopt it to prove communication lower bounds for search problems with many solutions. To the best of our knowledge, existing lower-bound methods could not address communication problems in these two settings. We believe these two settings could provide many applications.
1.3 Subsequent works and future directions
A late work by Sundaresan [43] further improved the lower bound of the Chained Index problem to via a reduction to a variant called the biased index problem.
Göös et al. [28] investigated quantum-classical separations in the communication model. To be more specific, they exhibit a total search problem whose communication complexity in the quantum simultaneous message passing model is exponentially smaller than in the classical two-way randomized model, which they call the Bipartite NullCodeword problem. To establish classical lower bounds, they employed a structure-vs-randomness approach, akin to the techniques used in [46] and this paper.
Another notable result was demonstrated by Mao, Yang, and Zhang in [38], where they improved the lower bound for a classical communication problem known as the -step pointer chasing problem by the structure-vs-randomness approach.
In [8], Beame and Whitmeyer established near-optimal lower bounds for the -party collision-finding problem of the strong bit pigeonhole principle which implies the tree-like semantic cutting-planes refutation lower bounds in proof complexity. However, their lower bounds only hold for the strong bit pigeonhole principle, they left lower bounds for the weak bit pigeonhole principle as an open question.
Paper organization
2 Preliminary
Definitions for set intersection problem
To begin with, we formally define the product distributions for set intersection problem adopted in this paper. For fixed parameters: is the number of parties, is the size of the domain, and are parameters indicating the size of each player’s set. We consider the following three types of hardness distributions in this paper (two of them have appeared in Section 1):
-
1.
Each player independently and uniformly samples elements from (may have duplicates), where equals .
-
2.
Each player independently and uniformly samples distinct elements from , where .
-
3.
Each player independently samples its set with that every element is contained in with probability .
We assume that , otherwise the existence of intersections can be not guaranteed. Furthermore, if holds for some constant , the common intersection of all players could be very large ().
The hardness distribution 3 is the Bernoulli product distribution with wide applications. Previous papers have mainly focused on this distribution. We prove the lower bound under distribution 1, and use two simple reductions to get the lower bound results for the hardness distributions 2 and 3. We refer to the two reductions to Section 3.3. In what follows, our discussion mainly focuses on distribution 1.
For a distribution and a communication protocol , we define the accuracy of on by:
For simplicity, we define this accuracy notion, which does not take the cases when sets are disjoint into consideration, differently from [7] in which they also consider the accuracy of distinguishing disjoint cases, namely, they define
Since we aim to establish lower bounds for those protocols achieving , we only consider the range of with222 is guaranteed by the definitions of hardness distribution 3 when .
In this paper, our lower bound result shows that achieving , where epsilon is a constant less than , requires large amounts of communication. This also implies a non-trivial hardness result to achieve since the disjoint cases could contribute at most to when holds. Hence, our results also imply hardness results under the [7] setting.
Next, we introduce some useful notions in communication complexity. In a -party communication problem, where each party holds an input from a domain , a rectangle is defined by ().
For a set , we denote as the uniform distribution on . In the set-intersection problem (particularly hard distribution 1), we consider the cases that each input is in where , and an instance can be transformed into a subset of by removing duplicate elements. Also, for two instances , we define by the intersection of the two subsets of deduced from and .
For a set of coordinates , we use to denote marginal distribution of on . For an instance and a set of coordinates , define to be an instance in by projecting on .
Structure-vs-pseudorandomness decomposition
We use capital letters to denote a set and bold symbols like to denote random variables. For a set , we use to denote the random variable uniformly distributed over the set . We introduce the formal definition of min-entropy.
Definition 9 (Min-entropy).
For a random variable taking value on , its min-entropy is defined as follows:
Definition 10 (Density function).
We define the one-side density function for a random variable on its support as:
Note that always holds by definitions and when is a uniform distribution.
The density function is also known as the entropy deficiency in lifting theorem papers, and we design the -side density function in order to extend the two-party results to the -party setting.
Definition 11 (-side density function).
For a structured rectangle , where each is subset of and associated with a set , we define its -side density function as:
In structure-vs-pseudorandomness decomposition, one of the most important notions, which captures the pseudorandomness, is the block-wise density.
Definition 12 (Block-wise density [29]).
For . A random variable supported on is said to be -dense if for all nonempty , we have that , here is the marginal distribution of on the set .
The definition of -dense measures the pseudorandomness of a random variable. In our proof, a typical choice of for set intersection and for the Chain Index problem.333 in previous structure-vs-pseudorandomness decomposition [30, 20, 19, 35].
The following lemma tells us that a random variable could be decomposed by a combination of random variables with dense properties by fixing some positions:
Lemma 13 (Density-restoring partition [30]).
Let be a subset of and be a subset of , and there exists an such that . Then, there exists a partition of :
such that every is associated with a set and a value . Then, they satisfy the following properties:
-
1.
;
-
2.
is -dense;
-
3.
.
Here, we define .
We also use the following simple version of Lemma 13 for some proofs.
Proposition 14.
Let be a partition of the set . Then
For dense random variables, we also have the following useful lemma.
Lemma 15.
If are independent -dense random variables and each takes value from with , where is a constant and , then for any element , it holds
here denotes the Euler’s number.
Proof.
We know that all ’s are independent. Thus, we first bound the probability that . Assuming that , we have the following argument
where the last inequality comes from the definition of -dense. Hence, we know that
3 Lower Bounds for Set Intersection
In this section, we prove the communication lower bound for the hardness distribution 1 (where each player gets independent and uniform samples from ). Then, in Section 3.3, we use reductions to obtain lower bounds for hardness distributions 2 and 3. Formally, we prove that:
Theorem 16.
If a communication protocol solves -party set-intersection problem under the hardness distribution 1 with accuracy bigger than , the communication complexity is
3.1 The Decomposition and Sampling Process
The key idea of this proof, as we introduce in Section 1, is to decompose rectangles (nodes444Note that a node of the protocol tree is a rectangle.) of the protocol tree into structured rectangles and analyze the accuracy of the protocol could achieve in those decomposed structured rectangles. We design a decomposition and sampling process in this section to
-
decompose the rectangles of the protocol tree into structured rectangles;
-
sample a decomposed rectangle with respect to its size.
We define the root rectangle of the protocol tree to be , which contains all valid inputs. is also a structured rectangle by definitions. We start from and begin our decomposition and sampling process, which uses a random walk on the protocol tree from the root to a leaf, and do the decomposition along the path. See Algorithm 1 for the formal decomposition process.
We use to denote the current rectangle of the decomposition and sampling process. It begins with , and at each step is partitioned into two subrectangles by the protocol. Then, we replace with or with probability or (which also equals to or as we defined in Algorithm 1), and reach a new rectangle. After reaching the new rectangle, the structured property of may get destroyed, and our decomposition works here to maintain the structured property. We use the density-restoring partition (Lemma 13) to further decompose the current rectangle into subrectangles , and each is a structured rectangle. Again, we choose to be our next rectangle with probability , and do the process above recursively until reaching a leaf rectangle. As shown in the decomposition and sampling process, we eventually sample a structured rectangle in the leaf level with respect to its size.
Note that at some point of the random walk, the current rectangle may not exist on the protocol tree since we do the density-restoring partition to further decompose the rectangles. However, every that potentially appears in the random walk must be fully contained in a rectangle of the protocol tree. Thus, the protocol also partitions into two sub-rectangles if is not in the leaf level of the protocol tree.
Note that the output of the process above is a random variable over rectangles. We define to be the random variables over decomposed structured rectangles in the leaf level (not leaf rectangles of the protocol tree, but sub-rectangles of those leaves after decomposition) sampled by the process above, and is associated with random sets s. For convenience, we define the support of to be . One may see the two important properties of the decomposition and sampling process:
-
Every rectangle is a structured rectangle;
-
For a rectangle , we have that
The verification of the two properties is straightforward from the definition of our decomposition and sampling process. The first statement offers a structured property that makes it easier to analyze the rectangles. The second statement tells us that the probability that equals the probability that the input lies in . This is crucial in later bounding the accuracy of .
Next, we bound the accuracy of . For every structured rectangle associated with , we define as , namely the fixed parts of . Hence, for each , it holds since is a structured rectangle. We can then divide all the rectangles in into two types:
-
1.
is a bad structured rectangle if ;
-
2.
is a good structured rectangle if .
Assume is a bad structured rectangle. Then, there exists a universal common element 555We can choose any element that lies in here. such that for any possible instance in . The protocol is thus able to achieve perfect correctness by outputting when the input lies in . Hence, we need to show with a low probability that is a bad rectangle, namely the probability that the input lies in bad rectangles is small. To be more specific, we prove the following lemma:
Lemma 17.
If , it holds that .
For those good structured rectangles, we show the following facts: For a good structured rectangle, the protocol cannot achieve high accuracy since there is no intersection on the fixed parts, while the other parts are dense. Formally, we prove the following lemma:
Lemma 18.
For a good structured rectangle , it holds that for any ,
Combining the three lemmas above, we can easily prove Theorem 16.
Proof of Theorem 16.
We prove the theorem by showing that communication protocol with can achieve at most accuracy.
It is well known that a communication protocol partitions the whole input domain into several leaf rectangles and assigns an answer to each leaf rectangle. With our decomposition and sampling process, original leaf rectangles are further decomposed into the two types of structured rectangles mentioned above. The accuracy of comes from the following two parts:
-
1.
The probability .
-
2.
The probability that the protocol outputs the correct answer in a good structured rectangle is .
From Lemma 17 and 18, we know that . By a union bound, the total accuracy is thus no more than as desired. It suffices to prove the two important lemmas above.
3.2 Proofs of Technical Lemmas
We first prove Lemma 17 by the following round-by-round analysis.
Proof of Lemma 17.
Recall the decomposition process from line 4 to line 12. In each communication round, player sends one bit, and partitions into two parts . Then, is replaced by (or ) with probability (or ). In this process, the density function would increase since the size of decreases. This contributes to the density function with an increment of:
-
with probability ;
-
with probability .
Thus, in expectation, the density function of after partitioning will increase
| (1) |
where denotes the size of before partitioning. Furthermore, if is no longer -dense, we partition by Lemma 13 and get and with for all . We use Lemma 13, where we take , and get:
| (2) |
Recall that here. In the decomposition process, is replaced with with probability . Hence, taking the expectation in one communication round, we have
| (3) |
Thus, combining (1), (2) and (3) and taking expectations, we know that after rounds of communication (where each round communicates exact one bit message), it holds:
Here, the comes from (1) and (3). We know that from definitions. Hence, we have
| (4) |
We can bound the probability that the bad structured rectangle appears round by round. At each round of communication, if we choose to replace , then we will fix more positions for . We then consider the probability that this new fixed part contributes to forming a bad structured rectangle with future fixed positions.
Let , for any , we label it as a error term if , 666 is a fixed subset of with size at most since is fixed on . Input may be labeled many times during the decomposition process.. By Lemma 15, for any ,
By a union bound, the probability that error terms appear in is
Also, we know that the total number of fixed elements equals , which is identical to the summation of of every step, thus, the average probability of error terms at the end of the decomposition process is at most
We note that for any , if is bad, then all instances have been labeled as an error term in the decomposition process, together with (4), we have
The last inequality holds since and . Next, we show that in the good structured rectangles, the protocol cannot achieve large accuracy in finding the common element. This also comes from the structured properties of the rectangles:
Proof of Lemma 18.
Notice that we consider the rectangle associated with that has no common elements on fixed parts . Thus, for any element , there exists at least a party which does not contain on its fixed part. Thus, we use Lemma 15 for with , and get
3.3 Lower Bounds for Other Hardness Distributions
In this section, we first establish a reduction from Bernoulli hardness distribution (hardness distribution 3) to hardness distribution 2 by the following lemma:
Lemma 19.
Proof.
We first use Chernoff bound to bound the probability of the size of set of each player exceeding or less than under the hardness distribution 3:
We use to denote the event that . Then, by a union bound, we know that:
Then, condition on , we have the success probability of in finding set intersection under hardness distribution 3 is bigger than . Furthermore, condition on , the hardness distribution 3 can be represented by a combination of product distributions:
where denotes the hardness distribution 2 with parameters . Then, the lemma follows from an averaging argument. It suffices to construct a reduction from hardness distribution 2 to hardness distribution 1.
Lemma 20.
If there exists a communication protocol with communication complexity which solves set-intersection under hardness distribution 2 with accuracy , there exists a communication protocol with communication complexity which solves set-intersection under hardness distribution 1 with accuracy when holds.
Proof.
We construct the communication protocol as follows:
-
1.
For each player , remove the duplicate elements of its input and get a .
-
2.
Randomly sample elements from , fail if .
-
3.
Run the communication protocol on s to find intersection.
We know that the successful probability of under hardness distribution 1 is bigger than
It suffices to bound . From the union bound, we have:
We know that
From Markov’s Inequality, we have
If holds, which is guaranteed by the constraints, also holds. This concludes the lemma.
3.4 Efficient Protocols for the Hardness Distribution
In this section, we first explain an efficient protocol for the hardness distribution 3, where we use to denote the distribution, showing that our lower bound result is almost tight for this distribution. Also, this protocol can be easily extended to some more general product distributions sharing "similarities" with the Bernoulli product distribution. Formally, we prove:
Theorem 21.
There is a protocol , which solves the hardness distribution 3, with and
Furthermore, this protocol can be extended to more general distributions. Let be any distribution that satisfies the following properties:
-
1.
each party holds a set of size ;
-
2.
the size of intersecting part of all parties is ;
there exists a protocol with communication cost that achieves accuracy under .
Proof.
To begin with, we first propose an efficient protocol to solve . Without loss of generality, we assume and each party gets a subset . Then, the communication protocol proceeds as follows:
-
1.
The first party uniformly and randomly picks elements from and sends them, denoted by , to the second party.
-
2.
The second party receives the message from the first one, and sends to the third party.
-
3.
The process goes on, and the last party computes . If it is not empty, the last party outputs any element in it. Otherwise, the protocol fails.
Then, we bound and its communication complexity to show is highly efficient. From the definitions, we know that
Also, we have that
The last inequality holds since . From Chernoff bound, we know that the probability that . The last inequality is from the constraint of . Furthermore, when
it holds that
Combining the facts above, we have .
On the other hand, we bound the communication complexity by bounding the expected size of . holds from definitions. Furthermore, we have
Then, follows by . This concludes the first statement.
Next, we slightly change the protocol above to match the second statement. The protocol proceeds as follows:
-
1.
The first party uniformly and randomly picks elements from and sends them, denoted by , to the second party.
-
2.
The second party receives the message from the first one, and sends to the third party.
-
3.
The process goes on, and the last party computes . If it is not empty, the last party outputs any element in it.
Obviously, the communication complexity of this protocol is . Also, we know the accuracy is bigger than
Thus, our lower bounds show that those trivial protocols are nearly optimal.
4 Lower Bounds for Chained Index
Recall that in the Chained Index problem, the player receives an input . The players aim to compute through a one-way communication. In this section, we show an improved lower bound for the Chained Index problem. In light of Yao’s principle, we consider the following hard distribution.
The distribution .
-
1.
Uniformly sample .
-
2.
Sample conditioned on .
-
3.
Output where for every .
For a subset , define the weight of under as
We prove the following lower bound. We say a one-way -party protocol has signature if, for each , the -th party sends at most bits (on all inputs).
Theorem 22.
Let be a constant. Let be a protocol for the -party chained index problem with signature . If has advantage, i.e.,
Then .
We use a decomposition and sampling process , as shown in Algorithm 2, in our analysis. takes as input a protocol , and samples a rectangle that is contained in for some leaf node . Our proof proceeds in two steps:
-
1.
Section 4.1 shows that the accuracy of is captured by a quantity called average fixed size, which is a natural quantity that arises in the running of .
-
2.
Section 4.2 proves that the average fixed size can be bounded from above by . Consequently, if enjoys high accuracy, we get a lower bound of .
We first recall some basic definitions.
-party one-way protocols
A deterministic -party one-way communication protocol is specified by a rooted binary tree. For every internal vertex ,
-
it has 2 children, denoted by and ;
-
is owned by some party – we denote the owner by ;
-
every leaf node specifies an output.
Starting from the root, the owner of the current node partitions its input space into two parts and , and sets the current node to if its input belongs to .
The communication complexity of , denoted by , is the depth of the tree. On a path from root to some leaf, each time the owner switches, we call it a new round; in a one-way protocol, the label of the owner is non-decreasing.
Fact 23.
The set of all inputs that lead to an internal vertex is a rectangle, denoted by .
Normalized protocols
We normalized a protocol as follows so as to make it defined on all inputs, including those not in . For the -th party, given input and previous transcripts , output if the input is invalid, i.e., given , there is no input in matches . Otherwise, the -th party outputs and proceeds as . Clearly, by normalizing, we communicate more bits.
Lemma 24 (Loop invariant).
After each iteration in algorithm 2,
-
;
-
for all , is -dense;
-
for all , there exists such that for all .
Proof.
The first item is true because every time is updated, is updated accordingly to a sub-rectangle of and updating into its sub-rectangles does not violate this condition.
Since we applied density restoring partition at the end of each iteration, the second and the third items are guaranteed by Lemma 13 and the way that are updated.
4.1 Relating Accuracy and Average Fixed Size
As shown in Lemma 24, during the execution of , for every , the set is “fixed” on in the sense that all strings in share the same value on coordinates in . So we call the expected size of average fixed size. However, in order to relate the accuracy of to average fixed size, we need to consider the expectation of in a slightly different distribution.
Definition 25 (Average fixed size).
Let denote the uniform distribution over the input space . Let and consider the following process, denoted by :
-
1.
run until the -th round;
- 2.
-
3.
upon entering the -th round (i.e., until Line 17 is reached with ), return .
The average fixed size of the -th party is defined as
Lemma 26 (Relating accuracy and average fixed size).
Assume that . Then
Remark 27.
satisfies the condition. Indeed,
where the first inequality is by , and the second is by for and .
The proof of the lemma is obtained through the following two lemmas. The first lemma readily says that conditioned on the flag is not raised, has little advantage in the rectangle output by . The second lemma shows the probability that the flag is raised is bounded in terms of the average fixed size.
Lemma 28.
If outputs and in the end, then
Lemma 29.
Next, we first prove Lemma 26 using the above two lemmas.
Proof of Lemma 26.
Note that in the running of , we first sample and then always update to a randomly chosen rectangle; the probability of each rectangle being chosen is proportional to its weight under . Consequently,
It remains to prove the two lemmas.
Proof of Lemma 28.
Say . Since in the end, we have for all . By Lemma 24, we have for all . Since is contained in some leaf node of , output the same answer in , say . Note that for ,
since we must have . Write . Then we have
Since for all , we have , which implies
Since we assumed , it holds that , concluding the proof.
Proof of Lemma 29.
Let denote the event that the flag is raised when (i.e., when the -th round ends) for the first time. Clearly, It suffices to show for each .
Fix and the random coins used for the first rounds, i.e., until Line 17 is reached with ). Let be the value of rectangle when running using until the -th round begins. The core of our proof is to compare the process with one that runs under uniform weight instead of the weight under ; this is why we can deal with the promise.
-
Let be the process that runs until the -th round begins with , then run the -th round with fresh random coins.
-
Let be the process that runs until the -th round begins with , then runs the -th round with replaced by .
Note that during the execution of and , the partitions are the same, and the only difference is that when choosing , the probabilities are different. Let be a possible value of at the end of the -th round. In we update according to , and thus the probability that in the end of equals
Similarly, the probability that in the end of equals
The next claim reveals a connection between the two probabilities, whose proof is by direct calculation and is deferred to the appendix.
Claim 30.
For all possible value , .
Since is determined by the value of and the event is determined by and , the above claim implies that Note that in , is chosen uniformly at random, and thus
Taking expectation over we get , as desired.
4.2 Average Fixed Size is Bounded by Communication
Now that the accuracy of a protocol is bounded from above by the average fixed size (i.e., ), in what follows we show that the average fixed size is at most . Formally, we prove that
Lemma 31.
Assume that is a normalized protocol with signature . Then
Proof.
The proof strategy is similar to the proof of Lemma 29. Fix and consider . Fix the random coins used for the first rounds (i.e., until Line 17 is reached with ). Let and be defined as in the proof of Lemma 29. Moreover, let denote the number of bits sent by the -th party, i.e., the number of iterations in the -th round. By a standard density increment argument, we have
Claim 32.
.
Proof of Claim 32.
We shall prove this lemma by a density increment argument. That is, we study the change of the density function in each iteration. Let be the value of at the end of the -th iteration.
We fix the random coins used for the first iterations and consider the updates in the current iteration.
-
1.
First, is partitioned into according to . Then, is updated to with probability . Consequently, will increase as shrinks, and in expectation (over the random choice of ) the increment is
(5) -
2.
Next, we further partition according to Lemma 13. Say is partitioned into and let be the index sets promised by Lemma 13; and for all we have
where . With probability , we update and . Therefore, taking expectation over the random choice of , the density function will decrease by
(6) Note that and thus
(7)
Let be the -algebra generated by the random coins used for the first iterations. Let be the increment of in the -th iteration. Observe that by definition. By Equation 6 and Equation 7, taking expectation over random choice of , decrease by at least due to the density restoring partition. Then
| (8) |
In the beginning, . Since the density function is always non-negative by definition, we have and thus On the other hand, by telescoping,
where the inequality follows from Equation 8. Observe that by definition. We conclude that
as desired.
4.3 Putting Things Together
Now we are prepared to prove Theorem 22.
Proof of Theorem 22.
We first normalize so as to make it accept all inputs in . Denoted by the normalized protocol, then has signature .
Set . One can check that satisfies the requirement in Lemma 26. By Lemma 26 and Lemma 31, we have
| (9) |
Since have the same output on valid inputs and we assumed , we get . Combining with Equation 9 and rearranging, we have
| (10) |
The above lower bound is vacuous when . Next, we strengthen the lower bound to via simple reductions. Consider two cases.
-
Case 1: . Equation 10 implies that
-
Case 2: . Let be the set of talking parties. If , we are done. Otherwise, we construct a protocol for where , described below.
-
–
Say the talking parties are . Let emulate , and emulate by doing nothing. Note that on receiving an input sampled from , the parties can imagine they are given inputs for all from . Since never talks, perfect emulate them by doing nothing. In sum,
Observe that has signature . Applying Equation 10 to , we get
-
–
Therefore, we conclude that , regardless of the value of .
References
- [1] Ryan Alweiss, Shachar Lovett, Kewen Wu, and Jiapeng Zhang. Improved bounds for the sunflower lemma. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 624–630, 2020. doi:10.1145/3357713.3384234.
- [2] Alexandr Andoni, Andrew McGregor, Krzysztof Onak, and Rina Panigrahy. Better bounds for frequency moments in random-order streams. arXiv preprint, 2008. arXiv:0808.2222.
- [3] Sepehr Assadi, Yu Chen, and Sanjeev Khanna. Polynomial pass lower bounds for graph streaming algorithms. In Proceedings of the 51st Annual ACM SIGACT Symposium on theory of computing, pages 265–276, 2019. doi:10.1145/3313276.3316361.
- [4] Laszlo Babai, Peter Frankl, and Janos Simon. Complexity classes in communication complexity theory. In 27th Annual Symposium on Foundations of Computer Science (sfcs 1986), pages 337–347, 1986. doi:10.1109/SFCS.1986.15.
- [5] Ziv Bar-Yossef, Thathachar S Jayram, Ravi Kumar, and D Sivakumar. An information statistics approach to data stream and communication complexity. Journal of Computer and System Sciences, 68(4):702–732, 2004. doi:10.1016/J.JCSS.2003.11.006.
- [6] Boaz Barak, Mark Braverman, Xi Chen, and Anup Rao. How to compress interactive communication. In Proceedings of the forty-second ACM symposium on Theory of computing, pages 67–76, 2010. doi:10.1145/1806689.1806701.
- [7] Balthazar Bauer, Pooya Farshim, and Sogol Mazaheri. Combiners for backdoored random oracles. In Advances in Cryptology–CRYPTO 2018: 38th Annual International Cryptology Conference, Santa Barbara, CA, USA, August 19–23, 2018, Proceedings, Part II 38, pages 272–302. Springer, 2018. doi:10.1007/978-3-319-96881-0_10.
- [8] Paul Beame and Michael Whitmeyer. Multiparty communication complexity of collision finding, 2024. doi:10.48550/arXiv.2411.07400.
- [9] Anup Bhattacharya, Sourav Chakraborty, Arijit Ghosh, Gopinath Mishra, and Manaswi Paraashar. Disjointness through the lens of vapnik–chervonenkis dimension: Sparsity and beyond. Comput. Complex., 31(2), December 2022. doi:10.1007/s00037-022-00225-6.
- [10] Sujoy Bhore, Fabian Klute, and Jelle J Oostveen. On streaming algorithms for geometric independent set and clique. In International Workshop on Approximation and Online Algorithms, pages 211–224. Springer, 2022. doi:10.1007/978-3-031-18367-6_11.
- [11] Mark Braverman, Ankit Garg, Denis Pankratov, and Omri Weinstein. From information to exact communication. In Proceedings of the Forty-Fifth Annual ACM Symposium on Theory of Computing, STOC ’13, pages 151–160, New York, NY, USA, 2013. Association for Computing Machinery. doi:10.1145/2488608.2488628.
- [12] Mark Braverman, Sumegha Garg, Qian Li, Shuo Wang, David P Woodruff, and Jiapeng Zhang. A new information complexity measure for multi-pass streaming with applications. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, pages 1781–1792, 2024. doi:10.1145/3618260.3649672.
- [13] Mark Braverman, Sumegha Garg, and David P Woodruff. The coin problem with applications to data streams. In 2020 ieee 61st annual symposium on foundations of computer science (focs), pages 318–329. IEEE, 2020. doi:10.1109/FOCS46700.2020.00038.
- [14] Joshua Brody, Amit Chakrabarti, Ranganath Kondapally, David P. Woodruff, and Grigory Yaroslavtsev. Beyond set disjointness: The communication complexity of finding the intersection. In Proceedings of the 2014 ACM Symposium on Principles of Distributed Computing, PODC ’14, pages 106–113, New York, NY, USA, 2014. Association for Computing Machinery. doi:10.1145/2611462.2611501.
- [15] Amit Chakrabarti. Lower bounds for multi-player pointer jumping. In Twenty-Second Annual IEEE Conference on Computational Complexity (CCC’07), pages 33–45. IEEE, 2007. doi:10.1109/CCC.2007.14.
- [16] Amit Chakrabarti, Graham Cormode, and Andrew McGregor. Robust lower bounds for communication and stream computation. In Proceedings of the fortieth annual ACM symposium on Theory of computing, pages 641–650, 2008. doi:10.1145/1374376.1374470.
- [17] Amit Chakrabarti, Yaoyun Shi, Anthony Wirth, and Andrew Yao. Informational complexity and the direct sum problem for simultaneous message complexity. In Proceedings 42nd IEEE Symposium on Foundations of Computer Science, pages 270–278. IEEE, 2001.
- [18] Amit Chakrabarti and Anthony Wirth. Incidence geometries and the pass complexity of semi-streaming set cover. In Proceedings of the twenty-seventh annual ACM-SIAM symposium on Discrete algorithms, pages 1365–1373. SIAM, 2016. doi:10.1137/1.9781611974331.CH94.
- [19] Arkadev Chattopadhyay, Yuval Filmus, Sajin Koroth, Or Meir, and Toniann Pitassi. Query-To-Communication Lifting for BPP Using Inner Product. In Christel Baier, Ioannis Chatzigiannakis, Paola Flocchini, and Stefano Leonardi, editors, 46th International Colloquium on Automata, Languages, and Programming (ICALP 2019), volume 132 of Leibniz International Proceedings in Informatics (LIPIcs), pages 35:1–35:15, Dagstuhl, Germany, 2019. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ICALP.2019.35.
- [20] Sandro Coretti, Yevgeniy Dodis, Siyao Guo, and John Steinberger. Random oracles and non-uniformity. In Annual International Conference on the Theory and Applications of Cryptographic Techniques, pages 227–258. Springer, 2018. doi:10.1007/978-3-319-78381-9_9.
- [21] Graham Cormode, Jacques Dark, and Christian Konrad. Independent sets in vertex-arrival streams. arXiv preprint, 2018. arXiv:1807.08331.
- [22] Jacques Dark, Adithya Diddapur, and Christian Konrad. Interval selection in data streams: Weighted intervals and the insertion-deletion setting. In 43rd IARCS Annual Conference on Foundations of Software Technology and Theoretical Computer Science (FSTTCS 2023), pages 24:1–24:17. Schloss-Dagstuhl-Leibniz Zentrum für Informatik, 2023. doi:10.4230/LIPIcs.FSTTCS.2023.24.
- [23] Nachum Dershowitz, Rotem Oshman, and Tal Roth. The communication complexity of multiparty set disjointness under product distributions. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, pages 1194–1207, 2021. doi:10.1145/3406325.3451106.
- [24] Moran Feldman, Ashkan Norouzi-Fard, Ola Svensson, and Rico Zenklusen. The one-way communication complexity of submodular maximization with applications to streaming and robustness. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 1363–1374, 2020. doi:10.1145/3357713.3384286.
- [25] Moran Feldman, Ashkan Norouzi-Fard, Ola Svensson, and Rico Zenklusen. Submodular maximization subject to matroid intersection on the fly. In 30th Annual European Symposium on Algorithms (ESA 2022), pages 52:1–52:14. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2022. doi:10.4230/LIPIcs.ESA.2022.52.
- [26] Dmitry Gavinsky. The communication complexity of the inevitable intersection problem. Chicago Journal of Theoretical Computer Science, 2020(3), August 2020. URL: http://cjtcs.cs.uchicago.edu/articles/2020/3/contents.html.
- [27] Satrajit Ghosh and Mark Simkin. The communication complexity of threshold private set intersection. In Alexandra Boldyreva and Daniele Micciancio, editors, Advances in Cryptology – CRYPTO 2019, pages 3–29, Cham, 2019. Springer International Publishing. doi:10.1007/978-3-030-26951-7_1.
- [28] Mika Göös, Tom Gur, Siddhartha Jain, and Jiawei Li. Quantum communication advantage in tfnp, 2024. doi:10.48550/arXiv.2411.03296.
- [29] Mika Göös, Shachar Lovett, Raghu Meka, Thomas Watson, and David Zuckerman. Rectangles are nonnegative juntas. SIAM Journal on Computing, 45(5):1835–1869, 2016. doi:10.1137/15M103145X.
- [30] Mika Göös, Toniann Pitassi, and Thomas Watson. Query-to-communication lifting for bpp. In 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS), pages 132–143, 2017. doi:10.1109/FOCS.2017.21.
- [31] Venkatesan Guruswami and Krzysztof Onak. Superlinear lower bounds for multipass graph processing. Algorithmica, 76:654–683, 2016. doi:10.1007/S00453-016-0138-7.
- [32] Dawei Huang, Seth Pettie, Yixiang Zhang, and Zhijun Zhang. The communication complexity of set intersection and multiple equality testing. SIAM Journal on Computing, 50(2):674–717, 2021. doi:10.1137/20M1326040.
- [33] Mi-Ying(Miryam) Huang, Xinyu Mao, Guangxu Yang, and Jiapeng Zhang. Breaking square-root loss barriers via min-entropy. In In Electronic Colloquium on Computational Complexity (ECCC) (TR24-067), 2024. URL: https://eccc.weizmann.ac.il/report/2024/067/.
- [34] Bala Kalyanasundaram and Georg Schintger. The probabilistic communication complexity of set intersection. SIAM Journal on Discrete Mathematics, 5(4):545–557, 1992. doi:10.1137/0405044.
- [35] Shachar Lovett, Raghu Meka, Ian Mertz, Toniann Pitassi, and Jiapeng Zhang. Lifting with sunflowers. Leibniz international proceedings in informatics, 215, 2022. doi:10.4230/LIPICS.ITCS.2022.104.
- [36] Shachar Lovett, Noam Solomon, and Jiapeng Zhang. From dnf compression to sunflower theorems via regularity. In Proceedings of the 34th Computational Complexity Conference, pages 1–14, 2019. doi:10.4230/LIPICS.CCC.2019.5.
- [37] Shachar Lovett and Jiapeng Zhang. Streaming lower bounds and asymmetric set-disjointness. In 2023 IEEE 64th Annual Symposium on Foundations of Computer Science (FOCS), pages 871–882. IEEE, 2023. doi:10.1109/FOCS57990.2023.00056.
- [38] Xinyu Mao, Guangxu Yang, and Jiapeng Zhang. Gadgetless lifting beats round elimination: Improved lower bounds for pointer chasing. arXiv preprint, 2024. doi:10.48550/arXiv.2411.10996.
- [39] Noam Nisan and Avi Widgerson. Rounds in communication complexity revisited. In Proceedings of the twenty-third annual ACM symposium on Theory of computing, pages 419–429, 1991. doi:10.1145/103418.103463.
- [40] Rotem Oshman and Tal Roth. The Communication Complexity of Set Intersection Under Product Distributions. In Kousha Etessami, Uriel Feige, and Gabriele Puppis, editors, 50th International Colloquium on Automata, Languages, and Programming (ICALP 2023), volume 261 of Leibniz International Proceedings in Informatics (LIPIcs), pages 95:1–95:20, Dagstuhl, Germany, 2023. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ICALP.2023.95.
- [41] Jeff M. Phillips, Elad Verbin, and Qin Zhang. Lower bounds for number-in-hand multiparty communication complexity, made easy. SIAM Journal on Computing, 45(1):174–196, 2016. doi:10.1137/15M1007525.
- [42] M. Saglam and G. Tardos. On the communication complexity of sparse set disjointness and exists-equal problems. In 2013 IEEE 54th Annual Symposium on Foundations of Computer Science (FOCS), pages 678–687, Los Alamitos, CA, USA, October 2013. IEEE Computer Society. doi:10.1109/FOCS.2013.78.
- [43] Janani Sundaresan. Optimal communication complexity of chained index. ITCS, 2025.
- [44] Shuo Wang, Guangxu Yang, and Jiapeng Zhang. Communication complexity of set-intersection problems and its applications. In In Electronic Colloquium on Computational Complexity (ECCC) (TR23-164), 2023.
- [45] Thomas Watson. Communication Complexity with Small Advantage. In Rocco A. Servedio, editor, 33rd Computational Complexity Conference (CCC 2018), volume 102 of Leibniz International Proceedings in Informatics (LIPIcs), pages 9:1–9:17, Dagstuhl, Germany, 2018. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.CCC.2018.9.
- [46] Guangxu Yang and Jiapeng Zhang. Communication lower bounds for collision problems via density increment arguments. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, pages 630–639, 2024. doi:10.1145/3618260.3649607.
Appendix A Proof of Claim 30
Claim 33 (Claim 30 restated).
Let . Let . For , define
Then
Proof of Claim 30.
To start with, observe that
| (11) |
We claim that for ,
| (12) |
Then we have
where the first and the third equality is from Equation 12. Combining with Equation 11 we have the desired result.
It remains to show Equation 12. Suppose that satisfies . Then we and
For every , there exists exactly possible values for each with (with one bit fixed to be ) and possible values for (which is not used at all). Therefore,
which is exactly what we wanted.
Appendix B Proof of Lemma 13
The following lemma and proof are from Lemma 5 in [30].
Lemma 34 (Lemma 13 restated).
Let . Let be a subset of and . Suppose that there exists an such that . Then, there exists a partition and every is associated with a set and a value that satisfy the following properties.
-
1.
;
-
2.
is -dense;
-
3.
, where .
Proof.
We prove it by a greedy algorithm as follows.
Item 1 is guaranteed by the construction of and .
We prove Item 2 by contradiction. Assume towards contradiction that is not -dense for some . By definition, there is a nonempty set and violating the min-entropy condition, namely, Write . Then
where the first equality holds as . However, this means at moment that is chosen, the set also violates the min-entropy condition (witnessed by ), contradicting the maximality of .
Finally, Item 3 is proved by straightforward calculation:
