Communication Complexity of Equality and Error-Correcting Codes
Abstract
We study the public-coin randomized communication complexity of the equality function. The communication complexity of this function is known to be low when the error probability is constant and the players have access to many random bits. The complexity grows, however, if the allowed error probability and the amount of randomness are restricted.
We show that public-coin randomized protocols for equality and error-correcting codes are essentially the same object. That is, given a protocol for equality, we can construct a code, and vice versa.
We substantially extend the protocol-implies-code direction: any protocol computing a function with a large fooling set can be converted into an error-correcting code. As a corollary, we show that among functions with a fooling set of size , equality on bits has the least randomized communication complexity, regardless of the restrictions on the error probability and the amount of randomness.
Finally, we use the connection to error-correcting codes to analyze the randomized communication complexity of equality for varying restrictions on the error probability and the amount of randomness. In most cases, we provide tight bounds. We pinpoint the setting in which tight bounds are still unknown.
Keywords and phrases:
communication complexity, randomized communication complexity, error-correcting codesCopyright and License:
2012 ACM Subject Classification:
Theory of computation Communication complexity ; Theory of computation Error-correcting codes ; Theory of computation Randomness, geometry and discrete structuresEditors:
C. Aiswarya, Ruta Mehta, and Subhajit RoySeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
In the standard model of communication complexity, two players, Alice and Bob, are tasked with computing the value of a fixed function on some given input . However, Alice only knows and Bob only knows , so they must communicate to compute . Communication complexity studies the amount of communication between players in terms of the number of bits exchanged. In recent decades, communication complexity has become one of the central areas in computational complexity, with numerous applications to other areas, including circuit complexity, proof complexity, online algorithms, and data structures [11, 19, 22, 10].
The equality function is one of the most fundamental functions in communication complexity. On input , this function outputs iff . This was the first function for which an exponential separation between deterministic and randomized communication complexity was shown (see [23]). Later, played a key role in many results in communication complexity, including direct sum results for the randomized setting [7], protocols for threshold functions [17], lifting theorems [14, 6], and communication complexity with an equality oracle [5, 18].
It is well known that the randomized communication complexity of is low for constant error probability. We can consider both the private-coin and public-coin models, which we sometimes refer to as the private randomness setting and the public randomness setting, respectively. In the private randomness setting, there is an protocol, usually attributed to Rabin and Yao (unpublished, see [11, 20]). For the case of public randomness, there is an protocol. The general idea of this protocol is that players can first use their shared randomness to sample a random hash function . Then Alice sends Bob , and Bob compares it to . The most standard protocol of this type is for players to pick a random vector and compare the inner products and over . In the case that and are not equal, there is a probability that the compared inner products are also not equal.
Although the public-coin randomized complexity of is for constant error probability and unlimited access to randomness, the complexity can increase when we impose stronger restrictions on these parameters. The error probability for (as well as for any other function) can be reduced to any at the cost of a factor in complexity by a standard error reduction procedure. Note that the protocol above uses a large amount of randomness, which grows as the error probability is reduced. The number of random bits, which we denote by , can be substantially reduced using the idea from Newman’s result on the connection between communication complexity in the public-coin and private-coin models [16, 13]. Another known construction of randomized protocols for with restricted randomness is through error-correcting codes. In such a protocol, players privately encode their inputs via an error-correcting code and compare them in a random coordinate (see Section 2.4 or [19, Chapter 3]).
As for lower bounds, Canetti and Goldreich [4] obtained general lower bounds in the setting with restricted randomness (by a reduction to the deterministic case; see Fact 14). For , their result gives a lower bound of . They also provided examples of matching upper bounds for the case of constant error probability. Moran et al. [15] (see also [13]) gave a lower bound on communication complexity in the private-coin model in the small error case for functions with large fooling sets. In particular, for , their bound implies a lower bound of in the private randomness setting. For public randomness, this implies a lower bound . (Note that to reduce a protocol with private randomness to one with public randomness, one of the players can generate random bits and send them to another player.) Despite these results, the entire landscape of communication complexity of for varying and is not fully clear.
Our Results
As mentioned above, error-correcting codes give communication protocols for . In our first result, we observe that actually, the converse is also true: communication protocols for give rise to error-correcting codes. Thus, error-correcting codes and protocols for are essentially the same object.
Theorem 1 (Informal version of Theorem 30 and Corollary 31).
Given a protocol for , we can construct an error-correcting code with equivalent parameters (up to a constant factor), and vice versa. Furthermore, this connection is precise for one-sided error protocols; in this case, protocols and error-correcting codes are in one-to-one correspondence.
We obtain the following corollary:
Corollary 2 (Informal version of Corollary 32).
Any randomized public-coin protocol for can be converted into a one-sided false-positive error protocol with no loss in parameters.
In particular, any communication protocol for with false positive error and false negative error can be converted into a one-sided error protocol with false positive error (the communication complexity and the number of random bits used remain the same).
Next, we show that one direction of the result above generalizes to arbitrary functions in the following way:
Theorem 3 (Informal version of Theorem 33).
Any randomized public-coin protocol for a function with a large fooling set can be converted into an error-correcting code with the same parameters as for .
As a consequence, we obtain the following corollary:
Corollary 4 (Informal version of Corollary 34).
Among all functions with a fooling set of size , on bits has the least randomized public-coin communication complexity for all values of error probability and number of random bits.
In other words, this corollary says that of all functions, provides the largest separation between fooling set size and public-coin randomized communication complexity.
We next use the established connection between communication complexity of and error-correcting codes to provide a detailed analysis of the communication complexity of on -bit inputs for varying amounts of random bits and error . We start by translating bounds for error-correcting codes to lower bounds for communication complexity. In particular, based on Plotkin bound for codes, we obtain the following lower bound on the randomized communication complexity of in the public-coin model (up to a constant factor):
Theorem 5 (Informal version of Theorem 35).
Any randomized public-coin protocol for satisfies the following lower bound for the communication complexity :
Furthermore, this lower bound is tight for many regimes of the parameters .
It is noted in [1] that Plotkin bound is the best known for the codes over large alphabet. We confirm this comparing it with Singleton bound, Elias-Bassalygo bound, Griesmer bound and MRRW bound. We observe that actually the combination of Singleton bound and Elias-Bassalygo bound gives the full lower bound in the theorem above (up to a constant factor).
Interestingly, the two known lower bounds for mentioned above have natural interpretations in terms of codes. The result of [15] corresponds to the Elias-Bassalygo bound (after translation of their result to the public randomness setting) and the result of [4] matches the Singleton bound.
Next, we further analyze the landscape of communication complexity of . Depending on values of the parameters and , the expression in the theorem above can resolve (up to constant factors) to , , or . In the first case, the trivial deterministic protocol for (see Section 2.2) is clearly optimal. In the second case, we show that for a wide range of parameters, the protocol for obtained from Reed-Solomon codes is optimal. However, this code does not apply in a certain range of parameters, specifically when . In this case, we devise a protocol that is suboptimal, but close to the lower bound (specifically, we iterate the Reed-Solomon code with itself). Finally, in the third case, we show that the standard protocol combined with an improvement of Newman’s theorem (see Fact 15) for reducing the number of random bits gives an upper bound that is tight, up to additive lower order terms. Previously this was known only for the case of constant [4].
Other related work
Fleischer et al. [8] studied the relation between the number of random bits and the zero-error communication complexity for the list-non-disjointness function. Ball et al. [2] gave a lower bound on randomized communication complexity with defective randomness through the size of the fooling set. The paper [15] analyses communication protocols of functions by looking at the statistical distance of the protocol on distinct inputs from the fooling set. The difference with our approach is that we look at Hamming distance between the vectors of outcomes instead.
One could apply round elimination techniques (e.g., [3]) to transform an arbitrary protocol for into a one-way protocol, thus obtaining an alternative proof of Corollary 2 (or Corollary 32). This allows for an alternative prove of the translation from protocols to codes in Theorem 1: we can prove this translation for one-way protocol and use round elimination to extend the result to general protocols. However, since the proof of this direction of the theorem does not differ that much for one-way and for general communication protocols, we decided to give a direct proof for the general case.
The rest of the paper is organized as follows. In Section 2, we give the necessary definitions and state known results. In Section 3, we prove our results on the connection between communication complexity and error-correcting codes. In Section 4, we use the connection established in Section 3 to deduce lower bounds on the communication complexity of equality for various amounts of allowed error and randomness (the analysis of suboptimal bounds on the codes is omitted and can be found in the full version of the paper [9]). In Section 5, we analyze the landscape of communication complexity of equality by comparing our lower bounds against upper bounds for various regimes of parameters.
2 Preliminaries
We make frequent use of the following fact:
Fact 6.
For two non-negative functions such that ,
2.1 Notation and Conventions
All logarithms are base 2 unless otherwise specified. We use the notation to denote the function composed with itself times.
For brevity, unless specified otherwise, when we refer to the “complexity” of a function, we are referring to the randomized communication complexity in the public-coin model. Similarly, when we refer to a “protocol”, we mean a randomized communication protocol in the public-coin model. By “naive protocol” for , we mean the deterministic protocol in which Alice sends her entire input, and by the “standard protocol” for , we mean the standard randomized protocol used in the public-coin model (see Chapter 3 of [11]. Finally, when we refer to a “code”, we are referring to an error-correcting code.
We use to denote the amount (in bits) of public randomness in a communication protocol, and we use to denote the allowed error. When protocols have two-sided error, we use to denote false positive error and to denote false negative error of a protocol. That is, is an upper bound on the probability that the protocol outputs 1 on an input where the function outputs 0. The false negative error, , is an upper bound on the probability that the protocol outputs 0 on an input where the function outputs 1. For convenience, we often use to denote public-coin randomized communication complexity. That is, we let denote . We use the notation code to refer to a code with block length , message length , and distance (see Section 2.3 for definitions of these parameters).
2.2 Communication Complexity
In the standard model of communication complexity, two players, Alice and Bob, wish to compute the value of a function on some given input . The challenge is that Alice only knows and Bob only knows , and thus the players need to communicate (i.e., send bits) to compute .
Definition 7 (protocol, value, cost, one-way).
A communication protocol is a binary tree where each internal node is labeled either by a function or by a function (denoting Alice or Bob’s message respectively) and each leaf node is labeled with an element in .
The value of the protocol on input is the label reached by traversing the tree according to the following rules: start at the root, and for each internal node labeled by (resp ), move left if (resp ) and right otherwise. The cost of on input is the length of the path taken on input . The cost of is the height of the tree. We call a protocol one-way if the entire protocol consists of just one message.
Definition 8 (deterministic communication complexity).
The deterministic communication complexity of is the minimum cost of , over all protocols that compute .
Fact 9 (deterministic communication complexity of ).
For , the equality function, , is defined to be 1 if and 0 otherwise. We have that .
The lower bound on can be proved using the fooling set method (see [11]). We define a fooling set below:
Definition 10 (fooling set [11]).
A fooling set for is a set and a value such that
-
1.
For every ,
-
2.
For all distinct pairs , either or .
While Definition 7 defines protocols for deterministic communication complexity, this paper focuses on randomized communication complexity, wherein Alice and Bob have a source of public randomness. Formally, there is a random string of bits available to both players. Where in the deterministic setting we had internal nodes in the protocol tree labeled with and , we now have internal nodes labeled with and ; that is, Alice and Bob’s communication can depend on . We can also view this model as a distribution on deterministic protocols:
Definition 11 (randomized public-coin protocol [11]).
A (randomized) public-coin protocol is a probability distribution over deterministic protocols. The success probability of a public-coin protocol on input is the probability of choosing a (deterministic) protocol, according to the probability distribution , that computes correctly.
Definition 12 (randomized communication complexity [11]).
The randomized communication complexity is the minimum cost of a public-coin protocol that computes with an error of at most (for every input ).
Fact 13 (randomized communication complexity of ).
The randomized communication complexity (in the public randomness model) of with constant error is .
Randomized protocols can be converted to deterministic protocols, with some overhead on the communication complexity, by repeating the randomized procedure and taking a majority vote. The following upper bound is proved in [4]:
Fact 14 ([4]).
Let , be a function computable via a randomized communication protocol using bits of communication, bits of randomness, and with error probability . Then the deterministic communication complexity is bounded by
| (1) |
The following fact improves upon Newman’s Theorem [16], and states that a public-coin protocol can be converted into a private-coin protocol with only logarithmic additive overhead in the complexity. Similarly, the amount of randomness used by a public-coin protocol can be reduced by incurring a small amount of extra error.
Fact 15 ([13]).
Let be a function. Then Furthermore, any public-coin protocol for which uses random bits and has error can be converted into a protocol using random bits with error , at no penalty to the communication complexity.
Interestingly, this improvement over original Newman’s Theorem is crucial for tight bounds in one of the cases in our analysis in Section 5.3.
See [11] for more on private and public-coin settings in communication complexity.
2.3 Error-Correcting Codes and Bounds
Definition 16 (code, block length, codewords).
A code of block length over alphabet is . The elements are called codewords.
Definition 17 (distance, relative distance).
The distance of a code is , where is the Hamming distance function. In other words, the distance of a code is the minimum Hamming distance between any pair of codewords. The relative distance of is the ratio of distance to block length, .
Definition 18 (message length).
The message length (sometimes called dimension) of a code over alphabet with is .
We call a code with block length , message length , and distance an code.
Definition 19 (rate).
The rate of a code with block length and message length is .
High distance and high rate are desirable properties for codes; however, there is a trade-off between these two parameters. The bounds stated below offer some insight into this trade-off.
Fact 20 (Singleton Bound).
Let be an code with alphabet size . Then
Equivalently, we can express the Singleton Bound in terms of rate and relative distance:
The Singleton Bound is tight when . That is, there exist codes for which under this condition. We call such codes MDS (Maximum Distance Separable) codes. See Section 2.3 below.
Definition 21 (-ary entropy).
The -ary entropy function , a generalization of binary entropy, is defined as follows:
As the block length of a code tends to infinity, we have the relationship , where is the volume of the Hamming Ball with strings of length , alphabet size , and radius . Hence appears in the asymptotic statement of various bounds for error-correcting codes, such as the Elias-Bassalygo Bound below:
Fact 22 (Elias-Bassalygo Bound).
Let be an code with alphabet size . Then
Fact 23 (Griesmer Bound).
For a linear code with alphabet size , we have:
For more information on linear codes, see [21].
Fact 24 (Plotkin Bound).
The Plotkin bound strengthens the Singleton bound. For a code with message length , block length , relative distance , and alphabet size , we have:
Fact 25 (MRRW Bound for large alphabet [1]).
Let be an code with alphabet size and . Then
Our upper bounds use protocols based on Reed-Solomon codes, which we define below. For more information about these codes, see [21].
Reed-Solomon codes are a family of codes which achieve the Singleton bound (i.e., for which ).
Formally, Reed-Solomon codes are defined as follows:
Definition 26 (Reed-Solomon codes).
Let . Let be distinct. The Reed-Solomon code over with evaluation points , is the code given by:
2.4 Codes Give Communication Protocols for
It is known (see [19, beginning of Chapter 3]) that codes give communication protocols for . In particular, we have the following fact:
Fact 27.
For every code with alphabet of size , there exists a communication protocol, , computing , such that , using public random bits, and bits of communication, with error . Moreover, this protocol is one-way and has one-sided false-positive error.
3 Communication Complexity vs. Error-Correcting Codes
In this section, we show that randomized communication protocols for give rise to error-correcting codes. In fact, we prove a generalization of this statement which applies to functions that are close to in some sense, which we call “-like” functions.
Definition 28 (EQ-like).
We say that a function is -like if for all , and for all pairs such that , either or .
Note that is a particular example of an -like function, where for all pairs such that , and . Another example of an -like function is the well-known Greater-Than function, defined as follows: where is the integer given by as its binary representation.
3.1 Protocols for Imply Codes
Consider a protocol for . Alice holds , Bob holds , and there is public randomness . Let be the transcript of Alice and Bob’s communication. That is,
where each is a function of all previous messages, , and their shared randomness , and each is a function of all previous messages, , and . Without loss of generality, at the end of the protocol, Alice outputs . Let and denote the false positive and false negative error of the protocol, respectively. That is,
We show that the transcript of Alice and Bob’s messages forms a code. An encoding of is simply the concatenation of for all choices of .
Lemma 29.
Let be -like, and let be the transcript for Alice and Bob’s communication in a randomized protocol for as given above. Then for
Proof.
Let . Without loss of generality suppose 111If instead , we replace with in the first two lines of Equation and the proof holds.. Then we have
where the substitution follows from the fact that and that the communication tree partitions the inputs into combinatorial rectangles. Hence,
Using Lemma 29, we obtain the following theorem:
Theorem 30.
Let be -like, and suppose there exists a randomized communication protocol for which has communication complexity , uses bits of public randomness, and has two-sided error probabilities and . Then there exists an error-correcting code with message length , codeword length , relative distance , and alphabet size .
Proof.
The proof follows directly from Lemma 29. Concretely, the codeword for is given by the concatenation of for all possible values of randomness . Each symbol in the codeword corresponds to for one choice of , and since each transcript uses bits of communication, there are possible symbols in the code alphabet. Finally, by Lemma 29, the probability that two different codewords agree on any given coordinate is bounded above by . Thus the relative distance of the code is .
If we consider a protocol with one-sided error, that is, a protocol with either or , we immediately obtain the following corollary:
Corollary 31 (One-sided error).
Let be -like, and suppose there exists a randomized communication protocol for which has communication complexity , uses bits of public randomness, and has one-sided error probability . Then there exists a code with message length , codeword length , relative distance , and alphabet size .
We remark that since codes give protocols (Fact 27), this corollary shows that one-way protocols for with one-sided false-positive error are in one-to-one correspondence with error-correcting codes. That is, we can translate from protocols to codes and vice versa with no loss in parameters. Another consequence of Corollary 31 is the following corollary:
Corollary 32.
Any randomized protocol for can be converted to a one-way protocol with one-sided false-positive error with no loss in parameters.
The following table summarizes the relationship between the communication complexity of a protocol for and the obtained code:
| Communication Complexity | Code |
|---|---|
| size of input = | # of messages = |
| # of random bits = | length of code = |
| error = | relative distance = |
| complexity = | alphabet size = |
3.2 Codes from Functions Based on Fooling Sets
We have seen that protocols for -like functions give codes. We can actually make a more general statement which applies to all functions. In particular, for any function , a protocol for gives a code whose parameters depend on the parameters of the protocol and the size of the largest fooling set for .
Theorem 33.
Let be any function with fooling set , and suppose there exists a randomized communication protocol for which has communication complexity , uses bits of randomness, and has error probabilities . Then there exists a code with block length , message length , relative distance ), and alphabet size .
Proof.
To prove the theorem, we show how to construct an error-correcting code from the function with fooling set . Let and denote the set of and inputs respectively which appear in elements of the fooling set , and define to be restricted to the inputs . Without loss of generality, assume that for , . By the definition of fooling set (see Section 2.2), is -like, up to a permutation on one half of the input. Therefore, a protocol for computing gives a protocol for , which by Lemma 29 gives an error-correcting code for messages of length . As a consequence of Theorem 33, we obtain the following corollary, which says that in some sense, represents the largest gap between fooling set size and randomized communication complexity in the public-coin model.
Corollary 34.
Let be any function with fooling set , and suppose there exists a randomized communication protocol for which has communication complexity , uses bits of randomness, and has error probabilities . Then there exists a protocol for on bits which has communication complexity , uses bits of randomness, and has one-sided false-positive error probability .
This follows because, given a protocol for , we can construct a code with message length , from which we obtain a protocol for on bits.
Informally, Corollary 34 says that among all functions with a fooling set of size , on bits has the smallest public-coin randomized complexity for all values of and . For protocols with two-sided error, this interpretation is true up to constant factors. For protocols with one-sided error, it is precise.
4 Communication Lower Bounds Implied by Code Bounds
The equivalence between communication protocol and codes allows us to view bounds on codes as bounds on communication complexity. In particular, since we showed that existence of protocols implies existence of codes, nonexistence of codes implies nonexistence of protocols. Using this connection, we can translate known bounds for codes (Section 2.3) into bounds on communication complexity using the correspondence given by Table 1222Note that these lower bounds also hold for functions with large fooling sets.. Below we describe the bound that can be obtained from Plotkin bound, that turns out to be the best known bound for large size of the alphabet.
The Plotkin bound (Fact 24) can be expressed in terms of communication protocols as:
| (2) |
In terms of error rate , this implies
| (3) |
From this we get the system
For any we have , thus solving for we get the following (up to constant factors)
Depending on and , either 1 or is the dominant term in the denominator of the first inequality. Based on these two cases, this bound can be equivalently stated (up to constant factors) as .
Combining two bounds together we get the following theorem.
Theorem 35.
For any randomized public-coin protocol for on bits with complexity , bits of randomness, and error probability ,
| (4) |
up to constant factors.
It was noted (without details) in [1] that Plotkin bound is the best known bound for the case of error-correcting codes with large alphabet. For the sake of completeness, we provide a detailed analysis of communication complexity bounds we can obtain from Singleton bound, Elias-Bassalygo bound, Griesmer bound and MRRW bound. The analysis for these bounds can be found in the full version of our paper [9]. We also note that the combination of Singleton bound, Elias-Bassalygo bound matches Plotkin bound up to the constant factor (see [9]).
Note that Theorem 35 gives a different lower bound (one of , or ) for different regimes of the parameters and . In Section 5, we examine the necessary relationship between , and for achieving each of the three possible lower bounds. We also analyze when these bounds are tight by giving matching upper bounds for most regimes of the parameters. This analysis is summarized in the table below.
| Description of the Regime | Lower Bound | Upper Bound |
|---|---|---|
| , | ||
| if | ||
| , | ||
| if | ||
| and constant and | ||
| , | ||
| if and is constant | ||
| , | ||
| if |
5 Analyzing the Landscape of Communication Complexity of
In this section, we analyze the various regimes of parameters which give rise to different lower bounds on communication complexity. We further examine when these bounds are tight by giving upper bounds. Now we consider the three cases based on which of the three terms the maximum in Theorem 35 evaluates to.
5.1 Case 1
In this case, we have
or, after rearranging, . In this case, , giving the lower bound . Clearly, the naive protocol for gives us the matching upper bound.
Remark 36.
This lower bound is actually easy to see, since for only the deterministic protocol is possible (because of the granularity of error probability).
5.2 Case 2
In this case, we have
The lower bound in this case is .
Remark 37.
Note that this case shows that we cannot study the value of up to a constant factor. The reason is that the bound depends on and multiplicative changes for may affect the bound dramatically. Note that we never allowed extra multiplicative constant factors for .
5.2.1 Upper Bound with Reed-Solomon Protocol
For the upper bound in this case, we can use the communication protocol derived from Reed-Solomon codes (Section 2.3). In this protocol, Alice and Bob encode their inputs using a Reed-Solomon code and check the value of a single random index in a subset of their respective codewords. For some values of parameters, the communication complexity of this protocol matches the lower bound given by the Singleton bound.
Lemma 38.
For any , there exists a public-coin randomized communication protocol that computes using bits of communication, bits of randomness, and error rate .
Proof.
Let be the largest prime number such that . To compute , we use a code to get an error rate of
| (5) |
but we know that (the naive protocol for gives this upper bound), giving . Note that Reed-Solomon codes are only defined for . For this protocol, . Thus, we have a protocol only in the regime
Hence, this protocol does not give an upper bound when
| (6) |
An example parameters in this regime is and . The lower bound in this case is . We partially address this issue in the next section.
5.2.2 Upper bound with iterated Reed-Solomon
The Reed-Solomon code allows us to do the following transformation of parameters:
This notation means that when we need to send a message of length , we can use random bits, introduce an error , and reduce the problem to sending a message of size .
This protocol gives a tight upper bound, but it has the restriction . For the range of parameters
we do not have a tight upper bound.
Restating in terms of and using Fact 6, Reed-Solomon codes give tight upper bounds for the regime of parameters satisfying
| (7) |
Thus, the following range of the lower bound is not matched with a tight upper bound:
| (8) |
Our plan is to apply the Reed-Solomon code iteratively. That is, in the first step, we apply the code with some parameters:
but then instead of sending bits, we apply the Reed-Solomon code again (with some new values of parameters).
We note that the idea to apply codes iteratively is well known in coding theory under the name of concatenated codes [21]. In particular, it was previously observed in communication complexity language that applying Reed-Solomon codes iteratively helps to improve multiplicative constant and lower order terms in communication complexity of the function in the regime of constant error probability and unlimited randomness [12]. However, we need a more detailed analysis.
For a real nonnegative , we introduce the function . We now describe the values of parameters that we can cover with iterations of Reed-Solomon protocol. The proof of the following lemma can be found in Section A.1.
Lemma 39.
For any integer constant we have that the communication protocol derived from iterations of Reed-Solomon has complexity and is valid for all
| (9) |
(up to an additive constant), where is the function applied to itself times.
For example, for constant the lower bound we get is still tight up to the constant factor. It is not hard to see in this case that the function is equal to up to an additive constant, and thus the range of parameters for which we do not have an upper bound reduces to
For non-constant , our upper bound is not tight. To analyze the interval in which it applies, observe that is equal to (up to an additive constant 1) for and is equal to (up to an additive constant 1) for .
5.3 Case 3
This case applies when or The lower bound in this case is .
Note that the first of the subcases here corresponds to , in which case the communication complexity bound is . We know that this bound is tight.
The remaining case to consider is which, after the rearranging gives This is equivalent to
As for the upper bound, we use the standard randomized protocol and apply Fact 15. The standard protocol gives us the parameters
Applying Fact 15 with error parameter gives
Setting yields
and this bound achieves optimal up to an additive factor in lower order terms, and optimal up to a constant multiplicative factor.
Remark 40.
We observe that this public-coin protocol implies a private-coin protocol, where the randomness is simply shared at the start of the protocol at the cost of bits of communication. This resulting private-coin protocol has error using bits of communication, essentially matching the lower bound of observed in [13]. This indicates that in this range of parameters, the optimal private-coin protocol is to simply share the private randomness and then simulate the standard public-coin protocol.
6 Discussion & Future Directions
We identify an equivalence between randomized public-coin protocols for equality and error-correcting codes, and generalize one direction of this equivalence to functions with large fooling sets. We use this connection to study the randomized communication complexity of for various regimes of the error and randomness parameters. We give lower bounds for protocols from lower bounds from codes, and analyze when these lower bounds are tight. An immediate question is whether we can tighten our bounds in the regime in which they are not tight.
One way to view our result on fooling sets is that the maximal separation between the fooling set size and randomized communication complexity with public randomness is achieved by the equality function (for all regimes of parameters). A natural future direction is to explore whether similar separations can be found between randomized communication complexity and other complexity measures, such as deterministic communication complexity, rectangle size, or rank. Partial progress in this direction was obtained by Canetti and Goldreich [4].
References
- [1] Matti J. Aaltonen. Linear programming bounds for tree codes (corresp.). IEEE Trans. Inf. Theory, 25(1):85–90, 1979. doi:10.1109/TIT.1979.1056004.
- [2] Marshall Ball, Oded Goldreich, and Tal Malkin. Communication complexity with defective randomness. In Valentine Kabanets, editor, 36th Computational Complexity Conference, CCC 2021, July 20-23, 2021, Toronto, Ontario, Canada (Virtual Conference), volume 200 of LIPIcs, pages 14:1–14:10. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2021. doi:10.4230/LIPICS.CCC.2021.14.
- [3] Jop Briët, Harry Buhrman, Debbie W. Leung, Teresa Piovesan, and Florian Speelman. Round elimination in exact communication complexity. In Salman Beigi and Robert König, editors, 10th Conference on the Theory of Quantum Computation, Communication and Cryptography, TQC 2015, May 20-22, 2015, Brussels, Belgium, volume 44 of LIPIcs, pages 206–225. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2015. doi:10.4230/LIPICS.TQC.2015.206.
- [4] Ran Canetti and Oded Goldreich. Bounds on tradeoffs between randomness and communication complexity. Comput. Complex., 3:141–167, 1993. doi:10.1007/BF01200118.
- [5] Arkadev Chattopadhyay, Shachar Lovett, and Marc Vinyals. Equality alone does not simulate randomness. In Amir Shpilka, editor, 34th Computational Complexity Conference, CCC 2019, July 18-20, 2019, New Brunswick, NJ, USA, volume 137 of LIPIcs, pages 14:1–14:11. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPICS.CCC.2019.14.
- [6] Susanna F. de Rezende, Or Meir, Jakob Nordström, Toniann Pitassi, Robert Robere, and Marc Vinyals. Lifting with simple gadgets and applications to circuit and proof complexity. In Sandy Irani, editor, 61st IEEE Annual Symposium on Foundations of Computer Science, FOCS 2020, Durham, NC, USA, November 16-19, 2020, pages 24–30. IEEE, 2020. doi:10.1109/FOCS46700.2020.00011.
- [7] Tomás Feder, Eyal Kushilevitz, Moni Naor, and Noam Nisan. Amortized communication complexity. SIAM J. Comput., 24(4):736–750, 1995. doi:10.1137/S0097539792235864.
- [8] Rudolf Fleischer, Hermann Jung, and Kurt Mehlhorn. A communication-randomness tradeoff for two-processor systems. Inf. Comput., 116(2):155–161, 1995. doi:10.1006/INCO.1995.1011.
- [9] Dale Jacobs, John Jeang, Vladimir Podolskii, Morgan Prior, and Ilya Volkovich. Communication complexity of equality and error correcting codes. Technical Report TR25-068, Electronic Colloquium on Computational Complexity (ECCC), 2025. URL: https://eccc.weizmann.ac.il/report/2025/068/.
- [10] Stasys Jukna. Boolean Function Complexity - Advances and Frontiers, volume 27 of Algorithms and combinatorics. Springer, 2012. doi:10.1007/978-3-642-24508-4.
- [11] Eyal Kushilevitz and Noam Nisan. Communication complexity. Cambridge University Press, 1997.
- [12] Olivier Lalonde. Entanglement-assisted communication complexity and nonlocal games. Master’s Thesis, Université de Montréal, 2023. URL: https://umontreal.scholaris.ca/items/d507c06f-33dc-49f5-ab91-31c2bd60c6bc.
- [13] Olivier Lalonde, Nikhil S Mande, and Ronald de Wolf. Tight bounds for the randomized and quantum communication complexities of equality with small error. arXiv preprint arXiv:2107.11806, 2021.
- [14] Bruno Loff and Sagnik Mukhopadhyay. Lifting theorems for equality. In Rolf Niedermeier and Christophe Paul, editors, 36th International Symposium on Theoretical Aspects of Computer Science, STACS 2019, March 13-16, 2019, Berlin, Germany, volume 126 of LIPIcs, pages 50:1–50:19. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPICS.STACS.2019.50.
- [15] Shay Moran, Makrand Sinha, and Amir Yehudayoff. Fooling pairs in randomized communication complexity. In Jukka Suomela, editor, Structural Information and Communication Complexity - 23rd International Colloquium, SIROCCO 2016, Helsinki, Finland, July 19-21, 2016, Revised Selected Papers, volume 9988 of Lecture Notes in Computer Science, pages 49–59, 2016. doi:10.1007/978-3-319-48314-6_4.
- [16] Ilan Newman. Private vs. common random bits in communication complexity. Information Processing Letters, 39(2):67–71, 1991. doi:10.1016/0020-0190(91)90157-D.
- [17] Noam Nisan. The communication complexity of threshold gates. Combinatorics, Paul Erdős Is Eighty, pages 301–315, 1993.
- [18] Toniann Pitassi, Morgan Shirley, and Adi Shraibman. The strength of equality oracles in communication. In Yael Tauman Kalai, editor, 14th Innovations in Theoretical Computer Science Conference, ITCS 2023, January 10-13, 2023, MIT, Cambridge, Massachusetts, USA, volume 251 of LIPIcs, pages 89:1–89:19. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2023. doi:10.4230/LIPICS.ITCS.2023.89.
- [19] Anup Rao and Amir Yehudayoff. Communication Complexity: and Applications. Cambridge University Press, 2020.
- [20] Alexander A. Razborov. Communication Complexity, pages 97–117. Springer Berlin Heidelberg, Berlin, Heidelberg, 2011. doi:10.1007/978-3-642-19533-4_8.
- [21] Ron M. Roth. Introduction to coding theory. Cambridge University Press, 2006.
- [22] Tim Roughgarden. Communication complexity (for algorithm designers). Found. Trends Theor. Comput. Sci., 11(3-4):217–404, 2016. doi:10.1561/0400000076.
- [23] Andrew Chi-Chih Yao. Some complexity questions related to distributive computing (preliminary report). In Michael J. Fischer, Richard A. DeMillo, Nancy A. Lynch, Walter A. Burkhard, and Alfred V. Aho, editors, Proceedings of the 11h Annual ACM Symposium on Theory of Computing, April 30 - May 2, 1979, Atlanta, Georgia, USA, pages 209–213. ACM, 1979. doi:10.1145/800135.804414.
Appendix A Appendix
A.1 Proof of Lemma 39
We start by illustrating the case of two iterations. In the first step, we apply the Reed-Solomon code with some parameters:
and then instead of sending bits, we apply the Reed-Solomon code (with new values of parameters) to this message again:
Notice that in both iterations we use random bits, but only communicate according to the result of the final iteration. The resulting procedure for two iterations gives us the following parameters
and has the following restrictions
Observe that we are interested in and communication complexity only up to a constant factor. Further, notice that our errors are only additive from the first to the second iteration. Finally, we see that allowing a higher error rate in a given round will never increase the amount of randomness or communication required. Therefore, we can assume that in each iteration we allow the same error ; otherwise, we could simply choose to increase the allowed error in the iteration with a smaller error rate.
With this assumption, on an arbitrary iteration we will have parameters
with the restriction
| (10) |
The total values of parameters after iterations (we assume that is a constant) are
Recall that our protocol using single-iteration Reed-Solomon works only for . Thus our goal is to develop a protocol that is valid for larger values of . Notice that only the final restriction directly contains a restriction on the total number of random bits used in the protocol:
| (11) |
Rearranging this expression, we get
We can see that minimizing the randomness used in the last iteration, , while keeping the same maximizes the range in which this protocol is valid. Naturally to minimize , we need to maximize the randomness used in all previous iterations. Thus, to optimize the parameters we need to fix
on intermediate iterations.
We prove the following stronger version of Lemma 39.
Lemma 41.
Recall that . For any integer constant we have that the communication protocol derived from iterations of Reed-Solomon has complexity and is valid for all
| (12) |
(up to an additive constant).
Proof.
We prove this by induction on . The base case corresponds to the standard Reed-Solomon protocol, which we already covered.
For the induction step, we do one extra step of the iteration and obtain the restriction
Since for the previous iterations we need to maximize , we must fix them so that all restrictions (10) are saturated. By the inductive hypothesis we have that (12) is saturated. Substituting from this equality to above yields
Rearranging this inequality we obtain, using Fact 6,
Adding this to the expression for gives the desired inequality.
Moreover, it is easy to see that if (10) is saturated for as well, the resulting inequality is saturated.
