Round-Vs-Resilience Tradeoffs for Binary Feedback Channels
Abstract
In a celebrated result from the ’s, Berlekamp showed that feedback can be used to increase the maximum fraction of adversarial noise that can be tolerated by binary error correcting codes from to . However, his result relies on the assumption that feedback is “continuous”, i.e., after every utilization of the channel, the sender gets the symbol received by the receiver. While this assumption is natural in some settings, in other settings it may be unreasonable or too costly to maintain.
In this work, we initiate the study of round-restricted feedback channels, where the number of feedback rounds is possibly much smaller than the number of utilizations of the channel. Error correcting codes for such channels are protocols where the sender can ask for feedback at most times, and, upon a feedback request, it obtains all the symbols received since its last feedback request. We design such error correcting protocols for both the adversarial binary erasure channel and for the adversarial binary corruption (bit flip) channel. For the erasure channel, we give an exact characterization of the round-vs-resilience tradeoff by designing a (constant rate) protocol with feedback rounds, for every , and proving that its noise resilience is optimal.
Designing such error correcting protocols for the corruption channel is substantially more involved. We show that obtaining the optimal resilience, even with one feedback round (), requires settling (proving or disproving) a new, seemingly unrelated, “clean” combinatorial conjecture, about the maximum cut in weighted graphs versus the “imbalance” of an average cut. Specifically, we prove an upper bound on the optimal resilience (impossibility result), and show that the existence of a matching lower bound (a protocol) is equivalent to the correctness of our conjecture.
Keywords and phrases:
Round-restricted feedback channel, error correcting code, noise resilienceFunding:
Mark Braverman: Supported in part by the NSF Alan T. Waterman Award, Grant No. 1933331, a Packard Fellowship in Science and Engineering, and the Simons Collaboration on Algorithms and Geometry.Copyright and License:
![[Uncaptioned image]](x1.png)
2012 ACM Subject Classification:
Theory of computation Error-correcting codesAcknowledgements:
We thank Noga Alon for discussions regarding 3.Editors:
Raghu MekaSeries and Publisher:

1 Introduction
Cybernetics.
Consider the following two scenarios. Scenario one: a steersperson wishes to steer a longship to shore. She maintains a steady course in a changing environment (wind, waves, storms, currents, tides, etc.) by adjusting her steering in continual response to the effect it is observed as having. Scenario two: a teacher has a semester-worth of topics he wishes to teach to his class. He schedules exams throughout the semester to help him adapt his pace and determine what material should be repeated.
The above two scenarios are examples of cybernetics, a field that studies self-regulating processes. A core concept in cybernetics is circular causality, which is typically implemented using feedback mechanisms, where the observed outcomes of actions are taken as inputs for further actions. This is the case for, e.g., spacecraft navigators, artificial limbs, and our bodies’ regulation of hormone and blood sugar levels. The term Cybernetics111Interestingly, Cybernetics comes from the Greek word “Kubernetes”, which means steersperson. was coined in 1948 by the mathematician and philosopher Norbert Wiener for “the science of control and communication in the animal and the machine” [37], following exchanges between numerous fields during the 1940s, including anthropology, mathematics, neuroscience, psychology, and engineering.
Feedback in information theory.
Cybernetics grew alongside and built on Claude Shannon’s information theory, that was developed to improve the transmission of information and introduced the notion of error correcting codes. Shannon was interested in knowing whether the existence of a “feedback link” in the channel, where after every utilization of the channel, the (possibly incorrect) symbol obtained by the receiver is also given to the sender, allows for better codes. A discouraging early result by Shannon showed that feedback does not improve the capacity of memoryless channels [31]. It would be another decade or so before Berlekamp proves that feedback can, in fact, increase the maximum fraction of adversarial errors that can be tolerated. Specifically, Berlekamp showed that the maximum noise resilience of the (adversarial) binary channel increases from to given feedback [4, 5] (also see [39, 35, 1]).
A key property of the feedback channel exploited by Berlekamp’s result, as well as by follow up work, is that it supports “continuous” feedback – after every communication round, the sender gets the symbol received by the receiver. This assumption is natural in some settings, e.g., in scenario one, the steersperson continuously watches the ship’s motion as she steers. However, this assumption may be unreasonable or too costly to maintain in other settings, e.g., in scenario two, the teacher may not want to continuously quiz his students.
This work: round-restricted feedback.
Motivated by such examples, in this work, we initiate the study of round-restricted feedback channels, where the number of feedback rounds is possibly much smaller than the number of communication rounds. Specifically, we wish to design protocols with optimal noise resilience that allow the sender (Alice) to transmit a message to the receiver (Bob), where during the execution of the protocol, the sender can ask for feedback at most times. Upon such a request, the sender obtains all the bits received by the receiver from the last time feedback was solicited.
One can consider two models for scheduling the feedback rounds: the adaptive and the non-adaptive models. In the non-adaptive model, the sender decides ahead of time (before the protocol is run and before the input is known) when to schedule the feedback rounds, while in the adaptive model, the timing of each feedback request may depend on the previously received feedback. In the second scenario, for example, the non-adaptive setting corresponds to scheduling all exams at the beginning of the semester, while the adaptive setting corresponds to scheduling the next exam after the previous one was given. While our techniques hold for both the adaptive and non-adaptive settings, we choose to present our results for the non-adaptive setting. See Section 1.3 and Section 2.1 for a discussion of the implication of our techniques for the adaptive setting.
We consider such message transmission protocols with feedback rounds over both the (adversarial) binary erasure channel, that erases some of the sent bits (those bits are received as “”), and over the (adversarial) binary corruption channel, that flips some of the sent bits. As was mentioned before, classical results in information theory show that with no feedback the maximum noise resilience of the binary corruption channel is [25], while with continuous feedback, the maximum resilience improves to [4, 5]. For the binary erasure channel, it is known that with no feedback the maximum resilience is , and it is easy to see that with continuous feedback it approaches : the sender re-transmits each symbol until the receiver receives it.
We mention that rounds (or passes) are often considered to be a scarce resource and that round-restricted algorithms are extensively studied in other communication settings, e.g., communication complexity, distributed computing, streaming algorithms, and cryptographic protocols, and that we draw inspiration from these settings.
1.1 Our Results and Conjecture
Due to page limit, sections containing formal discussion on the corruption case and all proofs are deferred to the full version [8]. Here, we focus on the erasure case and only provide high level ideas for the corruption case.
1.1.1 The (Adversarial) Binary Erasure Channel
As discussed above, the maximum resilience of the erasure channel is known for the extreme cases of no feedback and of continuous feedback. Our first result is an optimal round-vs-resilience tradeoff for the erasure channel with any number of non-adaptive feedback rounds.
Theorem 1.
The maximum noise resilience of the (adversarial) binary erasure channel with rounds of feedback is if and if . Furthermore, the maximum noise resilience can be obtained by a deterministic, constant-rate protocol.
Theorem 1 can be viewed as a “hierarchy theorem”, showing that more feedback rounds allow for strictly better resilience. On the other hand, Theorem 1 also shows that a constant number of feedback rounds already suffices to get a noise resilience of for the erasure channel.
Techniques.
The main ingredient in our proof of Theorem 1 is the construction of a list decodable code for the binary erasure channel with codewords, for all (not necessarily asymptotic) values of . Our code is optimal in the sense that it achieves the maximum error resilience for every list size simultaneously. We emphasize that for our protocols, we need such a code for all possible , which corresponds to all possible “block sizes”. We call codes with small ’s “small codes”. Given these codes, the protocols we use to prove Theorem 1 are rather simple – after every feedback round, Alice and Bob agree on a (smaller, unless there was a lot of noise) set of candidate inputs and Alice encodes with our optimal list decodable code with codewords. On the analysis front, we are able to argue that, unless the adversary erases many of the sent bits, the size of the candidate set shrinks substantially between feedback rounds, and measure this shrinkage exactly. See Section 2.1 for a detailed overview.
1.1.2 The (Adversarial) Binary Corruption Channel
Theorem 1 gives a complete characterization of the noise resilience of the erasure feedback channel as a function of the number of feedback rounds. However, as will be explained next, the case of corruptions is much more involved, and we will focus on protocols with one round of feedback. We mention that since the adaptive and non-adaptive models are the same for protocols with one feedback round, the results in this section hold for both the adaptive and non-adaptive settings. Our next theorem gives an upper bound on the noise resilience of such one-round protocols.
Theorem 2.
The maximum noise resilience of the (adversarial) binary corruption channel with one round of feedback is at most .
We conjecture that the upper bound of on the noise resilience in Theorem 2 is tight, and that it can be achieved by a constant-rate protocol. Perhaps surprisingly, proving this is equivalent to showing the following combinatorial conjecture about the existence of large cuts in graphs.
Conjecture 3.
Let be a graph with vertices and non-negative edge weights summing up to . Let be the sum of weights of all the edges with both endpoints in the subset of vertices , and let be the maximum total weight of all the edges across any cut in . Then,222As the expectations of and , for a uniformly random , are , Equation 1 can be equivalently written as , where the term inside the expectation is the “imbalance” of a random cut.
(1) |
We prove 3 for (large enough) graphs where all edges have equal weight, i.e., “unweighted” graphs. However, the case for general weighted graphs seems much harder, and, despite our best effort, we were unable to prove (or disprove) it. We also mention that 3 is tight for some graphs (e.g., cliques of size and with edges of equal weight), and related bounds on were studied in other contexts, e.g., [26, 2, 22].
Theorem 4.
In essence, Theorem 4 connects the problem of designing optimal error correcting protocols with one round of feedback to a combinatorial question about graphs. As we discuss later in Section 2.2, our techniques can also be used to connect the problem of designing optimal error correcting protocols with multiple rounds of feedback to similar questions about graphs.
Techniques.
The proof of Theorem 4 is technically involved and a detailed overview can be found in Section 2.2. At a high level, the main ingredient in designing our protocol is the construction of a special type of “weighted” codes, called -codes. A -code is parameterized by a “distance contribution function” that assigns a value in to each possible message . We require that for all , the codewords and are at least (relative) Hamming distance apart. Equivalently, we ask that the balls of radii around are all disjoint.333We mention that -codes are an example of non-equally spaced codes defined in [11]. We note that unlike traditional error correcting codes that have only one distance guarantee for all pairs of codewords (i.e., the minimum distance), the distance guarantees for different pairs of codewords in a -code are different. In fact, traditional codes can be viewed as -codes for a constant function.
-codes for non-constant functions are useful for our protocol as if the adversary already used up many of its corruptions before the feedback round, Alice knows she can afford to send her message encoded with an error correcting code that does not guarantee a large distance between and the other codewords. Geometrically, designing a -code is a sphere packing problem where we need to pack spheres of different radii . As for some ’s a small radius suffices, some of the spheres are small, which allows the other spheres being packed to be larger.
The proof of Theorem 4 shows that 3 implies the existence of -codes that are needed for our protocol to work. We assume that Alice uses a uniformly random code to encode her message before the feedback. The codeword sent by Alice can be corrupted by the channel in many ways, and each such way would imply a function such that Alice would like to use a -code to encode her message after the feedback. We denote by the set of functions for which the corresponding -codes are needed by our protocol. We also denote by the set of functions for which -codes exist. We wish to show . To this end, we show that both and are closed and convex, and that in every direction , the extremal point of in direction is “farther” than the extremal point of in direction . We then recast this geometric problem as a combinatorial problem by interpreting the direction vector as a weighted graph , and show that the extremal point of in direction corresponds to a in (as in the left hand side of 3), while the extremal point of in direction corresponds to the right hand side of 3.
For the converse direction of Theorem 4, we show that the arguments in the above paragraph are actually equivalences, except for the assumption that Alice uses a randomly sampled code to encode her message before the feedback. At a high level, we use Ramsey theory to show that the assumption that this code is a random code is, at least in some sense, without loss of generality (see Section 2.2.2 for a more precise statement).
1.2 Related Work
Feedback channels were studied since the early days of information theory and are still actively studied [31, 24, 14, 5, 9, 27, 32, 13, 33, 34, to cite a few]. While feedback does not increase the capacity of discrete memoryless channels with vanishing error, there are settings where feedback is known to allow improvement, like in the -error capacity case [31], and under variable decision time [9].
Partial feedback.
Haeupler, Kamath, and Velingker [23] considered the setting where the feedback is partial, and showed that even if Alice receives feedback bits from Bob for an arbitrarily small constant fraction of her transmissions, resilience close to (the optimal resilience of) is possible using a randomized protocol. However, the number of feedback rounds their protocol needs grows linearly with , the length of Alice’s input. See [36] for a subsequent result.
Independently and concurrently to our work, [15] improved [23] and showed a deterministic protocol that uses feedback bits over feedback rounds to get resilience approaching , along with a similar result for the erasure channel showing that the resilience approaches for this channel. The main difference between [15] and the current work is that we focus on finding the optimal resilience for any given number of feedback rounds whereas [15] focuses on showing that the resilience approaches the optimal value as the constant increases. Additionally, their work measures both the number of feedback rounds and the number of feedback bits, while we only focus on the number of rounds.
Two-way codes and interactive codes.
As discussed above, feedback is also known to increase the noise resilience of the adversarial binary corruption channel [4, 5], and this result played a big role in recent work in interactive coding [10, 18, 17] and two-way coding [16, 19, 11]. In interactive coding [28, 29, 30], we wish to simulate a communication protocol that was designed to work over the noiseless channel, by a protocol that works over a noisy channel. In the setting of two-way codes, like in the setting of traditional error correcting codes, Alice wishes to transmit a message to Bob over a noisy channel. However, unlike the case of traditional codes, where Alice is the only party that can transmit messages, in two-way codes Bob can also use the (noisy) channel to transmit messages back to Alice.
Observe that since Bob has no input, any two-way code can be run over the feedback channel and thus two-way error correcting codes can be viewed as protocols over a noisy feedback channel. In particular, since the noise tolerance of the binary corruption channel is only , the noise resilience of binary two-way codes over the binary corruption channel is at most . In the same way, results for the bounded round feedback channel give upper bounds on the noise resilience of the corresponding two-way channels.
Gupta, Kalai, and Zhang [16, 19] studied two-way error correcting codes over the binary erasure channel. Their main result is a code that is resilient to a fraction of adversarial errors, improving on the noise tolerance of the one-way binary erasure channel that is known to be . We mention that the two-way coding schemes of [16, 19] exchange (almost) linear number of messages. The work of [16] also gives an upper bound of on the maximum tolerance of the two-way binary erasure channel, and an upper bound of on the maximum tolerance of the two-way binary corruption channel. Given those upper bounds, a corollary of our results is that even a single round of noiseless feedback allows for a better error tolerance than any number of noisy feedback rounds over both the erasure and corruption channels444To see why, observe that if Bob’s messages are noiseless, we can assume without loss of generality that Bob’s messages are much shorter, say at most an fraction, of Alice’s messages. Indeed, if not, consider a modified protocol where all messages from Alice are repeated times, for some large . For the erasure channel, either all the repetitions of a bit from Alice are erased or Bob knows the bit exactly. Thus, his communication does not grow with . For the corruption channel, it suffices for Bob to say how many of the repetitions were received as , which can be done using bits. Moreover, we mention that for this claim, we do not need to rely on 3, as a lower bound slightly smaller than (but greater than ) on the maximum error resilience of protocols with one feedback round over the binary corruption channel can be obtained unconditionally using our techniques (but is not included in the current work)..
The recent work of Efremenko, Kol, Saxena, and Zhang [11] shows that the maximum noise resilience of two-way error correcting codes for the binary corruption channel is strictly better than the noise resilience of traditional error correcting codes for this channel, which is known to be [25]. At a very high level, those results for two-way codes are obtained by implementing a (weak) feedback mechanism over channels with no built-in feedback. Related ideas were used in [10, 18, 17] to give interactive binary error correcting codes with high noise resilience.
List decodable codes.
List decodable codes were introduced in the 50’s [12, 38] and have been studied over numerous papers and found many applications since then. We next list the works most related to ours. Most of the work on list decoding was done in the asymptotic regime, where the number of codewords goes to infinity. In this work, we are interested in the optimal list decodable codes for any (potentially small) number of codewords. However, as an ingredient in our proof, we use the asymptotic results of [3] (see also [6, 21, 7]) for optimal list decoding of the corruption channel (see Lemma 11). The list decoding question was also considered for other channels, for example, over the corruption channel with feedback [32] and the erasure channel [20].
1.3 Open Problems
Our work suggests the study of feedback channels through a new lens, namely, their feedback round complexity. We next list some suggestions for future work in this direction.
Graph-theoretic conjectures.
The most immediate question we leave open is proving 3 for all weighted graphs. We also propose the following potentially related conjecture, which is tight for all odd cliques with edges of equal weight.
Conjecture 5.
Let be a graph with vertices and non-negative edge weights summing up to . Let be the sum of weights of all edges incident on vertex . Then,
Round-vs-resilience tradeoff for other channels.
Proving 3 would imply that our protocol in Theorem 4 has optimal noise resilience among protocols with one round of feedback over the corruption channel. Obtaining a general round-vs-resilience tradeoff for any number of feedback rounds for the corruption channel and for other well-studied channels (e.g., the binary insertion-deletion channel555We note that this first requires a suitable definition for the insertion-deletion channel with constant number of rounds., the binary deletion-only channel, and non-binary channels), would be interesting.
Adaptive corruptions over the erasure channel.
Theorem 1 considers the case of non-adaptive feedback rounds, where Alice decides ahead of time when to ask for feedback. It can be shown that the case of adaptive feedback rounds, where Alice chooses when to ask for another round of feedback after seeing the previous feedback, allows for (strictly) better round-vs-resilience tradeoffs. Our techniques can be used to write a recursive formula for the noise resilience in the adaptive case, and finding a “clean”, closed-form formula for this setting (if one exists) is left open (see Section 2.1).
2 Proof Overview
In this section, we overview the proofs of Theorems 1 and 4, starting with the relatively easier Theorem 1.
2.1 Result for the Erasure Channel – Theorem 1
The defining feature of the erasure channel is that the receiver (Bob) either receives the bit sent by Alice or receives a special erasure symbol . This means that in any round where Bob receives , he is certain that this is due to the erasures in the channel, while if he receives a symbol different from , he is certain that the symbol must be what Alice sent in that round. In turn, this means that Bob knows exactly the amount of erasures introduced by the channel and also means that Bob can (recall that he is trying to determine Alice’s input) remove from consideration any candidate input that is “inconsistent” and would make Alice send a different symbol in any such round.
The general format of a protocol.
The above observation implies that protocols for the erasure channel with rounds of feedback (and therefore messages from Alice) have the following format: Alice starts with an input . For her first message, she takes a code666At this point, it may be helpful to view this as a function instead of a code. We explain why we are calling it a code later. Also, a more precise way to state this would be to say that there exists an such that , as all codewords need to be of the same length to avoid the parties from signaling through the length of the codeword. Nonetheless, we stick with statements like throughout this sketch for simplicity. and sends to Bob. Some of the bits of are received correctly by Bob, while the remaining bits are erased and replaced with . Using the bits he received correctly, Bob can calculate the number of erasures introduced by the channel in this round and can identify a subset of inputs for Alice that are consistent with the message he received. Note that Alice’s input must be in .
Then, a feedback round takes place, and as Alice learns all the received symbols, she can also compute and . As both parties now know these values, they can now “forget” this round and “reduce”777We elaborate what this means exactly in the paragraph on adaptive feedback rounds below. to a smaller problem where Alice wants to transmit an element to Bob using a protocol with rounds of feedback and the maximum number of erasures the channel can insert is lower than what it was before. Continuing this way, the goal of the parties is to reduce to a problem with rounds of feedback, and set of inputs such that there exists a (standard) error correcting code for elements in resilient to the number of erasures that the channel can insert in the last round.
List-decodable small codes.
It is readily seen that the protocol format described above does not care about the exact strings in the sets , as long as their sizes stay the same. Thus, the question of whether or not the above protocol format can be instantiated to get a protocol that is resilient to fraction of adversarial erasures, for some , reduces to determining when to schedule the feedback rounds, and given two feedback rounds, determining the codes to be used by Alice between these rounds. The codes should be such that, given an initial set size and a target set size888Note that is not known to the parties in advance, and thus it will be ideal if the code used is optimal for all simultaneously. , the number of erasures required to reduce the set size from to is the highest. Using such codes, Alice ensures that unless the adversary invests many erasures, the set of candidates shrinks substantially between feedback rounds. We first focus on designing such codes.
Codes like the above are known as list decodable codes, and have been well studied in the asymptotic regime, where tends to infinity, and exact answers are known (see, e.g., [12, 38, 6, 20, 7, 3, 32] and Lemma 11). However, for our purposes, we need the exact answer for smaller values of as well. Codes with small , i.e., “small codes” or codes with few codewords, have recently received a lot of attention and have proven to be useful in designing binary protocols with high error resilience in several contexts [10, 11, 16, 18, 19]. In the current paper, we provide a complete analysis of the list-decodability of these codes for the erasure channel, giving a function that characterizes exactly the minimum amount of erasure noise needed such that for any code , one can erase fraction of the bits and ensure that Bob gets a list of candidates of length strictly smaller than .
The formula for is given in Equation 7. Proving that this formula is correct requires showing both a construction (of codes with resilience approaching ) and an impossibility result. Our construction has the nice property that the same code is tight simultaneously for all values of . Roughly speaking, our code achieves this optimal erasure noise resilience by ensuring that every coordinate is as differentiating as possible, i.e., we ensure that for all coordinates , exactly (uniformly chosen) codewords have in that coordinate, while the remaining codewords have (see Lemmas 9 and 10). This is as opposed to randomly sampled codes where, e.g., a fraction (which is large for small ) of the coordinates are expected to be for all the codewords, and therefore not differentiate between any pair of codewords.
Scheduling the feedback rounds.
Even with an exact formula for in hand, it still remains to schedule the feedback round correctly in order to maximize the overall noise resilience of the obtained protocol. The fact that our constructed code is tight simultaneously for all values of is of great help for this part, as the actual value of is determined by the erasures inserted by the channel and not in our control. This means that in order to schedule the feedback rounds optimally, one needs to go over all possible values of (across all rounds) that may happen over the channel and maximize the corresponding error resilience. This requires a careful analysis of the obtained formula for .
Adaptive feedback rounds.
We finish this section by briefly discussing the extension of our result to adaptive feedback rounds, as hinted in Section 1.3. Recall our reduction above from to feedback rounds, and note that this reduction is not perfect in the following sense: the erasures inserted by the adversary in Alice’s first message in the -round protocol dictate the set of candidates and the budget of the -round protocol. Observe that the -round protocol with maximal noise resilience for transmitting a message depends on the size of the set of candidates and on the erasure budget. Now, since our -round protocol is non-adaptive, meaning that the timing of all feedback rounds is fixed in advance and cannot be recalculated given the erasures in the first round, our -round protocol may reduce to a sub-optimal -round protocol. Therefore, when scheduling the feedback rounds for our -round protocol, one needs to consider the values of that are possible across all rounds in order to get the optimal schedule.
On the other hand, if the feedback rounds can be scheduled adaptively, the reduction is indeed perfect. In this case, one just needs to schedule the first feedback round beforehand based on the possible values of for this round alone, and then, upon seeing the and values, one can take the -feedback round protocol with the maximum error resilience (when Alice’s input is from and the number of erasures is lower) and schedule the remaining feedback rounds according to this protocol. Thus, our techniques also lead to a tight recursive formula for the maximum error resilience in the case of adaptive feedback rounds, but converting it to a “clean” closed form formula (if at all possible) is left open.
2.2 Result for the Corruption Channel – Theorem 4
Compared to the erasure channel, where Bob knows exactly the amount of noise inserted and can safely eliminate many candidate inputs for Alice, the corruption channel is much harder. Here, upon receiving a message from Alice, all Bob can compute is, given a candidate input for Alice, what is the number of corruptions the channel inserted assuming Alice’s input was indeed . Crucially, this value of may be very different for different , and unless it exceeds the maximum possible number of corruptions in the channel (which can only happen when the protocol is quite far advanced), it can never have Bob eliminate from consideration entirely.
Consider now a protocol over the corruption channel with one round of feedback (and therefore, two messages from Alice). Suppose that Alice’s input comes from the set . As explained above, after receiving the first message from Alice, Bob knows for all . By subtracting from the maximum possible number of corruptions, Bob can compute, for all , a number which is the leftover corruptions, or, equivalently, the degree to which the second message of Alice can be corrupted, assuming her input is . As Alice receives feedback from Bob, she can also compute the values for all . In the remainder of this sketch, we normalize by dividing it by the length of Alice’s second message. This will result in a value in .
dc-codes.
Using this feedback, Alice’s goal in her second message is to allow Bob to uniquely identify her input. If is the code used by Alice in her second message, the only way Bob can uniquely decode Alice’s input is if for all , the codewords and are at least (relative) Hamming distance apart. The reason is that if is Alice’s input, then the adversary has fractional budget that it can use to corrupt , and thus the codeword received by Bob can be any string of (relative) Hamming distance at most from . Similarly, if is Alice’s input, then the codeword received by Bob can be any string of Hamming distance at most from . Note that the adversary cannot arrange for the received encodings to be the same if and only if and are at least (relative) Hamming distance apart. We call a code that satisfies this (relative) Hamming distance property a -code and mention that the values can equivalently be seen as the “distance contributed” by in such a code.
We note that unlike traditional error correcting codes that have only one distance guarantee for all pairs of codewords (i.e., the minimum distance), for -codes, the distance between a pair of codewords may be different depending on the “compatibility” of the messages they encode. Specifically, we think of each codeword as having a different “radius” and the code needs to “pack” all the induced balls of different radii. We point out that -codes are an example of non-equally spaced codes defined in [11].
We also observe that the small code used in our protocol for erasures can be viewed as a -code where for all inputs that Bob has ruled out (and therefore, do not need any distance guarantees), and for all inputs that he has not ruled out, where is the best possible constant ( is determined by the function). We mention that for the erasure channel, our protocol also needed list-decoding guarantees that are not needed here as we are only attempting to get a one feedback round protocol.
The discussion so far shows that the existence of a protocol with a given error resilience amounts to determining whether or not it holds that for all functions that can be induced by the corruptions inserted in Alice’s first message, there exists a -code that Alice can use to compute her second message. Curiously, we show in the next subsection that this question is equivalent to our seemingly unrelated combinatorial conjecture (3) about the existence of large cuts in graphs.
Towards multiple rounds of feedback.
The above approach of designing -codes (that have no rounds of feedback) to construct protocols with one round of feedback can be generalized. One can similarly argue that, for any , -codes with rounds of feedback can be used to construct protocols with rounds of feedback. Analogously to the above, the “extra” round is the first round, and dictates which -code is used in the rest of the protocol. Moreover, questions about constructing -codes with rounds of feedback can be translated to questions about graphs. The case is explained next, but similar ideas may be used for general , with appropriate changes in the definitions of the set and (see below).
2.2.1 3 Implies a Tight Protocol
We first show why 3 implies the existence of a tight protocol. In fact, we shall show the existence of a protocol where Alice’s message in the first round is simply the encoding of her input using a randomly sampled code. Let . A distance function is a function , where is the set of all subsets of of size . For a code , we denote by the distance function induced by , i.e., is the (relative) Hamming distance between and . For a distance contribution function , we denote by the distance function induced by , i.e., . For simplicity, throughout this overview we assume that (recall that is actually the normalized leftover corruption count, but in this sketch we will ignore the exact multiplicative and additive constants in this function).
Recasting as a geometric problem.
We denote by the set of all distance functions that are induced by codes . We denote by the set of all distance functions induced by functions that can be obtained by the corruptions inserted in Alice’s first message (recall that depends on , which is a function of the corruptions inserted in Alice’s first message). In other words, is the set of distance functions that can be realized and is the set of distance functions required by our protocol. We wish to prove .
We view distance functions as -dimensional vectors. We observe that both and are closed and convex and that the set is “downwards-closed”, meaning that if then any that is coordinate wise at most is also in . This means that showing is equivalent to showing that for all -dimensional non-negative hyperplanes , it holds that:
(2) |
Recasting as a combinatorial problem.
By scaling, we can assume that the entries of sum to and view them as the weights on the edges of an -vertex graph as in 3. As both and are closed and convex, both the maximums are attained at one of their vertices.
To reason about Equation 2, it will be useful to represent a code as a sequence of one-bit functions (the first one-bit function corresponds to the first coordinate of , etc.). Observe that for the code (i.e., ), it holds that is a boolean function with if and only if .
The LHS of Equation 2.
Since a general code is a sequence of one-bit functions, it can be shown that the function is a convex combination of the functions that are induced by one-bit functions . In particular, this means that the vertices of are distance functions induced by one-bit functions. Using the expression above for for one-bit function , the value of is the value of the cut in the graph indicated by :
(3) |
Thus, the left hand side of Equation 2 is the maximum cut in , as in 3.
The RHS of Equation 2.
We view the code used by Alice in her first message as a sequence of one-bit functions. Since in our protocol this code is randomly sampled, each of the one-bit functions is expected to appear equally often in Alice’s message999We mention that this is only in expectation and the length of Alice’s message need not depend exponentially on .. As the channel can corrupt each of these one-bit functions independently of all the others, we get that a distance function can be induced by the corruptions inserted in Alice’s first message (i.e., ) if and only if it is the expectation (under the uniform distribution over one-bit functions) of the distance functions that can be induced by corrupting one-bit functions.
Now, if Alice is sending a one-bit function , there are only two possibilities for Bob: either he receives a or he receives a . Let be the distance contribution function dictated by Bob’s received bit. We next show that in the former case, where Bob receives , the value of is the value of the cut in indicated by plus twice the weight of all edges such that on both its endpoints. To see that, recall that and that we assume . In our case, if (Alice’s bit was not corrupted) and if (Alice’s bit was corrupted). This implies that if , and that if , and that if . Therefore,
(4) |
Similarly, it can be shown that in the latter case, where Bob gets , the value of is the value of the cut in indicated by plus twice the weight of all edges such that on both its endpoints.
Recall that the bit function is a uniformly random bit-function. Taking an expectation over one bit functions , the value of the cut in indicated by is exactly the constant and the other terms on the right hand side of Equation 3 and Equation 4 are exactly as on the right hand side of 3, where the maximum becomes minimum because of the constants involved. Equation 2 now directly follows from 3.
2.2.2 A Tight Protocol Implies 3
We now finish this sketch by arguing why a tight protocol implies 3. For this, we note that all the arguments in Section 2.2.1 were actually equivalences, except two, one of which was explicitly stated and one was not. The explicit one was our assumption that Alice’s first message is simply the encoding of her input using a randomly sampled code. The second one was that Alice gets feedback from Bob at round , where is the total number of rounds of the protocol. The constant may seem arbitrary, but it is the constant one gets when one tries to match the constants obtained in the analysis in Section 2.2.1 with the constants in 3.
Both these assumptions are actually without loss of generality. We start by arguing this for the second one, again ignoring the actual constants and only stating the high level idea. Roughly speaking, the second assumption is without loss of generality as 3 is tight for cliques of size and , and if the constant is anything other than , Equation 2 will fail to hold for corresponding to one of these cliques.
It remains to show why the first assumption is without loss of generality. For this, our approach is to take an arbitrary code that Alice may use for her first message, and in several steps, convert it to a code that looks more and more like a random code, at the cost of a smaller . In each step , we convert to a code that is -random, in the sense that any set of codewords of the new code looks like codewords from a randomly sampled code.
For , this means that we have to show that each codeword has an equal number of s and s, and this can be easily achieved by concatenating all codewords with their negations (which preserves the distance properties). We now show how to get a -random code from a -random code, noting that similar (but technically more involved) ideas allow us to get a -random code from a -random code, for any . To show that a code is -random, we need to show that it is -random and that the fractional distance between any pair of codewords is (roughly) .
For this, let be an error parameter and construct a complete graph with the codewords as the vertices, and color the edge between codewords and as (1) red, if the fractional distance between them is smaller than , (2) blue, if the fractional distance between them is between and , (3) green, if the fractional distance between them is larger than . As gets larger and larger, Ramsey theory tells us that there must exist a large (going to infinity with ) monochromatic clique in this graph. This clique cannot be red, as we show that a large number of pairwise close codewords can be used to break the protocol. It also cannot be green, as that would violate known distance bounds for error correcting codes. Thus, it must be blue, implying that restricting attention to this clique gives us our desired -random code.
3 Model and Preliminaries
3.1 Notation and Preliminaries
For , let be the vector (of appropriate dimension inferred from context) with all its coordinates being . Throughtout, all inequalities between vectors are coordinate-wise. For , denotes the -dimensional standard simplex. For and , we write as a shorthand for falling factorial .101010See also Falling factorials (Wikipedia) for the notation. For a set and , let be the collection of all subsets of of size . For a function and subset , denotes the restriction of onto . For , is the (two-color) Ramsey number for , which is well-known to be finite. For and two bit strings , their Hamming distance is .
3.2 Our Model: Round-Restricted Binary Feedback Channels
We now define (deterministic, binary) protocols with (non-adaptive) round-restricted feedback for the message transfer task, where Alice has an input and Bob’s goal is to learn this input. Such a protocol is defined by a tuple:
(5) |
where (1) is the set of all possible inputs for Alice; (2) is the number of feedback rounds. Equivalently, we can say that Alice speaks in rounds; (3) For all , is the length of Alice’s message in the -th round. Throughout, we use ; (4) For all , is the message function Alice uses in the -th round; (5) is the function Bob uses to compute the output.
Execution of a protocol.
Let be a protocol as above. An adversary for is defined by a function . For , we will use to denote the function that outputs the -th coordinate of . We next define an execution of in the presence of an adversary for : At the beginning of the execution, Alice starts with an input . The execution consists of rounds and before the -th round, for , Alice and Bob have the (same) transcript . In round , Alice computes the message and sends it to Bob bit by bit, while Bob receives the string . As we assume a feedback channel, if , Alice also receives the string and both the parties add to and continue executing the protocol.
If , the execution of the protocol terminates and Bob outputs . Observe that this execution is completely determined by , , and . We denote the output of on input in the presence of adversary by .
Counting the noise.
Let be a protocol as above and be an adversary for . For , the amount of noise added by in on input is the number of times Bobs’ received bit is different from the bit Alice sent. Formally, we have:
(6) |
For , we say that an adversary has budget if we have
Types of Adversaries.
Let be a protocol as above and be an adversary for . We say that is a corruption adversary if it never outputs the symbol , i.e., for all and all , we have . We say that is an erasure adversary if it only “erases” the symbols sent by Alice. More precisely, we say that is an erasure adversary if for all , all , and all , if , then we have .
Resilience of a protocol.
Let be a protocol as above and . We say that has resilience over the binary erasure channel if for all erasure adversaries with budget and all , it holds that . Resilience over the binary corruption channel is defined analogously.
4 Optimal List-Decodable Small Codes
In this section, we construct the codes used by our protocol.
4.1 Definitions of List Decodability
Codes for erasures.
We start by defining list decodability for erasures.
Definition 6.
Let and . We say that a code is less-than--list decodable for erasures up to radius if for all subsets , we have , where:
To get the intuition behind the definition of , observe that is the minimum fraction of erasures for which there exists such that for all , it is possible to erase symbols from and get . Observe that this is equal to the fraction of coordinates where the encodings are not all the same ( not same).
For , we define to be the supremum of all values for which there exists and a code that is less-than--list decodable for erasures up to radius .
Codes for corruptions.
Next, we define list decodability for corruptions:
Definition 7.
Let and . We say that a code is less-than--list decodable for corruptions up to radius if for all , we have
Analogous to , for , we define to be the supremum of all values for which there exists and a code that is less-than--list decodable for corruptions up to radius .
4.2 Lemmas about and
In this section, we show the results we need about and . First, we define a helper function :
(7) |
We now show some useful properties of the function defined in Equation 7.
Claim 8.
The following hold:
-
1.
For all , it holds that
-
2.
For all , it holds that , and moreover,
It follows that for all .
-
3.
For all , it holds that .
4.2.1 Lemmas about
We now show that the functions and are the exact same. Owing to this lemma, we omit writing in the subscript in the rest of this text.
Lemma 9.
For all , we have:
Lemma 10.
For all , there exists a constant such that for all and for all , there exists a code such that for all , the code is less-than--list decodable up to radius .
4.2.2 Lemmas about
Using the results of Section 4.2.1, we show the following lemma:
Lemma 11.
For all , we have:
5 Protocols Against Erasures
In this section, we show one direction of Theorem 1, as formalized below. Later, in Section 6, we prove the other direction.
Theorem 12.
For all and , there exists a constant-rate (polynomial in ) protocol for message transfer with rounds of feedback, input length , and the following resilience over the binary erasure channel:
We prove Theorem 12 in the rest of this section. Throughout, we fix and . We assume . This is without loss of generality as a protocol for large follows from a protocol for smaller .
5.1 Our Protocol
Let be the constant from Lemma 10 for . For all and all , let be as promised by Lemma 10. We will omit when it is clear from context. For a set of size , we will also view as a code . Our protocol is given in Algorithm 1, where the lengths of the rounds are given as follows:
(8) |
5.2 Analysis
We now analyze Algorithm 1 and finish proving Theorem 12. That the protocol is constant rate is clear from Algorithm 1. It remains to show that it has the claimed noise resilience. For this, we fix an input for Alice and an erasure adversary for the protocol with desired budget as in Theorem 12. Observe that fixing and fixes the value of all the variables in the execution of Algorithm 1. For the analysis, we first show that:
Lemma 13.
For all , we have .
Lemma 14.
For all such that , it holds that
At a high level, Lemma 14 will be applied as follows: Consider an adversary that shrinks the set from size to size in a given round and from size to size in the next round. Lemma 14 shows that, if the two rounds are of equal length (recall from Equation 8 that the round lengths are always the same except when ), then it is always better for the adversary to erase one of the rounds completely and shrink from size to size directly in the other round.
We now divide the proof into two cases based on whether or not .
5.2.1 Proof of Theorem 12 When
Let for . We prove the theorem by showing that . Together with Lemmas 13 and 9, this shows the correctness of Algorithm 1.
Observe that at the beginning of the -th round, Alice and Bob agree on , the subset of all remaining possibilities for from the perspective of Bob, that are still consistent with the partial transcript so far. In order to keep Bob confused among , the adversary has to erase at least a fraction of Alice’s -th message, due to Lemma 10. As this holds for all rounds, the overall fraction of erasures is lower bounded by
Now suppose . It is sufficient to show for a contradiction. Without loss of generality, we assume since only decrease as becomes smaller by Item 3 of 8. If , we have
by Item 2 of 8. Otherwise, by Lemma 14, we also get
(as is always upper bounded by ) | ||||
(by Item 2 of 8) | ||||
5.2.2 Proof of Theorem 12 When
Let for . Similarly to the proof in Section 5.2.1, we show that . This ensures Bob outputs the correct because of Lemmas 13 and 9. Using a similar argument to Section 5.2.1, we have that the overall fraction of erasures is lower bounded by
Now for the purpose of contradiction, suppose that . It is sufficient to show
In the following, we again assume without loss of generality that since decreases as becomes smaller by Item 3 of 8. Let . By repeatedly applying Lemma 14, we have
Since by definition of , either or .
6 Impossibility Result for Erasures
In this section, we show the other direction of Theorem 1, as formalized below.
Theorem 15.
For all , there exists an such that the resilience of any protocol for message transfer with rounds of feedback and input length over the binary erasure channel is at most:
We prove Theorem 15 in the rest of this section. Throughout, we work with a fixed and define to be large enough for asymptotic inequalities to hold. We now divide the proof into two cases based on whether or not .
6.1 Proof of Theorem 15 When
Fix a protocol with input length and one round of feedback. Recall Equation 5 and let be the lengths of Alice’s messages sent in the two rounds, and be the two message functions Alice uses in the two rounds.
First suppose that . By Lemma 9, there exists a subset such that . This implies the adversary is able to erase a fraction of Alice’s first message so that Bob’s view when Alice’s input is is identical to Bob’s view when Alice’s input is , and therefore Bob is forced to send the same feedback in both cases. Now the adversary simply erases Alice’s second message entirely implying that Bob can never output the correct answer. By Item 2 of 8, the overall fraction of erasures is upper bounded as
Now consider the other case where . Again by Lemma 9, there exists a subset such that . In this case, the adversary erases a fraction of Alice’s first message so that Bob’s view is the same when Alice’s input is any of . Bob must send the same feedback in all three cases. Note that can also be viewed as a valid code and thus Lemma 9 still applies. In particular, it is always possible to erase a fraction of Alice’s second message so that for at least two of , Bob’s view remains the same at the end of the protocol. This concludes the proof as the overall fraction of erasures is at most
6.2 Proof of Theorem 15 When
Fix a protocol with input length and rounds of feedback. Recall Equation 5 and for , let be the length of Alice’s message sent in the -th round. Let .
We prove the theorem using an approach similar to Section 6.2, i.e., the adversary is always able to erase Alice’s messages in such a way that Bob has the same view at the end of the protocol for at least two different inputs. In particular, the adversary erases the entire messages of all rounds except for , the longest round, and , the second longest round. Then we have
(9) | ||||
(10) |
First consider the case where . Since the first rounds are completely erased, Bob obviously has the same view for all possible inputs at the beginning of the -th round. By Lemma 9, the adversary can erase a fraction of Alice’s -th message so that Bob’s view is the same when Alice’s input is any of some subset . This remains true at the beginning of the -th round as all intermediate rounds are completely erased. Now again by Lemma 9, the adversary is able to erase a fraction of Alice’s -th message so that Bob still has the same view at the end of the -th round, for at least two of . As all remaining rounds are also completely erased, Bob can never output the correct answer at the end of the protocol. By Item 2 of 8, the overall fraction of erasures is upper bounded as
(by Equation 10) | |||
(as for , and by Equation 9) | |||
Now suppose that . In this case, a similar argument shows the adversary must be able to confuse Bob by erasing a fraction of Alice’s -th message as well as a fraction of Alice’s -th message (in addition to completely erasing all other rounds of messages). Observe that by definition and that . So the overall fraction of erasures is at most
which has the desired upper bound as already shown above.
References
- [1] Rudolf Ahlswede, Christian Deppe, and Vladimir S. Lebedev. Non-binary error correcting codes with noiseless feedback, localized errors, or both. In International Symposium on Information Theory (ISIT), pages 2486–2487, 2006. doi:10.1109/ISIT.2006.262057.
- [2] Noga Alon. Voting paradoxes and digraphs realizations. Adv. Appl. Math., 29(1):126–135, 2002. doi:10.1016/S0196-8858(02)00007-6.
- [3] Noga Alon, Boris Bukh, and Yury Polyanskiy. List-decodable zero-rate codes. IEEE Trans. Inf. Theory, 65(3):1657–1667, 2019. doi:10.1109/TIT.2018.2868957.
- [4] Elwyn R. Berlekamp. Block coding with noiseless feedback. PhD thesis, Massachusetts Institute of Technology, USA, 1964. URL: http://hdl.handle.net/1721.1/14783.
- [5] Elwyn R Berlekamp. Block coding for the binary symmetric channel with noiseless, delayless feedback. Error-correcting codes, pages 61–68, 1968.
- [6] Vladimir Markovich Blinovskii. Bounds for codes in the case of finite-volume list decoding. Problemy Peredachi Informatsii, 22(1):11–25, 1986.
- [7] Vladimir M Blinovsky. Plotkin bound generalization to the case of multiple packings. Problems of Information Transmission, 45(1):1–4, 2009. doi:10.1134/S0032946009010013.
- [8] Mark Braverman, Klim Efremenko, Gillat Kol, Raghuvansh Saxena, and Zhijun Zhang. Round-vs-resilience tradeoffs for binary feedback channels. Electron. Colloquium Comput. Complex., TR22-179, 2022. URL: https://eccc.weizmann.ac.il/report/2022/179, arXiv:TR22-179.
- [9] Marat Valievich Burnashev. Data transmission over a discrete channel with feedback. random transmission time. Problemy peredachi informatsii, 12(4):10–30, 1976.
- [10] Klim Efremenko, Gillat Kol, and Raghuvansh R. Saxena. Binary interactive error resilience beyond (or why ). In Sandy Irani, editor, 61st IEEE Annual Symposium on Foundations of Computer Science, FOCS 2020, Durham, NC, USA, November 16-19, 2020, pages 470–481. IEEE, 2020. doi:10.1109/FOCS46700.2020.00051.
- [11] Klim Efremenko, Gillat Kol, Raghuvansh R. Saxena, and Zhijun Zhang. Binary codes with resilience beyond 1/4 via interaction. In 63rd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2022, Denver, CO, USA, October 31 - November 3, 2022, pages 1–12. IEEE, 2022. doi:10.1109/FOCS54457.2022.00008.
- [12] Peter Elias. List decoding for noisy channels. Technical report, Research Laboratory of Electronics, Massachusetts Institute of Technology, 1957.
- [13] Krishnan Eswaran, Anand D. Sarwate, Anant Sahai, and Michael Gastpar. Zero-rate feedback can achieve the empirical capacity. IEEE Transactions on Information Theory, 56(1):25–39, 2010. doi:10.1109/TIT.2009.2034779.
- [14] David G. Forney. Exponential error bounds for erasure, list, and decision feedback schemes. IEEE Transactions on Information Theory, 14(2):206–220, 1968. doi:10.1109/TIT.1968.1054129.
- [15] Meghal Gupta, Venkatesan Guruswami, and Rachel Yun Zhang. Binary error-correcting codes with minimal noiseless feedback. In Barna Saha and Rocco A. Servedio, editors, Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, Orlando, FL, USA, June 20-23, 2023, pages 1475–1487. ACM, 2023. doi:10.1145/3564246.3585126.
- [16] Meghal Gupta, Yael Tauman Kalai, and Rachel Yun Zhang. Interactive error correcting codes over binary erasure channels resilient to > ½ adversarial corruption. In Stefano Leonardi and Anupam Gupta, editors, STOC ’22: 54th Annual ACM SIGACT Symposium on Theory of Computing, Rome, Italy, June 20 - 24, 2022, pages 609–622. ACM, 2022. doi:10.1145/3519935.3519980.
- [17] Meghal Gupta and Rachel Yun Zhang. Efficient interactive coding achieving optimal error resilience over the binary channel. CoRR, abs/2207.01144, 2022. doi:10.48550/arXiv.2207.01144.
- [18] Meghal Gupta and Rachel Yun Zhang. The optimal error resilience of interactive communication over binary channels. In Stefano Leonardi and Anupam Gupta, editors, STOC ’22: 54th Annual ACM SIGACT Symposium on Theory of Computing, Rome, Italy, June 20 - 24, 2022, pages 948–961. ACM, 2022. doi:10.1145/3519935.3519985.
- [19] Meghal Gupta and Rachel Yun Zhang. Positive rate binary interactive error correcting codes resilient to adversarial erasures. CoRR, abs/2201.11929, 2022. arXiv:2201.11929, doi:10.48550/arXiv.2201.11929.
- [20] Venkatesan Guruswami. List decoding from erasures: bounds and code constructions. IEEE Trans. Inf. Theory, 49(11):2826–2833, 2003. doi:10.1109/TIT.2003.815776.
- [21] Venkatesan Guruswami and Madhu Sudan. List decoding algorithms for certain concatenated codes. In Proceedings of the thirty-second annual ACM symposium on Theory of computing, pages 181–190, 2000. doi:10.1145/335305.335327.
- [22] Gregory Z. Gutin and Anders Yeo. Lower bounds for maximum weighted cut. CoRR, abs/2104.05536, 2021. doi:10.48550/arXiv.2104.05536.
- [23] Bernhard Haeupler, Pritish Kamath, and Ameya Velingker. Communication with partial noiseless feedback. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM), volume 40 of LIPIcs, pages 881–897, 2015. doi:10.4230/LIPIcs.APPROX-RANDOM.2015.881.
- [24] Michael Horstein. Sequential transmission using noiseless feedback. IEEE Transactions on Information Theory, 9(3):136–143, 1963. doi:10.1109/TIT.1963.1057832.
- [25] M. Plotkin. Binary codes with specified minimum distance. IRE Transactions on Information Theory, 6(4):445–450, 1960. doi:10.1109/TIT.1960.1057584.
- [26] Svatopluk Poljak and Daniel Turzík. A polynomial time heuristic for certain subgraph optimization problems with guaranteed worst case bound. Discrete Mathematics, 58(1):99–104, 1986. doi:10.1016/0012-365X(86)90192-5.
- [27] Anant Sahai. Why do block length and delay behave differently if feedback is present? IEEE Transactions on Information Theory, 54(5):1860–1886, 2008. doi:10.1109/TIT.2008.920339.
- [28] Leonard J Schulman. Communication on noisy channels: A coding theorem for computation. In Foundations of Computer Science (FOCS), pages 724–733, 1992. doi:10.1109/SFCS.1992.267778.
- [29] Leonard J Schulman. Deterministic coding for interactive communication. In Symposium on Theory of computing (STOC), pages 747–756, 1993. doi:10.1145/167088.167279.
- [30] Leonard J. Schulman. Coding for interactive communication. IEEE Transactions on Information Theory, 42(6):1745–1756, 1996. doi:10.1109/18.556671.
- [31] Claude E. Shannon. The zero error capacity of a noisy channel. IRE Transactions on Information Theory, 2(3):8–19, 1956. doi:10.1109/TIT.1956.1056798.
- [32] Ofer Shayevitz. On error correction with feedback under list decoding. In IEEE International Symposium on Information Theory, ISIT, pages 1253–1257, 2009. doi:10.1109/ISIT.2009.5205965.
- [33] Ofer Shayevitz and Meir Feder. Optimal feedback communication via posterior matching. IEEE Transactions on Information Theory, 57(3):1186–1222, 2011. doi:10.1109/TIT.2011.2104992.
- [34] Ofer Shayevitz and Michèle A. Wigger. On the capacity of the discrete memoryless broadcast channel with feedback. IEEE Transactions on Information Theory, 59(3):1329–1345, 2013. doi:10.1109/TIT.2012.2227670.
- [35] Joel Spencer, Peter Winkler, and South St. Three thresholds for a liar. Combinatorics, Probability and Computing, 1:81–93, 1992. doi:10.1017/S0963548300000080.
- [36] Gang Wang, Yanyuan Qin, and Chengjuan Chang. Communication with partial noisy feedback. In IEEE Symposium on Computers and Communications (ISCC), pages 602–607, 2017. doi:10.1109/ISCC.2017.8024594.
- [37] Norbert Wiener. Cybernetics or Control and Communication in the Animal and the Machine. MIT press, 2019.
- [38] John M Wozencraft. List decoding. Quarterly Progress Report, 48:90–95, 1958.
- [39] K.Sh. Zigangirov. On the number of correctable errors for transmission over a binary symmetrical channel with feedback. Problems of Information Transmission, 12:85–97, 1976.