Abstract 1 Introduction 2 Proof Overview 3 Preliminaries 4 Low-Error Affine Condensers 5 Extractors for OBF Sources References

Bit-Fixing Extractors for Almost-Logarithmic Entropy

Dean Doron ORCID Ben-Gurion University of the Negev, Beer-Sheva, Israel Ori Fridman ORCID Ben-Gurion University of the Negev, Beer-Sheva, Israel
Abstract

An oblivious bit-fixing source is a distribution over {0,1}n, where k bits are uniform and independent and the rest nk are fixed a priori to some constant value. Extracting (close to) true randomness from an oblivious bit-fixing source has been studied since the 1980s, with applications in cryptography and complexity theory.

We construct explicit extractors for oblivious bit-fixing source that support k=O~(logn), outputting almost all the entropy with low error. The previous state-of-the-art construction that outputs many bits is due to Rao [Rao, CCC ’09], and requires entropy klogcn for some large constant c. The two key components in our constructions are new low-error affine condensers for poly-logarithmic entropies (that we achieve using techniques from the nonmalleable extractors literature), and a dual use of linear condensers for OBF sources.

Keywords and phrases:
Seedless extractors, oblivious bit-fixing sources
Category:
RANDOM
Copyright and License:
[Uncaptioned image] © Dean Doron and Ori Fridman; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Expander graphs and randomness extractors
Acknowledgements:
We thank Jesse Goodman for many helpful and interesting conversations, and Mahdi Cheraghchi for helpful discussion on the cryptographic applications of OBF extractors.
Funding:
The work is supported in part by NSF-BSF grant #2022644.
Editors:
Alina Ene and Eshan Chattopadhyay

1 Introduction

Whenever randomness is used in computations–whether because it is necessary or just because it is faster and simpler in practice, access to unbiased bits is assumed. Because sources of perfect randomness are notoriously hard to come by, ad hoc, very weak sources of randomness are used instead. It is essential, therefore, that the crude randomness generated by such sources be purified, thus driving the development of the beautiful theory of randomness extractors. This problem of extracting randomness from imperfect sources can be traced back to von Neumann [54]. A randomness extractor is a deterministic procedure that converts a weak random source into a random source that is close to uniform.

Definition 1 (seedless extractor).

Let 𝒳 be a family of distributions over {0,1}n. We say that 𝖤𝗑𝗍:{0,1}n{0,1}m is an extractor for 𝒳 with error ε, if for every X𝒳, 𝖤𝗑𝗍(X) is ε-close, in total variation distance, to Um, the uniform distribution over m bits.

One common assumption is that each weak source X𝒳 has min-entropy.111We say that X{0,1}n has min-entropy k, H(X)k, if maxxXPr[X=x]2k. We call such X an (n,k) source. Unfortunately, if all we assume is an entropy guarantee, it is easy to show that such an 𝖤𝗑𝗍 does not exist. However, seedless extraction is possible for some restricted classes of sources,222To extract from general (n,k) sources, one can use an additional short uniform seed, driving the rich theory of seeded extractors. See Definition 8 for the definition. and indeed, studying the capabilities and limitations of extraction from natural families of structured sources has been a fruitful endeavor for the past four decades (see [27, Section 1.3] for a comprehensive up-to-date survey of seedless extraction results).

One of the simplest and most natural families of weak sources of randomness is oblivious bit fixing sources, wherein k bits are uniform and independent, while the rest nk bits are fixed.

Definition 2 (OBF source).

A distribution X{0,1}n is an (n,k) oblivious bit-fixing source if there exists a subset I[n] of size k of “good indices”, such that the {Xi}iI-s are uniform and independent, and the rest are fixed.333To clarify the terminology, the word “oblivious” refers to the fact that the fixed bits are chosen before the random bits are tossed. On the other hand, in nonoblivious bit fixing sources, the non-uniform bits can depend arbitrarily on the values of the good ones (see Section 3.4).

In addition to simulating true randomness, extractors have been found to be useful, even essential, in myriad applications, and extractors for oblivious bit-fixing sources are no exception.

In cryptography, especially in exposure resilient cryptography (see, e.g., [21]), there are several related primitives for which OBF extractors are useful. (Cryptographic) resilient functions (and almost perfect resilient functions) are essentially OBF extractors, possibly under a different choice of error measure [6, 53]. Exposure resilient functions are a weaker variant, wherein the security is guaranteed only on average over the uniform bits [9]. Both primitives are useful for wiretap protocols, where two parties wish to agree on a random string in the presence of an adversary that can learn some of the transmitted bits. Another related notion is that of all-or-nothing transforms, used for protection of block ciphers [49, 32, 21, 22] (see also [15], and the discussion in [15, Appendix A]). OBF extractors were also used for distributed generation of correlated randomness [23]. Outside of cryptography, OBF extractors have also found applications in complexity theory, mainly for lower bounds [35, 14].

Previous constructions

Extractors for OBF sources were first studied by Chor, Goldreich, Håstad, Freidmann, Rudich, and Smolensky [17] in the zero-error setting (indeed, this is one of the rare examples in which outputting exactly uniform bits is possible). They observed that the simple XOR function outputs a uniform bit for any k1, but also proved that kn/3 is necessary even for the m=2 case. More generally, via establishing connections to error correcting codes, Chor et al. showed that when k=nt, we can (explicitly) extract mntlog(n/t) bits, but cannot extract more than, roughly, n(t/2)log(n/t) bits. Further lower bounds in this regime were obtained by Friedman [24].

What if we allow error, and wish to support lower entropies? Kamp and Zuckerman were the first to go below linear entropy, achieving kn1/2+γ and m=Ω(n2γ), with error ε=2Ω(m). The entropy loss in the [34] construction was later improved by Gabizon, Raz, and Shaltiel [26], who achieved m=(1o(1))k and exponentially-small error. Gabizon et al. were also able to reduce the entropy down to kpoly(logn), but with a worse error of kΩ(1). But often, especially in cryptographic applications, a large error is prohibitive. The first low-error OBF extractor for poly-logarithmic entropy was constructed by Rao [47], outputting m=(1o(1))k bits with error 2kΩ(1). We will further discuss Rao’s construction in Section 2.1.

For even lower entropies, recent explicit constructions that support near-logarithmic entropy already work for the more general family of affine sources.444An (n,k) affine source is a distribution that is flat over some affine subspace of 𝔽2n of dimension k. Thus, an (n,k) OBF source is simply an affine source whose basis consists of only elementary vectors. Using techniques from the nonmalleable extractors literature, Chattopadhyay, Goodman, and Liao [10] achieved k=O~(logn), and this was later improved by a recent work of Li [44] that gets k=O(logn). Unfortunately, the latter two constructions work for constant error and one output bit, which is not relevant for OBF extractors. However, one can think of applying techniques similar to the ones in [50, 42]. But these, to the best of our knowledge, do not readily give a significant improvement on either m or ε.

Lastly, it is interesting to note that while a straightforward application of the probabilistic method gives a non-explicit construction for (roughly) klogn+2log(1/ε) and m=k2log(1/ε), logarithmic entropy is not a lower bound. In [34], they gave a construction that outputs only Ω(logk) bits, has error kΩ(1), but works for any k. In [20], Cohen and Shinkar gave a lower bound for extraction, and showed that when k is a small enough function of n, Ω(logk) output bits are all we can hope for. Moreover, they gave a “semi-explicit” construction, when the error is large enough, for sub-logarithmic entropy. Their construction supports k=Ω(loglogn), outputs m=kOε(1) bits, and runs in time 2O~ε(logn).

Our result

We give an explicit low-error construction for kO~(logn) that outputs almost all the entropy.

Theorem 3 (see also Theorem 24).

There exist constants c>1 and β(0,1) such that the following holds for any n and klogn(loglogn)c. There exists an explicit function

𝖮𝖡𝖥𝖤𝗑𝗍:{0,1}n{0,1}m=(1o(1))k

that is a (k,ε) extractor for OBF sources, where ε=2(k/logn)β.

While the error guarantee does not meet the optimal 2Ω(k) (or even 2kΩ(1)), both our construction and [47] achieve ε=nω(1) starting from k=poly(logn). Importantly, our construction is the first to extract almost all the entropy with vanishing error starting from k=O~(logn). In particular, our error guarantee surpasses a kΩ(1)-type dependence for all ranges of k.

The low-error challenge

The recent decade has seen exciting progress in extractors that support small entropies, most notably in constructions of two-source extractors555A two-source extractor for entropy k extracts from the family of sources 𝒳 in which each X𝒳 comprises two independent sources (X1,X2){0,1}n×{0,1}n, each Xi satisfies H(Xi)k. ([13, 4, 19, 43, 44]), and the use of similar techniques to construct affine extractors [42, 10, 44]. While these extractors, which generally follow the celebrated Chattopadhyay–Zuckerman framework [13], work for low entropies, their error is relatively high for reasons which will be discussed in Section 2.2. Low-error constructions are crucial for the security of cryptographic applications, as well as for good correlation bounds in lower bounds applications, but despite numerous attempts, recent constructions do not obtain negligible error in polynomial time.666For two-source extractors, the best low-error constructions require k(1/2α)n for a small constant α>0 [8, 36], whereas the [13] construction gets k=poly(logn) with ε=kΩ(1), and followup constructions for even lower entropies only work for constant error. For affine extractors, the best low-error constructions require k=ω(n/loglogn) [55, 37, 45], whereas, again, recent constructions in the near-logarithmic regime only work for constant error. It is worth noting that if one allows three independent sources, then efficient low-error three-source extractors exist already for poly-logarithmic min-entropy [40, 3].

Our construction uses key components of the [13] framework (mainly towards constructing new low-error affine condensers, see Section 2.2), and is able to outperform the error parameter in most related constructions. We thus view this work also as a proof-of-concept for utilizing more recent machinery to tackle low-error constructions.

2 Proof Overview

Our construction combines two new components: A low-error affine condenser with a small gap, and a dual use of linear condensers for OBF sources. In Section 2.1, we will revisit Rao’s construction [47] and his use of linear condensers for OBF sources. In Section 2.2, we will discuss the [13] framework, the low-error adaptation of [3] to two-source condensers, and our construction of affine condensers. In Section 2.3, we will give the full construction, after discussing our use of linear condensers for OBF sources.

2.1 Rao’s Construction

Given an (n,k) OBF source X, Rao’s construction [47] goes roughly as follows.777The [47] construction works not only for OBF sources, but more generally to low-weight affine sources. Our result captures low-weight affine sources as well, and we discuss this in Section 5.3.

  1. 1.

    Transform X into X=PX, where P is a linear transformation, X is much shorter than X, yet it preserves the entropy of X. Thus, this step can be seen as applying a linear condenser for OBF sources, and we discuss this step a bit more thoroughly below. Note that X is an affine source.

  2. 2.

    Transform X into an affine somewhere-random source. This is a source distributed over a table, where one row is uniform, and every other row depends on the uniform one in an affine way. Z has few rows, kΩ(1), but the length of each row is almost k. This transformation is done via applying a (linear) seeded extractor on X and every possible seed, i.e., Zi=𝖤𝗑𝗍(X,i).888More accurately, to make each row longer than kΩ(1), Rao injects more entropy to the table by performing an additional extraction step, this time with X itself as the source, and Zi as the seed. This is the part that requires X to have poly-logarithmic entropy.

  3. 3.

    Next, extract from Z. This step is done via repeatedly applying an affine somewhere-random condenser, that halves the number of rows in the table, while only shortening each row by a little bit.

Our construction makes use of Step (1) (in fact, more than once, see Section 2.3), and completely dispenses with Step (3). In our construction, once we have our table Z from Step (2), we condense it into a single string using a different mechanism. This is encapsulated under a new affine condenser, which we discuss soon in Section 2.2.

Linear Condensers for OBF Sources

In [47], Rao observed that parity check matrices of binary error correcting codes are good linear condensers for OBF sources, and low-weight affine sources in general.999It is interesting to note that relying on the distance property of error correcting codes alone cannot give optimal linear OBF condensers. One can show that there exist condensers with output length t=O(k). Specifically, let P𝔽2n×t be the parity check matrix of a linear code 𝒞𝔽2n with co-dimension t and distance k+1 (see Section 3.5 for the relevant definitions). Then, given an (n,k) OBF source X, the affine source PX𝔽2t still has entropy k (see Lemma 22 for the easy proof).

This (lossless) condensing allows Rao, and us, to apply affine primitives on sources of much shorter lengths. Indeed, working with BCH codes, one can get t=O(klogn).101010[47] uses an error correcting code that comes from small-bias sets, and achieves a worse distance-to-codimension tradeoff. In fact, for our construction, BCH codes are necessary. In our construction, we will also use a lossy instantiation of these linear OBF condensers, where tk. We discuss this further in Section 2.3.

2.2 Low-Error Affine Condensers

The [13] Framework

We start by briefly describing the [13] framework, originally geared towards obtaining an extractor for two independent sources. The CZ construction uses two main ingredients: A “correlation breaking” primitive,111111The idea of constructing extractors via correlation breaking appeared prior to [13], notably in [38, 39, 41, 18]. and a resilient function. For the former, we will consider here a t-correlation breaker with advice,121212The actual definition offers much more flexibility than what we describe here, but it suffices for this discussion. Also, the original CZ construction uses non-malleable extractors as the correlation breaking primitive, but there are known reductions between the two objects. which is a function 𝖢𝖡:{0,1}n×{0,1}d{0,1}m that takes inputs from a weak source X1, a uniform seed Y, and a fixed advice string α, and the guarantee is that

𝖢𝖡(X1,Y,α)Um|{𝖢𝖡(X1,Y(1),α(1)),,𝖢𝖡(X1,Y(t),α(t))},

where Y(1),,Y(t) are independent of X1, and the α(i)-s are all different from α. Roughly speaking, for a typical yY, 𝖢𝖡(X1,y,α) is close to uniform even given the value of 𝖢𝖡 on t other adversarially chosen y(i)-s. Hence, if we were to build a table with D=2d rows, and put, say, 𝖢𝖡(X1,i,i) in the i-th row, then rows of good seeds were not only close to uniform, but they’re also close to being t-wise independent.

Yet, there are no one-source extractors, and [13] need to use a second weak source X2 to sub-sample from the seeds of 𝖢𝖡. Specifically, take a seeded extractor 𝖤𝗑𝗍:{0,1}n×{0,1}d0{0,1}d,131313See Definition 8 for the formal definition. and consider the table Z with r=2d0D rows, in which each row is given by

Zi=𝖢𝖡(X1,𝖤𝗑𝗍(X2,i),i). (1)

An analogous way to view the construction of Z, which would be more beneficial towards our construction, is to first create the table with r rows wherein the i-th row is simply 𝖤𝗑𝗍(X2,i). While most of the rows in that table are marginally close to uniform, they are arbitrarily correlated. We then use 𝖢𝖡 with an independent source X1 to (partially) break correlations.141414This viewpoint was taken, e.g., in [18, 42, 5].

It turns out that Z itself, when the parameters are set correctly, is close to a table in which rq “good” rows are truly t-wise independent, and there are q “bad” rows that can depend arbitrarily on the good ones. Indeed, many constructions following [13] can also be divided into two steps, where the first one transforms X1 and X2 into Z=Z(X1,X2) with the above structure, known as a non-oblivious bit-fixing source, and the second one, which we now discuss, is where resilient functions comes into play.

Concisely, a resilient function f:{0,1}r{0,1} is an extractor for non-oblivious bit fixing sources. That is, we want a nearly-balanced f whose output cannot be heavily influenced by any small set of q bad bits (we’re considering m=1 for now), and in our case f needs to be resilient even when the good bits are only t-wise independent (and uniform), unlike the more standard case in which the good bits are completely independent. The second step of the CZ framework then amounts to applying a resilient function on Z, outputting W=f(Z).

While this beautiful approach does give us that WεU1, it is inherently bound to have runtime which is polynomial in 1/ε. The Kahn–Kalai–Linial theorem [33] tells us that no matter what f is, there will always be a single bad bit that has influence p=Ω(logr/r), i.e., with probability p over the good bits, the single corrupt bit can fully determine the result.151515In terms of explicit f-s, [13] derandomized the Ajtai-Linial [1] randomized construction, supporting q=r1δ for any constant δ>0. In a subsequent work, Meka [46] obtained a derandomized version of [1] that matches the randomized construction and can support q=O(rlog2r) bad bits (see also [30, 31] for followup constructions with smaller bias). Thus, the running time, which is at least r, is also at least 1/ε. This is a common feature of almost all constructions that follow the CZ approach.

The [3] Low-Error Adaptation

In [3], Ben-Aroya, Cohen, Doron, and Ta-Shma, evaded the above resilient functions barrier by aiming for a weaker object – a two source condenser. Namely, the output W is not ε-close to uniform, but ε-close to some random variable W{0,1}m that has min-entropy mg, where we call g the entropy gap (note that an extractor has gap g=0). While when m=1, a single malicious bit can bias the result pretty significantly, the key observation in [3] is that when m becomes large, the probability that the bad bits can reduce g bits of entropy from the output can be exponentially-small in g. We call such a function f𝗌:ΣrΣ, Σ={0,1}m, an entropy-resilient function, which is essentially a condenser for non-oblivious symbol-fixing sources (that is, the natural extension of bit-fixing sources to arbitrary alphabets). In [3], they constructed an explicit such f𝗌, and were able to conclude that W=f𝗌(Z) is ε-close to having gap g=o(log(1/ε)) provided that k is at least polynomial in log(n/ε), and m=kΩ(1). Importantly, the construction’s runtime is polynomial in n, and not in 1/ε.

Teleporting to the affine setting

Starting from the work of Li [42], and throughout a sequence of followup works [11, 10, 12, 44], the correlation breaking framework was used to extract from affine sources and other related families of weak sources such as sumset sources, small-space sources, and interleaved sources. While in the affine case we don’t have two sources at our disposal, our single source has structure we can utilize.

Specifically, recall our table {𝖤𝗑𝗍(X,i)}i[r] above, and assume that 𝖤𝗑𝗍 is linear, meaning that for any fixed seed, the output is a linear function of the source. Here too, our table has many good rows (in fact, the good rows are exactly uniform), but they are, again, arbitrarily correlated. It turns out that we can use the same source X to break the correlation and make the table close to a t-non-oblivious bit-fixing source. This heavily uses the (by now) standard “affine conditioning” technique (see Lemma 7), in which given a linear function L:{0,1}n{0,1}m, we can decompose X=A+B, where both A and B are affine, and there is a bijection between A and L(X), and B is independent of L(X).161616In our case, the function L is chosen according to our linear seeded extractor 𝖤𝗑𝗍, and in fact bundles up t of them. For a more detailed description of the correlation breaking mechanism, see, for example, [42]. Conveniently, the intricate correlation breaking constructions for linearly-correlated source and seed was made explicit in [12], coined affine correlation breakers. The state-of-the-art affine correlation breakers follow from a recent work of Li [44] (see Definitions 13 and 14).

Our Low-Error Affine Condenser

Armed with a low-error two-source condenser, and a method to adapt the two-source setting to the affine world, a low-error affine condenser follows rather easily. Consider the table Z from Equation 1, but this time we form it as follows:

Zi=𝖠𝖿𝖿𝖢𝖡(X,𝖫𝖤𝗑𝗍(X,i),i), (2)

where 𝖠𝖿𝖿𝖢𝖡 is an affine correlation breaker, and 𝖫𝖤𝗑𝗍 is a linear seeded extractor. We then output

W=f𝗌(Z),

where f𝗌 is the above entropy resilient function. We can then show that if our affine source X𝔽2n has at least kpoly(log(n/ε)) entropy, W{0,1}kΩ(1) is ε-close to having entropy gap o(log(1/ε)). This is precisely a condensing guarantee, and the formal construction is given in Section 4.

2.3 Our OBF Extractor

Equipped with a linear condenser for OBF sources P, and a low-error affine condenser for poly-logarithmic entropies 𝖠𝖿𝖿𝖢𝗈𝗇𝖽, one natural attempt would be to try and output

𝖫𝖤𝗑𝗍(X,𝖠𝖿𝖿𝖢𝗈𝗇𝖽(P(X))),

where 𝖫𝖤𝗑𝗍 is a linear seeded extractor. Note that P(X) is a short affine source with entropy k, so the entropy lower bound of 𝖠𝖿𝖿𝖢𝗈𝗇𝖽 is now much lower! There are three potential problems with the above construction:

  1. 1.

    We want 𝖠𝖿𝖿𝖢𝗈𝗇𝖽 to supply the seed for 𝖫𝖤𝗑𝗍, but Y=𝖠𝖿𝖿𝖢𝗈𝗇𝖽(P(X)) is not uniform – it is only ε-close to having some entropy gap g. But since g is tiny, we can in fact use Y as a seed, suffering only a small loss in the error. (This observation is by now standard.)

  2. 2.

    X has length n, and we don’t have linear seeded extractors outputting many bits with optimal seed length. In particular, the entropy in Y cannot be greater than k, but all known extractors require seed log2n.

    To try and resolve the issue, one may change the construction to 𝖫𝖤𝗑𝗍(P(X),𝖠𝖿𝖿𝖢𝗈𝗇𝖽(P(X))), thereby only working with the short P(X). Even then, we are still left with the challenge of handling the correlations between the source and the seed.

  3. 3.

    Indeed, Y depends on P(X) (or X) – it’s a deterministic function of it! As mentioned above, we will use the affine conditioning technique in order to handle the correlations. We decompose X=A+B, where A and B are affine and independent, A is a deterministic function of P(X), and B is independent of P(X).

    Morally, after an appropriate “fixing”, the source we use for 𝖫𝖤𝗑𝗍 is the affine source B. But we can only guarantee that H(B)k|P(X)|, which provides no useful bound if we want to retain most of the entropy of X.

In order to make X shorter and guarantee that some entropy is left even after the affine conditioning, we make use of linear OBF condensers in two ways:

𝖮𝖡𝖥𝖤𝗑𝗍(X)=𝖫𝖤𝗑𝗍(P1(X),𝖠𝖿𝖿𝖢𝗈𝗇𝖽(P2(X)))

Above,

  • P1 is a lossless condenser for OBF sources, mapping n bits to n=O(klogn) bits.

  • P2 is a lossy condenser for OBF sources, mapping n bits to γk bits for a small γ(0,1), while retaining kklogn entropy. We construct P2 using error correcting codes as well, only with different parameters.

    For 𝖠𝖿𝖿𝖢𝗈𝗇𝖽 to condense, we need kpoly(log(n/ε)), and this sets our lower bound on k and the bound on the error.

Finally, we need to show that we can apply 𝖫𝖤𝗑𝗍. Towards this end, we “affine condition” with respect to P2(X), and so we can write

𝖮𝖡𝖥𝖤𝗑𝗍(X)=𝖫𝖤𝗑𝗍(P1(A),Y)+𝖫𝖤𝗑𝗍(P1(B),Y), (3)

where Y=𝖠𝖿𝖿𝖢𝗈𝗇𝖽(P2(X)), A and B are independent, and P2(X) is a deterministic function of A. Thus, B and Y are also independent, and furthermore, B still has enough entropy. This, in turn, implies that P1(B) has enough entropy, and we can safely fix a good seed yY, making the first term in Equation 3 fixed, and the second one close to uniform. The full details appear in Section 5.2.

3 Preliminaries

We use log(x) for log2(x). For an integer n, we denote by [n] the set {1,,n}. The density of a subset BA is denoted by μ(B)=|B||A|. For a function f:Ω1Ω2, we say that f is explicit if there exists a deterministic procedure that runs in time poly(log(|Ω1|)) and computes f.

Random variables, entropy

The support of a random variable X distributed over some domain Ω is the set xΩ for which Pr[X=x]0, which we denote by Supp(X). The total variation distance (or, statistical distance) between two random variables X and Y over the same domain Ω is defined as |XY|=maxAΩ(Pr[XA]Pr[YA]). Whenever |XY|ε we say that X is ε-close to Y and denote it by XεY. We denote by Un the random variable distributed uniformly over {0,1}n. We say a random variable is flat if it is uniform over its support. Whenever we write xA for A being a set, we mean x is sampled uniformly at random from the flat distribution over A.

For a function f:Ω1Ω2 and a random variable X distributed over Ω1, f(X) is the random variable distributed over Ω2 obtained by choosing x according to X and computing f(x). For a set AΩ1, f(A)={f(x):xA}. For every f:Ω1Ω2 and two random variables X and Y distributed over Ω1 it holds that |f(X)f(Y)||XY|, and is often referred to as a data-processing inequality. Another property of statistical distance is the triangle inequality, which states that for all distributions X,Y,Z, we have |XY||XZ|+|YZ|.

The min-entropy of X is defined by

H(X)=minxSupp(X)log1Pr[X=x],

and it always holds that H(X)H(X), where H() is Shannon’s entropy. A random variable X is an (n,k) source if X is distributed over {0,1}n and has min-entropy at least k.

Limited Independence

We say that a distribution X{0,1}n is (t,γ)-wise independent, if the restriction of X to any t coordinates is γ-close to Ut. When γ=0, we simply say that X is t-wise independent.

Lemma 4 ([2]).

Let X{0,1}n be (t,γ)-wise independent. Then, X is (nγt)-close to a t-wise independent distribution.

3.1 OBF and Affine Sources

We repeat the definition of affine and OBF sources.

Definition 5 (affine source).

An (n,k) affine source is a distribution X𝔽2n that is flat over some (unknown) affine subspace of dimension k.

In other words, for any such X there exist independent v0,v1,,vk𝔽2n such that sampling xX amounts to sampling α1,,αk𝔽2 uniformly at random, and outputting x=v0+i=1kαivi. Notice that when X is affine, H(X)=H(X).

Definition 6 (OBF source).

An (n,k) oblivious bit-fixing (OBF) source is a distribution X{0,1}n for which there exists an (unknown) subset I[n] of size k, and c{0,1}nk, such that XI=Uk, and X[n]I is fixed to c.

In other words, a bit-fixing source is a very structured affine source – one in which v1,,vk are indicator vectors. The following lemma will let us use linearly-correlated source and seed when using linear seeded extractors.

Lemma 7 (affine conditioning, [25, 47, 37]).

Let X be an (n,k) affine source, and let L:{0,1}n{0,1}m be a linear function. Then, there exist independent affine sources A and B, over {0,1}n, such that:

  • X=A+B.

  • There exists c{0,1}m such that for every bSupp(B) it holds that L(b)=c.

  • H(A)=H(L(A)) and there exists an affine function L𝗂𝗇𝗏:{0,1}n{0,1}m such that A=L𝗂𝗇𝗏(L(A)). In particular, H(A)m and H(B)km.

  • For every Supp(L(X)), H(X|{L(X)=})=H(B).

3.2 Seeded Extractors

Definition 8 (seeded extractor).

A function 𝖤𝗑𝗍:{0,1}n×{0,1}n{0,1}m is a (k,ε) (seeded) extractor if for every (n,k) source and an independent and uniform Y{0,1}d, it holds that 𝖤𝗑𝗍(X,Y)εUm. Furthermore, we say that 𝖤𝗑𝗍 is strong if (𝖤𝗑𝗍(X,Y),Y)ε(Um,Y).

We say that 𝖤𝗑𝗍 is linear if it is linear in the source, namely if for any x1,x2{0,1}n and y{0,1}d, it holds that 𝖤𝗑𝗍(x1+x2,y)=𝖤𝗑𝗍(x1,y)+𝖤𝗑𝗍(x2,y).

Combining a (linear variant) of the GUV seeded condenser [28, 16] with the Shaltiel–Umans extractor [51], we can get the following logarithmic-seed linear seeded extractor, that outputs a constant power of the entropy.

Theorem 9 (linear seeded extractor, I [42]).

There exists a constant c1 such that the following holds, for any positive integers n and c1log8nkn, and any εn2. There exists an explicit linear strong (k,ε) extractor 𝖫𝖤𝗑𝗍1:{0,1}n×{0,1}d{0,1}m with d=c1logn and m=k.

Replacing the SU extractor with Trevisan’s extractor [52], as analyzed in [48], we can output almost all the entropy, at the cost of a larger seed.

Theorem 10 (linear seeded extractor, II).

There exists a constant c2 such that the following holds, for any positive integers n and kn, and any ε,γ>0. There exist an explicit linear strong (k,ε) extractor 𝖫𝖤𝗑𝗍2:{0,1}n×{0,1}d{0,1}m where m=(1γ)k and

d=c2(logn+log2klog(1/ε)log(1/γ)).

We omit the easy proof, which is similar to the proof of Theorem 9 in [42].

When the extractor is linear and the source is affine, a good seed already implies perfect uniformity, for any nontrivial error.

Lemma 11 ([47]).

Let 𝖫𝖤𝗑𝗍:{0,1}n×{0,1}d{0,1}m be a linear strong (k,ε) extractor with ε<1/2. Then, for every (n,k) affine source, it holds that

PryUd[|𝖫𝖤𝗑𝗍(X,y)Um|=0]12ε.

We will also need the following claim, that is often used when one wishes to use seeded extractors with a non perfect seed.

Claim 12 (strong extractors with weak seeds).

Let 𝖤𝗑𝗍:{0,1}n×{0,1}d{0,1}m be a strong (k,ε) extractor, let X be an (n,k) source, and let Y be δ-close to a (d,dg) source which is independent of X. Then, with probability at least 12gεδ over yY, it holds that 𝖤𝗑𝗍(X,y)εUm.

Proof.

By Markov’s inequality, there exists a set BAD{0,1}d of density ρ(BAD)ε such that for every zBAD it holds that 𝖤𝗑𝗍(X,z)εUm. Let Y be a (d,dg) source that is δ-close to Y. Then,

Pr[YBAD] Pr[YBAD]+δ
=zBADPr[Y=z]+δ|BAD|2(dg)+δ2gε+δ.

3.3 Affine Correlation Breakers

We proceed with formally defining affine correlation breakers (with advice).

Definition 13 (affine correlation breaker).

We say that 𝖠𝖿𝖿𝖢𝖡:{0,1}n×{0,1}d×{0,1}a{0,1}m is a (t,k,ε) affine correlation breaker if for all distributions X,X(1),,X(t){0,1}n, B,B(1),,B(t){0,1}n, A,A(1),,A(t){0,1}n, Y,Y(1),,Y(t){0,1}d, and all strings α,α(1),,α(t){0,1}a such that:

  • X=A+B and X(i)=A(i)+B(i) for all i[t],

  • H(B)k and Y is uniform,

  • (B,B(1),,B(t)) is independent of (A,A(1),,A(t),Y,Y(1),,Y(t)), and,

  • For all i[t], αα(i),

it holds that

(𝖠𝖿𝖿𝖢𝖡(X,Y,α),{𝖠𝖿𝖿𝖢𝖡(X(i),Y(i),α(i))}i[t])ε(Um,{𝖠𝖿𝖿𝖢𝖡(X(i),Y(i),α(i))}i[t]).

We say that 𝖠𝖿𝖿𝖢𝖡 is strong if the above holds also when we add Y,Y(1),,Y(t).

We will use the following recent affine correlation breaker.

Theorem 14 ([44]).

For any positive integers n,m,t,a, and any ε>0, there exists an explicit strong (t,k,ε) affine correlation breaker 𝖠𝖿𝖿𝖢𝖡:{0,1}n×{0,1}d×{0,1}a{0,1}m, where k=O(tm+ta+t2log2(t+1)log(nt/ε)) and d=O(m+ta+tlog3(t+1)log(nt/ε)).

3.4 Entropy-Resilient Functions

We summarize the required definitions and results given in [3].

Definition 15 (NOSF sources).

Let Σ={0,1}m. A (q,t) non-oblivious Σ-fixing source X=(X1,,XD) is a random variable over ΣD={0,1}Dm for which there exists a set BAD[D] of cardinality qq such that:

  • The joint distribution of {(Xi)j:i[D]BAD,j[m]}, denoted by GX, is t-wise independent over {0,1}(Dq)m; and

  • Each of the random variables in BX{(Xi)j} with iBAD and j[m] may depend arbitrarily on all other random variables in GX and BX.

If t=(Dq)m, we say that X is a q-non-oblivious Σ-fixing source. If m=1, the definition coincides with the standard definition of non-oblivious bit-fixing sources.

In [3], Ben-Aroya et al. constructed seedless condensers for NOSF sources, also known as entropy-resilient functions.

Definition 16 (entropy-resilient functions).

Let Σ={0,1}m. A function f:ΣDΣ is a (q,t,g,ε) entropy-resilient function if for every (q,t) non-oblivious Σ-fixing source X over ΣD, the output f(X) is ε-close to an (m,mg)-source.

Theorem 17 ([3]).

For every constant 0<γ<1 there exist constants 0<α<1 and c1 such that the following holds. For all integers D, mDα/2, every ε>0, and for every integer tm(logD)c, there exists an explicit function 𝖤𝗇𝗍𝖱𝖾𝗌:ΣDΣ, for Σ={0,1}m, that is (q=D1γ,t,g,ε) entropy-resilient, with entropy gap g=o(log(1/ε)).

3.5 Error Correcting Codes

A binary code 𝒞𝔽2n is linear if 𝒞 is a linear subspace of 𝔽2n. The dimension of 𝒞 as a subspace is called the dimension of the code. We will identify a linear code 𝒞𝔽2n of dimension k with the image of its encoding function 𝒞:𝔽2k𝔽2n, which is given by a generator matrix A𝔽2n×k. A parity check matrix of 𝒞 is a generator matrix P𝔽2(nk)×n of the dual code 𝒞={x𝔽2n:x,c=0c𝒞}, and thus Px=0 if and only if x𝒞.

We say that an error correcting code 𝒞𝔽2n has distance d if for any distinct codewords x,y𝒞 it holds that xiyi in at least d i-s. We say that 𝒞 is an [n,k,d] code, if 𝒞𝔽2n is a linear code of dimension k and distance d. We will use the following explicit code, the BCH code.

Theorem 18 ([7, 29]).

There exists a constant c>0 such that the following holds. For every positive integers n and d<n, there exists an [n,r,d] code 𝒞𝖡𝖢𝖧 with co-dimension nr=clognd.171717The original BCH code assumes n=2m1 for some integer m, and then nrmd2. Standard manipulations let us work for any n. Moreover, the parity-check matrix of 𝒞𝖡𝖢𝖧 can be constructed in time poly(n).

4 Low-Error Affine Condensers

In this section, we give our construction of low-error affine condensers with tiny entropy gap.

Theorem 19.

There exist universal constants c1 and α(0,1) such that the following holds. For any positive integers n,k, and any ε>0 such that nklogc(nε), there exists an explicit function 𝖠𝖿𝖿𝖢𝗈𝗇𝖽:{0,1}n{0,1}m where mkα such that for any (n,k) affine source X, it holds that 𝖠𝖿𝖿𝖢𝗈𝗇𝖽(X) is ε-close to an (m,mg) source, where g=o(log(1/ε)).

We will start by reducing an affine source to a non-oblivious symbol-fixing (NOSF) source.

Lemma 20.

There exist constants c,c1>1 such that the following holds, for any positive integers n,t,m,k, and any ε>0, satisfying k(mtlog(n/ε))c. There exists an explicit function 𝖠𝖿𝖿𝖳𝗈𝖭𝖮𝖲𝖥:{0,1}nΣD with Σ={0,1}m and D=nc1 such that for every (n,k) affine source X, 𝖠𝖿𝖿𝖳𝗈𝖭𝖮𝖲𝖥(X) is ε-close to a (D11/c1,t) NOSF source.

Proof.

We use the following two building blocks.

  • Let 𝖫𝖤𝗑𝗍:{0,1}n×{0,1}d0{0,1}d be the linear strong (k𝖤𝗑𝗍,ε𝖤𝗑𝗍=1/2n) extractor given by Theorem 9, where k𝖤𝗑𝗍=d2c1log8n, for c1 being the constant given in Theorem 9. Note that we can take d0=c1logn.

  • Let 𝖠𝖿𝖿𝖢𝖡{0,1}n×{0,1}d×{0,1}d0{0,1}m be the (t1,k𝖢𝖡=ktd,ε𝖢𝖡) affine correlation breaker of Theorem 14 with ε𝖢𝖡=εt(mD)tm, and

    d=O(m+td0+tlog3(t+1)log(nt/ε𝖢𝖡))=O(t5mlogn+t4log(1/ε)).

    We need to make sure that kk𝖢𝖡+td where k𝖢𝖡=O(tm+td0+t2log2(t+1)log(nt/ε𝖢𝖡)), and indeed, this is satisfied by taking kC(t6mlogn+t5log(1/ε)) for a large enough constant C. We also need to make sure that kd2 and kclog8n, and both inequalities, together with our previous lower bound on k, indeed holds whenever

    kCt10m2log8nlog2(1/ε)

    for a large enough constant C.

Our construction goes as follows. For simplicity, identify {0,1}d0 with [D=nc1]. Given x{0,1}n:

  1. 1.

    For every i[D], compute yi=𝖫𝖤𝗑𝗍(x,i).

  2. 2.

    For every i[D], compute zi=𝖠𝖿𝖿𝖢𝖡(x,yi,i).

We’ll show that the table Z=(Z1,,ZD) satisfies the requirement of the lemma. First, by Lemma 11, we have a set BAD[D] of density ρ(BAD)1/n such that for any iBAD it holds that Yi{0,1}d is uniform. Next, fix t good rows {i1,,it}[D]BAD. Consider the linear function L:{0,1}n{0,1}td that is given by (𝖫𝖤𝗑𝗍(,i1),,𝖫𝖤𝗑𝗍(,it)). Via the affine conditioning lemma, Lemma 7, we can write X=A+B, where:

  • A and B are affine and independent.

  • H(B)H(X)td=k𝖢𝖡.

  • L(B) is constant, so Yi1,,Yit is independent of B, and moreover, B is independent of (A,Yi1,,Yit).

Recalling that each Yij is uniform, all conditions are met to apply are affine correlation breaker, which gives us

(Zij,{Zik}k[t]{j})ε𝖢𝖡(Um,{Zik}k[t]{j})

for every j[t]. Therefore,

(Zi1,,Zit)tε𝖢𝖡Utm,

and so by Lemma 4, Z itself is (mD)tmtε𝖢𝖡=ε close to a (q,t) non-oblivious {0,1}m-fixing source for q=Dn=D11c1.

Next, we can simply apply the low-error condenser for NOSF sources given by [3].

Proof of Theorem 19.

Given n and ε>0, let m be the number of bits we eventually output, and the entropy of X will be determined according to that later on. Set γ=1c1, and let α(0,1) and c1 be the constants that are set according to γ, guaranteed to us by Theorem 17.

Let 𝖠𝖿𝖿𝖳𝗈𝖭𝖮𝖲𝖥:{0,1}nΣD be the function from Lemma 20, where Σ={0,1}m and D=nc1, set with tmd0c and εε/2. By that lemma, there exists a constant C>1 such that whenever k(mlog(n/ε))C and X is an (n,k) affine source, it holds that 𝖠𝖿𝖿𝖳𝗈𝖭𝖮𝖲𝖥(X) is ε2-close to an (D1γ,t) NOSF source.

Now, let 𝖤𝗇𝗍𝖱𝖾𝗌:ΣDΣ be the (D1γ,t,g,ε/2) entropy resilient function from Theorem 17. Whenever mDα/2, we are guaranteed that 𝖤𝗇𝗍𝖱𝖾𝗌(𝖠𝖿𝖿𝖳𝗈𝖭𝖮𝖲𝖥(X)) is ε-close to an (m,mg) source with g=o(log(1/ε)). To conclude, we have that there exist constants C1 and δ(0,1) such that whenever mnδ and k(mlog(n/ε))C, the function

𝖠𝖿𝖿𝖢𝗈𝗇𝖽(x)=𝖤𝗇𝗍𝖱𝖾𝗌(𝖠𝖿𝖿𝖳𝗈𝖭𝖮𝖲𝖥(x))

is an affine condenser with entropy gap g=o(log(1/ε)). Putting it differently, there exist constants C1 and δ(0,1) such that whenever k(log(n/ε))C, our condenser outputs kδ bits.

5 Extractors for OBF Sources

5.1 Linear OBF Condensers

We begin by defining these condensers, and focus on the linear setting.

Definition 21 (OBF condenser).

A matrix L𝔽2m×n is an (n,k)(m,t) linear condenser for OBF sources if H(L(X))t for any (n,k) OBF source X.

Rao showed that linear condensers for OBF sources can be obtained by parity check matrices of binary codes. The next statement is a bit different than the one in [47], so we provide the proof for completeness.181818The connection between pseudorandomness for OBF sources and codes was established, at least for extractors, already in [17], and, possibly under different formulations, in the cryptography literature, e.g, in [15].

Lemma 22 (following [47]).

Let X be an (n,k) OBF source, let kk be any positive integer, and let 𝒞𝔽2n be an [n,r,d] code with a parity check matrix P𝔽2m×n, where m=nr and dk+1. Then, Z=PX{0,1}m satisfies H(Z)k.

Proof.

Letting i1,,ik[n] be the entropic coordinates of X, and let c𝔽2n be its constant shift.191919We can assume that c is zero on i1,,ik. Consider

Vk=Span({ei1,,eik})Supp(X)+c,

observing that each element of Vk has Hamming weight at most k. Since 𝒞 has distance greater than k, Vk𝒞={0} and so P is injective on Vk and thus also on Vk+c. Let W be the subspace for which Supp(X)+c=VkW, and also refer to Vk and W as the corresponding flat distributions. Thus, for each z{0,1}m,

Pr[Z=z]=Pr[PVk=z+PW+c]=𝔼wW[Pr[PVk=z+Pw+c]]2k,

and we are done.

Using the BCH codes from Theorem 18, we get the following linear condensers for OBF sources.

Corollary 23.

There exists a constant c>0 such that the following holds. For every positive integers n, kn, and kk, there exists an explicit linear function P:{0,1}n{0,1}m for m=cklogn, such that the following holds. For every (n,k) OBF source X, it holds that H(P(X))k. That is, P is an (n,k)(cklogn,k) linear condenser for OBF sources.

5.2 OBF Extractors for Almost-Logarithmic Entropy

We are given n,k, and γ>0. Let ε be our error guarantee, to be determined later on. We use the following ingredients:

  • Let P1:{0,1}n{0,1}n1 be the linear condenser for OBF sources given in Corollary 23, set with kk, so n1=cklogn for some universal constant c>0.

  • Let P2:{0,1}n{0,1}n2 be the linear condenser for OBF sources given in Corollary 23, set with kγk2clognk𝖢𝗈𝗇𝖽, so n2=γk/2.

  • Let 𝖠𝖿𝖿𝖢𝗈𝗇𝖽:{0,1}n2{0,1}d be the affine condenser from Theorem 19, set with kk𝖢𝗈𝗇𝖽 and εε/2.

  • Let 𝖫𝖤𝗑𝗍:{0,1}n1×{0,1}d{0,1}m be the (k𝖤𝗑𝗍=(1γ/2)k,ε𝖤𝗑𝗍) extractor of Theorem 10, where ε𝖤𝗑𝗍=ε4/16 and m=(1γ/2)k𝖤𝗑𝗍(1γ)k. There exists a constant c2 such that a seed of length d=c2(log(n1)+log2(k/ε)log(1/γ)) suffices.

Now, given x{0,1}n, our extractor 𝖮𝖡𝖥𝖤𝗑𝗍:{0,1}n{0,1}m is given by

𝖮𝖡𝖥𝖤𝗑𝗍(x)=𝖫𝖤𝗑𝗍(P1(x),𝖠𝖿𝖿𝖢𝗈𝗇𝖽(P2(x)))

The fact that 𝖮𝖡𝖥𝖤𝗑𝗍 is explicit readily follows from the explicitness of its components. The correctness is established in the following theorem.

Theorem 24.

There exist universal constants ck,cγ>1 and β(0,1) such that the following holds, assuming klognγ(loglognγ)ck and γlogcγkk. For every (n,k) OBF source X it holds that 𝖮𝖡𝖥𝖤𝗑𝗍(X)εUm for ε=2((γk)/logn)β, recalling that m=(1γ)k.

Proof.

Denote X1=P1(X), X2=P2(X), Y=𝖠𝖿𝖿𝖢𝗈𝗇𝖽(X2), and Z=𝖫𝖤𝗑𝗍(X1,Y). We “affine condition” (Theorem 19) according to P2, and so we can write X as X=A+B, where A and B are affine and independent and there exists a linear bijection between A and X2 (so B is independent of X2 and thus also of Y). We can then write

Z=𝖫𝖤𝗑𝗍(X1,Y)=𝖫𝖤𝗑𝗍(P1(A),Y)+𝖫𝖤𝗑𝗍(P1(B),Y).

We know that H(X1)k, H(A)n2, and H(B)H(X)n2(1γ/2)k=k𝖤𝗑𝗍. Also, X1=P1(A)+P1(B), and P1(A) and P1(B) are independent. Thus, H(P1(B))k𝖤𝗑𝗍 as well,202020To see this, recall that B is a subspace of X (up to a shift), and P1 is injective on X. and P1(B) is independent of Y.

Claim 25.

With probability at least 1ε/2 over yY it holds that 𝖫𝖤𝗑𝗍(P1(B),y)ε/2Um.

Proof.

To apply our affine condenser, Theorem 19, we need to make sure that:

  1. 1.

    k𝖢𝗈𝗇𝖽logc(n2/ε), where c>1 is the constant guaranteed to us by Theorem 19. Recall that

    logc(n2ε)=O(logc(kε)),

    and that

    k𝖢𝗈𝗇𝖽=Ω(γklogn),

    so the inequality is met as long as klognγ(loglognγ)ck and ε2((γk)/logn)β for some constants ck>1 and β(0,1).

  2. 2.

    dk𝖢𝗈𝗇𝖽α, where α(0,1) is the constant guaranteed to us by Theorem 19. Plugging in the expression for d above, we get that similarly, we need to satisfy:

    1. (a)

      klognγ(loglognγ)ck for some constants ck>1,

    2. (b)

      ε2((γk)/logn)β for some constant β(0,1), and,

    3. (c)

      γlogcγkk for some constant cγ>1.

So indeed, Theorem 19 tells us that Y itself is (ε/4)-close to a (d,dg)-source for g=o(log(1/ε)), and thus by Claim 12 we have that 𝖫𝖤𝗑𝗍(P1(B),y)ε𝖤𝗑𝗍Um except for probability

2gε𝖤𝗑𝗍+ε41εε24+ε4ε2.

Fix a good y (in the sense of the above claim), and observe that 𝖫𝖤𝗑𝗍(P1(A),y) is independent from 𝖫𝖤𝗑𝗍(P1(B),y), making

𝖫𝖤𝗑𝗍(X1,y)=𝖫𝖤𝗑𝗍(P1(A),y)+𝖫𝖤𝗑𝗍(P1(B),y)

close to uniform as well. Overall, Z2(ε/2)Um, and we are done.

Choosing γ=1logck for any constant c, we obtain our Theorem 3.

5.3 Handling Low-Weight Affine Sources

We say that an (n,k) affine source X has weight w if each basis element of X has weight at most w. This generalizes OBF sources (that have w=1), and [47] can handle w=kα for a small constant α>0. Note that the only place in the proof that we used the fact that X is an OBF source (rather than a full-fledged affine source) is Lemma 22, where we argued that codes with sufficiently large distance give linear condensers. This can easily be extended to the w>1 case.

Lemma 26 (following [47]).

Let X be an (n,k) affine source of weight w, let kk be any positive integer, and let 𝒞𝔽2n be an [n,r,d] code with a parity check matrix P𝔽2m×n, where m=nr and dwk+1. Then, Z=PX{0,1}m satisfies H(Z)k.

Thus, as w grows, the co-dimension of the code we need to take grows too, and so the condensing quality decreases. To modify our construction to handle w>1, we need to set n1=cwklogn and k𝖢𝗈𝗇𝖽=γk2cwlogn. Repeating the same analysis as in Section 5.2, we see that one can take w=poly(logk) without significant loss in parameters.

Theorem 27.

For any constant >0 there exist constants c>1 and β(0,1) such that the following holds for any n and klogn(loglogn)c. There exists an explicit function

𝖮𝖡𝖥𝖤𝗑𝗍:{0,1}n{0,1}m=(1o(1))k

that is a (k,ε) extractor for affine sources of weight w=logk, where ε=2(k/logn)β.

References

  • [1] Miklós Ajtai and Nathan Linial. The influence of large coalitions. Combinatorica, 13(2):129–145, 1993. doi:10.1007/BF01303199.
  • [2] Noga Alon, Oded Goldreich, and Yishay Mansour. Almost k-wise independence versus k-wise independence. Information Processing Letters, 88(3):107–110, 2003. doi:10.1016/S0020-0190(03)00359-4.
  • [3] Avraham Ben-Aroya, Gil Cohen, Dean Doron, and Amnon Ta-Shma. Two-source condensers with low error and small entropy gap via entropy-resilient functions. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM), pages 43:1–43:20. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPIcs.APPROX-RANDOM.2019.43.
  • [4] Avraham Ben-Aroya, Dean Doron, and Amnon Ta-Shma. An efficient reduction from two-source to nonmalleable extractors:: Achieving near-logarithmic min-entropy. SIAM Journal on Computing, 51(2):STOC17–31, 2019.
  • [5] Avraham Ben-Aroya, Dean Doron, and Amnon Ta-Shma. Near-optimal erasure list-decodable codes. In Computational Complexity Conference (CCC), pages 1:1–1:27. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2020. doi:10.4230/LIPIcs.CCC.2020.1.
  • [6] Charles H. Bennett, Gilles Brassard, and Jean-Marc Robert. How to reduce your enemy’s information. In Advances in Cryptology (CRYPTO), volume 218 of Lecture Notes in Computer Science, pages 468–476. Springer, 1985.
  • [7] Raj Chandra Bose and Dwijendra K. Ray-Chaudhuri. On a class of error correcting binary group codes. Information and control, 3(1):68–79, 1960. doi:10.1016/S0019-9958(60)90287-4.
  • [8] Jean Bourgain. More on the sum-product phenomenon in prime fields and its applications. International Journal of Number Theory, 1(01):1–32, 2005.
  • [9] Ran Canetti, Yevgeniy Dodis, Shai Halevi, Eyal Kushilevitz, and Amit Sahai. Exposure-resilient functions and all-or-nothing transforms. In Advances in Cryptology (EUROCRYPT), pages 453–469. Springer, 2000. doi:10.1007/3-540-45539-6_33.
  • [10] Eshan Chattopadhyay, Jesse Goodman, and Jyun-Jie Liao. Affine extractors for almost logarithmic entropy. In Annual Symposium on Foundations of Computer Science (FOCS), pages 622–633. IEEE, 2022.
  • [11] Eshan Chattopadhyay and Xin Li. Explicit non-malleable extractors, multi-source extractors, and almost optimal privacy amplification protocols. In Annual Symposium on the Foundations of Computer Science (FOCS), pages 158–167. IEEE, 2016. doi:10.1109/FOCS.2016.25.
  • [12] Eshan Chattopadhyay and Jyun-Jie Liao. Extractors for sum of two sources. In Annual Symposium on Theory of Computing (STOC), pages 1584–1597. ACM, 2022. doi:10.1145/3519935.3519963.
  • [13] Eshan Chattopadhyay and David Zuckerman. Explicit two-source extractors and resilient functions. Annals of Mathematics, 189(3):653–705, 2019.
  • [14] Ruiwen Chen, Valentine Kabanets, Antonina Kolokolova, Ronen Shaltiel, and David Zuckerman. Mining circuit lower bound proofs for meta-algorithms. computational complexity, 24:333–392, 2015. doi:10.1007/S00037-015-0100-0.
  • [15] Mahdi Cheraghchi, Fredric Didier, and Amin Shokrollahi. Invertible extractors and wiretap protocols. IEEE Transactions on Information Theory, 58(2):1254–1274, 2011. doi:10.1109/TIT.2011.2170660.
  • [16] Mahdi Cheraghchi and Piotr Indyk. Nearly optimal deterministic algorithm for sparse Walsh-Hadamard transform. ACM Transactions on Algorithms (TALG), 13(3):1–36, 2017. doi:10.1145/3029050.
  • [17] Benny Chor, Oded Goldreich, Johan Håstad, Joel Freidmann, Steven Rudich, and Roman Smolensky. The bit extraction problem or t-resilient functions. In Annual Symposium on Foundations of Computer Science (FOCS), pages 396–407. IEEE, 1985.
  • [18] Gil Cohen. Local correlation breakers and applications to three-source extractors and mergers. SIAM Journal on Computing, 45(4):1297–1338, 2016. doi:10.1137/15M1029837.
  • [19] Gil Cohen. Towards optimal two-source extractors and Ramsey graphs. In Annual Symposium on Theory of Computing (STOC), pages 1157–1170. ACM, 2017. doi:10.1145/3055399.3055429.
  • [20] Gil Cohen and Igor Shinkar. Zero-fixing extractors for sub-logarithmic entropy. In International Colloquium on Automata, Languages, and Programming (ICALP), pages 343–354. Springer, 2015. doi:10.1007/978-3-662-47672-7_28.
  • [21] Yevgeniy Dodis. Exposure-Resilient Cryptography. Phd thesis, MIT, 2000.
  • [22] Yevgeniy Dodis, Amit Sahai, and Adam Smith. On perfect and adaptive security in exposure-resilient cryptography. In Advances in Cryptology (EUROCRYPT), pages 301–324. Springer, 2001. doi:10.1007/3-540-44987-6_19.
  • [23] Cynthia Dwork, Krishnaram Kenthapadi, Frank McSherry, Ilya Mironov, and Moni Naor. Our data, ourselves: Privacy via distributed noise generation. In Advances in Cryptology (EUROCRYPT), pages 486–503. Springer, 2006. doi:10.1007/11761679_29.
  • [24] Joel Friedman. On the bit extraction problem. In Annual Symposium on Foundations of Computer Science (FOCS), volume 33, pages 314–314. IEEE, 1992.
  • [25] Ariel Gabizon and Ran Raz. Deterministic extractors for affine sources over large fields. Combinatorica, 28(4):415–440, 2008. doi:10.1007/S00493-008-2259-3.
  • [26] Ariel Gabizon, Ran Raz, and Ronen Shaltiel. Deterministic extractors for bit-fixing sources by obtaining an independent seed. SIAM Journal on Computing, 36(4):1072–1094, 2006. doi:10.1137/S0097539705447049.
  • [27] Jesse Patrick McGrenra Goodman. Seedless Extractors. Phd thesis, Cornell University, 2023. Available at https://jpmgoodman.com/thesis.pdf.
  • [28] Venkatesan Guruswami, Christopher Umans, and Salil Vadhan. Unbalanced expanders and randomness extractors from parvaresh–vardy codes. Journal of the ACM (JACM), 56(4):20, 2009.
  • [29] Alexis Hocquenghem. Codes correcteurs d’erreurs. Chiffers, 2:147–156, 1959.
  • [30] Peter Ivanov, Raghu Meka, and Emanuele Viola. Efficient resilient functions. In Annual Symposium on Discrete Algorithms (SODA), pages 2867–2874. ACM-SIAM, 2023. doi:10.1137/1.9781611977554.CH108.
  • [31] Peter Ivanov and Emanuele Viola. Resilient functions: Optimized, simplified, and generalized. arXiv preprint arXiv:2406.19467, 2024. doi:10.48550/arXiv.2406.19467.
  • [32] Markus Jakobsson, Julien P. Stern, and Moti Yung. Scramble all, encrypt small. In International Workshop on Fast Software Encryption, pages 95–111. Springer, 1999. doi:10.1007/3-540-48519-8_8.
  • [33] Jeff Kahn, Gil Kalai, and Nathan Linial. The influence of variables on boolean functions. In Annual Symposium on Foundations of Computer Science (FOCS), pages 68–80. IEEE, 1988.
  • [34] Jesse Kamp and David Zuckerman. Deterministic extractors for bit-fixing sources and exposure-resilient cryptography. SIAM Journal on Computing, 36(5):1231–1247, 2006. doi:10.1137/S0097539705446846.
  • [35] Ilan Komargodski, Ran Raz, and Avishay Tal. Improved average-case lower bounds for demorgan formula size. In Annual Symposium on Foundations of Computer Science (FOCS), pages 588–597. IEEE, 2013. doi:10.1109/FOCS.2013.69.
  • [36] Mark Lewko. An explicit two-source extractor with min-entropy rate near 4/9. Mathematika, 65(4):950–957, 2019.
  • [37] Xin Li. A new approach to affine extractors and dispersers. In Computational Complexity Conference (CCC), pages 137–147. IEEE, 2011. doi:10.1109/CCC.2011.27.
  • [38] Xin Li. Extractors for a constant number of independent sources with polylogarithmic min-entropy. In Annual Symposium on the Foundations of Computer Science (FOCS), pages 100–109. IEEE, 2013. doi:10.1109/FOCS.2013.19.
  • [39] Xin Li. New independent source extractors with exponential improvement. In Annual Symposium on Theory of Computing (STOC), pages 783–792. ACM, 2013. doi:10.1145/2488608.2488708.
  • [40] Xin Li. Three-source extractors for polylogarithmic min-entropy. In Annual Symposium on Foundations of Computer Science (FOCS), pages 863–882. IEEE, 2015. doi:10.1109/FOCS.2015.58.
  • [41] Xin Li. Three-source extractors for polylogarithmic min-entropy. In 2015 IEEE 56th Annual Symposium on Foundations of Computer Science, pages 863–882. IEEE, 2015. doi:10.1109/FOCS.2015.58.
  • [42] Xin Li. Improved two-source extractors, and affine extractors for polylogarithmic entropy. In Annual Symposium on Foundations of Computer Science (FOCS), pages 168–177. IEEE, 2016. doi:10.1109/FOCS.2016.26.
  • [43] Xin Li. Non-malleable extractors and non-malleable codes: Partially optimal constructions. In Computational Complexity Conference (CCC), pages 28:1–28:49. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPICS.CCC.2019.28.
  • [44] Xin Li. Two source extractors for asymptotically optimal entropy, and (many) more. In Annual Symposium on Foundations of Computer Science (FOCS), pages 1271–1281. IEEE, 2023. doi:10.1109/FOCS57990.2023.00075.
  • [45] Xin Li and Yan Zhong. Explicit directional affine extractors and improved hardness for linear branching programs. In Computational Complexity Conference (CCC), pages 10:1–10:14. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024. doi:10.4230/LIPIcs.CCC.2024.10.
  • [46] Raghu Meka. Explicit resilient functions matching Ajtai-Linial. In Annual Symposium on Discrete Algorithms (SODA), pages 1132–1148. ACM-SIAM, 2017. doi:10.1137/1.9781611974782.73.
  • [47] Anup Rao. Extractors for low-weight affine sources. In Computational Complexity Conference (CCC), pages 95–101. IEEE, 2009. doi:10.1109/CCC.2009.36.
  • [48] Ran Raz, Omer Reingold, and Salil Vadhan. Extracting all the randomness and reducing the error in trevisan’s extractors. Journal of Computer and System Sciences, 65(1):97–128, 2002. doi:10.1006/JCSS.2002.1824.
  • [49] Ronald L Rivest. All-or-nothing encryption and the package transform. In Fast Software Encryption: 4th International Workshop, FSE’97 Haifa, Israel, January 20–22 1997 Proceedings 4, pages 210–218. Springer, 1997. doi:10.1007/BFB0052348.
  • [50] Ronen Shaltiel. How to get more mileage from randomness extractors. Random Structures & Algorithms, 33(2):157–186, 2008. doi:10.1002/RSA.20207.
  • [51] Ronen Shaltiel and Christopher Umans. Simple extractors for all min-entropies and a new pseudorandom generator. Journal of the ACM (JACM), 52(2):172–216, 2005. doi:10.1145/1059513.1059516.
  • [52] Luca Trevisan. Extractors and pseudorandom generators. Journal of the ACM (JACM), 48(4):860–879, 2001. doi:10.1145/502090.502099.
  • [53] Umesh V. Vazirani. Strong communication complexity or generating quasi-random sequences from two communicating semi-random sources. Combinatorica, 7(4):375–392, 1987. doi:10.1007/BF02579325.
  • [54] John von Neumann. Various techniques used in connection with random digits. Applied Math Series, 12(36-38):5, 1951.
  • [55] Amir Yehudayoff. Affine extractors over prime fields. Combinatorica, 31(2):245–256, 2011. doi:10.1007/S00493-011-2604-9.