Abstract 1 Introduction 2 Preliminaries 3 Upper Bound for 3-restricted Matching Vector Families References Appendix A Proof of Lemma 11

Improved Lower Bounds for 3-Query Matching Vector Codes

Divesh Aggarwal ORCID Centre for Quantum Technologies, National University of Singapore, Singapore Pranjal Dutta ORCID National University of Singapore, Singapore Zeyong Li ORCID Centre for Quantum Technologies, National University of Singapore, Singapore Maciej Obremski ORCID National University of Singapore, Singapore Sidhant Saraogi ORCID Georgetown University, Washington, DC, USA
Abstract

A Matching Vector (𝐌𝐕) family modulo a positive integer m2 is a pair of ordered lists 𝒰=(𝒖1,,𝒖K) and 𝒱=(𝒗1,,𝒗K) where 𝒖i,𝒗jmn with the following property: for any i[K], the inner product 𝒖i,𝒗i=0(modm), and for any ij, 𝒖i,𝒗j0(modm). An 𝐌𝐕 family is called r-restricted if inner products ui,vj, for all i,j, take at most r different values. The r-restricted 𝐌𝐕 families are extremely important since the only known construction of constant-query subexponential locally decodable codes (LDCs) are based on them. Such LDCs constructed via matching vector families are called matching vector codes. Let 𝐌𝐕(m,n) (respectively 𝐌𝐕(m,n,r)) denote the largest K such that there exists an 𝐌𝐕 family (respectively r-restricted 𝐌𝐕 family) of size K in mn. Such a 𝐌𝐕 family can be transformed in a black-box manner to a good r-query locally decodable code taking messages of length K to codewords of length N=mn.

For small prime m, an almost tight bound 𝐌𝐕(m,n)O(mn/2) was first shown by Dvir, Gopalan, Yekhanin (FOCS’10, SICOMP’11), while for general m, the same paper established an upper bound of O(mn1+om(1)), with om(1) denoting a function that goes to zero when m grows. For any arbitrary constant r3 and composite m, the best upper bound till date on 𝐌𝐕(m,n,r) is O(mn/2), is due to Bhowmick, Dvir and Lovett (STOC’13, SICOMP’14).In a breakthrough work, Alrabiah, Guruswami, Kothari and Manohar (STOC’23) implicitly improve this bound for 3-restricted families to 𝐌𝐕(m,n,3)O(mn/3).

In this work, we present an upper bound for r=3 where 𝐌𝐕(m,n,3)mn/6+O(logn), and as a result, any 3-query matching vector code must have codeword length of NK6o(1).

Keywords and phrases:
Locally Decodable Codes, Matching Vector Families
Funding:
Divesh Aggarwal: Funded by the bridging grant at Centre for Quantum Technologies titled “Quantum algorithms, complexity, and communication”.
Pranjal Dutta: Funded by NRF Singapore under the project “Computational Hardness of Lattice Problems and Implications” [NRF-NRFI09-0005].
Zeyong Li: Funded by NRF Singapore under the project “Computational Hardness of Lattice Problems and Implications” [NRF-NRFI09-0005].
Maciej Obremski: Funded by MOE tier 2 grant T2EP20223-0024, Breaking the Box – On Security of Cryptographic Devices.
Sidhant Saraogi: Funded in part by the National Science Foundation CAREER award to Justin Thaler (grant CCF-1845125).
Copyright and License:
[Uncaptioned image] © Divesh Aggarwal, Pranjal Dutta, Zeyong Li, Maciej Obremski, and Sidhant Saraogi; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Error-correcting codes
Related Version:
Previous Version: https://eccc.weizmann.ac.il/report/2024/061/
Acknowledgements:
The authors would like to thank Sivakanth Gopi and Anup Rao for sharing a proof of Theorem 21. They would also like to thanks anonymous reviewers for their helpful comments.
Editor:
Raghu Meka

1 Introduction

1.1 Locally Decodable Codes

A code C:ΣKΣN is said to be (r,δ,ε)-locally decodable if for every i[K], the i-th coordinate of the message 𝒙iΣ can be recovered with probability at least 1ε by a randomized decoding procedure that makes only r queries to the codeword, i.e., reads the codeword at r positions even if the codeword is corrupted on up to δN locations.

Locally decodable codes were formally introduced in [19] but had already been studied informally in the context of the PCP theorem [3, 4]. Locally decodable codes have found applications in many areas of computer science such as worst-case to average-case reductions, private information retrieval, secure multi-party computation, derandomization, matrix rigidity, data structures, and fault-tolerant computation. See [25] for a detailed survey.

A central research question is to understand the largest possible message length K (as a function of N) that can be achieved by a r-query locally decodable code. For the simplest non-trivial setting of r=2, we have a nearly complete understanding – the Hadamard code achieves K=log2N, and it was shown in [20, 12] that this is the best possible up to a constant factor, i.e., K=O(log2N).

Much less is understood about the dependence between K and N for the number of queries r3. In particular, the only known constructions for r-query locally decodable codes for a constant r3 are based on families of matching vector codes [24, 10, 9]. These constructions achieve K=(logN)Ω(loglogN). On the other hand, the known upper bounds on K are very far from what is achieved by these constructions. In a series of works [19, 20, 22, 23, 6], it has been shown that K=N12/(r+1)polylogN, when r is odd, and K=N12/rpolylogN, when r is even. For the special case where r=3, it was shown that K=O(N1/3log2N) in a celebrated recent result [2].

1.2 Matching Vector Families and Application to Locally Decodable Codes

A matching vector (𝐌𝐕) family is a combinatorial object that has been used in different contexts such as Ramsey graphs and weak representation of OR polynomials[11, 17, 1, 13], but the most prominent application of matching vector families is in the construction of constant-query locally decodable codes [9, 25]. A matching vector family is defined by two ordered lists 𝒰=(𝒖1,,𝒖K) and 𝒱=(𝒗1,,𝒗K) where for all i[K], 𝒖i,𝒗i are in mn for some positive integers m,n>1. An S-matching vector family can be defined as follows.

Definition 1 (Matching Vector Family).

Let Sm{0}. An S-matching vector family is a set of vectors 𝒰={u1,,uK}, 𝒱={v1,,vK}, 𝒰,𝒱mn such that:

  • i[K],ui,vi=0.

  • ij, i,j[K],ui,vjS.

An S-matching vector family is often called (|S|+1)-restricted matching vector family, and such a family gives a locally decodable code for messages of block length K, codeword length N=mn, and number of queries r=|S|+1. See Theorem 15 ([25]) for the construction of a locally decodable code from a matching vector family.

Finally, we denote 𝐌𝐕(m,n) (respectively 𝐌𝐕(m,n,r)), as the largest K such that there exists an 𝐌𝐕 family (respectively r-restricted 𝐌𝐕 family) of size K in mn.

1.3 Upper Bounds for Matching Vector Families

Since the only known approach that has led to constructions of constant-query locally decodable codes is via constructing a large matching vector family, it is natural to ask what is the largest K as a function of m,n for which there is a r-restricted matching vector family of size K. This in particular will give an upper bound for the rate of a r-query locally decodable code via matching vector codes, i.e., any r-query locally decodable code achieving a better rate must be via some novel approach that does not require a matching vector family.

This problem was first studied in [9], where the authors showed that when m is a large prime, 𝐌𝐕(m,n)O(mn/2). On the other hand, the same paper established a general upper bound of O(mn1+om(1)), with om(1) denoting a function that goes to zero when a composite m grows. In [8], the authors improved this bound to poly(m)m.625n, for a very special case when m=pq, for distinct primes p and q such that pq, and nm. On the other hand, when mn is a prime, it can be shown via linear algebraic methods that 𝐌𝐕(m,n)O(nm1) [5].

Bhowmick, Dvir and Lovett [7] showed how to extend the results of [9] to the case of m being a composite integer. They showed that 𝐌𝐕(m,n,r)rO(rlogr)mn/2, i.e., for constant query codes, they showed that K=O(N).

Additionally, in the same paper, Bhowmick, Dvir and Lovett also showed that under the (recently proved) polynomial Freiman Ruzsa (PFR) theorem [16], for m constant, 𝐌𝐕(m,n)mO(n/logn).

1.4 Our Result

The main contribution of the paper is to give a better upper bound for 3-restricted matching vector families.

Theorem 2.

Let n,m2 be positive integers with n>10m3. Suppose m has atleast two distinct prime factors, then

𝐌𝐕(m,n,3)mn+2log(n+1)2q+2m2,

where q is the second-smallest prime dividing m.

We already know a polynomial bound for prime powers due to [15] which we state below and prove for completeness in Section 2.4.

Theorem 3.

Suppose n,m2 be positive integers where m is a constant prime power. Then 𝐌𝐕(m,n,3)(2emn)m1.

This allows us to conclude an upper bound for arbitrary m:

 Remark 4.

When m has at least one factor which is not 2, that immediately gives a mn/6+O(logn) upper bound, while if the constant m is a power of 2, the bound is just poly(n), as mentioned above.

Finally, we can also derive a lower bound for the 3-query LDCs derived from matching vector families:

Corollary 5.

Suppose C:ΣKΣN is a 3-query locally decodable code constructed from a 3-restricted matching vector family in a black-box manner (as in Theorem 15), then NK6oK(1) where oK(1) vanishes as K increases.

1.5 Concurrent Work

Concurrent to our work, the Polynomial Freiman Ruzsa conjecture was proven by Gowers, Green, Manners and Tao [16] for all abelian groups with bounded torsion! As a corollary (via Theorem 6.4 of [7]), one could conclude that

𝐌𝐕(m,n)21200c(m)m6logmn/logn,

where c(m) is some explicit constant depending only on m. Infact, if we plug in the bound of [16] to Lemma 6.3 from [7], we can conclude that c(m)m3. Note that this bound is non-trivial (i.e., less than mn) only when n2120063+6log6log6>2256 (plugging in m=6) and beats our bound only when n2258.

While this result supersedes ours in many parameter regimes, the techniques and approaches are very different. And our approach might have the potential of breaking the 2n/logn barrier. (See discussions below.)

1.6 Our Techniques

In the discussion below, we want to prove upper bounds for a matching vector family 𝒰,𝒱 of size K over mn for some positive integers m and n.

Techniques from [7].

We begin by describing Bhowmick, Dvir, and Lovett’s ([7]) approach towards getting an upper bound of mn/2. They observe that if we consider the random variables 𝑼,𝑽 as sampled independently and randomly from 𝒰={𝒖1,,𝒖K} and 𝒱={𝒗1,,𝒗K}, respectively, then we have that Pr[𝑼,𝑽=0]=1K, and hence 𝑼,𝑽modm is far from uniform.

If m is a prime, then standard properties of the inner product two-source extractor imply that 𝐇(𝑼)+𝐇(𝑽)=2logK is not much bigger than nlogm, else the inner product extractor’s output would be very close to uniform. This straightaway yields the desired bound of Kmn/2.

For composite m, this approach does not work directly, since m is not a field. In particular, say for m=6, it is possible to have two random variables distributed uniformly over {0,2,4}n such that their inner product is far from uniform modulo 6, but the sum of the min-entropies is 2nlog23nlog6. The authors overcome this by showing that this is essentially the only bad case, i.e., it can be attributed to some “bad” factor s|m (in the example above, s=2). Hence, with some careful inductive argument on the modulus m/s, one can still force out the fundamental observation: 𝐇(𝑼)+𝐇(𝑽)=2logK, which cannot be too large due to the biased behavior of 𝑼,𝑽, showing the desired upper bound on K.

Going beyond 𝒎𝒏/𝟐.

A natural attempt towards improving the upper bound for K would be to consider the inner product 𝑼,𝑽 where 𝑼=𝑼1++𝑼 for some integer with each 𝑼i sampled uniformly and independently from 𝒰 (and 𝑽 similarly defined), and hope for two nice events:

  1. (a)

    𝐇(𝑼)+𝐇(𝑽) 2logK2logK;

  2. (b)

    𝑼,𝑽 remains far from uniform just like 𝑼,𝑽. And hence 𝐇(𝑼)+𝐇(𝑽)mlogn.

In other words, we are trying to boost the lower bound on min-entropy by adding independent samples of 𝑼, and at the same time maintain the upper bound on min-entropy. And if the above is true, one would obtain a substantial improvement on the upper bound (e.g. Kmn/4 even for the smallest non-trivial =2).

In the rest of this proof overview, we will walk through how we show the two events above are true for any {α,β}-matching vector family. For simplicity, let m=pq for distinct primes p and q, with q>p. And one could further assume that α=0(modp),α0(modq), and β=0(modq),β0(modp); see Lemma 25 for more details.

Event (𝒂)

In order to show that 𝐇(𝑼)logK, it is sufficient to show that 𝑼=i=1𝑼i does not generate too many collisions.

Omitting most technical details, we show that if 𝒖1++𝒖=𝒖1++𝒖 (i.e. a collision), then it must be the case that 𝒖i,𝒗1=α for all 1iq, where 𝒗1 is the matching vector for 𝒖1. This is due to the heavy constraint that – (1) any inner product can only take values from {0,α,β}, and (2) inner product with 𝒗1 gives an expression of the form x1α+x2β, on the LHS, while y1α+y2β, on the RHS, for nonnegative integers xi,yi such that x1+x21. and y1+y2. (Remember 𝒖1,𝒗1=0.)

This anomaly (that there is plausibly 1 less term in the LHS) is sufficient to conclude that if there are too many collisions, one can easily extract a large {α}-matching vector family, which would contradict a known linear upper bound, since a simple pigeon-hole principle argument shows that 𝐌𝐕(pq,n,2)n+1; see Lemma 18.

In the actual proof, for technical reasons, we work with a slightly different distribution 𝑼, which adds up distinct random vectors from 𝒰. In other words, one may view 𝑼 as sampling with replacement and 𝑼 as sampling without replacement from 𝒰. These two distributions are so close to each other that we are able to substitute one with the other without incurring much loss in the final argument. Moreover, this distinction also helps us to argue that the sum of the coefficients in LHS is exactly 1, while it is exactly in RHS. For the detailed proof on lower bounding the min-entropy, see Section 3.2.

Event (𝒃)

To show that 𝑼,𝑽 is far from uniform, we directly examine its bias |𝔼[ω𝑼,𝑽]| in terms of Fourier coefficients (where ω is a primitive m-th root of unity). Via standard Fourier-analytic techniques, and using the “almost” multiplicative property of the bias (Lemma 11), we can almost show that

|𝔼[ω𝑼,𝑽]||𝔼[ω𝑼,𝑽]|2.

That is, adding independent samples of 𝑼 (and 𝑽) will not smooth out the inner product for not too large (see Lemmas 11 and 13).

On the other hand, since 𝑼,𝑽 takes only 3 possible values with 0 showing up with negligible probability 1K, we can show that |𝔼[ω𝑼,𝑽]| has to be large! And this is essentially all we need. For the detailed proof on upper bounding the min-entropy, see Section 3.1.

Combining all of the above and choosing =q, we get 2qlogK𝐇(𝑼)+𝐇(𝑽)nlogm which gives us the desired upper bound of Kmn/2q. For a general m one can always reduce it to working with m=psqt, for primes p and q; see Lemmas 24 and 25. And for m=psqt, the argument is similar to the above.

Lastly, it is worth noting that via a more direct analysis on the min-entropy of carefully chosen random variables, we are able to avoid the inductive argument in [7], which was tailored for composite m.

1.7 Conclusions and Open Questions

Our approach can be broadly summarized as considering 𝑼:=𝑼1++𝑼, and similarly 𝑽:=𝑽1++𝑽, and then proving that

  1. 1.

    ω𝑼,𝑽 is far from uniform, where ω is the m-th root of unity, and

  2. 2.

    𝐇(𝑼)𝐇(𝑼) for =q.

The first item implies that 𝐇(𝑼)+𝐇(𝑽) is not much larger than mlogn, and this combined with the second item implies that 𝐇(𝑼)=𝐇(𝑽) is small, giving an upper bound on K.

We prove these two items for =q, and for r-restricted families for r=3. Notice that the approach is much more general, and it is just a shortcoming of our proof techniques that doesn’t allow us to go beyond >q, or r>3. In particular, for r=3, the statement (1) above holds for q, and we leave it as an open question whether one can prove that for q, 𝐇(𝑼)q𝐇(𝑼). This will immediately imply a better upper bound on K for 3-query matching vector codes.

Similarly, we leave it as an open question whether the same approach can be extended to a larger number of queries.

2 Preliminaries

2.1 Random Variable and Entropy

Definition 6 (Min-Entropy).

Let X be a random variable over a set 𝒳, the Min-Entropy is defined as:

𝐇(X)=minx𝒳log(Pr[X=x]).
Definition 7 (Collision Entropy).

Let X and Y be independent and identically distributed random variables, the Collision Entropy is defined as:

𝐇2(X)=logPr[X=Y].

Alternatively,

𝐇2(X)=logx(Pr[X=x])2.
Observation 8 (Relation Between Collision and Min-Entropy).

For any random variable X, it holds that

𝐇2(X)𝐇(X).
Lemma 9.

Let X and Y be independent random variables of the same domain 𝒳 closed under addition. It holds that

𝐇(X)𝐇(X+Y).
Proof.

For any z𝒳, we have:

Pr[X+Y=z] =y𝒳Pr[X=zy|Y=y]Pr[Y=y]
=y𝒳Pr[X=zy]Pr[Y=y]
y𝒳(maxxPr[X=x])Pr[Y=y]
maxxPr[X=x].

Therefore,

𝐇(X)=log1maxxPr[X=x]log1maxzPr[X+Y=z]=𝐇(X+Y),

as desired.

Let ω be a primitive root of unity of order m. Let X and Y be two probability distributions over mn. The bias of the output distribution wrt ω, denoted 𝔼[ωX,Y], can be defined as follows:

𝔼[ωX,Y]=𝔼xX,yY[ωx,y]=x,yX(x)Y(y)ωx,y.
Lemma 10 (Lemma 2.14 from [7]).

Let ω be a primitive root of unity of order m. Let X and Y be two probability distributions over mn. If |𝔼[ωX,Y]|ϵ, then 𝐇2(X)+𝐇2(Y)nlogm+2log(1/ϵ).

Lemma 11 (Lemma 3.3 from [21]).

Let ω be a primitive root of unity of order m. Let X and Y be two probability distributions over mn, then |𝔼[ωXX,Y]||𝔼[ωX,Y]|2. Here XX is shorthand for taking the difference of two samples drawn independently according to the distribution X.

 Remark 12.

Lemma 11 was stated for distributions over 𝔽n for any finite field 𝔽 in [21], but the proof trivially extends to distributions over mn. For completeness, we reproduce the proof from [21] in the appendix.

Lemma 13 ([21]).

Let ω be a primitive root of unity of order m. Let X and Y be two probability distributions over mn. For any integer x,y0, |𝔼[ω2xX2xX,2yY2yY]||𝔼[ωX,Y]|2x+y+2. Here 2xX represents the distribution of adding 2x independent samples of X.

Proof.

Notice that (XX)(XX)=2X2X. Hence this follows from applying Lemma 11 (x+1)+(y+1)=x+y+2 times.

2.2 Matching Vector Families and Codes

Much of these definitions follow the same notation as Yekhanin’s survey [25].

Definition 14 (Locally Decodable Codes).

A q-ary code C:ΣKΣN with |Σ|=q is (r,δ,ε)-locally decodable if there exists a randomized decoding algorithm 𝒜 such that

  1. 1.

    For all xΣK, i[K] and all yΣN such that Δ(C(x),y)δ:

    Pr𝒜[𝒜y(i)=x(i)]1ε
  2. 2.

    𝒜 makes atmost r queries to y.

The best-known constructions of locally decodable codes are derived from constructions of matching vector families over m for composite m’s. The following theorem demonstrates the parameters of locally decodable codes obtained from given parameters of a matching vector family.

Theorem 15 ([25]).

Let 𝒰,𝒱 be a S-matching vector family over mn with |𝒰|=|𝒱|=K. Suppose m|q1 for some prime power q, then there exists a linear code C:𝔽qK𝔽qmn that is (|S|+1,δ,(|S|+1)δ)-locally decodable for every δ.

In the low-query complexity regime, r=O(1), the best-known matching vector families are based on Grolmusz’s construction of set systems.

Theorem 16 ([25, Lemma 4.8]).

Let m=i=1tpi be a product of distinct primes. Let w be a positive integer. Let {ei}i=1t be integers such that pieiw1/t for all i. Let d=maxipiei and hw be arbitrary. Then there exists an (hw)-sized σm-bounded family of matching vectors in mn where n=(hd) and σ is an arbitrary real larger than i[t]1pi.

Then, using Theorem 15, this matching vector family construction can be transformed into a linear binary LDC with the following parameters:

Theorem 17 ([18], as stated in [25], Theorem 3.11).

For every integer t2 and for all K2, there exists an r=32t2-query (binary) linear locally decodable code over 𝔽2t encoding K-long messages to 2t(logK)1/t(loglogK)11/t-long codewords and allowing δ=O(1/r) fraction of errors.

In particular, when r=3, there exists a code with codeword length N=22O~(logK)

2.3 Upper Bounds for 2-restricted Matching Vector Families

In this section, we prove an upper bound on any {α}-matching vector family over mn. Without loss of generality, one can assume that m is a prime power ps, because otherwise if m=psqt, then without loss of generality α0(modps), and therefore the same set of vectors form a matching vector family over psn.

Lemma 18.

For any prime p and positive integer s, 𝐌𝐕(ps,n,2)ns+1.

Proof.

Suppose α0(modps). Assume towards contradiction that 𝒰,𝒱 is a {α}-matching vector family over psn with |𝒰|=|𝒱|=Kns+2.

Define 𝒖i:=𝒖ns+2𝒖i for 𝒖i𝒰. By definition of the {α}-matching vector family, for i,jns+1 we have:

𝒖i,𝒗j={α,i=j0,ij.

We claim that there exist integer coefficients c1,,cns+1, not all zero, such that (p1)cip1, and

i=1ns+1ci𝒖i=0.

To see this, consider all possible sums i=1ns+1di𝒖ipsn, for di{0,1,,p1}. There are pns+1 such sums that can each take one of (ps)n=pns values. By the pigeon-hole principle, there exist two distinct sums that are equal, i.e.,

i=1ns+1di𝒖i=i=1ns+1di𝒖i,

for some tuples (d1,,dns+1)(d1,,dns+1). The claim follows by taking ci=didi for i{1,,ns+1}.

We have that at least one of c1,,cns+1 is non-zero. Without loss of generality, let cns+10. Notice that cns+1{(p1),,1}{1,,p1}, and hence has a multiplicative inverse modulo ps. Let ci=cns+11ci(modps) for i[ns+1]. Thus, we have that

𝒖ns+1=i=1nsci𝒖i.

Now take the inner product on both sides with 𝒗ns+1. Note that, 𝒖ns+1,𝒗ns+1=α, while

i=1nsci𝒖i,𝒗ns+1=i=1nsci𝒖i,𝒗ns+1= 0.

Therefore, we obtain that α=0 which is a contradiction.

 Remark 19.

By the above lemma, we get 𝐌𝐕(m,n,2)nlogm+1, since trivially, slogm/logplogm.

2.4 Upper Bounds for Prime Powers

In this section, we provide a polynomial upper bound for matching vector families over m where m is a prime power. The sketch was personally communicated to us by Sivakanth Gopi. To prove this, we need the following lemma from [14]. We state it without proof, but it follows from Lucas’ Theorem.

Lemma 20.

Let p be a prime and s+. There exists a polynomial f(x1,,xn) of degree ps1 such that for x{0,1}n, f(x)0(modp) iff i=1nxi is divisible by ps. Specifically, f(x)=j=1s1(1(i=1nxipj)p1).

Here (i=1nxik)=i1<<ikxi1xik is always a valid integer polynomial for kn. Note that in [14], the lemma is technically stated with 1f as opposed to f. We are now ready to show the upper bound. Also, note that f takes only 2 values 1 and 0 (modp), depending on the divisibility property.

Theorem 21 (Upper bound for prime-power [15]).

Let m=ps for some prime p and s+. Then, for any matching vector family 𝒰,𝒱mn of size t=|𝒰|=|𝒱| we have that t(2emn)m1.

Proof.

We will first convert the matching vector family 𝒰,𝒱 into a matching vector family consisting of vectors with only zero or one coordinates, we stress that the matching vector family is still considered over m and 0,1 are interpreted as elements of m, not 𝔽2. Thereafter, we will argue by showing an upper bound on the rank of a specially constructed matrix, via the polynomial defined in Lemma 20.

Converting into a binary matching vector family.

We will transform vectors in 𝒰,𝒱 into vectors in {0,1}nmn for n=nm2, such that the pairwise inner-product 𝒖,𝒗 for 𝒖𝒰,𝒗𝒱 is preserved over m. For any 𝒖𝒰 (resp. 𝒗𝒱), map each coordinate 𝒖i (resp. 𝒗i) into the m×m matrix111Matrix is interpreted as a vector simply by reading entries of the matrix row by row. 𝒖i (resp. 𝒗i) with all 1’s in the first 𝒖i rows (resp. first 𝒗i columns). It is easy to see that 𝒖i𝒗i=𝒖i,𝒗i and

𝒖,𝒗=i[n]𝒖i𝒗i=i[n]𝒖i,𝒗i=𝒖,𝒗

As a result, we can turn any matching vector family 𝒰,𝒱mn into one with only binary vectors in mmn2. For the rest of the argument, denote 𝒰,𝒱mn to be the binary vectors obtained like above from the original 𝒰,𝒱 where n=nm2.

Constructing a useful matrix.

Suppose 𝒰={𝒖(1),,𝒖(t)},𝒱={𝒗(1),,𝒗(t)} where all
𝒖(i),𝒗(j){0,1}nmn. Recall m=ps, let f(x) be such a polynomial as defined in Lemma 20. By the same lemma, 𝒖,𝒗=i=1n𝒖i𝒗i=0(modm) iff f(𝒖1𝒗1,,𝒖n𝒗n)0(modp). Now, we define a matrix Mpt×t whose rows are indexed by vectors in 𝒰 and whose columns are indexed by vectors in 𝒱 such that:

Mi,j=M(𝒖(i),𝒗(j)):=f(𝒖1(i)𝒗1(j),,𝒖n(i)𝒗n(j)).

Note that M(𝒖(i),𝒗(j))0 iff i=j. Hence, M is full rank diagonal matrix (infact, it is the identity matrix!).

On the other hand, by Lucas’ Theorem, f has degree at most ps1=m1 and n variables. Notice that there are at most (nm1)+(nm2)++(n0)(n+m1m1) possible monomials up to degree m1. For any polynomial f of degree m1 we can consider the vector 𝒇~m(n+m1m1), a vector of coefficients of each of (n+m1m1) monomials. One can also take any input x1,,xn and consider a vector 𝒙~m(n+m1m1) consisting of evaluations of each of the monomials on x1,,xn. Notice that then f(x1,,xn)=𝒙~,𝒇~. Similarly, for each 𝒖𝒰,𝒗𝒱 we can find 𝒖~,𝒗~m(n+m1m1), such that f(𝒖1𝒗1,,𝒖n𝒗n)=𝒖~,𝒗~, for example 𝒖~ consists of evaluations of monomials on input u1,,un while 𝒗~ combines evaluations of monomials on v1,,vn with coefficients of polynomial f. Thus, an obvious counting argument gives the desired upper bound:

t=rank(M)(n+m1n)=(nm2+m1m1)(2enm)m1.

Note that the above proof fails immediately for composites. For one, we don’t have a version of the Lucas’ Theorem. Moreover, there is no unified notion of rank that can be utilized, for example by reducing it to working over a prime field.

 Remark 22.

For any constant m=ps for some prime p and positive integer s, Theorem 21 implies 𝐌𝐕(m,n,r)(2emn)m1=poly(n) for any 0<rm.

2.5 Some properties of 𝟑-restricted matching vector families

We will consider matching vector families in m where m is a composite positive integer. All algebraic operations in this section, unless otherwise stated, are modulo m.

Let α,βm{0}. Consider a matching vector family (𝒰,𝒱) such that 𝒰={u1,,uK},V={v1,,vK}mn that satisfy the following properties:

  • For all i[K], 𝒖i,𝒗i=0.

  • For all i,j[K] with ij, we have that 𝒖i,𝒗j{α,β}.

We call (𝒰,𝒱) an {α,β}-matching vector family of size K.

 Remark 23.

For the rest of the paper until the end of Section 3.2, we will assume that (𝒰,𝒱) is not a matching vector family modulo m for any m|m. We can make this assumption without loss of generality. This is because we will prove an upper bound on K that increases as a function of m, and we get a better bound if there exists such an m.

In our argument, it will be useful to work with composites that are exactly the product of two prime powers. This is clearly the first non-trivial case since we have a polynomial upper bound (in n) for the case of prime powers. In fact, for 3-restricted matching vector families, [7] demonstrated that this is the only non-trivial case. We include the proof below for completeness.

Lemma 24 ([7]).

Let (𝒰,𝒱) be a (q+1)-restricted matching vector family in mn. Then m has at most q prime factors.

Proof.

Suppose m=ipiri for some primes pi’s and corresponding positive integers ri’s. Suppose the possible non-zero values of inner products are α1,,αq(modm). For each αj, there must exist i such that αj0(modpiri). We can take m to be the product of atmost q prime powers that satisfy the above condition for the respective αj’s. Thus the (q+1)-restricted matching vector family in mn induces (via the modulo m operation) a matching vector family in mn of the same size, and with the same number of restrictions.

From the previous lemma, we can assume m=psqt for any {α,β}-matching vector family (𝒰,𝒱) where p,q are some primes and s,t are some positive integers. In the next lemma, we can show some more restrictions on the values taken by α and β.

Lemma 25.

Let m=psqt be a composite positive integer such that for any m which is a divisor of m, (𝒰,𝒱) is not a matching vector family modulo m. Then we have α=0(modpsqt1), α0(modqt), β=0(modps1qt), β0(modps).

Proof.

By the same argument as Lemma 24, we can assume without loss of generality that α0(modqt) and β0(modps). Now, suppose for a contradiction β0(modps1qt). Clearly, α0(modps1qt) Then (𝒰,𝒱) is a {α,β}-matching vector family modulo m=ps1qt|m which is a contradiction. Similarly, we can show α=0(modpsqt1).

3 Upper Bound for 3-restricted Matching Vector Families

Let 𝒰,𝒱 be a matching vector family of size K in mn with non-zero residues α and β. By Lemma 24, we have that m=psqt for primes p and q and integers s,t1 with p<q. Further, using Lemma 25, we have that α=0(modpsqt1), α0(modqt), β=0(modps1qt), β0(modps). Define, a parameter τ (which is a small constant) as follows:

τ:=12pq.

Let 𝒮q,K={A[K]|A|=q} be the set of all subsets of [K] of size q. Let ω=e2πim be a primitive root of unity of order m.

For analysis, we define a few random variables as follows: Let 𝑼i be uniform and independent vectors drawn from 𝒰. Let be the smallest integer such that q2. We have:

  1. 1.

    𝑼:=𝑼i.

  2. 2.

    𝑼§:=i=12𝑼ii=2+12+1𝑼i. That is, sample with replacement 22 independent vectors from 𝒰, add up the first 2 vectors and subtract away the remaining 2 vectors.

  3. 3.

    𝑼:=i=1q𝑼i. That is, sample with replacement q independent vectors from 𝒰 and add them up.

  4. 4.

    𝑼:=i=q+12𝑼ii=2+12+1𝑼i. Here we define 𝑼 such that 𝑼§=𝑼+𝑼.

  5. 5.

    𝑼 is distributed as follows: pick a subset A𝒮q,K uniformly at random and 𝑼:=iA𝒖i. In other words, sample without replacement q vectors from 𝒰 and add them up. This is in contrast to 𝑼 which is sampling with replacement.

Random variables for 𝒱 are defined similarly.

Sampling without replacement helps in lower bounding the min-entropy of 𝑼. On the other hand, we upper bound the min-entropy of 𝑼 where sampling with replacement (and hence full independence) is useful. And we are able to relate these two quantities as these two distributions are really close to each other.

The introduction of 𝑼§ (and hence 𝑼=𝑼§𝑼) serves as a tool for lower bounding the bias of 𝑼 via Lemma 13 because we are restricted to sums containing powers of 2 elements.

3.1 Upper Bound on Min-Entropy

In the following lemma, we prove that since 𝒰,𝒱 constitute a 3-restriced 𝐌𝐕 family over mn, then 𝑼,𝑽 has to be far from uniform. This is captured via Fourier analysis.

Lemma 26 (Large bias for 𝐌𝐕 family).

Let K10pq. Then, we have that |𝔼[ω𝐔,𝐕]|τ.

Proof.

This proof relies on the fact that the random variable ω𝑼,𝑽 is essentially supported on two values, {ωα,ωβ}, since the mass on ω0 (corresponding to 𝑼,𝑽=0) is very small. To complete the argument, we show that αβ is an integer different from m/2, and thus the mass on ωα doesn’t cancel that on ωβ. The detailed argument is given below.

Recall that 𝑼,𝑽{0,α,β} and Pr[𝑼,𝑽=0]=1K. Let λ:=Pr[𝑼,𝑽=α], λ[0,11K]. Then Pr[𝑼,𝑽=β]=1λ1K. Since α=0(modpsqt1) and β=0(modps1qt), let αm=cq and βm=dp, for some 0<c<q and 0<d<p. By definition of the bias, we have

|𝔼[ω𝑼,𝑽]| =|u,vmn𝑼(u)𝑽(v)ωu,v|
minλ[0,11K]|1K+λωα+(11/Kλ)ωβ|
minλ[0,11K]|λωα+(11/Kλ)ωβ|1K
=minλ[0,11K]|λ+(11/Kλ)ωβα|1K
=minλ[0,11K]λ2+(11/Kλ)2+2λ(11/Kλ)cos(2π(cqdp))1K
minλ[0,11K]λ2+(11/Kλ)2+2λ(11/Kλ)cos((pq1)πpq)1K
(K1K)sin(π2pq)1K
τ.

The third equality follows from the identity that for constants ci, and θ,

|c1+c2eiθ|2=|(c1+c2cosθ)+ic2sinθ|2=c12+c22+2c1c2cosθ.

The third last inequality follows from the observation (which we discuss below) that |2πcpdqpqπ|πpq, and that cos θ for θ(0,2π) is minimized when |θπ| is minimized. If pq is odd, the above is trivially true since after multiplying left-hand-side by pqπ we obtain |2(cpdq)pq| and this is a difference of odd and even number and thus the absolute value has to be at least 1. Now, suppose p=2, via similar argument we have

|2cdqq1|1q>12q,

since c0(modq) and q is an odd prime.

The second last inequality follows from taking λ=K12K which minimizes the expression, and using 1cosθ=2sin2(θ/2), giving a value of the square root to be

(2(K12K)2 2(K12K)2cos(πpq))1/2=(K1K)sin(π2pq).

Finally, the last inequality follows from sinxx2 for x[0,1] and that K>10pq.

Next, we will show that 𝑼 has low min-entropy.

Lemma 27.

|𝔼[ω𝑼+𝑼,𝑽+𝑽]|τ16q2.

Proof.

Recall that is the smallest integer such that q2. Thus, 21<q, i.e., 2<2q. By Lemma 13, we can control the bias of 𝑼§,𝑽§:

|𝔼[ω𝑼+𝑼,𝑽+𝑽]| =|𝔼[ω𝑼§,𝑽§]|
|𝔼[ω𝑼,𝑽]|22(+1)
|𝔼[ω𝑼,𝑽]|4(2q)2
=τ16q2.

Define the event EU (resp. EV) where each 𝑼i (resp. 𝑽i) is unique when 𝑼 (resp. 𝑽) is sampled. Notice that 𝑼|EU is distributed exactly as 𝑼. Also, Pr[¬EU]=Pr[¬EV]q2K and so, by the union bound Pr[¬(EUEV)]2q2K.

Lemma 28.

|𝔼[ω𝑼+𝑼,𝑽+𝑽]|τ16q22q2K.

Proof.

The proof uses the above observation that 𝑼|EU (resp. 𝑽|EV) is distributed exactly as 𝑼 (resp. 𝑽), and that both EU and EV hold simultaneously with probability at least 12q2K. In particular,

|𝔼[ω𝑼+𝑼,𝑽+𝑽]| =|𝔼[ω𝑼+𝑼,𝑽+𝑽EUEV]Pr[EUEV]
+𝔼[ω𝑼+𝑼,𝑽+𝑽¬(EUEV)]Pr[¬(EUEV)]|
|𝔼[ω𝑼+𝑼,𝑽+𝑽EUEV]|Pr[EUEV]
+|𝔼[ω𝑼+𝑼,𝑽+𝑽¬(EUEV)]|Pr[¬(EUEV)]
=|𝔼[ω𝑼+𝑼,𝑽+𝑽]|Pr[EUEV]
+|𝔼[ω𝑼+𝑼,𝑽+𝑽¬(EUEV)]|Pr[¬(EUEV)]
|𝔼[ω𝑼+𝑼,𝑽+𝑽]|+2q2K.

It follows that |𝔼[ω𝑼+𝑼,𝑽+𝑽]|τ16q22q2K.

Lemma 29.

𝐇(𝑼)+𝐇(𝑽)nlogm2log(τ16q22q2K).

Proof.

By Lemma 28 and Lemma 10, we have that

𝐇2(𝑼+𝑼)+𝐇2(𝑽+𝑽)nlogm2log(τ16q22q2K).

Finally, notice that 𝑼 is independent of 𝑼 and 𝑽 is independent of 𝑽. By Lemma 9, we have:

𝐇(𝑼)+𝐇(𝑽)𝐇(𝑼+𝑼)+𝐇(𝑽+𝑽)nlogm2log(τ16q22q2K),

where 𝐇(X)𝐇2(X) for any random variable X.

3.2 Lower Bound on Min-Entropy

In this subsection, we will talk about sums of vectors from 𝒰,𝒱. For each A[K], we can naturally associate a sum of vectors iA𝒖i from 𝒰 (and similarly for 𝒱). Note that we only care about unique sums where each vector from 𝒰 (or 𝒱) appears only once. In this view, a collision corresponds to pair of sets A,B,AB such that iA𝒖i=jB𝒖j (or similarly for vectors from 𝒱). In the following we give a lower bound on the min-entropy of 𝑼. By symmetry, the statement holds for 𝑽 as well.

We start with a lemma that discusses the behaviour of a collision.

Lemma 30.

Let m=psqt where q>p and 𝒰,𝒱 be a {0,α,β}-restricted matching vector family of size K such that α=0(modpsqt1), α0(modqt) and β=0(modps1qt), β0(modps). Let A,B𝒮q,K,AB be such that iA𝐮i=iB𝐮i where each 𝐮i𝒰. Then:

  1. (i)

    For each iA,jB, 𝒖i𝒖j (AB=).

  2. (ii)

    For each iA,jB, 𝒖j,𝒗i=α(modqt).

Proof.
Proof of Part (i).

For any iA, we can write jB𝒖j,𝒗ijA𝒖j,𝒗i as a linear combination of α’s and β’s such that

aiα+biβ=jB𝒖j,𝒗ijA𝒖j,𝒗i=jB𝒖jjA𝒖j,𝒗i=0(modm) (1)

for some integers ai,bi.

As 𝒖i,𝒗i=0(modm), there are q terms in the first sum and q1 terms in the second sum; this is where we use the fact that 𝒖i were picked without replacement. Therefore, we have the obvious inequalities: (q1)ai,biq. Since, β=0(modqt), and aiα+biβ=0(modm), it follows that aiα+biβ=0(modqt)aiα=0(modqt). Since α is divisible by qt1, and not qt, and ai is between (q1) and q, it must happen that either ai=0 or ai=q.

Suppose for a contradiction that iAB. Let kA\B (non-empty since AB). Then,

akα+bkβ=jB𝒖j,𝒗kjA𝒖j,𝒗k=jB\A𝒖j,𝒗kjA\(B{k})𝒖j,𝒗k.

By our choice of 𝒖k, fact that all inner products on the right-hand side above are non-zero, and since |A|=|B|=q, we know that:

  1. 1.

    |B\A|=|A\(B{k})|+1, and therefore ak+bk=|B\A||A\(B{k})|=1.

  2. 2.

    (q1)|A\(B{k})|bk|B\A|q1

If ak=0, then bk=1 implies that akα+bkβ=β0(modm), a contradiction to Equation 1. Also, notice that akq since for the same reason as above ak|B\A|q1. This completes the proof of Part (i).

Proof of Part (ii).

Now, consider any iA, clearly iB from our proof above. From a similar argument as above, we know ai+bi=1. And further ai=q and bi=1ai=(q1). However the only way we can have ai=q is if each term in the sum jB𝒖j,𝒗i is α which completes the proof. Note that we can trivially reduce from modm to modqt, since 𝒖j,𝒗i=α0(modqt).

Theorem 31.

Let m=psqt where q>p and let 𝐔 be defined as in the beginning of this section. Then 𝐇(𝐔)qlog(Kq)log(nt+1).

Proof.

We will upper bound the probability Pr𝑼[𝑼=𝒖], where 𝒖mn is an arbitrary vector. Let

C𝒖:={A𝒮q,KiA𝒖i=𝒖},

be the set of sums of length q from 𝒮q,K such that its sum is equal to 𝒖. Note that each term in the sum is unique by how we sample 𝑼.

Now, consider any pair of collisions A,BC𝒖. From Lemma 30, we know that for each iA,jB, ij and 𝒖j,𝒗i=α, and all the sets in C𝒖 are disjoint. In particular, if we pick exactly one vector from each sum in C𝒖 (and their corresponding matching vectors in 𝒱), the set of vectors form a {α}-matching vector family in qtn. Formally, 𝒰:={𝒖ii=min(A),AC𝒖}𝒰, and 𝒱𝒱 with the corresponding matching vectors forms a {α}-matching vector family in qtn of size |C𝒖|.

By Lemma 18, we have |C𝒖|nt+1. Therefore, we have

𝐇(𝑼)log((Kq)nt+1)qlog(Kq)log(nt+1).

 Remark 32.

As pointed out by a helpful anonymous reviewer, we can extend Lemma 30 to work for A,B𝒮q,K where q>q in certain cases. Suppose we again write jB𝐮j,𝐯ijA𝐮j,𝐯i as a linear combination of ai,qα+bi,qβ=0(modm). The proof only requires two properties from q:

  1. 1.

    For the first part, we require that for every q′′<q the corresponding equation ai,q′′α+bi,q′′β=0(modm) has only trivial solutions (ai,q′′=bi,q′′=0).

  2. 2.

    For the second part, we require that ai,q=q.

For m=6, q=3 but by a calculation it can be shown that for m=15, we can have q=6>5=q. This implies that we can improve the conclusion of Theorem 31 to 𝐇(𝐔)qlog(Kq)log(nt+1) which eventually gives us a better bound for Theorem 33. However, we omit this for clarity.

3.3 Main Theorem

We are now ready to prove our main theorem.

Theorem 33.

Let n,m1 be positive integers with n>10m3. Suppose 𝒰,𝒱 is a 3-restricted 𝐌𝐕 family of size K over mn. Then K(2emn)m1 if m is a prime power, and otherwise

Kmn+2log(n+1)2q+2m2,

where q is the second-smallest prime dividing m.

Proof.

Let m be the smallest factor of m such that 𝒰,𝒱 is a 3-restricted 𝐌𝐕 family of size K over mn. If m is a prime power, then by Theorem 21,

K(2emn)m1(2emn)m1.

Suppose m is not a prime power, then K(2emn)m1mn+2log(n+1)2q+2m2; the second inequality follows, since n>10m3.

Now we consider the case when m is not a prime power. Then, by Lemma 24, we have that m=p1sp2t for primes p1 and p2 and integers s,t1 with p1<p2. Further, using Lemma 25, we have that α=0(modp1sp2t1), α0(modp2t), β=0(modp1s1p2t), β0(modp1s).

Let q be the second-smallest prime dividing m, and we assume that K>mn+2log(n+1)2q+2m2, since otherwise we are already done. Also, note that by assumption, p2q. Since n10m3, and 2p1p2m, this implies that

Km7m2m2(2m)4m2 4p22(2p1p2)16p22.

By Lemma 29, we have without loss of generality (between 𝑼 and 𝑽) that

𝐇(𝑼) nlogm2log((12p1p2)16p222p22K)
nlogm2log(12(12p1p2)16p22)
nlogm2+16p22log(2p1p2)+1
nlogm2+4m2logm+4m2+1.

Here we used that 2p2p1p2m. On the other hand, by Theorem 31, we know that

𝐇(𝑼)p2log(Kp2)log(nt+1).

Putting them together we have:

p2log(Kp2)log(nt+1) nlogm2+4m2logm+4m2+1.

This gives us that

K p2mn2p2m4m2p224m2+1p22log(nt+1)p2
mn2p2m4m2p2+1m1.9m2p22log(nt+1)p2
mn2p2m6m2p22log(nt+1)p2
mn2p2m2m22log(nt+1)p2
mn2qm2m22log(nt+1)p2
=mn2qm2m2mlog(nt+1)p2logm;
mn2qm2m2mlog(nt+1)p2tlogp2;
mn2qm2m2mlog(nt+1)p2t
mn2qm2m2mlog(nt+1)1/tp2
mn2qm2m2mlog(n+1)p2
mn+2log(n+1)2qm2m2.

In the second inequality, we used p2m, and also, 24m2+1<m1.9m2, which is true since m6. In the third inequality, we used the fact that 6m2p25.9m2p2+1, which is equivalent to showing m210p21, which is true since m210p2=(m2p2)(m5)>1, because m2p22q6. In the seventh inequality, we used that mp2t. Finally, the second to last inequality follows from Bernoulli’s inequality: (nt+1)(n+1)t.

This is a contradiction to our assumption that K>mn+2log(n+1)2q+2m2. This finishes the proof.

Corollary 34.

For an arbitrary positive integer m, consider a 3-query matching vector codes C:ΣKΣN constructed from 3-restricted matching vector families using the construction in Theorem 15, then NΩ(K6oK(1)).

The proof follows straightforwardly from the Theorem 15 combined with Theorem 33.

References

  • [1] Noga Alon. The shannon capacity of a union. Combinatorica, 18(3):301–310, 1998. doi:10.1007/PL00009824.
  • [2] Omar Alrabiah, Venkatesan Guruswami, Pravesh K Kothari, and Peter Manohar. A near-cubic lower bound for 3-query locally decodable codes from semirandom csp refutation. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing, pages 1438–1448, 2023. doi:10.1145/3564246.3585143.
  • [3] Sanjeev Arora, Carsten Lund, Rajeev Motwani, Madhu Sudan, and Mario Szegedy. Proof verification and the hardness of approximation problems. Journal of the ACM (JACM), 45(3):501–555, 1998. doi:10.1145/278298.278306.
  • [4] Sanjeev Arora and Shmuel Safra. Probabilistic checking of proofs: A new characterization of np. Journal of the ACM (JACM), 45(1):70–122, 1998. doi:10.1145/273865.273901.
  • [5] László Babai and Péter Frankl. Linear algebra methods in combinatorics. to appear, 2020.
  • [6] Arnab Bhattacharyya, L. Sunil Chandran, and Suprovat Ghoshal. Combinatorial Lower Bounds for 3-Query LDCs. In 11th Innovations in Theoretical Computer Science Conference (ITCS 2020), volume 151, pages 85:1–85:8, 2020. doi:10.4230/LIPIcs.ITCS.2020.85.
  • [7] Abhishek Bhowmick, Zeev Dvir, and Shachar Lovett. New bounds for matching vector families. SIAM Journal on Computing, 43(5):1654–1683, 2014. doi:10.1137/130932296.
  • [8] Yeow Meng Chee, San Ling, Huaxiong Wang, and Liang Feng Zhang. Upper bounds on matching families in pqn. IEEE transactions on information theory, 59(8):5131–5139, 2013.
  • [9] Zeev Dvir, Parikshit Gopalan, and Sergey Yekhanin. Matching vector codes. SIAM Journal on Computing, 40(4):1154–1178, 2011. doi:10.1137/100804322.
  • [10] Klim Efremenko. 3-query locally decodable codes of subexponential length. In Proceedings of the forty-first annual ACM symposium on Theory of computing, pages 39–44, 2009. doi:10.1145/1536414.1536422.
  • [11] P. Frankl and R. M. Wilson. Intersection theorems with geometric consequences. Combinatorica, 1(4):357–368, December 1981. doi:10.1007/bf02579457.
  • [12] Oded Goldreich, Howard Karloff, Leonard J Schulman, and Luca Trevisan. Lower bounds for linear locally decodable codes and private information retrieval. computational complexity, 15:263–296, 2006. doi:10.1007/S00037-006-0216-3.
  • [13] P. Gopalan. Constructing ramsey graphs from boolean function representations. In 21st Annual IEEE Conference on Computational Complexity (CCC’06), pages 14 pp.–128, 2006. doi:10.1109/CCC.2006.14.
  • [14] Sivakanth Gopi. Lecture notes: Polynomial representing or mod m. https://homes.cs.washington.edu/˜anuprao/pubs/codingtheory/lecture12.pdf, 2019.
  • [15] Sivakanth Gopi. Upper bound for matching vector families for prime-powers. Personal Communication, 2024.
  • [16] W. T. Gowers, Ben Green, Freddie Manners, and Terence Tao. Marton’s conjecture in abelian groups with bounded torsion, 2024. arXiv:2404.02244.
  • [17] Vince Grolmusz. On the weak mod m representation of boolean functions. Chicago Journal of Theoretical Computer Science, 1995.
  • [18] Toshiya Itoh and Yasuhiro Suzuki. Improved constructions for query-efficient locally decodable codes of subexponential length. IEICE Transactions on Information and Systems, 93(2):263–270, 2010. doi:10.1587/TRANSINF.E93.D.263.
  • [19] Jonathan Katz and Luca Trevisan. On the efficiency of local decoding procedures for error-correcting codes. In Proceedings of the thirty-second annual ACM symposium on Theory of computing, pages 80–86, 2000. doi:10.1145/335305.335315.
  • [20] Iordanis Kerenidis and Ronald de Wolf. Exponential lower bound for 2-query locally decodable codes via a quantum argument. Journal of Computer and System Sciences, 3(69):395–420, 2004. doi:10.1016/J.JCSS.2004.04.007.
  • [21] Anup Rao. An exposition of bourgain’s 2-source extractor. In Electronic Colloquium on Computational Complexity (ECCC), volume 14. Citeseer, 2007.
  • [22] David P Woodruff. New lower bounds for general locally decodable codes. In Electronic Colloquium on Computational Complexity (ECCC), 2007.
  • [23] David P Woodruff. A quadratic lower bound for three-query linear locally decodable codes over any field. Journal of Computer Science and Technology, 27(4):678–686, 2012. doi:10.1007/S11390-012-1254-8.
  • [24] Sergey Yekhanin. Towards 3-query locally decodable codes of subexponential length. Journal of the ACM (JACM), 55(1):1–16, 2008. doi:10.1145/1326554.1326555.
  • [25] Sergey Yekhanin. Locally decodable codes. Foundations and Trends® in Theoretical Computer Science, 6(3):139–255, 2012. doi:10.1561/0400000030.

Appendix A Proof of Lemma 11

We include the proof of completeness.

Proof.
|𝔼[ωX,Y]| =|ymnY(y)xmnX(x)ωx,y|
ymnY(y)|xmnX(x)ωx,y|.

On the other hand, since the square function is convex, by Jensen’s inequality, one have

|𝔼[ωX,Y]|2 ymnY(y)|xmnX(x)ωx,y|2
=|ymnY(y)x1,x2mnX(x1)X(x2)ωx1,yωx2,y|
=|ymnY(y)x1,x2mnX(x1)X(x2)ωx1x2,y|
=|𝔼[ωX1X2,Y]|.