Polynomials, Divided Differences, and Codes
Abstract
Multiplicity codes (Kopparty et al., J. ACM 2014) are multivariate polynomial codes where the codewords are described by evaluations of polynomials (with a degree bound) and their derivatives up to some order (the multiplicity parameter), on a suitably chosen affine set of points. While efficient decoding algorithms were known in some special cases of point sets, by a reduction to univariate multiplicity codes, a general algorithm for list decoding up to the distance of the code when the point set is an arbitrary finite grid, was obtained only recently (Bhandari et al., IEEE TIT 2023). This required the characteristic of the field to be zero or larger than the degree bound, which is somewhat necessary since list decoding up to distance with small output list size is not possible when the characteristic is significantly smaller than the degree.
In this work, we present an alternative construction based on divided differences of polynomials, that conceptually resembles the classical multiplicity codes but has good list decodability “insensitive to the field characteristic”. We obtain a simple algorithm that list decodes this code up to distance for arbitrary finite grids over all finite fields. Our construction can also be interpreted as a folded Reed-Muller code, which may be of independent interest.
Keywords and phrases:
Error-correcting code, polynomial code, Reed-Solomon code, Reed-Muller code, folded Reed-Solomon code, folded Reed-Muller code, multiplicity code, divided difference, q-derivative, polynomial method, list decoding, list decoding capacity, linear algebraic list decodingFunding:
S. Venkitesh: Len Blavatnik and the Blavatnik Family Foundation.2012 ACM Subject Classification:
Mathematics of computing Coding theory ; Theory of computation Error-correcting codesAcknowledgements:
The author thanks Dean Doron, Mrinal Kumar, Noga Ron-Zewi, Amnon Ta-Shma, and Mary Wootters for patiently listening to and giving feedback on presentations of this work at different times. A part of this work was done while the author was a postdoc at the Department of Computer Science, University of Haifa, and was participating in the program on Error Correcting Codes and Computation (2024) at the Simons Institute for the Theory of Computing, UC Berkeley.Funding:
A part of this work was done while the author was supported in part by BSF grant 2021683, ISF grant 735/20, and by the European Union (ERC, ECCC, 101076663).Editors:
Raghu MekaSeries and Publisher:

1 Introduction and overview
Codes based on polynomial evaluations have found widespread applications, both implicitly and explicitly, in theoretical computer science, combinatorics, and allied areas. The prototypical univariate polynomial code is the Reed-Solomon (RS) code [57, 58], wherein the codewords are evaluations of degree bounded univariate polynomials on a set of points, and the prototypical multivariate polynomial code is the analogously defined Reed-Muller (RM) code [52, 57, 39, 68, 14]. 111The RM code was considered over the binary field by [52, 57], and over larger fields by [39, 68, 14]. Despite several decades of research, the decodability properties of these codes are far from being well-understood.
1.1 The list decoding problem
In this work, we are specifically interested in the list decoding problem [16, 69], where the aim is to output a list of close codewords for any given received word efficiently. The main parameter of interest is the list decoding radius, which represents the fraction of disagreements that we can decode from, and we would like it to be as large as possible.
Unique decoding of polynomial codes
To begin with, consider the easier setting where we fix the list decoding radius to be half the minimum distance of the code, which is precisely the setting of the unique decoding problem – in this case, there could only be at most one codeword this close to any given received word, and the aim is to efficiently find it if it exists. This problem has now been solved completely for RS codes [56, 49, 67], for univariate multiplicity codes [53], for RM codes in a special case [40, 5], and quite recently for RM codes in general [41] and for multivariate multiplicity codes [8]. What is noteworthy is that all of these algorithms are insensitive to the field characteristic. We now move on to consider larger list decoding radius.
List decoding of univariate polynomial codes
In the last decade of the 20th century, two classic works [65, 27] showed that all RS codes can be list decoded up to the Johnson bound (which is relative radius , where is the rate of the code), and also laid an algebraic algorithmic framework that has since been exploited over and over again for several other polynomial codes. Further, we do not know of explicit RS codes that can be list decoded beyond the Johnson bound, but we do know of explicit RS codes that are not list decodable significantly beyond the Johnson bound [7]. Due to recent breakthroughs [11, 26, 2], we now know that random RS codes are list decodable up to (information theoretic) capacity (that is, relative radius for arbitrarily small but fixed , which represents getting approximately close to the information theoretic limit of list decoding), but we neither know of an explicit construction nor know of an algorithm to achieve this.
It is in this context of list decoding up to the information theoretic capacity that a folding trick has been enlightening. The works of [28, 29] showed that folded Reed-Solomon (FRS) codes (where the bits of an RS codeword are bunched into blocks of large constant size), and univariate multiplicity codes [43] (wherein the codewords are evaluations of polynomials and their derivatives up to a large constant multiplicity) can be algorithmically list decoded up to capacity! Further improvements to the parameters have been achieved in [44, 61, 63, 12] in different settings.
List decoding of multivariate polynomial codes
Moving on to multivariate polynomial codes, while the list decoding algorithms of [65, 27] generalize easily to RM codes, the decoding radius now worsens as a function of the number of variables, and attaining Johnson bound is no longer possible with these algorithms. While we know of a Johnson bound attaining algorithm [55] for RM codes when the finite grid is the full vector space over the base field , designing a more general algorithm for arbitrary finite grids is still open!
As it turns out, the folding trick helps in the multivariate setting too. The multivariate multiplicity codes [45], which are the obviously defined multivariate analogues of univariate multiplicity codes, were shown to be list decodable up to distance (that is, up to radius , where is the minimum distance of the code) by [43] when the set of evaluation points is the full vector space . However, just as in the case of RM codes, an analogous result for arbitrary finite grids evaded us for quite some time. Very recently, [9] finally bolstered the algebraic algorithmic framework with some simple additional ingredients and managed to design an algorithm for list decoding multivariate multiplicity codes over arbitrary finite grids up to distance.
Sensitivity of list decodability to the field characteristic
Finally, yet another crucial component that could influence the performance of the code (specifically a polynomial code) is the most fundamental underlying algebraic object – the base field over which the polynomials are considered, which is always a finite field in our discussion.
Unlike the unique decoding setting that we mentioned earlier, we do observe sensitivity to the field characteristic in the setting of the list decoding problem. Let us fix the notation that denotes the base field, and denotes the characteristic of . On one hand, we know that when is a large power of , there are RS codes that cannot achieve anywhere close to list decoding capacity [7] since we can show instances of received words having superpolynomial sized output lists. On the other hand, upon large constant folding, the univariate multiplicity codes achieve capacity when is larger than the degree parameter [29, 43], or smaller but linear in the degree [44], and the FRS codes achieve list decoding capacity [28, 29] insensitive to . Thus, the list decodability of the FRS code is insensitive to the field characteristic, whereas the list decodability of the multiplicity code is sensitive to the field characteristic. Indeed, in the case when is a large power of , the large output lists shown by [7] for RS codes can be used to construct large output lists for multiplicity codes, and so the list decodability is provably poor.
Things get even more interesting in the multivariate setting, since the multivariate multiplicity codes achieve distance when is larger than the degree parameter [55, 9], but we do not know of a field characteristic insensitive construction of a multivariate polynomial code that algorithmically achieves distance. Furthermore, we also do not know of an appropriate multivariate analogue of the field characteristic insensitive univariate FRS code. In this work, we answer both questions with a single code construction, along with a list decoding algorithm. In particular, our code subsumes the list decoding performance of the field characteristic sensitive multivariate multiplicity code [9], and also provides a natural folded Reed-Muller code construction. The motivation for our construction comes from an extremely simple yet foundational observation within the univariate world itself – the FRS code is also a multiplicity code!
1.2 FRS codes, univariate multiplicity codes, and divided differences
Fix a field . Consider distinct nonzero points . Let be a multiplicative generator, and suppose such that the points are all distinct. For , the degree- Folded Reed-Solomon (FRS) code is defined by
Here, the distance function is the Hamming distance on alphabet , that is, the Hamming weight 222We are only concerned with codes that are linear over the base field , and so the Hamming distance between two codewords is always equal to the Hamming weight of the codeword . of a codeword is the number of such that the block
is a nonzero vector in .
On the other hand, simply assuming are distinct and nothing more about these points, for , the degree- (univariate) multiplicity code is defined by
Here for all . 333Typically, since we are working over finite fields, there will also be a scaling factor multiplied with the derivative, to prevent some pathologies arising in finite fields about higher multiplicity vanishing at a point. With this scaling factor, the derivative is usually called the Hasse derivative. We do not consider this, and implicitly assume that the characteristic is large enough so that these pathologies do not arise. Indeed, this is precisely the setting where the multiplicity codes have good list decodability. Once again, the Hamming weight of a codeword is the number of such that the block is a nonzero vector in .
Let us begin by looking at an algebraic feature that is intrinsic to multiplicity codes. Consider any with . For any , we get the standard Taylor expansion
provided . Is there an analogous Taylor expansion in the context of the FRS code? As it turns out, there is such an expansion, and it is even simpler – it is the expansion in terms of the Newton forward differences. Precisely, we have
where
Notably, since is a multiplicative generator, we always have , and therefore, this expansion is valid unconditional on .
In the first decade of the 20th century, Jackson [31, 32, 33] (also see [35, 34]) had defined the following specific divided difference operator on the polynomial ring,
In terms of this operator, the above expansion takes the form
where , and . Returning to the code , it is then immediate by the definition of that there are invertible lower triangular matrices such that for any codeword , we have
In other words, after a specific change of basis (see, for instance, [42, 4] for a proof), the FRS code can be interpreted as a multiplicity code with respect to the derivative-like operator .
Note that we have list decoding algorithms for FRS codes [28, 29], with the first one being algebraic and the second one being linear algebraic. If we are given the -multiplicity encoding, we may just change the basis to the usual FRS encoding and run one of these two FRS decoders. However, what is really encouraging is that we can straightforwardly adapt the linear algebraic multiplicity decoder [29] to our -multiplicity encoding, and the analysis will have a strong resemblance with that in the case of the classical derivatives. In this work, our main contribution is a construction that pushes this resemblance to the multivariate setting.
Our main algorithm will, in fact, work in the more general paradigm of list recovery, as was the case with the previous algorithms mentioned here.
Assumption on the field
Since our intention is to present the basic ideas as clearly and quickly as possible, we will stick to a simpler setting henceforth, where we assume that the evaluation points are contained in a finite field , but the polynomials are over a degree-3 field extension . (See Remark 4.) We can relax this setting to some extent by tinkering with the parameters involved, but we will not make this effort.
The -derivative
We will now forego the use of the symbol for the multiplicative generator. Instead we will denote to be a multiplicative generator, and henceforth refer to the operator as the -derivative. This is the more standard terminology in the literature where this operator has appeared before. The -derivative 444We use the notation ‘’ in -derivative in this work instead of the more popular ‘’, since we use to denote the field size. is also called the Jackson derivative, and is a special case of the Hahn derivative [30], as well as the Newton forward difference [36, Chapter 1, Section 9]. See [37, Chapter 26] for a discussion on a slightly more general quantum differential, as well as some symmetrized variants. These are all instances of a broader notion of divided difference in finite calculus [46, 10, 54, 64, 36, 59]. The -derivative seems to have been used so far only over fields of characteristic zero, particularly in -combinatorics [18, 3, 21, 60, 20] and quantum calculus [37, 17, 47]. Further, the -derivative does not seem to have made any explicit appearance so far in the polynomial method literature. However, as we see in this work, this notion turns out to be useful over fields of positive characteristic, and indeed, as part of the polynomial method.
1.3 Multivariate -derivatives, multivariate -multiplicity code, and folded RM code
Our extension of -derivatives from the univariate to multivariate setting is natural, and mimicks the extension of the classical derivatives. Assume indeterminates , and for any and , we define the -th -derivative as the iterated operator
where denotes the univariate -derivative in the variable . It is easy to see that the operator is well-defined and invariant under the order of iterations, just as in the case of the classical partial derivatives.
Now consider indeterminates . For any , denote , and further, for any , denote . Also denote , and . Then for any and , there exists (see Proposition 9) an invertible matrix such that
(1) |
Fix any , and consider any nonempty . For any , we define the -variate multiplicity- degree- -multiplicity code by
Note that the distance function is now the Hamming distance on the alphabet . Further, by the change of basis that we just saw above, we can define a natural multivariate analogue of the univariate FRS codes. For any , we define the -folded degree folded Reed-Muller (FRM) code by
So by (1), we clearly have the change of basis
The rate and distance of our code constructions can be given as follows, which will be immediate from Corollary 11 and Theorem 13.
Proposition 1.
For any , the codes and have
1.4 Our result
Our main result is that over any field with sufficiently large , and for sufficiently large constant multiplicity , any constant rate multivariate -multiplicity code , can be efficiently list decoded up to distance.
Theorem 2.
Consider any . For any and for the choices and , the code is efficiently list decodable up to radius with output list contained in a -affine space of dimension atmost .
Our list decoding algorithm is conceptually simpler than that by [9] for the classical multivariate multiplicity codes. Their algorithm in the classical setting entailed recovering the close enough polynomial messages one homogeneous component at a time, while using the classical Euler’s formula for homogeneous polynomials (that relates a homogeneous polynomial to its first order derivatives) 555Euler’s formula states that for a homogeneous degree- polynomial over a field of characteristic greater than , we have . to fully recover a homogeneous component (or a partial derivative of it) from its further first order partial derivatives. This formula, as well as the Taylor expansion for classical derivatives employed within the recovery step, both require the field characteristic to be large. In our -setting, firstly there is no Euler’s formula involving -derivatives, and interestingly we don’t need one! Secondly, the Taylor expansion for -derivatives is clearly field characteristic insensitive. These are the only two places in the analysis where conceptually, the field characteristic was relevant in the classical setting, and is now irrelevant in the -setting. Further, the short and clean analysis that we obtain encourages us to believe that this construction is the correct way to obtain a folded version of the RM code.
Let us give a quick outline of our algorithm. We make use of a clean gluing trick from [9] 666As mentioned in [9], this gluing trick seems to have first appeared in the construction of a hitting set generator in [25]. and then proceed to perform a simpler analysis that is extremely close to the univariate analysis of [29]. The informal takeaway of what we show is that
Suppose we are given the code as in the statement of Theorem 2. Consider any received word
Consider two fresh sets of indeterminates and . For a suitably chosen parameter , we will find a nonzero interpolating polynomial with coefficients in , having the form
such that the following hold.
- (Z1)
-
(Z2)
For any , define a projection
Further, for every , define a -linear operator (on an appropriate subspace of ) that satisfies the condition . The polynomial must now satisfy the vanishing conditions
Note that the vanishing conditions in (Z2) define a linear system on the coefficients of , and we obtain the interpolating polynomial as a nontrivial solution of this linear system, by having large enough degree bounds as in (Z1). So for any close enough polynomial (that is, having a large number of agreements with ), the conditions in (Z2) will imply that vanishes at the agreement points with multiplicity at least (in the sense of multivariate -derivatives, which we will define later), and this will imply that by a suitable version of Polynomial Identity Lemma for multivariate -derivatives. We can then solve this equation to recover . Importantly, the list decoding radius is dependent on the choice of . We can achieve list decoding up to distance, that is, the claim of Theorem 2 by choosing and . In fact, as in the analyses of [29, 9], we will show that the collection of all close enough polynomials is contained in a -affine subspace having dimension at most .
This is a fairly simple strategy, very much within the algebraic algorithmic framework started employed in previous list decoding algorithms [65, 27, 28, 29, 43, 9]. Further, employing the gluing trick with the additional -variables as in [9] allows us to formulate what we believe is the correct way to extend the linear algebraic decoding framework from the univariate to the multivariate setting, and our extension turns out to be simpler than the analysis in [9].
There is a subtlety in the recovery step in our multivariate setting vis-à-vis the univariate setting. When we solve the equation to recover a close enough polynomial , we will see that the coefficients of will be captured within the coefficients of some linear relations between monomials in the -variables. Using a large enough efficient hitting set (an interpolating set of affine points that is large enough to certify all polynomials with a degree bound) in the -variables, we can then recover each of these coefficients. This is similar to what was done in [9] in an analogous situation, but our overall analysis is a bit different and simpler. We will outline this difference now.
Formally, for a vector space of all polynomials having degree at most , a hitting set is a finite set such that for any , if then there exists such that . For instance, by the Combinatorial Nullstellensatz [1, Theorem 1.1], every finite grid with is a hitting set for . (We will precisely use this obvious hitting set in our analysis.) Our departure from the analysis of [9] is that they need to recover the coefficients of one homogeneous component at a time, by employing the Euler’s formula. In our analysis, we will directly recover one coefficient at a time (with respect to a suitable choice of basis), and this is extremely similar to the corresponding step in the linear algebraic list decoding algorithm [29] for classical univariate multiplicity codes. Recall that we intend to obtain an affine subspace of small dimension that contains all close enough polynomials. Assume the form for all target polynomials in the affine subspace. We will recover the target polynomials inductively from lower coefficients to higher coefficients. In order to kick-start the recovery procedure, as a base case of the induction, we set the coefficients arbitrarily. Then for any , the equation would imply an equation that has the form
where the polynomials comprising the above -linear combinations all have a good -degree bound due to the -degree bound on . Note that each of the coefficients in the R.H.S. above is strictly below (in the usual componentwise partial order) some coefficient in the L.H.S. above. Since by induction hypothesis, we have already recovered all the coefficients in the R.H.S, we have the -linear combination comprising the L.H.S. above also determined. We then instantiate a hitting set in the -variables to recover the individual coefficients in the L.H.S. above, in much the same way as [9] instantiate for their linear combinations.
Thus, the coefficients are the only possible free coefficients in the above procedure, and therefore, the output list is contained in an affine -subspace having dimension at most . We can further adapt the off-the-shelf pruning algorithm by [44] and its improved analysis by [66] to the multivariate setting, nearly identically as in [9], to finally end up with a randomized algorithm that guarantees constant output list size. Very recently, two concurrent works [63, 12] improved the output list size guarantee for list decoding FRS codes and univariate multiplicity codes from exponential in to and the optimal respectively. It is possible to extend their argument in the obvious way in the multivariate setting to give an even smaller output list size guarantee in the case of list decoding multivariate -multiplicity codes. We do not mention the details in this presentation.
We will, in fact, prove a more general list recovery result as in Theorem 19, but stick to only showing polynomial output list size guarantee, for simplicity.
1.5 Further questions
Multivariate multiplicity codes evaluated on the vector space are known to have good locality properties [45], especially when the field characteristic is larger than the degree [43, 44]. It will be interesting to see if similar locality properties hold for the multivariate -multiplicity code in a field characteristic insensitive sense.
2 Preliminaries
List decoding and list recovery
Let be a finite alphabet. An input list is any tuple of subsets . Further, for any , the (relative) Hamming distance between and is defined by
We say an input list is -sized if for all .
Let be a field, and consider an -linear code , that is, is an -linear space.777The more conventional definition defines to be a linear code if is linear over , but our definition is more relaxed where we allow to be an -linear space. Our relaxed definition holds for our codes of interest, whereas the conventional definition does not hold. Indeed, we heavily use the relaxed linearity properties of our code constructions, often implicitly. For and , we say is -list recoverable (LR) if there does not exist any -sized sequence of input lists that admits distinct codewords such that
The case corresponds to list decoding with identical definitions.
Field extensions
Fix a finite field , and a finite degree extension having extension degree , that is, . Also denote . Fix a multiplicative generator . We immediately note the following basic observations (see, for instance, [48, Chapter 2]), and also add quick proofs for completeness.
Proof.
-
(a)
Since and , we have . Since , we have .
-
(b)
Since is the multiplicative generator of , and , the multiplicative order of is . Obviously, for all . Now suppose for some . Then , which means , that is, .
Remark 4.
Note that in this work, we will assume that the field size . We will also consider an additional parameter , and always work in the case where . By Proposition 3(b), this is true if . Further, we will have the degree of all polynomials that we consider to satisfy . Therefore, we will also need . Both these requirements are satisfied if we take and , and we assume these throughout the rest of this work. In fact, we will only take to be a constant relative to . So henceforth, we have , and is a multiplicative generator of .
For any , denote
Note that for all , and . The quantity is called the Gaussian binomial coefficient [22]. We will also have the convention that if , and if .
2.1 Univariate -derivatives
For any , the -derivative [31, 32, 33, 34, 35] is defined by
Further, we are interested in iterated applications of the operator , and so we denote , and for all .
Remark 5.
We have for all . Consider any nonconstant and . Then we immediately get by linearity of (see Proposition 6(a) below). More importantly, as a departure from classical derivatives, we get if .
Let us quickly collect a few basic properties of -derivatives. For an indeterminate and , denote . For any and , denote .
Proposition 6 ([37, 17]).
-
(a)
Linearity. is an -linear map on , for all .
-
(b)
Scaling. For any and ,
-
(c)
Taylor expansion. For any ,
In particular,
-
For any , we have .
-
For any , we have , where is the -th Hasse derivative888For , the Hasse derivatives satisfy: . of at .
-
-
(d)
Product rule. For any and ,
Change of basis
A simple change of basis shows that -derivatives are built out of simple evaluations at correlated points. Define an infinite matrix by
Clearly, whenever , that is, lower triangular. Further, we have the diagonal entries for all . So is invertible, and is lower triangular, given by
This can be proven by instantiating, for instance, [62, Theorem 1.2.3].
Proposition 7 (Change of basis for -derivatives).
For any and , we have
Proof.
2.2 Multivariate -derivatives
Now assume indeterminates , and for any subset , denote . For every , denote the -linear -derivative map on by , and note that it extends in an obvious (and unique) way to a -linear map on . It is also immediate that commute as -linear maps on . Then for any and , we define the -th -derivative by
For , denote
Also, note the usual binomial coefficient . For indeterminates and , denote . For any and , denote , and . Also denote . For any , denote .
We easily get the following basic properties of multivariate -derivatives.
Proposition 8.
-
(a)
Linearity. is a -linear map on , for all .
-
(b)
Scaling. For any and , we have
-
(c)
Taylor expansion. For any ,
In particular,
-
For any , we have .
-
For any , we have , where is the -th Hasse derivative999For , the Hasse derivatives satisfy: . of at .
-
-
(d)
Product rule. For any and ,
Proof.
Each item follows easily by induction on , with the base cases for Item (a), (b), (c), and (d) being Proposition 6(a), (b), (c), and (d) respectively.
Change of basis
Using the infinite matrices defined in Section 2.1, define two infinite matrices by
It is then easy to see (for instance, by induction on ) that is invertible and . Also, for all , and as well as for all . Further, we obtain the following change of basis, again by induction on .
Proposition 9 (Change of basis for multivariate -derivatives).
For any and , we have
A monomial basis for function spaces
We assume familiarity with Gröbner basis theory. For definitions and more details, see [13, Chapters 2–5]. For any , consider the ideal defined (by appealing to the change of basis in Proposition 9) as
The following is a characterization of a Gröbner basis of that is universal (invariant under choice of monomial order) and reduced (minimal with respect to divisibility of leading terms).101010Given any monomial order, a universal Gröbner basis and a reduced Gröbner basis for an ideal both always exist; however, a Gröbner basis that is both universal and reduced may not always exist. A universal reduced Gröbner basis, whenever it exists, is the closest to being the Gröbner basis for an ideal, and allows performing Euclidean division carelessly. This is an extension of [1, Theorem 1.1] and [6, Theorem 4.1], while being a special case of [23, Theorem 4.7].111111A related characterization of standard monomials, which is a strictly weaker notion than Gröbner basis, was obtained much earlier by [19, 50] via determination of winning strategies of a lex game, and a further reinterpretation of these winning strategies using the notion of compression for set systems. This reinterpretation was also rediscovered independently by [51]. A similar characterization in the case of classical multivariate derivatives was given in [43, Section 6], and can also be obtained by [23, Theorem 4.7].
Theorem 10 ([23]).
For any , the set of polynomials
is a universal reduced Gröbner basis for .
Further, for any , define the function spaces
As an immediate corollary of Theorem 10 and the simple properties of Euclidean division for polynomials, we can describe a monomial basis (a basis where each vector is the appropriate evaluation of a monomial) for the above vector spaces.
Corollary 11.
The collections of evaluation vectors
are -linear bases of and respectively, where
2.3 Counting -roots of polynomials with multiplicities
The discussion so far naturally leads us to the question of bounding the number of -roots of a polynomial with multiplicities.
For any nonzero univariate and , define the -multiplicity of at by . The following claim is easy, by appealing to univariate factorization.
Proposition 12 (Easy).
For any nonempty set , and nonzero with , we have . In particular, for any , we have .
In the multivariate setting, for any nonzero and , define the -multiplicity of at by
In this section, we will prove the following multivariate extension of Proposition 12, which is analogous to [15, Lemma 2.7] in the setting of classical derivatives.
Theorem 13.
For any nonempty set , and any nonzero with , we have
In particular, for any , we have .
Towards a proof of Theorem 13, we first consider some simple consequences of the definition of multivariate -multiplicity.
Lemma 14.
Consider any nonzero .
-
(a)
For any , if , then
-
(b)
For any and , if , then
Proof.
-
(a)
Since , we have
Consider any . So , and therefore . This means .
-
(b)
Suppose . This means
In particular, restricting our attention to of the form , we get
This means .
We can now prove Theorem 13.
Proof of Theorem 13.
We will prove by induction on . The base case is true by Proposition 12. Now suppose the assertion is true for some , and now consider indeterminates . Without loss of generality, assume , and write
For any , denote . So by induction hypothesis, we have
(2) |
Now fix , and let such that and
. This means
-
, and so
-
, and .
Then, for any , we get
by Lemma 14(a) | ||||
by Lemma 14(b) |
So by Proposition 12,
Then (2) implies
which completes the induction.
3 List decoding/recovery of -multiplicity codes
We will now move to our main result on list decoding multivariate -multiplicity codes. In fact, we will present the algorithm for the more general list recovery paradigm.
3.1 List decoding/recovery of univariate -multiplicity codes
Let us see that the classical univariate multiplicity code list decoder [29] can be suitably adapted to list decode , which therefore provides an alternative algorithm to list decode FRS codes. We will, in fact, consider list recovery.
Theorem 15.
Consider any . For any and , and for the choices and , the code is efficiently list recoverable up to radius with output list size at most .
Consider additional indeterminates , and the -linear subspace of the polynomial ring defined by
Also, denote . Define a -linear operator by
We will be interested in iterated applications of , and therefore, denote and for all . For any and , denote
. The following is immediate by the definition of .
Lemma 16.
Let .
-
(a)
If such that for all , then for all .
-
(b)
, for all .
Proof.
Let .
-
(a)
If such that for all , this means . Now we have
So, for all . This means for all .
-
(b)
We have
by Proposition 6(c)
It is worth noting that a close variant of the operator appears in [24, Definition 8.1] in the context of nearly linear-time list decoding of FRS codes, without the interpretation in terms of -derivatives. So it is reasonable to surmise that a nearly linear-time implementation of the list decoding algorithm à la [24] is possible for the univariate -multiplicity codes. In order to keep our main presentation short, we limit ourselves here to adapting the more conventional polynomial-time algorithm of [29].
The interpolation step of the list recovery algorithm, which will be used multiple times in this discussion, can be captured as follows.
Proposition 17.
Consider any , parameters , and . Let , and define
Then for any with , there exists a nonzero polynomial with
such that for any ,
Proof.
Assume the notation . We can, in fact, add additional arbitrary elements and assume , for all . Also enumerate . We will construct a nonzero polynomial such that
-
(a)
, that is, .
-
(b)
, and for all .
-
(c)
for all .
The number of linear constraints is , and the number of coefficients is . So for the choice
we indeed have , and this ensures a nontrivial solution, that is, .
Now consider any , and suppose . By Lemma 16, this implies . But we also have . Therefore, by Proposition 12, we can conclude that if . This is equivalent to
The above is true if
Once we have an interpolating polynomial that captures the correct codewords, we can extract the output list as follows.
Proposition 18.
Consider parameters , and such that . If is a nonzero polynomial with
then the solution space
is an affine -linear space with dimension at most .
Proof.
Consider any satisfying . Note that we have . If for all , then , which means , a contradiction. So there exists such that . Without loss of generality, we assume . (Otherwise, we work with the largest such that .)
Let us also quickly consider another notation. For any , and , denote
by Proposition 6(b) |
Recall that . Since , there exists such that for all . Further, since , we will work with the basis of monomials . Now we have
So by Proposition 6(c), for every , we have
Now for every , since , we can rewrite the above
So we can set the undetermined coefficients among freely, and the remaining coefficients of are uniquely determined the above equation. This means the solution space is an affine -linear space having dimension at most .
The final list recovery result can now be proven.
Proof of Theorem 15.
Choosing , and plugging in all the parameters into Proposition 17 and Proposition 18 immediately implies that the output list is contained in an affine -linear space with dimension at most . The pruning algorithm with the improved analysis [44, 66] then asserts that the output list size is at most .
3.2 List decoding/recovery of multivariate -multiplicity codes
We will now see an algorithm to list decode . Even though it is inspired by the classical multivariate multiplicity code list decoder [9], it turns out that we can give a simpler algorithm and analysis that is perhaps the correct multivariate extension of our list decoder in the univariate case. We will, in fact, perform list recovery, and the main theorem statement is the following.
Theorem 19.
Consider any . For any and , and for the choices and , the code is efficiently list recoverable up to radius with output list contained in a -affine space of dimension atmost .
Consider additional indeterminates , and the -linear subspace of the polynomial ring defined by
Also, denote for all . For any , define a -linear operator by
For any and , denote
The following is then immediate.
Lemma 20.
Let .
-
(a)
If such that for all , then for all .
-
(b)
, for all .
Remark 21.
A consequence of Lemma 20 is that for all . We do not explicitly use this elsewhere.
Proof of Lemma 20.
Let .
-
(a)
If such that for all , this means for all . Now we have
So, for all . This means for all .
-
(b)
We have
by Proposition 8(d)
The interpolation step of the list recovery algorithm, which will be used multiple times in this discussion, can be captured as follows.
Proposition 22.
Consider any , parameters , and . Let , and define
Then for any with , there exists a nonzero polynomial with
such that for any ,
Moreover, we can ensure that is a polynomial in with .
Proof.
Assume the notation . We can, in fact, add additional arbitrary elements and assume , for all . Also enumerate . We will construct a nonzero polynomial such that
-
(a)
, that is, .
-
(b)
, and for all .
-
(c)
for all with .
The number of linear constraints is
and the number of coefficients is
So for the choice
we indeed get a nontrivial solution, that is, . Further, since each linear constraint is a polynomial in having degree at most , we can ensure that is a polynomial in having (see [38] or [9, Lemma 3].)
Now consider any , and suppose . By Lemma 20, this implies . But we also have . Therefore, by Theorem 13, we can conclude that if , which holds if
The above is true if
Once we have an interpolating polynomial that captures the correct codewords, we can extract the output list as follows.
Proposition 23.
Consider parameters , and such that . If is a nonzero polynomial with
then the solution space
is an affine -linear space with dimension at most .
Proof.
Consider any satisfying . Note that we have . If for all , then , which means , a contradiction. So there exists such that . Without loss of generality, we assume . (Otherwise, we work with the largest such that .)
Let us also quickly consider another notation. For any , and , denote
by Proposition 8(b) |
Recall that . Since , there exists such that for all . Further, since , we will work with the basis of monomials . Now we have
So for every , we have
Now for every , since , we can rewrite the above as
(3) |
So we can set the undetermined coefficients among freely, and the remaining coefficients of are uniquely determined the above equation. This means the solution space is an affine -linear space having dimension at most . Further, at any instance of (3), if we know the R.H.S. and determine the L.H.S., then each coefficient can be recovered by instantiating an arbitrary finite grid with as a hitting set for the -variables, since we have .
We are now ready to complete the proof of our main theorem.
Proof of Theorem 19.
Choosing , and plugging in all the parameters into Proposition 17 and Proposition 18 immediately proves the claim.
References
- [1] Noga Alon. Combinatorial Nullstellensatz. Combinatorics, Probability and Computing, 8(1-2):7–29, 1999. doi:10.1017/S0963548398003411.
- [2] Omar Alrabiah, Venkatesan Guruswami, and Ray Li. Randomly Punctured Reed–Solomon Codes Achieve List-Decoding Capacity over Linear-Sized Fields. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, STOC 2024, pages 1458–1469, New York, NY, USA, 2024. Association for Computing Machinery. doi:10.1145/3618260.3649634.
- [3] G.E. Andrews. -Series: Their Development and Application in Analysis, Number Theory, Combinatorics, Physics and Computer Algebra, volume 66 of CBMS Regional Conference Lecture Series in Mathematics. AMS, 1986.
- [4] M.H. Annaby and Z.S. Mansour. -Taylor and interpolation series for Jackson -difference operators. Journal of Mathematical Analysis and Applications, 344(1):472–483, 2008. doi:10.1016/j.jmaa.2008.02.033.
- [5] László Babai, Lance Fortnow, Leonid A. Levin, and Mario Szegedy. Checking computations in polylogarithmic time. In Proceedings of the Twenty-Third Annual ACM Symposium on Theory of Computing, STOC ’91, pages 21–32, New York, NY, USA, 1991. Association for Computing Machinery. doi:10.1145/103418.103428.
- [6] Simeon Ball and Oriol Serra. Punctured combinatorial nullstellensätze. Combinatorica, 29(5):511–522, 2009. doi:10.1007/s00493-009-2509-z.
- [7] Eli Ben-Sasson, Swastik Kopparty, and Jaikumar Radhakrishnan. Subspace Polynomials and Limits to List Decoding of Reed–Solomon Codes. IEEE Transactions on Information Theory, 56(1):113–120, 2010. doi:10.1109/TIT.2009.2034780.
- [8] Siddharth Bhandari, Prahladh Harsha, Mrinal Kumar, and Ashutosh Shankar. Algorithmizing the Multiplicity Schwartz-Zippel Lemma. In Nikhil Bansal and Viswanath Nagarajan, editors, Proceedings of the 2023 ACM-SIAM Symposium on Discrete Algorithms, SODA 2023, Florence, Italy, January 22-25, 2023, pages 2816–2835. SIAM, 2023. doi:10.1137/1.9781611977554.ch106.
- [9] Siddharth Bhandari, Prahladh Harsha, Mrinal Kumar, and Madhu Sudan. Decoding Multivariate Multiplicity Codes on Product Sets. IEEE Transactions on Information Theory, 70(1):154–169, 2024. doi:10.1109/TIT.2023.3306849.
- [10] George Boole. A Treatise on the Calculus of Finite Differences. MacMillan and Co., 1860. Reprint: Cambridge University Press, 2009. doi:10.1017/CBO9780511693014.
- [11] Joshua Brakensiek, Sivakanth Gopi, and Visu Makam. Generic Reed-Solomon Codes Achieve List-Decoding Capacity. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, pages 1488–1501, New York, NY, USA, 2023. Association for Computing Machinery. doi:10.1145/3564246.3585128.
- [12] Yeyuan Chen and Zihan Zhang. Explicit Folded Reed-Solomon and Multiplicity Codes Achieve Relaxed Generalized Singleton Bound. arXiv Preprint, 2024. doi:10.48550/arXiv.2408.15925.
- [13] David Cox, John Little, and Donald O’Shea. Ideals, varieties, and algorithms: an introduction to computational algebraic geometry and commutative algebra. Springer Science & Business Media, 2013.
- [14] P. Delsarte, J.M. Goethals, and F.J. Mac Williams. On generalized Reed Muller codes and their relatives. Information and Control, 16(5):403–442, 1970. doi:10.1016/S0019-9958(70)90214-7.
- [15] Zeev Dvir, Swastik Kopparty, Shubhangi Saraf, and Madhu Sudan. Extensions to the Method of Multiplicities, with Applications to Kakeya Sets and Mergers. SIAM Journal on Computing, 42(6):2305–2328, 2013. doi:10.1137/100783704.
- [16] Peter Elias. List Decoding for Noisy Channels. Ph.d. thesis, Massachusetts Institute of Technology, 1957. RLE Technical Report No. 335. http://hdl.handle.net/1721.1/4484.
- [17] Thomas Ernst. A Comprehensive Treatment of -Calculus. Birkhäuser Basel, 2012. doi:10.1007/978-3-0348-0431-8.
- [18] H. Exton. -Hypergeometric Functions and Applications. Halsted Press, Chichister, 1983.
- [19] Bálint Felszeghy, Balázs Ráth, and Lajos Rónyai. The lex game and some applications. Journal of Symbolic Computation, 41(6):663–681, 2006. doi:10.1016/j.jsc.2005.11.003.
- [20] Dominique Foata and Guo-Niu Han. The -Series in Combinatorics: Permutation Statistics (Lecture addendums, revised version). Unpublished, 2021. Available at https://irma.math.unistra.fr/~foata/paper/pub91.pdf.
- [21] George Gasper and Mizan Rahman. Basic Hypergeometric Series. Encyclopedia of Mathematics and its Applications. Cambridge University Press, 2 edition, 2004. doi:10.1017/CBO9780511526251.
- [22] Carl Friedrich Gauß. Summatio quarumdam serierum singularium. Dieterich, 1808. Digitized version available at http://eudml.org/doc/203313.
- [23] O. Geil and U. Martínez-Peñas. Bounding the Number of Common Zeros of Multivariate Polynomials and their Consecutive Derivatives. Combinatorics, Probability and Computing, 28(2):253–279, 2019. doi:10.1017/S0963548318000342.
- [24] Rohan Goyal, Prahladh Harsha, Mrinal Kumar, and Ashutosh Shankar. Fast list-decoding of univariate multiplicity and folded Reed-Solomon codes. arXiv Preprint, 2023. doi:10.48550/arXiv.2311.17841.
- [25] Zeyu Guo, Mrinal Kumar, Ramprasad Saptharishi, and Noam Solomon. Derandomization from algebraic hardness. SIAM Journal on Computing, 51(2):315–335, 2022. doi:10.1137/20M1347395.
- [26] Zeyu Guo and Zihan Zhang. Randomly Punctured Reed-Solomon Codes Achieve the List Decoding Capacity over Polynomial-Size Alphabets. arXiv Preprint, 2023. doi:10.48550/arXiv.2304.01403.
- [27] V. Guruswami and M. Sudan. Improved decoding of Reed-Solomon and algebraic-geometric codes. In Proceedings 39th Annual Symposium on Foundations of Computer Science (Cat. No.98CB36280), pages 28–37, 1998. doi:10.1109/SFCS.1998.743426.
- [28] Venkatesan Guruswami and Atri Rudra. Explicit codes achieving list decoding capacity: Error-correction with optimal redundancy. IEEE Transactions on Information Theory, 54(1):135–150, 2008. doi:10.1109/TIT.2007.911222.
- [29] Venkatesan Guruswami and Carol Wang. Linear-algebraic list decoding for variants of reed–solomon codes. IEEE Transactions on Information Theory, 59(6):3257–3268, 2013. doi:10.1109/TIT.2013.2246813.
- [30] Wolfgang Hahn. Über Orthogonalpolynome, die -Differenzengleichungen genügen. Mathematische Nachrichten, 2(1-2):4–34, 1949. doi:10.1002/mana.19490020103.
- [31] F. H. Jackson. A -form of Taylor’s theorem. Messenger of Mathematics, 38:62–64, 1909.
- [32] F. H. Jackson. A -series corresponding to Taylor’s series. Messenger of Mathematics, 39:26–28, 1909.
- [33] F. H. Jackson. On -Functions and a certain Difference Operator. Transactions of the Royal Society of Edinburgh, 46(2):253–281, 1909. doi:10.1017/S0080456800002751.
- [34] F. H. Jackson. On -Definite Integrals. The Quarterly Journal of Pure and Applied Mathematics, 41:193–203, 1910.
- [35] F. H. Jackson. -Difference Equations. American Journal of Mathematics, 32(4):305–314, 1910. doi:10.2307/2370183.
- [36] Charles Jordan. Calculus of Finite Differences. Chelsea Publishing Company, New York, 2nd edition, 1950.
- [37] Victor Kac and Pokman Cheung. Quantum Calculus. Springer New York, NY, 2002. doi:10.1007/978-1-4613-0071-7.
- [38] R. Kannan. Solving systems of linear equations over polynomials. Theoretical Computer Science, 39:69–88, 1985. Third Conference on Foundations of Software Technology and Theoretical Computer Science. doi:10.1016/0304-3975(85)90131-8.
- [39] T. Kasami, Shu Lin, and W. Peterson. New generalizations of the Reed-Muller codes–I: Primitive codes. IEEE Transactions on Information Theory, 14(2):189–199, 1968. doi:10.1109/TIT.1968.1054127.
- [40] T. Kasami, Shu Lin, and W. Peterson. Polynomial codes. IEEE Transactions on Information Theory, 14(6):807–814, 1968. doi:10.1109/TIT.1968.1054226.
- [41] John Y. Kim and Swastik Kopparty. Decoding Reed-Muller Codes over Product Sets. Theory of Computing, 13(21):1–38, 2017. doi:10.4086/toc.2017.v013a021.
- [42] Wolfram Koepf, Predrag M. Rajković, and Sladjana D. Marinković. Properties of -holonomic functions. Journal of Difference Equations and Applications, 13(7):621–638, 2007. doi:10.1080/10236190701264925.
- [43] Swastik Kopparty. List-Decoding Multiplicity Codes. Theory of Computing, 11(5):149–182, 2015. URL: 10.4086/toc.2015.v011a005, doi:10.4086/TOC.2015.V011A005.
- [44] Swastik Kopparty, Noga Ron-Zewi, Shubhangi Saraf, and Mary Wootters. Improved List Decoding of Folded Reed-Solomon and Multiplicity Codes. SIAM Journal on Computing, 52(3):794–840, 2023. doi:10.1137/20M1370215.
- [45] Swastik Kopparty, Shubhangi Saraf, and Sergey Yekhanin. High-Rate Codes with Sublinear-Time Decoding. J. ACM, 61(5), 2014. doi:10.1145/2629416.
- [46] S.F. Lacroix. Traité des Différences et des Séries. Vol. 3 of Traité du Calcul Differentiel et du Calcul Intégral. Courcier, Paris, 1819.
- [47] D. Larsson and S.D. Silvestrov. Burchnall-Chaundy Theory for -Difference Operators and -Deformed Heisenberg Algebras. Journal of Nonlinear Mathematical Physics, 10(sup2):95–106, 2003. doi:10.2991/jnmp.2003.10.s2.7.
- [48] Rudolf Lidl and Harald Niederreiter. Finite Fields. Cambridge University Press, 2nd edition, 1997.
- [49] J. Massey. Shift-register synthesis and BCH decoding. IEEE Transactions on Information Theory, 15(1):122–127, 1969. doi:10.1109/TIT.1969.1054260.
- [50] Tamás Mészáros. S-extremal set systems and Gröbner bases (Diploma Thesis). PhD thesis, Budapest University of Technology and Economics, 2005. Available at https://sites.google.com/view/tamas-meszaros/.
- [51] Shay Moran and Cyrus Rashtchian. Shattered Sets and the Hilbert Function. In Piotr Faliszewski, Anca Muscholl, and Rolf Niedermeier, editors, 41st International Symposium on Mathematical Foundations of Computer Science (MFCS 2016), volume 58 of Leibniz International Proceedings in Informatics (LIPIcs), pages 70:1–70:14, Dagstuhl, Germany, 2016. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.MFCS.2016.70.
- [52] D.E. Muller. Application of Boolean algebra to switching circuit design and to error detection. Transactions of the I.R.E. Professional Group on Electronic Computers, EC-3(3):6–12, 1954. doi:10.1109/IREPGELC.1954.6499441.
- [53] Rasmus Refslund Nielsen. List decoding of linear block codes. Ph.d. thesis, Technical University of Denmark, 2001. Available at https://orbit.dtu.dk/files/3028444/List%20decoding%20of%20linear%20block%20codes.pdf.
- [54] Niels Erik Nörlund. Vorlesungen über Differenzenrechnung. Springer Berlin, 1924.
- [55] R. Pellikaan and Xin-Wen Wu. List decoding of -ary Reed-Muller codes. IEEE Transactions on Information Theory, 50(4):679–682, 2004. doi:10.1109/TIT.2004.825043.
- [56] W. Peterson. Encoding and error-correction procedures for the Bose-Chaudhuri codes. IRE Transactions on Information Theory, 6(4):459–470, 1960. doi:10.1109/TIT.1960.1057586.
- [57] I. Reed. A class of multiple-error-correcting codes and the decoding scheme. Transactions of the IRE Professional Group on Information Theory, 4(4):38–49, 1954. doi:10.1109/TIT.1954.1057465.
- [58] I. S. Reed and G. Solomon. Polynomial Codes Over Certain Finite Fields. Journal of the Society for Industrial and Applied Mathematics, 8(2):300–304, 1960. doi:10.1137/0108018.
- [59] C.H. Richardson. An Introduction to the Calculus of Finite Differences. D. Van Nostrand Company, Inc., New York, 1954.
- [60] Steven Roman. The Umbral Calculus. Dover Publications, 2005.
- [61] Noga Ron-Zewi, S. Venkitesh, and Mary Wootters. Efficient List Decoding of Polynomial Ideal Codes with Optimal List Size. arXiv Preprint, 2024. doi:10.48550/arXiv.2401.14517.
- [62] Eugene Spiegel and Christopher O’Donnell. Incidence Algebras. CRC Press, 1997.
- [63] Shashank Srivastava. Improved List Size for Folded Reed-Solomon Codes. arXiv Preprint, 2024. doi:10.48550/arXiv.2410.09031.
- [64] J.F. Steffensen. Interpolation. The Williams & Wilkins Company, Baltimore, 1927.
- [65] Madhu Sudan. Decoding of Reed Solomon Codes beyond the Error-Correction Bound. Journal of Complexity, 13(1):180–193, 1997. doi:10.1006/jcom.1997.0439.
- [66] Itzhak Tamo. Tighter list-size bounds for list-decoding and recovery of folded Reed-Solomon and multiplicity codes. IEEE Transactions on Information Theory (Early Access), pages 1–1, 2024. doi:10.1109/TIT.2024.3402171.
- [67] L. Welch. Error Correction for Algebraic Block Codes. IEEE Int. Symp. Inform. Theory, 1983. URL: https://cir.nii.ac.jp/crid/1573668924281711104.
- [68] E. Weldon. New generalizations of the Reed-Muller codes–II: Nonprimitive codes. IEEE Transactions on Information Theory, 14(2):199–205, 1968. doi:10.1109/TIT.1968.1054128.
- [69] J.M. Wozencraft. List Decoding. Quarterly Progress Report, Research Lab of Electronics, MIT, 48:90–95, 1958.