A Simplified Reduction for Error Correcting Matrix Multiplication Algorithms
Abstract
We study the problem of transforming an algorithm for matrix multiplication, whose output has a small fraction of the entries correct into a matrix multiplication algorithm, whose output is fully correct for all inputs. In this work, we provide a new and simple way to transform an average-case algorithm that takes two matrices for a prime , and outputs a matrix that agrees with the matrix product on a fraction of entries on average for a small , into a worst-case algorithm that correctly computes the matrix product for all possible inputs.
Our reduction employs list-decodable codes to transform an average-case algorithm into an algorithm with one-sided error, which are known to admit efficient reductions from the work of Gola, Shinkar, and Singh [12]. Our reduction is more concise and straightforward compared to the recent work of Hirahara and Shimizu [18], and improves the overhead in the running time incurred during the reduction.
Keywords and phrases:
Matrix Multiplication, Reductions, Worst case to average case reductionsCategory:
RANDOMCopyright and License:
2012 ACM Subject Classification:
Theory of computation Problems, reductions and completenessAcknowledgements:
We thank Valentine Kabanets for suggesting to extend the result of [12] to zero error model over arbitrary finite fields. We are also thankful to the anonymous reviewers for their valuable comments.Editors:
Alina Ene and Eshan ChattopadhyaySeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl β Leibniz-Zentrum fΓΌr Informatik
1 Introduction
The problem of efficiently multiplying two matrices is one of the most fundamental problems in computer science. The naive approach for Matrix Multiplication takes time, and there is a lot of ongoing research attempting to improve the running time. Strassen [29] first gave a well-known and widely adopted algorithm, with the running time of . Since then, a long line of work ([24, 5, 26, 25, 28, 8, 27, 31, 22, 3, 9, 32]) has achieved improvement in the exponent, with the current best known algorithm having the runtime of [2]. It is still an open problem to find the optimal running time for Matrix Multiplication.
A natural and relaxed version of the above question is to come up with an algorithm which outputs the product of two matrices, but is allowed to have some incorrect entries. In this work, we address the problem of correcting the output of such an algorithm. Assume that we have an algorithm for matrix multiplication which gets two matrices and outputs a matrix which has a fraction of its entries correct, is it possible to correct the incorrect entries efficiently, i.e., without significantly increasing the running time? For two matrices and , define their agreement, denoted by as
Hence, the problem can be stated as correcting the output of the matrix multiplication algorithm which has some agreement with the correct product. It is perhaps more convenient to view this question as the problem of getting a worst-case to average-case approximate reduction for matrix multiplication. Worst-case to Average-case reductions could help us to get fast algorithms which work in the worst-case, if we could come up with fast average-case algorithms. On the opposite side, such reductions allow us to show that if a problem is known to be worst-case hard, then it is also average-case hard. So our problem could be restated as β Given an algorithm which gets two matrices , and outputs a matrix such that has a non-trivial agreement with in expectation, can we efficiently transform it into an algorithm which outputs the correct matrix product for all possible inputs?
1.1 Previous Work
Worst-Case to Average-case reductions for matrix multiplication have received considerable attention lately. Asadi, Golovnev, Gur, and Shinkar [4] and Hirahara and Shimizu [16] developed Worst-case to Average-case reductions for matrix multiplications. Specifically, they showed that if there exists an algorithm which, on input two matrices over a finite field, returns the correct matrix product on a small fraction of all possible inputs, then it is possible to transform it into an algorithm that computes the matrix product correctly on all the inputs.
Gola, Shinkar, and Singh [12] addressed the related problem of correcting matrix product efficiently, i.e., given an algorithm which on input two matrices over a finite field, outputs a matrix which agrees with the correct matrix product on a non-trivial fraction of entries, can it be reduced to an algorithm that computes the matrix product correctly on all the inputs? Note that this is a weaker assumption than the one used in previous works, since it only assumes a fraction of the output entries to be correct in expectation, whereas the previous works assumed that the output is fully correct for a small fraction of inputs. They showed that over a finite field when the output has a high agreement () with the correct matrix product, a reduction can be obtained using standard self-correction techniques for linear functions [6]. In the same work, using techniques from additive combinatorics, it was shown that over , an optimal reduction can be performed in the case of one-sided error, i.e., when all the entries in the output are correct, and the expected agreement is , for any .
Hirahara and Shimizu [18] improved upon the previous results by providing a reduction from an algorithm which has a non-trivial agreement with the correct matrix product over a finite field (specifically, an agreement of for a field of size ), to a worst-case algorithm. Their key idea was to apply a left-right encoding to the output, using error correcting codes that are encodable and list-decodable in nearly linear time, allowing them to list-decode the left-right encoding efficiently, thereby correcting the matrix product with the help of a matrix-vector product verification algorithm to identify the correct item from the list. Over larger fields, Reed-Solomon Codes satisfied these requirements and allowed for optimal reductions. However, for small fields, they used Ta-Shma codes [30] which are based on random walks over expander graphs, and were shown to be nearly linear time list-decodable by Jeronimo, Srivastava and Tulsiani [21] and Jeronimo [20]. Using efficient list-decoding properties of Ta-Shma codes, a reduction (optimal upto a factor of 2) over small fields was obtained. By using the same codes and modifying the approach through the use of the notion of approximate list-decoding, an optimal reduction, i.e. from an agreement of for a field of size , was obtained in [19].
1.2 Our Contributions
We provide a new worst-case to average-case approximate reduction over finite fields. Over the binary field , we prove the following.
Theorem 1.
Let be sufficiently large, and let be a small constant. Let be an algorithm that gets as input two matrices and outputs a matrix in time such that
Then, there exists an algorithm running in time that gets as input two matrices and outputs a matrix such that for all
Β Remark 2.
Theorem 1 works for all for some absolute constant . The exact limitations on come from Theorem 15. If we construct a list decodable code that supports smaller (say, all ), then we can allow a smaller in the reduction in Theorem 1.
It is an open problem to show a reduction whose running time is .
Note that any trivial algorithm which returns all or all will have an expected agreement of . Hence, this result is optimal in terms of agreement, since it allows us to reduce any non-trivial algorithm doing slightly better than a trivial algorithm. Similarly, over a finite field of prime order , a worst-case to average-case reduction for an expected agreement of would be optimal.
Theorem 3.
Let be sufficiently large, and let be a small constant. Let be an algorithm that gets as input two matrices and outputs a matrix in time such that
Then, there exists an algorithm running in time that gets as input two matrices and outputs a matrix such that for all
The overhead in our reduction for a finite field of order is exponential in . This improves upon the overhead of the reduction given by [18] over small fields, which is double-exponential in .
We believe that our reduction is conceptually simpler as compared to the reduction of [18]. For example, while both reductions use list decodable codes, in this paper we need the list decodable codes to have rather standard properties (which we construct in the Appendix), while [18] requires explicit expansion properties like splittablity of a collection of -tuples. Our construction of the list-decodable code relies on simple code concatenation techniques, and does not require Walk-Amplified codes based on direct sum encodings of a collection of -tuples obtained by taking random walks over expander graphs of large girth. In addition, the code we construct has a full rank partition (Definition 6), which is a more natural notion than the -partition defined in [18]. Furthermore, we do not require the notions of left-right encoding and approximate list-decoding for obtaining optimal reductions.
Finally, in the process of obtaining an optimal reduction, we also provide a way to generalize the one-sided error reduction over the binary field [12] to any finite field of prime order.
2 Preliminaries
We begin by recalling some basic coding theory terminology. A code over the Finite Field is a subset . If is a linear subspace of , it is called a linear code. When the field is , we call it a binary code. Elements of the code are referred to as codewords. A code can be equivalently described by an encoding, which is an injective mapping , whose range gives the code . For a linear code, the encoding can be expressed as , where is called a generator matrix. The rate of the code is given by , or equivalently as . The distance of the code is defined as , where . 111Strictly speaking, it is called the relative distance of the code. However, we shall simply refer to it as distance in this work. We refer the reader to [15] for further details on this subject.
A well-known linear code which we use in this work is the Reed-Solomon code.
Definition 4 (Reed-Solomon codes).
For and a prime , a Reed-Solomon Encoding over a finite field with the message and the set of evaluation points , denoted by , is given by , where .
We now define the notion of list-decoding.
Definition 5 (List-decodable codes).
A code given by the encoding is -list-decodable if for all , . A list-decoding algorithm is an algorithm which takes as input and with high probability, outputs all the vectors such that .
Next we define the notion of full-rank partition. This notion is a special case of the notion of full-rank partition used in [16]. Our definition corresponds to and , which (in our opinion) is the most natural choice of parameters.
Definition 6 (Full-rank partition).
Let be integers such that for some . We say that a matrix admits a full-rank partition if there is a partition of such that:
-
for all .
-
For all , the rows indexed by are linearly independent. Equivalently, the submatrix is full rank.
We proceed to mention the linear list-decodable code which we crucially use in our reduction.
Lemma 7.
Fix and let be sufficiently large, such that for some absolute constant . There exists a randomized algorithm that runs in time and with probability at least outputs an advice , such that given we have the following.
There exists a linear code , where is an integer, and it satisfies the following properties.
-
1.
is encodable in time .
-
2.
is -list decodable for , and the running time of the decoder is .
-
3.
The generating matrix for has a full-rank partition. Furthermore, there exists an algorithm that outputs the generating matrix and the partition in time .
We are certain that such codes are already known (see, e.g., [13]), however we could not find the exact statement in the literature. In particular, the notion of full-rank partition is a standard notion studied in the literature. We refer the reader to Appendix A for the proof of Lemma 7.
We now state a variant of Markovβs inequality which we use repeatedly in this work.
Lemma 8 (Markovβs Inequality).
Let be a random variable taking values in the interval that has non-zero mean . Then, for any ,
Proof.
Let . Then,
This gives us that .
We now state the verification algorithms we use in the reduction. Firstly, the well-known Freivaldβs Randomized Algorithm for verifying matrix multiplication. Secondly, the data structure algorithm given by Hirahara and Shimizu [16] for verifying matrix-vector multiplication.
Lemma 9 (Freivaldβs Algorithm [11]).
There exists a randomized algorithm with runtime which, when given as input three matrices , outputs either YES or NO such that:
-
.
-
.
Lemma 10 (Lemma 6.2 from [17]).
There exists a data structure algorithm such that:
-
is a randomized algorithm which gets as input , and outputs a string in time.
-
is a deterministic oracle algorithm which, when given oracle access to and inputs , verifies whether in time, and is correct with probability over the internal coin tosses of .
Finally, we mention the one-sided error result from [12] which we use in our reduction.
Lemma 11 (Theorem 3 from [12]).
Let be a constant. If there exists an algorithm running in time such that
-
-
If , then .
Then, there exists an algorithm with runtime , such that for all ,
3 Reduction to the one-sided error case
In this section, we present a reduction of matrix multiplication over the binary field to the one-sided error case of Lemma 11, using a binary linear code whose generating matrix has a full-rank partition. Specifically, we prove the following lemma.
Lemma 12.
For some constant , let be a binary linear code which can be encoded in time and is -list decodable within radius in time . Furthermore, assume that the generating matrix of has a full-rank partition, such that the generating matrix and the partition are computable in time .
Suppose there exists an algorithm running in time such that for inputs
Then, there exists an algorithm , such that for inputs it holds that
-
If , then , and
-
,
and the running time of is .
The reduction is outlined in Algorithm 1. We now describe the steps of the reduction.
Let be the generator matrix representing the binary linear code . For a matrix , define , i.e. is a function applying the binary linear code to every row of .
Since has a full-rank partition, its rows can be partitioned into sets , such that for every between and , and the sub matrix is full rank.
Now, for , we construct a matrix such that in expectation. To this end, we note that if is chosen uniformly at random, then the matrix given by will also be a uniformly random matrix, since is full-rank. Our idea is to apply the encoding given by the sub-matrices and then constructing by concatenating the columns of all the resultant matrices. Formally, we write C as:
where denotes the column concatenation. Now, we observe that
where the last inequality follows from the fact that for every , is a uniformly random matrix. In other words, all the columns are generated by taking the product of two uniformly random matrices, for which has an agreement of .
Since , by Markovβs inequality (Lemma 8), there must exist at least fraction of rows of such that their agreement with the corresponding rows of is at least in expectation. Formally, there is a set of rows of size such that for all , where is the row indexed by in and is the row indexed by in .
Now, for each row , by another application of Markovβs inequality (Lemma 8) we can imply that there exists a subset of good inputs of density at least such that if , . Therefore, with probability at least , the distance of the row of from the encoding of the row of the correct product is at most .
Hence, for random choices of A and B, we can list decode these rows in nearly linear time and identify the correct row using an efficient Matrix-Vector product verification algorithm. Furthermore, the verification algorithm (Lemma 10) also lets us identify these good rows.
The Matrix-Vector product verification algorithm gets each vector from the list obtained by decoding the row , and checks whether , where is the vector representing the row indexed by in . If the output is YES, it implies that is the row of the correct matrix product .
Finally, fixing all entries in the undecoded rows by takes us to the one-sided error setting of [12], as stated in Lemma 11. That is,
-
If , then , and
-
For at least fraction of inputs , there are rows such that holds for all .
This implies the following.
Proof.
The first item follows directly from the reduction. For the second item, observe that the reduction guarantees that , since for fraction of rows we have at least fraction of the inputs such that agree with on the βth row.
For item 2, note that if for each we had the βth row of correct for all , then we would have , since is equally likely to be or for any unless the βth row of is all zeros, which happens with probability .
However, by the argument above, for each we have only fraction of the inputs for which the βth row of is correct. By the standard concentration inequality, the fraction of ones in each such row is at least for all except for fraction of the inputs, and thus , as required.
Time Complexity of Algorithm 1
Encoding the rows of can be done in time and partitioning the rows of can be done in time . Steps 3 and 4 can be performed in time. Step 5 takes time. From Lemma 10, step 6 takes time and Steps 7 to 9 can be performed in time . The rest of the steps take time.
This completes the proof of Lemma 12.
4 Worst-Case to Average-Case reduction
We now use Lemma 12 and Lemma 11 to obtain a worst case to average case reduction. In particular, we provide the proof of Theorem 1.
Theorem 1. [Restated, see original statement.]
Let be sufficiently large, and let be a small constant. Let be an algorithm that gets as input two matrices and outputs a matrix in time such that
Then, there exists an algorithm running in time that gets as input two matrices and outputs a matrix such that for all
Proof.
We start by applying the reduction from Lemma 12 with the list-decodable code from Lemma 7. Assuming that the code from Lemma 7 satisfies the required properties, which happens with probability at least , we get an algorithm whose running time is and it satisfies the conditions of Lemma 11 with .
Then, by applying Lemma 11, we get an algorithm whose running time is , and it solves the matrix multiplication problem in worst case with high probability.
Finally, we can run , and verify the resulting product using Freivaldβs Matrix Multiplication Algorithm (Lemma 9). Since the code in Lemma 7 satisfies the required properties with probability at least , we can repeat the procedure times to get the correct output with high probability. Therefore, the total running time of the algorithm we get is , and since , it implies that the total running time is .
5 Reduction for a Finite Field of Prime order
In this section, we show that our reduction can be generalized to any finite field , where is a prime.
This can be achieved by extending the one-sided error model of [12] to a zero-error model over . In the one sided error condition of Lemma 11, the output can be viewed as corresponding to the correct entry, while can be viewed as corresponding to or βDonβt knowβ.
The proof of Lemma 11 relied on the techniques from Additive Combinatorics. The reduction given by [12] for the one-sided error model proceeded by defining the good coordinates, which output with a significant probability on random input matrices. Thereafter, they expressed the input matrices as the sum of random matrices, and used the Probabilistic Bogolyubov Lemma to retrieve all the entries in the correct output. Although they proved the Probabilistic Bogolyobuv lemma for the binary field , their proof can be extended through similar techniques for Fourier analysis over . Hence, the worst-case algorithm in Lemma 14 can be obtained in the same way as the reduction for the one-sided error model, by tweaking the definition of good coordinates to be the coordinates which are not with a significant probability.
Hence, the lemma can be generalized as follows.
Lemma 14.
Let be a constant and be a prime. If there exists an algorithm running in time such that for all
-
-
.
Then, there exists an algorithm running in time such that for all it holds that
Below we sketch the proof of the lemma, mostly referencing the reduction analyzed in [12].
Proof.
The lemma follows by slightly tweaking the reduction for the one-sided error regime, given in [12]. Similar to [12, Definition 12], we define the set of good coordinates to be
Using the algorithm for approximating good coordinates [12, Algorithm 2] and modifying the if-condition in line 12 to take into account the new definition of good coordinates G, we get a zero-error algorithm such that for , [12, Claim 19]. This is the most technical step in the proof, relying on the Probabilistic Bogolyubov Lemma. Finally, using random permutations and running several times [12, Algorithm 3], we get the desired result.
For generalizing Theorem 1 to , we can proceed in the same way as in Algorithm 1 by initializing all the entries of to be rather than zero. This enables us to get a reduction to the zero-error case of Lemma 14. For this reduction, we need a linear code over which is list-decodable from a radius of and admits a full-rank partition. This can be achieved by generalizing the construction of the binary linear code in Lemma 7 to a linear code with alphabet size , which is addressed in Remark 20 in the appendix.
Hence, we get the following generalization for a field of prime order.
Theorem 3. [Restated, see original statement.]
Let be sufficiently large, and let be a small constant. Let be an algorithm that gets as input two matrices and outputs a matrix in time such that
Then, there exists an algorithm running in time that gets as input two matrices and outputs a matrix such that for all
References
- [1] M. Alekhnovich. Linear diophantine equations over polynomials and soft decoding of reed-solomon codes. IEEE Transactions on Information Theory, 51(7):2257β2265, 2005. doi:10.1109/TIT.2005.850097.
- [2] Josh Alman, Ran Duan, Virginia Vassilevska Williams, Yinzhan Xu, Zixuan Xu, and Renfei Zhou. More asymmetry yields faster matrix multiplication. In Proceedings of the 2025 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 2005β2039. SIAM, 2025. doi:10.1137/1.9781611978322.63.
- [3] Josh Alman and Virginia Vassilevska Williams. A refined laser method and faster matrix multiplication. In SODA 2021, pages 522β539. SIAM, 2021. doi:10.1137/1.9781611976465.32.
- [4] Vahid R. Asadi, Alexander Golovnev, Tom Gur, and Igor Shinkar. Worst-case to average-case reductions via additive combinatorics. In Proceedings of the 54th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2022, pages 1566β1574, New York, NY, USA, 2022. Association for Computing Machinery. doi:10.1145/3519935.3520041.
- [5] Dario Bini, Milvio Capovani, Francesco Romani, and Grazia Lotti. O(n2.7799) complexity for n Γ n approximate matrix multiplication. Information Processing Letters, 8(5):234β235, 1979. doi:10.1016/0020-0190(79)90113-3.
- [6] Manuel Blum, Michael Luby, and Ronitt Rubinfeld. Self-testing/correcting with applications to numerical problems. Journal of Computer and System Sciences, 47(3):549β595, 1993. doi:10.1016/0022-0000(93)90044-W.
- [7] A. Borodin and R. Moenck. Fast modular transforms. Journal of Computer and System Sciences, 8(3):366β386, 1974. doi:10.1016/S0022-0000(74)80029-2.
- [8] Don Coppersmith and Shmuel Winograd. Matrix multiplication via arithmetic progressions. Journal of Symbolic Computation, 9(3):251β280, 1990. Computational algebraic complexity editorial. doi:10.1016/S0747-7171(08)80013-2.
- [9] Ran Duan, Hongxun Wu, and Renfei Zhou. Faster matrix multiplication via asymmetric hashing. In 64th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2023, Santa Cruz, CA, USA, November 6-9, 2023, pages 2129β2138. IEEE, 2023. doi:10.1109/FOCS57990.2023.00130.
- [10] Charles M. Fiduccia. Polynomial evaluation via the division algorithm the fast fourier transform revisited. In Proceedings of the Fourth Annual ACM Symposium on Theory of Computing, STOC β72, pages 88β93, New York, NY, USA, 1972. Association for Computing Machinery. doi:10.1145/800152.804900.
- [11] Rusins Freivalds. Probabilistic machines can use less running time. In IFIP congress, volume 839, page 842, 1977.
- [12] Ashish Gola, Igor Shinkar, and Harsimran Singh. Matrix Multiplication Reductions. In Amit Kumar and Noga Ron-Zewi, editors, Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2024), volume 317 of Leibniz International Proceedings in Informatics (LIPIcs), pages 34:1β34:15, Dagstuhl, Germany, 2024. Schloss Dagstuhl β Leibniz-Zentrum fΓΌr Informatik. doi:10.4230/LIPIcs.APPROX/RANDOM.2024.34.
- [13] Venkatesan Guruswami. List decoding of binary codesβa brief survey of some recent results. In Yeow Meng Chee, Chao Li, San Ling, Huaxiong Wang, and Chaoping Xing, editors, Coding and Cryptology, pages 97β106, Berlin, Heidelberg, 2009. Springer Berlin Heidelberg. doi:10.1007/978-3-642-01877-0_10.
- [14] Venkatesan Guruswami, Johan HΓ₯stad, and Swastik Kopparty. On the list-decodability of random linear codes. IEEE Transactions on Information Theory, 57(2):718β725, 2011. doi:10.1109/TIT.2010.2095170.
- [15] Venkatesan Guruswami, Atri Rudra, and Madhu Sudan. Essential coding theory, 2023. Draft available at https://cse.buffalo.edu/faculty/atri/courses/coding-theory/book/.
- [16] Shuichi Hirahara and Nobutaka Shimizu. Hardness self-amplification: Simplified, optimized, and unified. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, pages 70β83, New York, NY, USA, 2023. Association for Computing Machinery. doi:10.1145/3564246.3585189.
- [17] Shuichi Hirahara and Nobutaka Shimizu. Hardness self-amplification: Simplified, optimized, and unified, 2023. Full version of [16]. URL: https://eccc.weizmann.ac.il/report/2023/026/.
- [18] Shuichi Hirahara and Nobutaka Shimizu. Error-correction of matrix multiplication algorithms. In Proceedings of the 57th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2025. Association for Computing Machinery, 2025.
- [19] Shuichi Hirahara and Nobutaka Shimizu. An optimal error-correcting reduction for matrix multiplication. In Proceedings of the 52nd EATCS International Colloquium on Automata, Languages, and Programming, ICALP, pages 97:1β97:17, 2025. doi:10.4230/LIPIcs.ICALP.2025.97.
- [20] Fernando Granha Jeronimo. Fast Decoding of Explicit Almost Optimal -Balanced q-Ary Codes And Fast Approximation of Expanding k-CSPs. In Nicole Megow and Adam Smith, editors, Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2023), volume 275 of Leibniz International Proceedings in Informatics (LIPIcs), pages 60:1β60:16, Dagstuhl, Germany, 2023. Schloss Dagstuhl β Leibniz-Zentrum fΓΌr Informatik. doi:10.4230/LIPIcs.APPROX/RANDOM.2023.60.
- [21] Fernando Granha Jeronimo, Shashank Srivastava, and Madhur Tulsiani. Near-linear time decoding of ta-shmaβs codes via splittable regularity. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, STOC 2021, pages 1527β1536, New York, NY, USA, 2021. Association for Computing Machinery. doi:10.1145/3406325.3451126.
- [22] FranΓ§ois Le Gall. Powers of tensors and fast matrix multiplication. In Proceedings of the 39th International Symposium on Symbolic and Algebraic Computation, ISSAC β14, pages 296β303, New York, NY, USA, 2014. Association for Computing Machinery. doi:10.1145/2608628.2608664.
- [23] Ray Li and Mary Wootters. Improved List-Decodability of Random Linear Binary Codes. In Eric Blais, Klaus Jansen, JosΓ© D. P. Rolim, and David Steurer, editors, Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2018), volume 116 of Leibniz International Proceedings in Informatics (LIPIcs), pages 50:1β50:19, Dagstuhl, Germany, 2018. Schloss Dagstuhl β Leibniz-Zentrum fΓΌr Informatik. doi:10.4230/LIPIcs.APPROX-RANDOM.2018.50.
- [24] V. Ya. Pan. Strassenβs algorithm is not optimal trilinear technique of aggregating, uniting and canceling for constructing fast algorithms for matrix operations. In 19th Annual Symposium on Foundations of Computer Science (SFCS 1978), pages 166β176, 1978. doi:10.1109/SFCS.1978.34.
- [25] Francesco Romani. Some properties of disjoint sums of tensors related to matrix multiplication. SIAM Journal on Computing, 11(2):263β267, 1982. doi:10.1137/0211020.
- [26] A. SchΓΆnhage. Partial and total matrix multiplication. SIAM Journal on Computing, 10(3):434β455, 1981. doi:10.1137/0210032.
- [27] Andrew James Stothers. On the complexity of matrix multiplication, 2010. URL: https://api.semanticscholar.org/CorpusID:262795811.
- [28] V. Strassen. The asymptotic spectrum of tensors and the exponent of matrix multiplication. In 27th Annual Symposium on Foundations of Computer Science (sfcs 1986), pages 49β54, 1986. doi:10.1109/SFCS.1986.52.
- [29] Volker Strassen. Gaussian elimination is not optimal. Numerische Mathematik, 13(4):354β356, 1969.
- [30] Amnon Ta-Shma. Explicit, almost optimal, epsilon-balanced codes. In Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2017, pages 238β251, New York, NY, USA, 2017. Association for Computing Machinery. doi:10.1145/3055399.3055408.
- [31] Virginia Vassilevska Williams. Multiplying matrices faster than coppersmith-winograd. In Proceedings of the Forty-Fourth Annual ACM Symposium on Theory of Computing, STOC β12, pages 887β898, New York, NY, USA, 2012. Association for Computing Machinery. doi:10.1145/2213977.2214056.
- [32] Virginia Vassilevska Williams, Yinzhan Xu, Zixuan Xu, and Renfei Zhou. New bounds for matrix multiplication: from alpha to omega. In Proceedings of the 2024 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 3792β3835. SIAM, 2024. doi:10.1137/1.9781611977912.134.
Appendix A List decodable codes
Theorem 15.
Fix and let be sufficiently large, such that for some absolute constant . There exists a randomized algorithm that runs in time and with probability at least outputs an advice , such that given we have the following.
There exists a linear code , where is an integer, and it satisfies the following properties.
-
1.
is encodable in time .
-
2.
is -list decodable for , and the running time of the decoder is .
-
3.
The generating matrix for has a full-rank partition, and there exists an algorithm that outputs the partition in time .
Β Remark 16.
We make two comments about the advice in Theorem 15.
-
1.
The code in Theorem 15 depends on , and the running times of the algorithms for encoding and list-decoding are when given the advice .
-
2.
Given we can verify that it satisfies the desired properties in time . In particular, by sampling independent advices and verifying them, we can find the advice with the desired properties in time with probability 0.99.
Our construction will use the following result on list decoding of Reed-Solomon codes.
Lemma 17 (List Decoding Reed-Solomon Codes [1]).
Let be a finite field, and let . Consider the Reed-Solomon code of degree . For the code is -list decodable, and the list-decoding algorithm runs in time
We start the proof of Theorem 15 with the following simple claim.
Claim 18.
Fix , and let . For a sufficiently large such that there exists a linear code defined as for some generating matrix such that
-
is -list decodable, and
-
has a full-rank partition defined as .
A uniformly random matrix satisfies the desired properties with probability at least . Furthermore, given a matrix , we can verify that it satisfies the desired properties in time .
Proof.
Let be a uniformly random matrix. Let be the linear code corresponding to the generating matrix .
By the result of Li and Wootters [23, Theorem 5], with probability the code is -list decodable.
Next we show that with probability at least the random matrix has the full-rank partition . Note that in each block of rows are linearly independent with probability
By independence, between blocks we get that has a full-rank partition with probability at least . Since is sufficiently large, both events happen with probability at least .
For the βfurthermoreβ statement of the claim, note that if we are given a matrix , then we can check that for all possible strings there are at most codewords at distance in total time , as required.
Next we prove Theorem 15 by concatenating Reed-Solomon code as the outer code with the code from Claim 18 as the inner code. Specifically, we use the following lemma.
Lemma 19 (Concatenating codes).
Suppose is a linear code defined as for some matrix such that
-
has an encoding algorithm whose running time is ,
-
is -list decodable, and the running time of the list decoding algorithm is ,
-
has a full-rank partition defined as .
Let be a power of 2.
Then, there exists a linear code with and , such that
-
1.
There is an encoding algorithm for with running time ,
-
2.
is -list decodable,
-
3.
The list decoding algorithm runs in time ,
-
4.
has a full-rank partition.
Proof.
We define as the concatenation of Reed-Solomon code as the outer code and as the inner code. Specifically, the encoding of works as follows.
Given a message of length view it as blocks of length each. Treating each block as an element of the finite field with we first apply the Reed-Solomon code on to get an encoding in . Then, we apply the code to each symbol of the Reed-Solomon codeword. Therefore, is a linear code that takes bits and outputs bits.
Next we prove that satisfies the properties stated in the lemma.
- 1.
-
2.
For the decoding algorithm, suppose we get a word that is close to an encoding of some message . By Markovβs inequality there are at least fraction of blocks that are close to .
We run the -list decoding algorithm for on each block of , and get a list of size at most for each block. By choosing a uniformly random element in each list, we obtain a word of length over the alphabet that agrees with the Reed-Solomon encoding of in at least fraction of symbols in expectation. Furthermore, by Chernoff bound, with probability the word we obtain agrees with the Reed-Solomon encoding of in at least fraction of symbols. Let us assume from now on the event that at least fraction of symbols are correct. Since , we can apply Lemma 17 to obtain a list of length at most that contains our message , as required.
Therefore, any string belongs to the list of outputs with probability at least . Therefore, by union bound the output of the decoding algorithm contains all strings with probability at least , where the notations might hide different constants.
-
3.
For the running time of the list decoder, we first run the decoder for on each of the blocks, which take total time. After that we sample one element from each list, and run the list decoder for Reed-Solomon, which runs in time .
-
4.
Consider first the generating matrix of the Reed-Solomon code. By the interpolation properties of polynomials, it is easy to see that any partition of the rows of is a full-rank partition. In particular, the restriction of to any blocks is a bijection mapping of the bits to bits, and hence the corresponding submatrix is invertible.
Next, consider only the first bits of the Reed-Solomon encoding corresponding to blocks. (All other blocks are handled similarly.)
Note that applying on each of the blocks, corresponds to multiplying by a block-diagonal matrix , where each block of is the matrix . Consider the partition of , and denote .
We claim that for each , if we take the union of the rows of corresponding to βs in each of the blocks, then we get a full rank submatrix. That is because the row are divided into blocks on the diagonal of , such that each of consists of a full-rank submatrix (one for each ), and the rows are linearly independent of each other. Multiplying the first rows of (which is a submatrix of full-rank) by that submatrix of (which is also full-rank) gives a full-rank submatrix. This defines a full-rank partition of .
This completes the proof of Lemma 19.
Proof of Theorem 15.
Consider the code and the matrix from Claim 18 with and . By the guarantee of Claim 18, satisfies the desired properties with probability at least . Assuming that we have such , the encoding algorithm of runs in time , and the trivial list-decoding works in time by simply encoding all possible messages and checking their distance from the given word.
We construct in two steps.
-
1.
First, we apply Lemma 19 with as the inner code, obtaining the code
such that
-
(i)
is -list decodable,
-
(ii)
the running time of the encoding algorithm is ,
-
(iii)
the decoding algorithm runs in time , and
-
(iv)
has a full-rank partition.
-
(i)
-
2.
Next we apply Lemma 19 with as the inner code. We obtain the code , where . By the guarantees of Lemma 19, satisfies the following properties:
-
(i)
it is -list decodable,
-
(ii)
the running time of the encoding algorithm is ,
-
(iii)
the running time of the list decoding algorithm is upper bounded by , and
-
(iv)
has a full-rank partition.
-
(i)
This completes the proof of Theorem 15.
Β Remark 20.
It is not difficult to generalize Theorem 15 to a finite field for any prime . To extend Claim 18 to a finite field of order p, apply the result of Guruswami, HΓ₯stad, Kopparty [14, Theorem 2.5] that the random linear code is -list decodable with probability . Its generating matrix will have a full-rank partition with probability greater than . The rest of the construction goes through in a similar manner.
