Document

# Optimal Coding Theorems in Time-Bounded Kolmogorov Complexity

## File

LIPIcs.ICALP.2022.92.pdf
• Filesize: 0.73 MB
• 14 pages

## Acknowledgements

We thank Bruno Bauwens for discussions and useful insights.

## Cite As

Zhenjian Lu, Igor C. Oliveira, and Marius Zimand. Optimal Coding Theorems in Time-Bounded Kolmogorov Complexity. In 49th International Colloquium on Automata, Languages, and Programming (ICALP 2022). Leibniz International Proceedings in Informatics (LIPIcs), Volume 229, pp. 92:1-92:14, Schloss Dagstuhl – Leibniz-Zentrum für Informatik (2022)
https://doi.org/10.4230/LIPIcs.ICALP.2022.92

## Abstract

The classical coding theorem in Kolmogorov complexity states that if an n-bit string x is sampled with probability δ by an algorithm with prefix-free domain then 𝖪(x) ≤ log(1/δ) + O(1). In a recent work, Lu and Oliveira [Zhenjian Lu and Igor C. Oliveira, 2021] established an unconditional time-bounded version of this result, by showing that if x can be efficiently sampled with probability δ then rKt(x) = O(log(1/δ)) + O(log n), where rKt denotes the randomized analogue of Levin’s Kt complexity. Unfortunately, this result is often insufficient when transferring applications of the classical coding theorem to the time-bounded setting, as it achieves a O(log(1/δ)) bound instead of the information-theoretic optimal log(1/δ). Motivated by this discrepancy, we investigate optimal coding theorems in the time-bounded setting. Our main contributions can be summarised as follows. • Efficient coding theorem for rKt with a factor of 2. Addressing a question from [Zhenjian Lu and Igor C. Oliveira, 2021], we show that if x can be efficiently sampled with probability at least δ then rKt(x) ≤ (2 + o(1)) ⋅ log(1/δ) + O(log n). As in previous work, our coding theorem is efficient in the sense that it provides a polynomial-time probabilistic algorithm that, when given x, the code of the sampler, and δ, it outputs, with probability ≥ 0.99, a probabilistic representation of x that certifies this rKt complexity bound. • Optimality under a cryptographic assumption. Under a hypothesis about the security of cryptographic pseudorandom generators, we show that no efficient coding theorem can achieve a bound of the form rKt(x) ≤ (2 - o(1)) ⋅ log(1/δ) + poly(log n). Under a weaker assumption, we exhibit a gap between efficient coding theorems and existential coding theorems with near-optimal parameters. • Optimal coding theorem for pK^t and unconditional Antunes-Fortnow. We consider pK^t complexity [Halley Goldberg et al., 2022], a variant of rKt where the randomness is public and the time bound is fixed. We observe the existence of an optimal coding theorem for pK^t, and employ this result to establish an unconditional version of a theorem of Antunes and Fortnow [Luis Filipe Coelho Antunes and Lance Fortnow, 2009] which characterizes the worst-case running times of languages that are in average polynomial-time over all 𝖯-samplable distributions.

## Subject Classification

##### ACM Subject Classification
• Theory of computation
##### Keywords
• computational complexity
• randomized algorithms
• Kolmogorov complexity

## Metrics

• Access Statistics
• Total Accesses (updated on a weekly basis)
0

## References

1. Scott Aaronson. The equivalence of sampling and searching. Theory Comput. Syst., 55(2):281-298, 2014.
2. Eric Allender. Applications of time-bounded Kolmogorov complexity in complexity theory. In Kolmogorov complexity and computational complexity, pages 4-22. Springer, 1992.
3. Eric Allender. When worlds collide: Derandomization, lower bounds, and Kolmogorov complexity. In International Conference on Foundations of Software Technology and Theoretical Computer Science (FSTTCS), pages 1-15. Springer, 2001.
4. Eric Allender. The complexity of complexity. In Computability and Complexity, pages 79-94. Springer, 2017.
5. Luis Filipe Coelho Antunes and Lance Fortnow. Worst-case running times for average-case algorithms. In Conference on Computational Complexity (CCC), pages 298-303, 2009.
6. Bruno Bauwens and Marius Zimand. Universal almost optimal compression and Slepian-Wolf coding in probabilistic polynomial time. CoRR, abs/1911.04268, 2019. URL: http://arxiv.org/abs/1911.04268.
7. Shai Ben-David, Benny Chor, Oded Goldreich, and Michael Luby. On the theory of average case complexity. J. Comput. Syst. Sci., 44(2):193-219, 1992. URL: https://doi.org/10.1016/0022-0000(92)90019-F.
8. Andrej Bogdanov and Luca Trevisan. Average-case complexity. Found. Trends Theor. Comput. Sci., 2(1), 2006.
9. Harry Buhrman, Lance Fortnow, and Sophie Laplante. Resource-bounded Kolmogorov complexity revisited. SIAM J. Comput., 31(3):887-905, 2001.
10. Harry Buhrman, Troy Lee, and Dieter van Melkebeek. Language compression and pseudorandom generators. Comput. Complex., 14(3):228-255, 2005.
11. M. R. Capalbo, O. Reingold, S. P. Vadhan, and A. Wigderson. Randomness conductors and constant-degree lossless expanders. In STOC, pages 659-668, 2002. URL: https://doi.org/10.1145/509907.510003.
12. Li Chen, Rasmus Kyng, Yang P. Liu, Richard Peng, Maximilian Probst Gutenberg, and Sushant Sachdeva. Maximum flow and minimum-cost flow in almost-linear time. CoRR, abs/2203.00671, 2022. URL: https://doi.org/10.48550/arXiv.2203.00671.
13. Kai-Min Chung, Siyao Guo, Qipeng Liu, and Luowen Qian. Tight quantum time-space tradeoffs for function inversion. In Symposium on Foundations of Computer Science (FOCS), pages 673-684, 2020.
14. Anindya De, Luca Trevisan, and Madhur Tulsiani. Time space tradeoffs for attacks against one-way functions and PRGs. In Annual International Cryptology Conference (CRYPTO), pages 649-665, 2010.
15. Amos Fiat and Moni Naor. Rigorous time/space trade-offs for inverting functions. SIAM J. Comput., 29(3):790-803, 1999. URL: https://doi.org/10.1137/S0097539795280512.
16. Lance Fortnow. Kolmogorov complexity and computational complexity. Complexity of Computations and Proofs. Quaderni di Matematica, 13, 2004.
17. Halley Goldberg, Valentine Kabanets, Zhenjian Lu, and Igor C. Oliveira. Probabilistic Kolmogorov complexity with applications to average-case complexity. Preprint, 2022.
18. Venkatesan Guruswami, Christopher Umans, and Salil P. Vadhan. Unbalanced expanders and randomness extractors from Parvaresh-Vardy codes. J. ACM, 56(4):20:1-20:34, 2009.
19. Iftach Haitner, Omer Reingold, and Salil P. Vadhan. Efficiency improvements in constructing pseudorandom generators from one-way functions. SIAM J. Comput., 42(3):1405-1430, 2013.
20. Shuichi Hirahara. Non-black-box worst-case to average-case reductions within NP. In Symposium on Foundations of Computer Science (FOCS), pages 247-258, 2018.
21. Shuichi Hirahara. Average-case hardness of NP from exponential worst-case hardness assumptions. In Symposium on Theory of Computing (STOC), pages 292-302, 2021.
22. S. Hoory, N. Linial, and A. Wigderson. Expander graphs and their applications. Bull. Amer. Math. Soc., 43:439-561, 2006.
23. Russell Impagliazzo and Ramamohan Paturi. On the complexity of k-SAT. J. Comput. Syst. Sci., 62(2):367-375, 2001. URL: https://doi.org/10.1006/jcss.2000.1727.
24. Jan Krajíček. Information in propositional proofs and algorithmic proof search. The Journal of Symbolic Logic, pages 1-22, 2021.
25. Troy Lee. Kolmogorov complexity and formula lower bounds. PhD thesis, University of Amsterdam, 2006.
26. Leonid A. Levin. Randomness conservation inequalities; information and independence in mathematical theories. Information and Control, 61(1):15-37, 1984.
27. Leonid A. Levin. Average case complete problems. SIAM J. Comput., 15(1):285-286, 1986. URL: https://doi.org/10.1137/0215020.
28. Ming Li and Paul M. B. Vitányi. Average case complexity under the universal distribution equals worst-case complexity. Inf. Process. Lett., 42(3):145-149, 1992.
29. Ming Li and Paul M. B. Vitányi. An introduction to Kolmogorov complexity and its applications. Springer-Verlag, 2019. 4th edition (1st edition in 1993).
30. Yanyi Liu and Rafael Pass. On one-way functions and Kolmogorov complexity. In Symposium on Foundations of Computer Science (FOCS), pages 1243-1254, 2020.
31. Zhenjian Lu and Igor C. Oliveira. An efficient coding theorem via probabilistic representations and its applications. In International Colloquium on Automata, Languages, and Programming (ICALP), pages 94:1-94:20, 2021.
32. Zhenjian Lu, Igor C. Oliveira, and Rahul Santhanam. Pseudodeterministic algorithms and the structure of probabilistic time. In Symposium on Theory of Computing (STOC), pages 303-316, 2021.
33. Zhenjian Lu, Igor C. Oliveira, and Marius Zimand. Optimal coding theorems in time-bounded Kolmogorov complexity, 2022. URL: https://doi.org/10.48550/ARXIV.2204.08312.
34. Aleksander Madry. Navigating central path with electrical flows: From flows to matchings, and back. In Symposium on Foundations of Computer Science (FOCS), pages 253-262, 2013. URL: https://doi.org/10.1109/FOCS.2013.35.
35. Igor C. Oliveira. Randomness and intractability in Kolmogorov complexity. In International Colloquium on Automata, Languages, and Programming (ICALP), pages 32:1-32:14, 2019.
36. Hanlin Ren and Rahul Santhanam. Hardness of KT characterizes parallel cryptography. In Computational Complexity Conference (CCC), pages 35:1-35:58, 2021.
37. Michael Sipser. A complexity theoretic approach to randomness. In Symposium on Theory of Computing (STOC), pages 330-335, 1983.
38. Salil P. Vadhan and Colin Jia Zheng. A uniform min-max theorem with applications in cryptography. In Annual Cryptology Conference (CRYPTO), pages 93-110, 2013.