On the Amortized Complexity of Approximate Counting

Authors Ishaq Aden-Ali , Yanjun Han , Jelani Nelson , Huacheng Yu



PDF
Thumbnail PDF

File

LIPIcs.APPROX-RANDOM.2024.33.pdf
  • Filesize: 0.78 MB
  • 17 pages

Document Identifiers

Author Details

Ishaq Aden-Ali
  • University of California, Berkeley, CA, USA
Yanjun Han
  • New York University, NY, USA
Jelani Nelson
  • University of California, Berkeley, CA, USA
Huacheng Yu
  • Princeton University, NJ, USA

Acknowledgements

We thank Sidhanth Mohanty for very enlightening discussions on unpredictable paths, half Cauchy random variables, and stochastic processes in general that ultimately led to the discovery of the first version of our main lower bound. We also thank Mark Sellke for answering a certain question regarding stochastic processes. Lastly, we thank Greg Valiant for raising the question of the amortized space complexity of approximate counting.

Cite AsGet BibTex

Ishaq Aden-Ali, Yanjun Han, Jelani Nelson, and Huacheng Yu. On the Amortized Complexity of Approximate Counting. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2024). Leibniz International Proceedings in Informatics (LIPIcs), Volume 317, pp. 33:1-33:17, Schloss Dagstuhl – Leibniz-Zentrum für Informatik (2024)
https://doi.org/10.4230/LIPIcs.APPROX/RANDOM.2024.33

Abstract

Naively storing a counter up to value n would require Ω(log n) bits of memory. Nelson and Yu [Jelani Nelson and Huacheng Yu, 2022], following work of Morris [Robert H. Morris, 1978], showed that if the query answers need only be (1+ε)-approximate with probability at least 1 - δ, then O(log log n + log log(1/δ) + log(1/ε)) bits suffice, and in fact this bound is tight. Morris' original motivation for studying this problem though, as well as modern applications, require not only maintaining one counter, but rather k counters for k large. This motivates the following question: for k large, can k counters be simultaneously maintained using asymptotically less memory than k times the cost of an individual counter? That is to say, does this problem benefit from an improved amortized space complexity bound? We answer this question in the negative. Specifically, we prove a lower bound for nearly the full range of parameters showing that, in terms of memory usage, there is no asymptotic benefit possible via amortization when storing multiple counters. Our main proof utilizes a certain notion of "information cost" recently introduced by Braverman, Garg and Woodruff [Mark Braverman et al., 2020] to prove lower bounds for streaming algorithms.

Subject Classification

ACM Subject Classification
  • Theory of computation → Lower bounds and information complexity
Keywords
  • streaming
  • approximate counting
  • information complexity
  • lower bounds

Metrics

  • Access Statistics
  • Total Accesses (updated on a weekly basis)
    0
    PDF Downloads

References

  1. Ziv Bar-Yossef, T. S. Jayram, Ravi Kumar, and D. Sivakumar. An information statistics approach to data stream and communication complexity. J. Comput. Syst. Sci., 68(4):702-732, 2004. Google Scholar
  2. Mark Braverman, Sumegha Garg, and David P. Woodruff. The coin problem with applications to data streams. In Proceedings of the 61st IEEE Annual Symposium on Foundations of Computer Science (FOCS), pages 318-329, 2020. Google Scholar
  3. Amit Chakrabarti, Yaoyun Shi, Anthony Wirth, and Andrew Chi-Chih Yao. Informational complexity and the direct sum problem for simultaneous message complexity. In Proceedings of the 42nd Annual Symposium on Foundations of Computer Science (FOCS), pages 270-278, 2001. Google Scholar
  4. Thomas M. Cover and Joy A. Thomas. Elements of information theory. Wiley, 2 edition, 2006. URL: http://www.elementsofinformationtheory.com/.
  5. Ronald L. Graham, Donald E. Knuth, and Oren Patashnik. Concrete Mathematics: A Foundation for Computer Science, 2nd Ed. Addison-Wesley, 1994. Google Scholar
  6. Elena Kolevska. What happens with Redis runs out of memory, December 2018. URL: https://www.youtube.com/watch?v=Xjq5XL2u3po.
  7. Jérémie Lumbroso. The story of HyperLogLog: How Flajolet processed streams with coin flips. CoRR, abs/1805.00612, 2018. URL: https://arxiv.org/abs/1805.00612.
  8. Robert H. Morris. Counting large numbers of events in small registers. Commun. ACM, 21(10):840-842, 1978. Google Scholar
  9. Jelani Nelson and Huacheng Yu. Optimal bounds for approximate counting. In Proceedings of the 41st ACM International Conference on Principles of Database Systems (PODS), pages 119-127, 2022. Google Scholar
  10. Michael Sipser. Introduction to the theory of computation. PWS Publishing Company, 1997. Google Scholar
Questions / Remarks / Feedback
X

Feedback for Dagstuhl Publishing


Thanks for your feedback!

Feedback submitted

Could not send message

Please try again later or send an E-mail