The πβΒ² min-sum k-clustering problem is to partition an input set into clusters C_1,β¦,C_k to minimize β_{i=1}^k β_{p,q β C_i} βp-qββΒ². Although πβΒ² min-sum k-clustering is NP-hard, it is not known whether it is NP-hard to approximate πβΒ² min-sum k-clustering beyond a certain factor. In this paper, we give the first hardness-of-approximation result for the πβΒ² min-sum k-clustering problem. We show that it is NP-hard to approximate the objective to a factor better than 1.056 and moreover, assuming a balanced variant of the Johnson Coverage Hypothesis, it is NP-hard to approximate the objective to a factor better than 1.327. We then complement our hardness result by giving a fast PTAS for πβΒ² min-sum k-clustering. Specifically, our algorithm runs in time O(n^{1+o(1)}dβ 2^{(k/Ξ΅)^O(1)}), which is the first nearly linear time algorithm for this problem. We also consider a learning-augmented setting, where the algorithm has access to an oracle that outputs a label i β [k] for input point, thereby implicitly partitioning the input dataset into k clusters that induce an approximately optimal solution, up to some amount of adversarial error Ξ± β [0,1/2). We give a polynomial-time algorithm that outputs a (1+Ξ³Ξ±)/(1-Ξ±)Β²-approximation to πβΒ² min-sum k-clustering, for a fixed constant Ξ³ > 0.
@InProceedings{karthikc.s._et_al:LIPIcs.SoCG.2025.62, author = {Karthik C. S. and Lee, Euiwoong and Rabani, Yuval and Schwiegelshohn, Chris and Zhou, Samson}, title = {{On Approximability of πβΒ² Min-Sum Clustering}}, booktitle = {41st International Symposium on Computational Geometry (SoCG 2025)}, pages = {62:1--62:18}, series = {Leibniz International Proceedings in Informatics (LIPIcs)}, ISBN = {978-3-95977-370-6}, ISSN = {1868-8969}, year = {2025}, volume = {332}, editor = {Aichholzer, Oswin and Wang, Haitao}, publisher = {Schloss Dagstuhl -- Leibniz-Zentrum f{\"u}r Informatik}, address = {Dagstuhl, Germany}, URL = {https://drops.dagstuhl.de/entities/document/10.4230/LIPIcs.SoCG.2025.62}, URN = {urn:nbn:de:0030-drops-232142}, doi = {10.4230/LIPIcs.SoCG.2025.62}, annote = {Keywords: Clustering, hardness of approximation, polynomial-time approximation schemes, learning-augmented algorithms} }
Feedback for Dagstuhl Publishing