Improved FPT Approximation for Sum of Radii Clustering with Mergeable Constraints
Abstract
In this work, we study -min-sum-of-radii (-MSR) clustering under mergeable constraints. -MSR seeks to group data points using a set of up to balls, such that the sum of the radii of the balls is minimized. A clustering constraint is called mergeable if merging two clusters satisfying the constraint, results in a cluster that also satisfies the constraint. Many popularly studied constraints are mergeable, including fairness constraints and lower bound constraints.
In our work, we design a -approximation for -MSR under any given mergeable constraint with runtime , i.e., fixed-parameter tractable in for constant . Our result directly improves upon the FPT -approximation by Carta et al. [10]. We also provide a hardness result that excludes the exact solvability of -MSR under any given mergeable constraint in time , assuming ETH is true.
Keywords and phrases:
sum-of-radii clustering, mergeable constraints, approximation algorithmCategory:
APPROXCopyright and License:
2012 ACM Subject Classification:
Theory of computation Approximation algorithms analysisFunding:
This work was supported by the National Science Foundation under Grant No. AF 2311397.Editors:
Alina Ene and Eshan ChattopadhyaySeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
Given a set of points in a metric space and an integer , the -min-sum-of-radii clustering problem (-MSR for short) seeks to group the points using a set of at most balls, called clusters, each centered at a point, minimizing the sum of the radii of those balls. Like most clustering objectives, finding an optimal -MSR clustering is NP-hard, as shown by Gibson et al. [21]. They also designed a quasi-polynomial-time approximation scheme (QPTAS) for -MSR in the same work, implying it is unlikely that -MSR is APX-hard under standard complexity-theoretic assumptions. Consequently, substantial attention has been directed toward designing approximation algorithms. Currently, the best-known polynomial-time approximation factor is [9].
To highlight the motivation for studying -MSR, we compare it to -center and -median, two popular clustering objectives. -center is similar to -MSR, but instead of minimizing the sum of the radii of balls, it minimizes the radius of the largest ball. Simple -approximation algorithms exist for -center [22, 23], and variants of -center are relatively easy to solve. However, as a trade-off, -center is prone to outliers, as well as artifacts such as the dissection effect shown in Figure 1, where clusters might be split unnecessarily. -median, on the other hand, seeks to minimize the sum of each point’s distance to the closest center, rather than the radii of a cluster. -median and its variants are relatively more difficult to solve, but its robustness has given rise to its popularity [3, 16]. Informally, the -MSR problem lies between -center and -median – it reduces the dissection effect seen in -center, while often allowing for simpler solutions than those typically admitted by -median, which motivates our study.
-MSR clustering has also been studied under additional constraints, popularly known as constrained variants. In our work, we focus on specific types of constraints, namely mergeable constraints. A clustering constraint is said to be mergeable if merging two clusters that satisfy the constraint, results in a cluster that also satisfies the constraint. The property has proven useful for analyzing constrained -MSR clustering, with a number of works dedicated to it [10, 18, 6]. A popular class of mergeable constraints includes many fairness constraints. First introduced by Chierichetti et al. [15], given a partition of the input dataset into groups based on some attribute (commonly referred to as colors), fair clustering aims to ensure that each group in the dataset is proportionally represented within every cluster. For instance, given a set of red and blue points, fair clustering seeks to have a balanced number of red and blue points in each cluster. Carta et al. showed that many common fairness constraints are mergeable [10]; we discuss these constraints further in Section 1.2. There are also non-fair constraints that are mergeable, such as the lower bound constraint, where each cluster is required to have at least points, for some given [1].
Constrained versions of -MSR are often difficult to approximate and lack tractability in polynomial time. A potential approach to address this issue is the design of fixed-parameter tractable (FPT) algorithms. FPT algorithms still have a runtime polynomial in the input size but do not have to be polynomial in their parameters. For example, an algorithm running in time is considered to be FPT in parameter . Fixed-parameter tractability essentially allows efficient computation, assuming the parameters are small in practice. For challenging constrained clustering problems such as constrained -MSR, FPT algorithms are often the preferred approach, as they offer computational tractability and can yield improved approximation guarantees. In our work, we also adopt this approach to address the complexity of the problem.
1.1 Our Contributions
Our work establishes two main results concerning -MSR under mergeable constraints: one demonstrating that the problem admits a -approximation in FPT time in for constant , and another proving its computational hardness. We begin by presenting the following theorem.
Theorem 1.
For any , there is a -approximation algorithm for the -MSR problem under mergeable constraints that runs in time.
Our result directly improves upon the FPT -approximation by Carta et al. [10]. Moreover, to solve -MSR under a given mergeable constraint, their algorithm requires the existence of a constant-factor approximation algorithm for -center under the same constraint, while ours does not have this restriction. There are constraints under which no constant-factor approximation is known for -center. For instance, no approximation algorithm is known for -center under what is called fair representational constraint, which is one of the more general definitions of fair constraints. Our algorithm also has the advantage of generality. While the FPT algorithms by Drexler et al. [18] and Banerjee et al. [6] achieve an approximation factor of , they are limited to specific metric spaces such as Euclidean spaces or those with bounded doubling dimension. In contrast, our algorithm applies to general metric spaces.
Next, we give the following theorem regarding the hardness of -MSR clustering under the fair representational constraint.
Theorem 2.
Fair Representational Clustering cannot be solved in time unless ETH is false.
This result also trivially extends to -MSR under mergeable constraints, as the fair representational constraint is known to be mergeable [18].
1.2 Related Work
Gibson et al. showed that the -MSR objective is NP-hard, even in planar graphs and metrics with doubling dimensions [21]; consequently, a line of work gave approximation algorithms for the problem. Charikar and Panigrahy first gave a polynomial time -approximation for the problem in their seminal work [12]. This result remained the best approximation until Friggstad and Jamshidian followed up with a -approximation nearly twenty years later [20]. Soon after, Buchem et al. improved the approximation factor to , which is the current best known in polynomial time [9].
-MSR has been studied under various constraints, most notably the capacitated constraint and the matroid constraint. A line of work on devising -approximation for capacitated -MSR led to an FPT -approximation [24, 19, 25, 5]. -MSR has also been studied under matroid constraints, leading to an FPT approximation by Inamdar and Varadarajan [24] and an improved approximation by Chen et al. [13]. Polynomial time -approximations for these variants of -MSR are not known. Interestingly, polynomial time -approximations exist for -MSR with lower bound [2, 9].
Another common -MSR constraint to study is the mergeable constraint. Most of the studies are geared towards fairness constraints, since many fairness constraints are mergeable. Drexler et al. first developed an FPT -approximation algorithm for -MSR under mergeable constraints in Euclidean metrics of arbitrary dimension with constant [18]. The result was subsequently extended to metrics with bounded doubling dimension by Banerjee et al. [6]. Carta et al. gave a -approximation algorithm for -MSR under mergeable constraints in general metrics [10].
Fair -MSR has also received attention from researchers. Carta et al. showed that several fairness constraints are mergeable [10]. We name a few of these constraints that are commonly used. The simplest definition is called exact fairness [8], where each cluster must have exactly the same number of points from each color. The next step up is Pairwise Fair Clustering (or sometimes called -fair clustering) [15]. Given a set of red points and blue points and integer , Pairwise Fair Clustering ensures that the ratio between blue and red points in each cluster is between and . This definition is sometimes extended to have more than two colors. Another common fairness definition is called Fair Representational Clustering [7, 8], where the fraction of points in some color allowed in each cluster is at most and at least , given parameters . Bandyapadhyay et al. gave polynomial-time -approximation algorithms for -MSR under the Pairwise Fair Clustering constraint with two groups and exact fair constraint for groups [4]. Chen et al. developed an FPT -approximation algorithm for -MSR under an alternate notion of fairness that is a special case of matroid -MSR [13]. In their definition of fairness, each color is allowed to have a pre-defined number of points that can be chosen as cluster centers. Banerjee et al. [6] also gave an FPT -approximation algorithm for the same problem in Euclidean metric. For more works on fair clustering, one can refer to the survey by Chhabra et al. [14].
Roadmap.
2 Preliminaries
In -MSR clustering, we are given a set of points in a metric space with metric . The goal is to find (i) a set of up to points called centers, and (ii) a function assigning each point to some center . We call a clustering of . For each , we call the cluster centered at , and we define the radius corresponding to as . We refer to the sum of radii of a clustering as the cost of the clustering, and we denote it by ; we use the same notation to denote a sum of a set of radii , or for a set of clusters .
We merge two clusters by setting for all points and some . We say that a constraint is mergeable if for any two clusters satisfying the constraint, merging the clusters results in a new cluster that also satisfies the constraint. We refer to a clustering that satisfies some (mergeable) constraint as a constrained clustering, and a clustering without any constraint as a vanilla clustering. We say that a clustering is feasible if every point in is assigned to some center, , and each cluster satisfies the desired constraint (if any). Moreover, we say that a clustering is optimal if it is feasible, and is minimized over all such clustering.
We denote a ball with center and radius by . Given a point and a ball , we say that is contained in if , denoted by ; note that a point may be contained in a ball , but not assigned to its center , as may be contained in multiple balls but only assigned to one cluster.
3 An FPT -approximation Algorithm
In this section, we describe an FPT -approximation algorithm for -MSR under mergeable constraints. We first describe the algorithm at a high level, then state the procedures that it calls, present the algorithm, and lastly analyze it.
Our algorithm stems from the idea that we can find a vanilla clustering and expand the balls such that each cluster of any constrained clustering is contained within one of the expanded balls. This allows us to transform a vanilla clustering into a constrained clustering in FPT time in . Our algorithm has four major steps: (1) Guess the radii of an optimal vanilla clustering up to a factor. (2) Compute a -approximate vanilla clustering using the guessed radii, (3) Expand the vanilla clusters, resulting in a -approximate set of balls, and (4) Merge the expanded balls to create clusters that satisfy the given mergeable constraint. These steps are accomplished through calls to a series of procedures, which we describe in the following subsections.
Similar to our algorithm, the algorithm of Carta et al. [10] also finds a -approximate set of balls with the aforementioned containment property. However, our algorithm finds the balls in a simpler and more general manner in that it does not rely on a -center algorithm under the same constraint to obtain the balls. Carta et al. also made the same observation as ours that one can satisfy the mergeable constraint by merging the -approximate balls, but our approach differs in how we choose the new center for each set of balls being merged. The algorithm of Carta et al. picks the center of the largest ball, which results in a factor increase in cost, but we show that one can always pick a center that results in at most a factor increase in cost.
3.1 Finding a Vanilla Clustering
We first describe a procedure that takes a radii vector , and a multiplicity vector as input. Moreover, if there is a feasible vanilla clustering that uses those radii with the respective multiplicities, the procedure outputs a vanilla clustering that uses those radii values up to a factor of 2 with the same multiplicities. We make use of the fact shown in Observation 3 that for any point in a cluster with radius , we can cover all points in the cluster with a ball centered at with a radius of . This allows us to recursively find a point that is not yet covered, and guess the cluster containing and the corresponding radius to cover all points of that cluster. We note that our procedure is essentially the same as the procedure used by Chen et al. in their work on -MSR [13], but restated to take in a pair of radii and multiplicity vectors as an additional input; this allows us to use the same set of radii and multiplicity vectors in both the vanilla clustering step and the cluster expansion step of our algorithm. The procedure is given as Algorithm 1.
Observation 3.
For any cluster with center and radius , we can cover all points in using a ball centered at some point with radius .
Proof.
Consider the point and some other point . By triangle inequality, we have that . Furthermore, since , . It follows that for any , therefore we can cover all points in with any point and a radius .
Lemma 4.
Suppose there is a feasible vanilla clustering of that uses exactly balls of radius for . Then Algorithm 1 outputs a feasible vanilla clustering of that uses at most balls of radius for .
Proof.
Let be a set of feasible vanilla clusters corresponding to the radii vector . We consider the following construction of . Let be some point in . Note that is in some with radius for some . We add to as a new center and set for all points such that , and remove and all such from . We also remove the cluster from consideration. We can do this because by Observation 3 we have that with a radius of covers all points in . We repeat this process until there are no points left in . We can always find a clustering within iterations this way because each time we repeat the process, we cover all points in some cluster . Specifically, there exists a recursion branch in Algorithm 1 that follows this construction. It follows that Algorithm 1 outputs the desired vanilla clustering of .
Lemma 5.
Given a vector with radii, Algorithm 1 runs in time.
Proof.
First, we consider the depth of the recursive calls in Algorithm 1. Since each clustering is restricted to having a maximum of clusters, and each recursive call decreases the number of clusters by one, each branch of recursive calls will have at most levels of recursion. Now, we consider how many additional calls each recursive call makes. By our assumption, we have radii in . Since each recursive call makes an additional call for each with , each recursive call can make additional calls. Since each call can be computed in time, it follows that Algorithm 1 runs in time.
3.2 Expanding the Vanilla Clustering
Given a feasible vanilla clustering, we describe a process to expand the vanilla balls so that each cluster of a fixed, constrained clustering is contained in one of the expanded balls. The idea is to guess how much to expand the vanilla clusters using a guessed set of radii for the constrained clustering. In particular, for each vanilla cluster , we guess the biggest constrained cluster whose center is contained in , and expand by its radius, thereby fully containing all the constrained clusters whose centers are in . We assume that the radii vector and a multiplicity vector of the constrained clustering are given. The procedure is given as Algorithm 2.
Now we show some properties of the output of Algorithm 2.
Lemma 6.
Given a vanilla clustering with balls , and the radii vector and a multiplicity vector of a fixed constrained clustering, the output of Algorithm 2 contains a vector such that each cluster of the constrained clustering is contained in for some . Moreover, .
Proof.
Let be a fixed constrained clustering that for each uses clusters of radii . Consider the following construction of a vector . For each center with radius , let be the largest cluster in whose center is in the vanilla cluster , and let be its radius. We set each to . Note that the expanded ball contains any cluster in whose center is in , satisfying the desired property. As the center of each constrained cluster in is contained in a unique vanilla cluster, the radius of the constrained cluster is used at most once to expand the radius of that vanilla cluster. In particular, each is used at most times for the expansion. It follows that is in . For the same reason, , which proves the moreover part.
Now we show the following two lemmas about the output size and runtime of Algorithm 2.
Lemma 7.
Given the vectors of size , the number of guesses that Algorithm 2 outputs is bounded by .
Proof.
Algorithm 2 computes by enumerating a set of guesses for how much to expand each vanilla cluster. There are vanilla clusters, and there are ways to expand each cluster, since there are distinct radii in . It follows that the number of entries in and are bounded by .
Since each entry in takes time to compute, we have the following lemma.
Lemma 8.
Given the vectors of size , Algorithm 2 runs in time.
3.3 Merging Balls
Suppose we are given a set of balls that cover a set of points . Moreover, we assume that there is a clustering of with a mergeable constraint such that each cluster is contained in one of the given balls. We design a procedure that helps find a clustering of with the mergeable constraint such that its cost is at most times the sum of the radii of the given balls.
Before describing the algorithm, we define a few notations. Let denote a set of balls. Let be a graph with the balls in as vertices and the set of edges . A connected component of therefore contains a maximal subset of balls where every pair is connected by a path.
The merging algorithm, Algorithm 3, is quite simple. Given a set of balls that cover a set of points , Algorithm 3 first construct the graph as defined above. For each connected component of balls in , Algorithm 3 finds a point contained in the balls of such that assigning all other points contained in to results in a cluster whose radius is minimized.
In the following, we show that the output clustering of Algorithm 3 satisfies a mergeable constraint based on an additional assumption.
Lemma 9.
Suppose there is a clustering of with a mergeable constraint such that each cluster is contained in a ball of . Then the output clustering of Algorithm 3 satisfies the mergeable constraint.
Proof.
Since each cluster in is in a ball of , the union of the balls in each connected component of is the disjoint union of a subset of clusters of . This means that when Algorithm 3 merges the balls in a connected component into a single cluster, the resulting cluster satisfies the constraint, as the constraint is mergeable. It follows that the output clustering also satisfies the mergeable constraint.
Lemma 10.
Algorithm 3 runs in time.
Proof.
The main bottlenecks of Algorithm 3 are the construction of the edges for graph , and finding a center in each connected component that minimizes the radius when merging the balls in . Since there are points in and balls, checking if each point is in both of each pair of to construct takes time. To find , we can simply probe all points in the connected component. Since there are components, and each component can contain up to points, we can find in time, resulting in an overall runtime of .
Next, we give an upper bound on the cost of the clustering computed by Algorithm 3. Consider any connected component of with a set of balls . Let be the set of points in the union of the balls of . Denote the sum of the radii of the balls in by . Then we have the following lemma.
Lemma 11.
There is a point such that for any , .
Before proving the lemma, we describe a major consequence of it. Recall that for each connected component of , Algorithm 3 assigns all points in the associated balls to a single cluster center . Moreover, this center is chosen to be a point that minimizes the maximum distance from all other points. Thus, by the above lemma, the resulting cluster has , where is the sum of the radii of the balls in the component. As is the disjoint union of balls in the components, we have the following corollary.
Corollary 12.
The cost of the output clustering of Algorithm 3 is at most times the sum of the radii of the input balls in .
Now, we move on towards proving Lemma 11. We need a few additional definitions and observations. Let be a spanning tree of the concerned connected component. We call a path in a maximum leaf-to-leaf path if is a simple path from a leaf node in to another leaf node, and the sum of the radii of the balls along is the maximum among all such paths. Now, consider any path between two leaves. Also, consider the graph , constructed by removing the vertices and edges of from . Note that is a collection of connected components each having exactly one ball connected to exactly one with an edge of . We say that such a component is connected to via . A set of balls is said to form a connected set if the subgraph of induced by the balls in is connected.
Observation 13.
Consider a connected set of balls with the sum of radii . For any two points contained in the balls of , .
Proof.
Let be the balls containing , respectively. Since is a connected set, there is a simple path between and in that uses only balls of . Thus, by triangle inequality, . Hence, .
Similarly, we also have the following observation.
Observation 14.
Consider a connected set of balls with the sum of radii . For any point in a ball of and a ball , .
Proof.
Let be a ball containing . Since is a connected set, there is a simple path between and in that uses only balls of . Thus, by triangle inequality, .
The proof of Lemma 11 is as follows.
Proof.
We prove the existence of the point with the desired property by considering Procedure 4 that explicitly computes a center. We clarify that this procedure is only for the purpose of analysis, and we do not run it as a part of our algorithm.
Procedure 4 picks the new center according to two main cases: (i) there exists a ball in with radius at least , (ii) all balls in have radii less than .
In case (i), there exists a ball such that . In this case, Procedure 4 picks as the new center. By Observation 14, we have that for any point , .
In case (ii), each ball in has a radius of less than . The center selection is split up into two subcases: (iia) , and (iib) .
We first consider case (iia), where a point contained in both and is chosen as the new center. As each ball has radius less than and is a leaf, while loop 7 cannot terminate with , i.e., at its termination . Thus, is well-defined. Consider the partition of into two parts and . is the union of the balls and any ball that lies in a component of that is connected to via for . Similarly, is the union of the balls and any ball that lies in a component of that is connected to via for . Let and . Note that . Next, we bound the distance between the chosen center and the points in and . Note that by our definition of and , . First, we consider for any . Note that at the end of while loop 7. Also, by our definition of , it is the sum of the radii of the balls in . Then, as is a connected set of balls and , by Observation 13, it follows that . Now we consider for any . Since in this subcase and is a partition of , we know that the sum of the radii of the balls in , . Then, as is a connected set of balls and , by Observation 13, it follows that . Thus, the lemma follows in this subcase.
Next, we consider case (iib), here, we pick center of ball as the new center. Similar to case (iia), we partition into three parts , , and . is the union of the balls and any ball that lies in a component of that is connected to via for . Similarly, is the union of the ball and any ball that lies in a component of that is connected to via . Lastly, is the union of the balls and any ball that lies in a component of that is connected to via for . Let , , and . Note that , so we can again bound the overall distance by considering and separately. Towards this end, note that by the end of Loop 5, and .
We first consider for any . Note that is a connected set of balls with a sum of radii of . Furthermore, recall that by the assumption of case (iib), and by the assumption of case (ii). By Observation 14, we have that .
Next, we consider for some . Note that . Since by the end of Loop 7, we have that . Note also that is a connected set of balls with a sum of radii of . Since by the assumption of case (ii), by Observation 14 we have that for any , .
Lastly, we consider for some . Let be a ball that contains . Consider any simple path in between any leaf ball in and such that also contains . Such a path exists, as lies in a component of that is connected to via . We denote by the path obtained by removing from . Note that
| (1) |
Since balls along the path have the maximum sum of radii, we have that
and therefore
| (2) |
Combining Inequalities 1 and 2 we have that
| (3) |
Now, note that is a connected set of balls including with the sum of radii . Thus, by Observation 14 we have that . Applying Inequality 3 we have that .
Since for all and , and covers all points in , is at a distance of at most from any point in in case (iib). This finishes the proof of the lemma.
3.4 Guessing Radii
Here we describe a procedure for guessing the radii of any -MSR clustering up to factor. Formally, one can enumerate a set of -sized vectors such that for any -MSR clustering with radii , there is a vector in where for all , and if , . Moreover, the size of is only . From now on, we call this type of enumeration process a guessing procedure, and we refer to each enumerated entry as a guess. Similar procedures have been used previously for guessing the radii of an optimal capacitated sum of radii clustering [24, 5, 19, 25, 13]. We restate those procedures to obtain a general algorithm that can enumerate the radii of all clusterings. In our algorithm, this procedure will be used to guess both a set of optimal vanilla radii and a set of optimal radii under any given mergeable constraint. Towards this end, we clarify that in the guessed radii vector , all radii might not be distinct, and specifically we have many distinct radii. So, we will enumerate -sized vectors instead while keeping track of the multiplicities.
The procedure works in the following manner. Given a set of points in a metric space, and parameters and , we first compute a radius that can potentially be the maximum radius of some clustering; we can do this in time by trying distances between each pair of points in . For each , we set as the range in which we guess the radii. The procedure is described as Algorithm 5.
Now we show some properties of the output size, runtime, and correctness of the procedure. First, we note the following observation about the number of distinct radii in each .
Observation 15.
Each from the output of Algorithm 5 has distinct radii.
Proof.
Consider the radii vector . Since is formed by picking powers of from the range , the number of radii in is bounded by .
Furthermore, we have the following observation about the size of , which will help us bound the runtime of both Algorithm 5 and some other algorithms.
Observation 16.
The size of is bounded by .
Proof.
First, note that there are a total of choices for . Next, we consider the number of distinct radii vectors that are computed for each . By Observation 15 we have that is bounded by . Since each , there are possible vectors for each choice of . It follows that there are a total of many vectors in .
Since Algorithm 5 enumerates each entry of , the next lemma follows.
Lemma 17.
Algorithm 5 runs in time.
Next, we show that for any clustering, Algorithm 5 outputs a solution with a correct guess.
Lemma 18.
For any -MSR clustering with radii , there are vectors in and in for some such that there exist exactly indices with for and indices with . Moreover, .
Proof.
Consider the choice of made in Algorithm 5 so that . Also consider the vector in corresponding to this choice. By definition, and . Define to be the number of radii in that are at most . For , define to be the number of radii in that are also in . By definition, , and so is a vector in which is considered w.r.t . Also by definition of for , there are indices with , i.e., as desired.
Next, we prove the moreover part. The sum of the radii is at most . Also, for , . Summing over all , . Hence, . Scaling down by a factor of 3, we obtain the lemma.
Corollary 19.
For any optimal sum of radii clustering (vanilla or constrained) , there are vectors in and in for some such that there is a feasible -approximate clustering that uses balls of radius for all .
Proof.
Let be the radii of an optimal clustering. Fix the vectors and given by Lemma 18. We show that each cluster of can be covered by a ball of radius for some such that balls of radius are used for all . As the clustering induced by these balls is essentially the same clustering , it remains feasible. To cover each cluster of that uses a ball of radius , if , use the same center, but with radius ; otherwise, as , there is with , and in this case use the same center, but with radius . By the correspondence of the radii and in Lemma 18, balls of radius are used for all . Lastly, as , the new set of balls induces a -approximate clustering.
3.5 The Algorithm and its Analysis
With all of the procedures defined, we state the main algorithm. Let and be some constants, our algorithm is given as Algorithm 6.
We now analyze Algorithm 6, starting with the correctness and approximation factor. We have the following Lemma.
Lemma 20.
Algorithm 6 produces a -approximate -MSR clustering satisfying a given mergeable constraint.
Proof.
Let and be the respective optimal cost of vanilla -MSR and -MSR with the given mergeable constraint. By Corollary 19, there are vectors in and in for some such that there is a feasible -approximate vanilla clustering that uses balls of radius for all . Fix this choice of and in Line 2. Then by Lemma 4, (Algorithm 1) outputs a feasible vanilla clustering . Let be the centers in . Similarly, by Corollary 19, there are vectors in and in for some such that there is a feasible -approximate clustering with the given mergeable constraint that uses balls of radius for all . Fix this choice of and in Line 7. Then by Lemma 6, the output of (Algorithm 2) and hence contains a vector such that each cluster of is contained in for some . Moreover, by the same lemma, . This is true, as the optimal vanilla clustering cost is at most the optimal cost of any constrained clustering.
Fix the above choice of in Line 10. Then , and hence by Lemma 9, the clustering returned by (Algorithm 3) satisfies the mergeable constraint. Moreover, by Corollary 12, the cost of is at most . Scaling by a constant, we obtain the lemma.
Now we analyze the runtime of Algorithm 6. We have the following lemma.
Lemma 21.
Algorithm 6 runs in time.
Proof.
We first show the runtime of each part of the algorithm. By Lemma 17 we have that the Guess-Radii procedure has a runtime of . Next, by Observation 16 we also have that Guess-Radii produces many vectors in . By Lemma 5 we have that the Vanilla-Clustering procedure runs in time; since we run Vanilla-Clustering on each pair and , this step of the algorithm has a runtime of
Moving on to the expand radii step, by Lemma 8 we have that the Expand-Balls procedure runs in time. Also, Expand-Balls runs for each of the pairs and , resulting in a runtime of . The next step of the algorithm, Merge-Balls, runs in time by Lemma 10. We also showed in Lemma 7 that Expand-Balls outputs pairs and , so the total number of guesses for expanded radii is bounded by . Since Algorithm 6 runs Merge-Balls for each set of guessed radii , this step has a runtime of
Adding everything together we have that Algorithm 6 runs in time
Theorem 1. [Restated, see original statement.]
For any , there is a -approximation algorithm for the -MSR problem under mergeable constraints that runs in time.
4 Hardness of Clustering with Mergeable Constraints
Definition 22 (Fair Representational Clustering).
We are given disjoint groups of points, denoted as , with a total of points with a distance metric . Each group is associated with fairness parameters , for . A clustering is said to be fair representational if, within every cluster, the proportion of points from group lies between and , for all . The objective is to find a fair representational clustering with clusters, that minimizes the sum of the radii of the clusters.
As shown by Carta et al. [11], the fair representational constraint is mergeable. We prove that Fair Representational Clustering cannot be solved in time unless ETH is false. Specifically, we show a reduction from the Dominating set problem.
Given an instance of Dominating set containing an -vertex graph and a parameter , we construct a new graph . Initially, . Add two sets of vertices each to and the edges for all . The instance of Fair Representational Clustering is defined as follows: , , , , is the shortest path metric in , , and for , and .
Lemma 23.
Dominating set admits a solution of size if and only if there is a fair representational clustering of with clusters of cost .
Proof.
The forward direction is easy to prove. Suppose Dominating set admits a solution . We construct a clustering of as follows. Use the points in as centers. For each , assign to a center that dominates . Assign the -th points of and to . It is easy to verify that each cluster has radius 1, so the total cost is . Also in each cluster, the proportion of points of is at least , the proportion of points of is at least , and the same for . So, is a fair representational clustering.
Now, suppose there is a fair representational clustering of with clusters of cost . First, we argue that each cluster has a non-zero radius, as otherwise, it is a singleton cluster, which is not fair. This implies that the radius of each cluster is 1. Now, each cluster center must be in , as the distance between any two points is 2, excluding them to be a center of a fair cluster of radius 1. It follows that any point is at a distance at most from the cluster centers. Hence, these vertices in form a dominating set of size .
Since it is not possible to solve Dominating set in time unless ETH is false [17], we obtain Theorem 2.
Theorem 2. [Restated, see original statement.]
Fair Representational Clustering cannot be solved in time unless ETH is false.
References
- [1] Sara Ahmadian and Chaitanya Swamy. Approximation algorithms for clustering problems with lower bounds and outliers. In Ioannis Chatzigiannakis, Michael Mitzenmacher, Yuval Rabani, and Davide Sangiorgi, editors, 43rd International Colloquium on Automata, Languages, and Programming, ICALP, volume 55 of LIPIcs, pages 69:1–69:15. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2016. doi:10.4230/LIPICS.ICALP.2016.69.
- [2] Sara Ahmadian and Chaitanya Swamy. Approximation Algorithms for Clustering Problems with Lower Bounds and Outliers. In Ioannis Chatzigiannakis, Michael Mitzenmacher, Yuval Rabani, and Davide Sangiorgi, editors, 43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016), volume 55 of Leibniz International Proceedings in Informatics (LIPIcs), pages 69:1–69:15, Dagstuhl, Germany, 2016. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ICALP.2016.69.
- [3] Vijay Arya, Naveen Garg, Rohit Khandekar, Adam Meyerson, Kamesh Munagala, and Vinayaka Pandit. Local search heuristics for k-median and facility location problems. SIAM J. Comput., 33(3):544–562, 2004. doi:10.1137/S0097539702416402.
- [4] Sayan Bandyapadhyay, Eden Chlamtáč, Yury Makarychev, and Ali Vakilian. A polynomial-time approximation for pairwise fair -median clustering. arXiv preprint arXiv:2405.10378, 2024. doi:10.48550/arXiv.2405.10378.
- [5] Sayan Bandyapadhyay, William Lochet, and Saket Saurabh. FPT constant-approximations for capacitated clustering to minimize the sum of cluster radii. In Erin W. Chambers and Joachim Gudmundsson, editors, 39th International Symposium on Computational Geometry, SoCG 2023, June 12-15, 2023, Dallas, Texas, USA, volume 258 of LIPIcs, pages 12:1–12:14. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2023. doi:10.4230/LIPICS.SOCG.2023.12.
- [6] Sandip Banerjee, Yair Bartal, Lee-Ad Gottlieb, and Alon Hovav. Improved fixed-parameter bounds for min-sum-radii and diameters k-clustering and their fair variants. Proceedings of the AAAI Conference on Artificial Intelligence, 39(15):15481–15488, April 2025. doi:10.1609/aaai.v39i15.33699.
- [7] Suman Bera, Deeparnab Chakrabarty, Nicolas Flores, and Maryam Negahbani. Fair algorithms for clustering. In Advances in Neural Information Processing Systems, pages 4954–4965, 2019.
- [8] Ioana O Bercea, Martin Groß, Samir Khuller, Aounon Kumar, Clemens Rösner, Daniel R Schmidt, and Melanie Schmidt. On the cost of essentially fair clusterings. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2019). Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPIcs.APPROX-RANDOM.2019.18.
- [9] Moritz Buchem, Katja Ettmayr, Hugo KK Rosado, and Andreas Wiese. A (3)-approximation algorithm for the minimum sum of radii problem with outliers and extensions for generalized lower bounds. In Proceedings of the 2024 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 1738–1765. SIAM, 2024.
- [10] Lena Carta, Lukas Drexler, Annika Hennes, Clemens Rösner, and Melanie Schmidt. FPT Approximations for Fair k-Min-Sum-Radii. In Julián Mestre and Anthony Wirth, editors, 35th International Symposium on Algorithms and Computation (ISAAC 2024), volume 322 of Leibniz International Proceedings in Informatics (LIPIcs), pages 16:1–16:18, Dagstuhl, Germany, 2024. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ISAAC.2024.16.
- [11] Lena Carta, Lukas Drexler, Annika Hennes, Clemens Rösner, and Melanie Schmidt. Fpt approximations for fair -min-sum-radii, 2024. doi:10.48550/arXiv.2410.00598.
- [12] Moses Charikar and Rina Panigrahy. Clustering to minimize the sum of cluster diameters. J. Comput. Syst. Sci., 68(2):417–441, 2004. doi:10.1016/j.jcss.2003.07.014.
- [13] Xianrun Chen, Dachuan Xu, Yicheng Xu, and Yong Zhang. Parameterized approximation algorithms for sum of radii clustering and variants. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 20666–20673, 2024. doi:10.1609/AAAI.V38I18.30053.
- [14] Anshuman Chhabra, Karina Masalkovaitė, and Prasant Mohapatra. An overview of fairness in clustering. IEEE Access, 2021. doi:10.1109/ACCESS.2021.3114099.
- [15] Flavio Chierichetti, Ravi Kumar, Silvio Lattanzi, and Sergei Vassilvitskii. Fair clustering through fairlets. In Advances in Neural Information Processing Systems, pages 5029–5037, 2017. URL: https://proceedings.neurips.cc/paper/2017/hash/978fce5bcc4eccc88ad48ce3914124a2-Abstract.html.
- [16] Vincent Cohen-Addad, Anupam Gupta, Lunjia Hu, Hoon Oh, and David Saulpic. An improved local search algorithm for k-median, 2021. arXiv:2111.04589.
- [17] Marek Cygan, Fedor V Fomin, Łukasz Kowalik, Daniel Lokshtanov, Dániel Marx, Marcin Pilipczuk, Michał Pilipczuk, and Saket Saurabh. Parameterized algorithms, volume 5. Springer, 2015. doi:10.1007/978-3-319-21275-3.
- [18] Lukas Drexler, Annika Hennes, Abhiruk Lahiri, Melanie Schmidt, and Julian Wargalla. Approximating fair k-min-sum-radii in euclidean space. In International Workshop on Approximation and Online Algorithms, pages 119–133. Springer, 2023. doi:10.1007/978-3-031-49815-2_9.
- [19] Arnold Filtser and Ameet Gadekar. Fpt approximations for capacitated sum of radii and diameters. arXiv preprint arXiv:2409.04984, 2024. doi:10.48550/arXiv.2409.04984.
- [20] Zachary Friggstad and Mahya Jamshidian. Improved polynomial-time approximations for clustering with minimum sum of radii or diameters. In Shiri Chechik, Gonzalo Navarro, Eva Rotenberg, and Grzegorz Herman, editors, 30th Annual European Symposium on Algorithms, ESA 2022, September 5-9, 2022, Berlin/Potsdam, Germany, volume 244 of LIPIcs, pages 56:1–56:14. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2022. doi:10.4230/LIPIcs.ESA.2022.56.
- [21] Matt Gibson, Gaurav Kanade, Erik Krohn, Imran A. Pirwani, and Kasturi R. Varadarajan. On clustering to minimize the sum of radii. SIAM J. Comput., 41(1):47–60, 2012. doi:10.1137/100798144.
- [22] Teofilo F Gonzalez. Clustering to minimize the maximum intercluster distance. Theoretical Computer Science, 38:293–306, 1985. doi:10.1016/0304-3975(85)90224-5.
- [23] Dorit S Hochbaum and David B Shmoys. A best possible heuristic for the k-center problem. Mathematics of operations research, 10(2):180–184, 1985. doi:10.1287/MOOR.10.2.180.
- [24] Tanmay Inamdar and Kasturi Varadarajan. Capacitated Sum-Of-Radii Clustering: An FPT Approximation. In Fabrizio Grandoni, Grzegorz Herman, and Peter Sanders, editors, 28th Annual European Symposium on Algorithms (ESA 2020), volume 173 of Leibniz International Proceedings in Informatics (LIPIcs), pages 62:1–62:17, Dagstuhl, Germany, 2020. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ESA.2020.62.
- [25] Ragesh Jaiswal, Amit Kumar, and Jatin Yadav. FPT approximation for capacitated sum of radii. In Venkatesan Guruswami, editor, 15th Innovations in Theoretical Computer Science Conference, ITCS 2024, January 30 to February 2, 2024, Berkeley, CA, USA, volume 287 of LIPIcs, pages 65:1–65:21. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024. doi:10.4230/LIPICS.ITCS.2024.65.
