New Results on a General Class of Minimum Norm Optimization Problems
Abstract
We study the general norm optimization for combinatorial problems, initiated by Chakrabarty and Swamy (STOC 2019). We propose a general formulation that captures a large class of combinatorial structures: we are given a set of weighted elements and a family of feasible subsets . Each subset is called a feasible solution/set of the problem. We denote the value vector by , where is the value of element . For any subset , we use to denote the -dimensional vector (i.e., we zero out all entries that are not in ). Let be a symmetric monotone norm function. Our goal is to minimize the norm objective over feasible subset . The problem significantly generalizes the corresponding min-sum and min-max problems.
We present a general equivalent reduction of the norm minimization problem to a multi-criteria optimization problem with logarithmic budget constraints, up to a constant approximation factor. Leveraging this reduction, we obtain constant factor approximation algorithms for the norm minimization versions of several covering problems, such as interval cover, multi-dimensional knapsack cover, and logarithmic factor approximation for set cover. We also study the norm minimization versions for perfect matching, - path and - cut. We show the natural linear programming relaxations for these problems have a large integrality gap. To complement the negative result, we show that, for perfect matching, it is possible to obtain a bi-criteria result: for any constant , we can find in polynomial time a nearly perfect matching (i.e., a matching that matches at least proportion of vertices) and its cost is at most times of the optimum for perfect matching. Moreover, we establish the existence of a polynomial-time -approximation algorithm for the norm minimization variant of the - path problem. Specifically, our algorithm achieves an -approximation with a time complexity of , where .
Keywords and phrases:
Approximation Algorithms, Minimum Norm Optimization, Linear ProgrammingCategory:
Track A: Algorithms, Complexity and GamesCopyright and License:
![[Uncaptioned image]](x1.png)
2012 ACM Subject Classification:
Mathematics of computing Approximation algorithms ; Mathematics of computing Combinatorial optimization ; Mathematics of computing Linear programmingEditors:
Keren Censor-Hillel, Fabrizio Grandoni, Joël Ouaknine, and Gabriele PuppisSeries and Publisher:

1 Introduction
In many optimization problems, a feasible solution typically induces a multi-dimensional value vector (e.g., by the subset of elements of the solution), and the objective of the optimization problem is to minimize either the total sum (i.e., norm) or the maximum (i.e., norm) of the vector entry. For example, in the minimum perfect matching problem, the solution is a subset of edges and the induced value vector is the weight vector of the matching (i.e., each entry of the vector is the weight of edge if the edge is in the matching and 0 for a non-matching edge) and we would like to minimize the total sum. Many of such problems are fundamental in combinatorial optimization but require different algorithms for their min-sum and min-max variants (and other possible variants). Recently there have been a rise of interests in developing algorithms for more general objectives, such as norms [4, 24], top- norms [22], ordered norms [9, 12] and more general norms [13, 14, 31, 19, 1, 33], as interpolation or generalization of min-sum and min-max objectives. The algorithmic study of such generalizations helps unify, interpolate and generalize classic objectives and algorithmic techniques.
The study of approximation algorithm for general norm minimization problems is initiated by Chakrabarty and Swamy [13]. They studied two fundamental problems, load balancing and -clustering, and provided constant factor approximation algorithm for these problems. For load balancing, the induced value vector is the vector of machine loads and for -clustering the vector is the vector of service costs. Subsequently, the norm minimization has been studied for a variety of other combinatorial problem such as general machine scheduling problem [19], stochastic optimization problems [31], online algorithms [38], parameterized algorithms [1] etc. In this paper, we study the norm optimization problem for a general set of combinatorial problems. In our problem, a feasible set is a subset of elements and the multi-dimensional value vector is induced by the subset of elements of the solution. Our problem is defined formally as follows:
Definition 1 (The Norm Minimization Problem (MinNorm)).
We are given a set of weighted elements and a family of feasible subsets . Each subset is called a feasible solution/set of the problem. We denote the value vector by , where is the value of element . We say a subset feasible if . For any subset , we use to denote the -dimensional vector (i.e., we zero out all entries that are not in ), and we call the value vector induced by . Let be a symmetric monotone norm function. Given the norm function , our goal is to find a feasible solution in such that the norm of the value vector induced by the solution is minimized, i.e., we aim to solve the following optimization problem
Note that the case is the most studied min-sum objective and we call the corresponding problem the original optimization problem. Other interesting norms include norms, Top- norms (the sum of top- entries), ordered norms (see its definition in Section 3). Note that our general framework covers the -clustering studied in [13]: in the -clustering problem, the universe is the set of edges and each feasible solution in is a subset of edges that corresponds to a -clustering. The load balancing problem does not directly fit into our framework, since one needs to first aggregate the processing times to machine loads, then apply the norm.
Before stating our results, we briefly mention some results that are either known or very simple to derive.
-
1.
(Matroid) Suppose the feasible set is a matroid and a feasible solution is a basis of this matroid. In fact, the greedy solution (i.e., the optimal min-sum solution) is the optimal solution for any monotone symmetric norm. This is a folklore result and can be easily seen as follows: First, it is easy to establish the following observation, using the exchange property of matroid: We use to denote the sum of largest elements of . For any and any basis , where is the basis obtained by the greedy algorithm. Then using the majorization lemma by Hardy, Littlewood and Pòlya (Lemma 3), we can conclude is optimal for any monotone symmetric norm.
-
2.
(Vertex Cover) We first relax the problem to the following convex program:
The objective is convex since is norm (in particular the triangle inequality of norm). Then, we solve the convex program and round all to and others to . It is easy to see this gives a 2-approximation (using the property for ).
-
3.
(Set Cover) The norm-minimization set cover problem is a special case of the generalized load balancing problem introduced in [19]. Here is the reduction: each element corresponds to a job and each subset to a machine; if element is in set , the processing time , otherwise ; the inner norm of each machine is the max norm (i.e., ) and the outer norm is . Hence, this implies an -approximation for norm-minimization set cover problem using the general result in [19]. The algorithm in [19] is based on a fairly involved configuration LP. In the full version [17], we also provide a much simpler randomized rounding algorithm that is also an -approximation. Note this is optimal up to a constant factor given the approximation hardness of set cover [23, 21].
- 4.
Our Contributions
Our technical contribution can be summarized as follows:
-
1.
(Theorem 5) We present a general reduction of the norm minimization problem to a multi-criteria optimization problem with logarithmic budget constraints, up to a constant approximation factor. This immediately implies an -approximation for the MinNorm problem if there is a poly-time -approximation for the corresponding weight minimization problem (See Theorem 6).
-
2.
Leveraging the reduction in Theorem 5, we obtain constant factor approximation algorithms for the norm minimization versions of several covering problems, such as interval covering (Theorem 17), multi-dimensional knapsack cover (Theorem 9 and Theorem 10). These algorithms are based on rounding the natural linear programming relaxation of the multi-criteria optimization problem, possibly with a careful enumeration of partial solutions. For set cover, we obtain a simple randomized approximation algorithm with approximation factor (See full version [17]), which is much simpler than the general algorithm in [19].
-
3.
We also study the norm minimization versions for perfect matching, - path and - cut. We show the natural linear programming relaxations for these problems have a large integrality gap (Theorem 21 and a second theorem provided only in the full version [17]). This indicates that it may be difficult to achieve constant approximation factors for these problems.
-
4.
To complement the above negative result, we show that, for perfect matching, it is possible to obtain a bi-criteria approximation: for any constant , we can find a nearly perfect matching that matches at least proportion of vertices and the norm of this solution is at most times of the optimum for perfect matching where is any positive real constant (Theorem 24).
-
5.
We present an approximate dynamic programming approach that yields a -approximation -time algorithm for the min-norm - path problem for (Theorem 22), demonstrating an alternative technique for solving norm minimization problems beyond LP rounding.
2 Related Work
Top- and Ordered Optimization.
As a special case of general norm optimization, ordered optimization for combinatorial optimization problems have received significant attention in the recent years. In fact, an ordered norm can be written as a conical combination of top- norms (see Claim 2). The ordered k-median problem was first studied by Byrka et al. [9] and Aouad and Segev [2]. Byrka et al. [9] obtained the first constant factor approximation algorithm (the factor is ). Independently, Chakrabarty and Swamy [12] obtained an algorithm with approximation factor for the top- norm), which can be combined with the enumeration procedure of Aouad and Segev [2] to get the same factor for the general ordered -median. The current best known approximation is , by Chakrabarty and Swamy [13]. Deng and Zhang [20] studied ordered -median with outliers and obtained a constant factor approximation algorithm. Maalouly and Wulf [22] studied the top- norm optimization for the matching problem and obtained an polynomial time exact algorithm (see full version [17]). Braverman et al. studied coreset construction for ordered clustering problems [7] which was motivated by applications in machine learning. Batra et al. [5] studied the ordered min-sum vertex cover problem and obtained the first poly-time approximation approximation with approximation factor .
General Symmetric Norm Optimization.
Chakrabarty and Swamy [13] first studied general monotone symmetric norm objectives for clustering and unrelated machine load balancing and obtained constant factor approximation algorithms, substantially generalizing the results for -Median and -Center and makespan minimization for unrelated machine scheduling. In a subsequent paper [14], they obtained a simpler algorithm for load balancing that achieves an approximation factor of . Abbasi et al. [1] studied the parametrized algorithms for the general norm clustering problems and provided the first EPAS (efficient parameterized approximation scheme). Deng et al. [19] introduced the generalized load balancing problem, which further generalizes the problem studied by [14]. In the generalized load balancing problem, the load of a machine is a symmetric, monotone (inner) norm of the vector of processing times of jobs assigned to . The generalized makespan is another (outer) norm aggregating the loads. The goal is to find an assignment of jobs to minimize the generalized makespan. They obtained a logarithmic factor approximation, which is optimal up to constant factor since the problem generalizes the set cover problem. For the special case where the inner norms are top- norms, Ayyadevara et al. [3] showed the natural configuration LP has a integrality gap.
Submodular/Supermodular Optimization.
Optimizing submodular/supermodular function under various combinatorial constraints is another important class of optimization problems with general objectives and has been studied extensively in the literature. See e.g., [10, 35, 15, 8] and the survey [34]. However, note that results for submodular functions does not imply results for general symmetric monotone norms, since a general symmetric monotone norm is not necessarily a submodular function (see e.g., [19]).
Patton et al. [38] studied submodular norm objectives (i.e., norms that also satisfies continuous submodular property). They showed that it can approximate well-known classes of norms, such as norms, ordered norms, and symmetric norms and applied it to a variety of problems such as online facility location, stochastic probing, and generalized load balancing. Recently, Kesselheim et al. [33] introduced the notion of p-supermodular norm and showed that every symmetric norm can be approximated by a p-supermodular norm. Leveraging the result, they obtain new algorithms online load-balancing and bandits with knapsacks, stochastic probing and so on.
Multi-budgeted Optimization.
There is a body of literature in the problem of optimizing linear or submodular objectives over a combinatorial structure with additional budget constraints (see e.g., [39, 11, 26, 27, 16, 6, 25]). For a single budget constraint, randomized or deterministic PTASes have been developed for various combinatorial optimization problems (e.g. spanning trees with a linear budget [39]). Assuming that a pseudopolynomial time algorithm for the exact version of the problems exists, Grandoni and Zenklusen showed that one can obtain a PTAS for the corresponding problem with any fixed number of linear budgets [27]. More powerful techniques such as randomized dependent rounding and iterative rounding have been developed to handle more general submodular objectives and/or other combinatorial structures such as matroid or intersection of matroid (e.g., [26, 27, 16, 25]). Iterative rounding technique [26, 37] has been used in general norm minimization problems [13, 14]. Our algorithms for matching (Section 9) and knapsack cover (Section 5) also adopt the technique.
3 Preliminaries
Throughout this paper, for vector , define as the non-increasingly sorted version of , and for any . Let be the top- norm that returns the sum of the largest absolute values of entries in any vector, . Denote as the set of positive integers no larger than , and .
We say function is a norm if: (i) if and only if , (ii) for all , (iii) for all . A norm is monotone if for all , and symmetric if for any permutation of . We are also interested in the following special monotone symmetric norms.
Top- norms.
Let . A function is a Top- norm, denoted by , if for each input vector it returns the sum of the largest absolute values of entries in . For non-negative vectors, it simply returns the sum of the largest entries. We notice that by letting , recovers the and norms, respectively, thus it generalizes the latter two.
Ordered norms.
Let be a non-increasing non-negative vector. For each vector , let denote its non-increasingly sorted version and define . A function is a -ordered norm (or simply an ordered norm), denoted by , if for each input vector it returns the inner product of and ; we obtain whenever . It is easy to see that, by having as a vector of 1s followed by 0s, recovers . On the other hand, it is known that each ordered norm can be written as a conical combination of Top- norms, as in the following claim.
Claim 2.
The following lemma is due to Hardy, Littlewood and Pòlya. [29].
Lemma 3.
([29]). If and satisfy for each , one has for any symmetric monotone norm .
4 A General Reduction to Multi-Budgeted Optimization Problem
In this section, we provide an equivalent formulation for the general symmetric norm minimization problem MinNorm (up to constant approximation factor). Recall that as defined in Section 1, we are given a set of elements, and represents a family of feasible subsets of . The goal of MinNorm is to find a feasible subset to minimize , where is a symmetric monotone norm function. We say that we find a -approximation for the problem for some , if we can find an such that , where is the optimal solution. Since a general norm function is quite abstract and hard to deal with, we formulate the following (equivalent, up to constant approximation factor) optimization problem which is more combinatorial in nature.
Definition 4 (Logarithmic Budgeted Optimization (LogBgt)).
The input of a Logarithmic Budgeted Optimization Problem is a tuple , where:
-
is a finite set with elements.
-
are disjoint subsets of , where is the number of sets. For , We refer to as the -th group, and for any , we call the group index of .
-
is a family of feasible subsets of . The size of may be exponentially large in , but we ensure that there exists a polynomial-time algorithm to decide whether for a subset .
For any , we say a subset is a -valid solution if and only if:
-
1.
satisfies the feasibility constraint, i.e., , and
-
2.
for all .
For any , we define -LogBgt problems as follows: Given an input , the goal is to find a -valid solution or certify that there is no -valid solution. In particular, we denote -LogBgtas -LogBgt.
Notice that the structure of a problem is defined by and (for example, the vertex cover problem is given by vertex set and contains all subsets of corresponding to a vertex cover), each problem corresponds to a MinNorm version and an LogBgt version. We show that solving LogBgt is equivalent to approximating MinNorm, up to constant approximation factors. In fact, the reduction from norm approximation to optimization problem with multiple budgets has been implicitly developed in prior work [13, 32]. For generality and ease of usage, we encapsulate the reduction in the following general theorem.
Theorem 5.
For any ( can depend on ) and , if we can solve -LogBgt in polynomial time, we can approximate the MinNorm problem within a factor of in polynomial time. On the other hand, if we can find a -approximation for MinNorm in polynomial time, we can solve the -LogBgt in polynomial time.
We defer the full details to the full version [17] and now outline the main ideas behind the proof.
Reducing MinNorm to LogBgt.
Suppose we are given an instance of the MinNorm problem and have access to an algorithm for LogBgt. Let denote an optimal solution for MinNorm, and let be the vector of values sorted in nonincreasing order. Our strategy is to “guess” an approximation of by enumerating candidate threshold vectors (each threshold vector is of the form , i.e., we guess the values at positions ) One can show (The details can be found in the full version [17]) that this enumeration can be done in poly-time and is guaranteed to include a candidate that is “close” to the true vector .
For each candidate enumeration , we construct an associated LogBgt instance as follows. For each index and for each candidate threshold indexed by (with ), if we assign index to a set . Additionally, every index for which is placed in the last set . With this construction, when the guessed threshold vector is “close” to , solving the resulting LogBgt instance produces a solution that is within a constant factor of the optimal MinNorm value.
Reducing LogBgt to MinNorm.
In the reverse direction, suppose we are given an instance of the LogBgt problem and a -approximation algorithm for MinNorm. In this case, we construct a corresponding MinNorm instance by setting, for each element , a value that reflects its membership in one of the sets (or its exclusion from all such sets). The idea is that the structure of the LogBgt instance is encoded in the chosen values so that a -approximation for the resulting MinNorm instance yields a solution for the original LogBgt problem. In particular, one can show that a -approximation for MinNorm leads to a -approximation for LogBgt.
A Logarithmic Approximation.
Based on Theorem 5, we can easily deduce the following general theorem. We use to denote a general combinatorial optimization problem with the min-sum objective function , where we write and is the set of feasible solutions.
Theorem 6.
If there is a poly-time approximation algorithm for the min-sum problem (with approximation factor ), there is a poly-time factor approximation algorithm for the corresponding MinNorm problem for any fixed constant .
Proof.
By Theorem 5, we just need to find a poly-time approximation algorithm for the LogBgt version. Consider the input where (recall the definitions in Section 4). Then we construct for each by:
-
1.
if ;
-
2.
if .
Then, if there is a -valid solution for the LogBgt problem, we can see that there is a feasible set with . Then, by the assumption of the theorem, the approximation algorithm for can output a feasible solution with . This further implies that is a -valid solution, because
This means that the -LogBgt problem can be solved in polynomial time. By Theorem 5, we complete the proof.
5 Multi-dimensional Knapsack Cover Problem
In this section, we consider the multi-dimensional knapsack cover problem defined as follows.
Definition 7 (Min-norm -dimensional Knapsack Cover Problem (MinNorm-KnapCov)).
Let be a positive integer. We are given a set of items , where each item has a weight vector . The feasible set is defined as:
Now, given a symmetric monotone norm and a value vector , we can define the norm minimization problem for -dimensional Knapsack Cover and denote it as MinNorm-KnapCov.
In light of Theorem 5, we introduce disjoint sets and consider the LogBgt problem with , which We denote as LogBgt-KnapCov. We consider LogBgt-KnapCov for two cases: (1) and (2) ( will be defined in Section 5.2). For both cases, we use the following natural linear programming formulation for LogBgt-KnapCov :
(LP-KnapCover-1) | ||||||||
For both cases, we develop a method called partial enumeration. Partial enumeration lists a subset of possible partial solutions for the first several groups. Here is the complete definition:
Definition 8 (Partial Enumeration).
For a -LogBgt problem with set , the partial enumeration algorithm first determine a quantity (depending on the problem at hand). The partial enumeration algorithm returns a subset . Each element of is a partial solution (Recall the definition of partial solution: ), and this algorithm ensures:
-
1.
If there exists a -valid solution, then at least one partial solution satisfies that there exists an -valid extended solution (A solution is called an extended solution of a partial solution if for all ).
-
2.
The size of is polynomial, and this partial enumeration algorithm runs in polynomial time.
5.1 An Algorithm for
When , for any , we choose and apply partial enumeration. Recall that a partial solution can be described as a vector , where for . Our partial enumeration algorithm in this section simply outputs the set of all possible partial solutions (i.e., where ).
For each partial solution, we use it to modify the Linear Program ˜LP-KnapCover-1 and perform a rounding algorithm. The rounding algorithm do the following steps:
-
1.
Modify LP-KnapCover-1 by removing according to the partial solution. Then confirm there is no solution or find an extreme point .
-
2.
Round all to for
We can prove that the rounding algorithm can obtain a -valid solution based on a specific partial solution. Ultimately, we obtain a -valid solution or ensure no -valid solution for this enumeration result, and this algorithm runs in -time. The details are provided in the full version [17].
Theorem 9.
If is a constant, then for any constant , there exists an polynomial-time algorithm which can solve -LogBgt-KnapCov. Thus we have a polynomial-time -approximation algorithm for MinNorm-KnapCov when .
5.2 An Algorithm for Larger d
In this subsection, we provide a polynomial-time constant-factor approximation algorithm for , where is defined as
To ensure there exist valid solutions, must be a non-empty set.
The second algorithm employs the same rounding procedure but modifies the partial enumeration method. The new algorithm choose . For , it partitions into multiple subsets based on vectors of size , which represent the logarithms of weights. Instead of enumerating all subsets, we only enumerate the number of elements within each subset and then take double the number of any elements in this subset. Further details are provided in the full version [17].
We ensure that this partial enumeration algorithm runs in polynomial time if . In addition, the rounding procedure can find a -valid solution based on at least one partial solution. The complete details are presented in the full version [17].
Theorem 10.
There exists a algorithm that can solve -LogBgt-KnapCov when . Thus we have a -approximation algorithm for MinNorm-KnapCov with .
6 Interval Cover Problem
In this section, we study the norm minimization for the interval cover problem, which is defined as follows:
Definition 11 (Min-norm Interval Cover Problem (MinNorm-IntCov)).
Given a set of intervals and a target interval on the real axis, a feasible solution of this problem is a subset such that fully covers the target interval (i.e., ). Suppose is a value vector and is a monotone symmetric norm function. Our goal is to find a feasible subset such that is minimized. We denote the problem as MinNorm-IntCov.
In light of Theorem 5, we can focus on obtaining a constant-factor approximation algorithm for -LogBgt-IntCov. The input of a LogBgt-IntCov problem is a tuple , which is the LogBgt problem with input where the set of feasible solutions is .
In this section, we begin by transforming the interval cover problem into a new problem called the tree cover problem. The definitions of both problems are provided in Section 6.1. These transformation results in only a constant-factor loss in the approximation factor (i.e., if a polynomial-time algorithm can solve -LogBgt-TreeCov for some constant , then there exists a polynomial-time constant-factor approximation algorithm for LogBgt-IntCov).
Next, we focus on LogBgt-TreeCov. We first employ the partial enumeration algorithm, as defined in Section 5, to list partial solutions for the first sets. The details of this process are provided in Section 6.2. Following partial enumeration, we apply a rounding algorithm to evaluate each partial solution. The entire rounding process is detailed in Section 6.3.
6.1 From Interval Cover to Tree Cover
We first introduce some notations for the tree cover problem. Denote a rooted tree as , where forms an undirected tree and is the root. For each node , let be the set of children of , and be the set of all descendants of (including ). It is easy to see that . For a subset of vertices , we define . We also define as the parent node of and define as the set of ancestors of (, and for any , ). In addition, we define the set of leaves .
Definition 12 (LogBgt Tree Cover Problem (LogBgt-TreeCov)).
We are given a tuple , where is a rooted tree and . , and is a partition of (and is called the th group), and is not in any group. The partition satisfies the following property: For any node and an arbitrary child of , belongs to group with . For each , we define if . In particular, we denote . So for all .
A feasible solution for the tree cover problem is a subset such that the descendants of covers all leaves. Formally, the feasible set is defined as
We prove the following theorem to reduce the interval cover problem to the tree cover problem. The proof of the theorem can be found in the full version [17].
Theorem 13.
If there exists a polynomial-time algorithm for the -LogBgt-TreeCov problem, then there exists a polynomial-time algorithm for the -LogBgt-IntCov problem.
Based on this theorem, we mainly need to deal with the tree cover problem in the following subsections.
6.2 Partial Enumeration Method for Tree Cover Problem
In this subsection, we present a partial enumeration algorithm for the LogBgt-TreeCov problem. Recall that we introduced the concept of partial enumeration in Section 5. For a LogBgt-TreeCov problem with input , where is a rooted tree, and . we set and perform partial enumeration for the first sets. The goal is to find a set such that there exists a partial solution satisfying: at least one of the partial solution can be extended to a -valid solution for some constant .
In this subsection, we define as the group index of for . Now we focus on the LogBgt-TreeCov problem. For each , define the first type of cost . We then define the second type of cost:
Intuitively, the cost represents the “cost” of selecting , as it indicates the proportion of the group that occupies. Meanwhile, denotes the minimum cost required to cover using its descendants.
We now present the partial enumeration algorithm. The pseudo-code can be found in the full version [17]. Here, we briefly describe the main idea of the partial enumeration algorithm.
We employ a depth-first search (DFS) strategy to explore most of the states in the search space. During the search process, we maintain two sets:
-
, representing the set of candidate elements that can still be explored, i.e., contains all uncovered leaves.
-
, storing the elements that have already been selected as part of the partial solution.
Initially, is the child set of the root, and . At each recursive step, we select with the smallest group index. The recursion proceeds by exploring two possibilities:
-
1.
Adding to the partial solution, i.e., including in and continuing the search.
-
2.
Excluding from the partial solution, i.e., replacing with its child nodes while keeping unchanged. (If is a leaf, this option is not applicable.)
The search terminates when fails to satisfy at least one of the following conditions:
-
1.
,
-
2.
,
-
3.
-
4.
The first and fourth conditions are derived from the objective of the Partial Enumeration Method. Regarding the second condition, we observe that for all , it holds that and . Furthermore, if , the impact of ignoring is negligible.
The third condition is based on the property that for any -valid solution , it satisfies:
Due to these conditions, for each group where , we only need to determine at most items. Consequently, we can establish that our partial enumeration algorithm runs in polynomial time.
We then present the following theorem:
Theorem 14.
There exists a polynomial-time partial enumeration algorithm for the LogBgt-TreeCov problem with the following guarantee: If the input has a -valid solution, then at least one of the output partial solutions has a -valid extended solution.
The complete proof can be found in the full version [17].
6.3 A Rounding Algorithm for Tree Cover Problem
We now focus on the -LogBgt-TreeCov problem with input , where is a rooted tree. Let be the set of leaves. Recall that represents the set of ancestors of node . For sets and , we express the formulation of the linear program as follows:
(LP-Tree-Cover(c,,)) | ||||||||
We call cardinality constraints, and call feasibility constraints. Recall that . Also, define .
The algorithm is as follows, and the pseudocode is provided in the full version [17]:
-
1.
Check if LP-Tree-Cover(, , ) has a feasible solution. If so, obtain an extreme point . Otherwise, confirm that there is no such integral solution.
-
2.
Remove the leaves with , and delete all the descendants of their parents. Then becomes a leaf. Repeat this process until for each leaf . Let the modified node set and leaf set be and , respectively.
-
3.
For , attempt to round . If , round it to . If , and is not a leaf in , also round it to . In all other cases, round to . Let be the set of nodes for which was rounded to . Note that may not cover .
-
4.
Remove all descendants in , and attempt to choose another set from to cover all leaves. Formalize this objective as LP-Tree-Cover. Specifically,
-
, and
-
.
To understand this, observe that consists of the nodes in the th to th groups that remain uncovered. The set includes nodes in that are either leaves or have at least one uncovered child with a group index greater than (i.e., at least one descendant leaf remains uncovered).
Then solve LP-Tree-Cover(, , ). The fact that this problem must have feasible solutions is proved later, so we do not need to consider the case of no solution.
-
-
5.
Let be an extreme point of LP-Tree-Cover(, , ). For each , round it to if and only if . Let , then covers .
-
6.
Combine the three parts of the solution. That is, return .
We prove the following lemma for the above algorithm:
Lemma 15.
There exists a polynomial-time algorithm such that, if and a partial solution has a -valid extended solution, then the algorithm returns a -valid solution.
By combining Theorem 14 and Lemma 15, we establish the following theorem:
Theorem 16.
For any , there exists a polynomial-time algorithm for -LogBgt-TreeCov.
Furthermore, applying Theorem 13 and Theorem 16, we obtain the following result:
Theorem 17.
There exists a polynomial-time algorithm that solves the -LogBgt-IntCov. Consequently, we obtain a polynomial-time constant-factor approximation algorithm for MinNorm-IntCov.
The complete proofs are provided in the full version [17].
7 Integrality Gap for Perfect Matching, s-t Path, and s-t Cut
In this section, we argue that it may be challenging to achieve constant approximations for the norm optimization problems for perfect matching, - path, and - cut just by LP rounding. We show that the natural linear programs have large integrality gaps.
Definition 18 (Min-Norm - Path Problem (MinNorm-Path)).
Given a directed graph (here is the set of vertices and is the set of edges) and nodes , define the feasible set:
For the MinNorm version, we are also given a monotone symmetric norm and a value vector . The goal is to select an - path that minimizes .
In light of Theorem 5, we can define the LogBgt-Path problem with input tuple , which is the LogBgt problem defined in Definition 4 with input .
Definition 19 (Min-Norm Perfect Matching Problem (MinNorm-PerMat)).
Given a bipartite graph with , define the feasible set:
For the MinNorm version, we are also given a monotone symmetric norm and a value vector . The goal is to select that minimizes .
We define the LogBgt-PerMat problem with as the LogBgt problem with .
Due to Theorem 5, we establish the equivalence between approximating the MinNorm problem and the LogBgt problem. Hence, if we can show that the LogBgt version is hard to approximate, then the same hardness (up to constant factor) also applies to the MinNorm version.
Now, we consider using the linear programming rounding approach to approximate the LogBgt problem with . Such an algorithm proceeds according to the following pipeline:
-
First, we formulate the natural linear program (for ) as the following:
(LP-LBO-Original(, )) satisfies -
Then, to solve the -LogBgt problem, we first check whether LP-LBO-Original(, ) has feasible solutions. If feasible, we use a rounding algorithm to find an integral solution for (LP-LBO-Original(, )).
Since the factor in LP-LBO-Original(, ) determines the approximation factor, we study the following linear program.
(LP-LBO()) | |||||||
satisfies | |||||||
Clearly, if LP-LBO-Original(, ) has a feasible solution, the optimal value of LP-LBO() is at most . Suppose we can round a fractional solution LP-LBO-Original(, ) to an integral feasible -valid solution for some constant (i.e., satisfies and ). Then, we get an integral solution for LP-LBO() with objective value , contradicting the fact that the integrality gap of LP-LBO() is . Hence, we can conclude that if the integrality gap of LP-LBO() is , it would be difficult to derive a constant factor approximation algorithm for both LogBgt and MinNorm-version of the problem using the LP LP-LBO-Original().
7.1 Reduction from Perfect Matching to s-t Path
For an LogBgt perfect matching problem with , where is a bipartite graph, we consider the following LP (on the left). The LP on the right is for LogBgt-Path with .
We have the following theorem showing that LogBgt-Path problem is not harder than LogBgt-PerMat problem.
Theorem 20.
For any arbitrary function , We have the following conclusions:
-
(a)
If the integrality gap of (LP-LBO-PM()) is no more than for all instances of the LogBgt-PerMat problem, then the integrality gap of (LP-LBO-Path()) is also for all instances of the LogBgt-Path problem.
-
(b)
If we have a polynomial-time -approximation algorithm for the MinNorm-PerMat problem, then we have a polynomial-time -approximation algorithm for the MinNorm-Path problem.
7.2 Integrality Gaps for Min-Norm s-t Path and Min-Norm Perfect Matching
Theorem 21.
For infinitely many , there exists an instance of size such that the integrality gap of (LP-LBO-Path()) can be . Thus, the relaxations of the natural linear programming of both LogBgt-Path and LogBgt-PerMat have integrality gaps of .
The proof of the above theorem can be found in the full version [17]. Also, we have the result about the integrality gap for - cut, which is also omitted and can be found in the full version.
Remark. In the example in the proof (see the full version [17] for details), the gap between any feasible subset of ( is the fractional solution) and the fractional solution is larger than any given constant. Thus any rounding algorithm that deletes zero-value variables (including the rounding algorithm we developed in this paper and the iterative rounding method in [13]) cannot successfully yield a constant-factor approximation.
8 An Algorithm for Min-Norm s-t Path Problem
Recall that we define the MinNorm-Path problem and prove that the natural linear program has a large integrality gap in Section 7. In this section, we provide a factor approximation algorithm that runs in time, for any . In particular, this implies an -factor polynomial-time approximation algorithm and a constant-factor quasi-polynomial -time algorithm for MinNorm-Path. Note that this does not contradict Theorem 21, since we do not use the LP rounding approach in this section.
In light of Theorem 5 with , we consider -LogBgt-Path problem, with input tuple , where and , where is the approximation factor we aim to achieve.
We first provide an overview of our main ideas. A natural approach to solve LogBgt-Path is to employ dynamic programming, in which the states keep track of the number of edges used from each group. However, since we have groups, the number of states may be as large as . To resolve this issue, we perform an approximation dynamic programming, in which we only approximate the number of edges in each group. In particular, the numbers are rounded to the nearest power of after each step, for carefully chosen value (to ensure that we do not lose too much in the rounding). This rounding technique is inspired by a classic approximation algorithm for the subset sum problem [30] Now, the dynamic programming state is a vector that approximates the number of edges used from each group in a path from to with at most edges, where and . The dynamic programming process involves storing the number of selected items in each group and rounding them to the nearest power of at each iteration. However, this method results in a state space of size , which is better than , but still super-polynomial. To resolve the above issue, we need the second idea, which is to trade off the approximation factor and the running time. In particular, we introduce an integer parameter , defined as for some . We then partition the original groups into supergroups, each containing groups. This reduces the number of states to , but incurs a loss of in the approximation factor.
Now, we present the details of our algorithm. Let , and for all . For and , define (specifically, ). Furthermore, we define the vector . It is important to notice that:
-
If is a -valid solution (), then for all .
-
If for all , then is a -valid solution.
In iteration (), for each pair of vertices , we define as a set of vectors in that encodes information about paths from to containing at most edges. Specifically, for any path from to with at most edges, the set includes a corresponding vector that approximates .
-
Initially, . For , it is set to if is an edge, and if is not an edge.
-
In the -th iteration (), we begin by initializing for all . Then, for each pair , we enumerate all vertices and add the sum of and to . Here, the sum of two sets is defined as the set of all pairwise sums of elements from the two sets.
-
To reduce the size of , we round the components of these vectors in to or powers of (recall that ).
Our main result in this section is the following theorem:
Theorem 22.
For any , there exists a -time algorithm for -LogBgt-Path. Thus we have an approximation approximation which runs in -time and achieves an approximation factor of for MinNorm-Path.
The details of the proof are provided in the full version [17].
9 A Bi-criterion Approximation for Matching
In this section, we consider the matching problem. While we do not know how to design a constant factor approximation algorithm for the MinNorm version of perfect matching problem yet, we demonstrate that it is possible to find a nearly perfect matching within a constant approximation factor – a bi-criterion approximation algorithm. In particular, for any given norm, we can find a matching that matches fraction of nodes and its corresponding norm is at most (a constant) times the norm of the optimal integral perfect matching. Note that a constant factor approximation algorithm (without relaxing the perfect matching requirement) is impossible using the natural linear program, due to Theorem 21.
First, we introduce some notations. We are given a bipartite graph , where is the set of nodes of the first color class, is the set of node of the other color class, and is the set of edges. Let . We define . We also study the LogBgt version and we let be the disjoint subsets of .
Definition 23 (-nearly Matching).
Let . We define the following problem as -nearly matching. Given a bipartite graph , a set is a nearly matching if it is a matching with at least edges.
Again, we use the natural linear program for the LogBgt version of the problem. Through an iterative rounding method (similar to that in [13]), we can obtain the following bi-criterion approximation result.
Theorem 24.
If there exists a 1-valid solution for a LogBgt-PerMat problem, then there exists an -time algorithm to obtain a -valid solution for the corresponding LogBgt -nearly matching problem for any .
This theorem implies the following result for MinNorm-PerMat.
Theorem 25.
Given a MinNorm-PerMat problem with a monotone symmetric norm and a value vector , let be an optimal solution for this problem. For any constants , there exists a polynomial-time algorithm to obtain a -nearly matching such that .
The details and variants of the problem are omitted and can be found in the full version [17].
10 Concluding Remarks
In this paper, we propose a general formulation for general norm minimization in combinatorial optimization. Our formulation captures a broad class of combinatorial structures, encompassing various fundamental problems in discrete optimization. Via a reduction of the norm minimization problem to a multi-criteria optimization problem with logarithmic budget constraints, we develop constant-factor approximation algorithms for multiple important covering problems, such as interval cover, multi-dimensional knapsack cover, and set cover (with logarithmic approximation factors). We also provide a bi-criteria approximation algorithm for min-norm perfect matching, and an -approximation algorithm for the min-norm - path problem, via a nontrivial approximate dynamic programming approach.
Our results open several intriguing directions for future research. First, one can explore other combinatorial optimization problems, such as Steiner trees and other network design problems, within our general framework. Additionally, our formulation could be extended to encompass the min-norm load balancing problem studied in [13] (where job processing times are first summed into machine loads before applying a norm), and even the generalized load balancing [19] and cascaded norm clustering problems [18, 1] (which allow for two levels of cost aggregation via norms). Second, obtaining a nontrivial true approximation algorithm for perfect matching – rather than a bi-criterion approximation – remains an important open problem. Third, it is an important open problem whether a polynomial-time constant-factor approximation exists for the min-norm - path problem. Lastly, it would be interesting to study other general objective functions beyond symmetric monotone norms and submodular functions (such as general subadditive functions [28] and those studied in [36]).
References
- [1] Fateme Abbasi, Sandip Banerjee, Jarosław Byrka, Parinya Chalermsook, Ameet Gadekar, Kamyar Khodamoradi, Dániel Marx, Roohani Sharma, and Joachim Spoerhase. Parameterized approximation schemes for clustering with general norm objectives. In 2023 IEEE 64th Annual Symposium on Foundations of Computer Science (FOCS), pages 1377–1399. IEEE, 2023. doi:10.1109/FOCS57990.2023.00085.
- [2] Ali Aouad and Danny Segev. The ordered -median problem: surrogate models and approximation algorithms. Math. Program., 177(1-2):55–83, 2019. doi:10.1007/S10107-018-1259-3.
- [3] Nikhil Ayyadevara, Nikhil Bansal, and Milind Prabhu. On minimizing generalized makespan on unrelated machines. Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, page 21, 2023.
- [4] Yossi Azar and Amir Epstein. Convex programming for scheduling unrelated parallel machines. In Proceedings of the 37th Annual ACM Symposium on Theory of Computing, pages 331–337, 2005. doi:10.1145/1060590.1060639.
- [5] Jatin Batra, Syamantak Das, and Agastya Vibhuti Jha. Tight approximation algorithms for ordered covering. In Algorithms and Data Structures Symposium, pages 120–135. Springer, 2023. doi:10.1007/978-3-031-38906-1_9.
- [6] André Berger, Vincenzo Bonifaci, Fabrizio Grandoni, and Guido Schäfer. Budgeted matching and budgeted matroid intersection via the gasoline puzzle. Mathematical Programming, 128:355–372, 2011. doi:10.1007/S10107-009-0307-4.
- [7] Vladimir Braverman, Shaofeng H-C Jiang, Robert Krauthgamer, and Xuan Wu. Coresets for ordered weighted clustering. In International Conference on Machine Learning, pages 744–753. PMLR, 2019. URL: http://proceedings.mlr.press/v97/braverman19a.html.
- [8] Niv Buchbinder, Moran Feldman, Joseph Naor, and Roy Schwartz. Submodular maximization with cardinality constraints. In Proceedings of the twenty-fifth annual ACM-SIAM symposium on Discrete algorithms, pages 1433–1452. SIAM, 2014. doi:10.1137/1.9781611973402.106.
- [9] Jaroslaw Byrka, Krzysztof Sornat, and Joachim Spoerhase. Constant-factor approximation for ordered -median. In Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, pages 620–631, 2018. doi:10.1145/3188745.3188930.
- [10] Gruia Calinescu, Chandra Chekuri, Martin Pál, and Jan Vondrák. Maximizing a submodular set function subject to a matroid constraint. In International Conference on Integer Programming and Combinatorial Optimization, pages 182–196. Springer, 2007.
- [11] Paolo M. Camerini, Giulia Galbiati, and Francesco Maffioli. Random pseudo-polynomial algorithms for exact matroid problems. Journal of Algorithms, 13(2):258–273, 1992. doi:10.1016/0196-6774(92)90018-8.
- [12] Deeparnab Chakrabarty and Chaitanya Swamy. Interpolating between -median and -center: Approximation algorithms for ordered -median. In 45th International Colloquium on Automata, Languages, and Programming, volume 107 of LIPIcs, pages 29:1–29:14, 2018. doi:10.4230/LIPICS.ICALP.2018.29.
- [13] Deeparnab Chakrabarty and Chaitanya Swamy. Approximation algorithms for minimum norm and ordered optimization problems. In Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, pages 126–137, 2019. doi:10.1145/3313276.3316322.
- [14] Deeparnab Chakrabarty and Chaitanya Swamy. Simpler and better algorithms for minimum-norm load balancing. In 27th Annual European Symposium on Algorithms, volume 144 of LIPIcs, pages 27:1–27:12, 2019. doi:10.4230/LIPICS.ESA.2019.27.
- [15] Chandra Chekuri and Alina Ene. Submodular cost allocation problem and applications. In Automata, Languages and Programming - 38th International Colloquium, volume 6755 of Lecture Notes in Computer Science, pages 354–366, 2011. doi:10.1007/978-3-642-22006-7_30.
- [16] Chandra Chekuri, Jan Vondrák, and Rico Zenklusen. Multi-budgeted matchings and matroid intersection via dependent rounding. In Proceedings of the twenty-second annual ACM-SIAM symposium on Discrete Algorithms, pages 1080–1097. SIAM, 2011. doi:10.1137/1.9781611973082.82.
- [17] Kuowen Chen, Jian Li, Yuval Rabani, and Yiran Zhang. New results on a general class of minimum norm optimization problems, 2025. arXiv:2504.13489.
- [18] Eden Chlamtáč, Yury Makarychev, and Ali Vakilian. Approximating fair clustering with cascaded norm objectives. In Proceedings of the 2022 annual ACM-SIAM symposium on discrete algorithms (SODA), pages 2664–2683. SIAM, 2022. doi:10.1137/1.9781611977073.104.
- [19] Shichuan Deng, Jian Li, and Yuval Rabani. Generalized unrelated machine scheduling problem. In Nikhil Bansal and Viswanath Nagarajan, editors, Proceedings of the 2023 ACM-SIAM Symposium on Discrete Algorithms, SODA 2023, Florence, Italy, January 22-25, 2023, pages 2898–2916. SIAM, 2023. doi:10.1137/1.9781611977554.CH110.
- [20] Shichuan Deng and Qianfan Zhang. Ordered k-median with outliers. In Amit Chakrabarti and Chaitanya Swamy, editors, Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, APPROX/RANDOM 2022, September 19-21, 2022, University of Illinois, Urbana-Champaign, USA (Virtual Conference), volume 245 of LIPIcs, pages 34:1–34:22. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2022. doi:10.4230/LIPICS.APPROX/RANDOM.2022.34.
- [21] Irit Dinur and David Steurer. Analytical approach to parallel repetition. In Symposium on Theory of Computing, pages 624–633, 2014. doi:10.1145/2591796.2591884.
- [22] Nicolas El Maalouly. Exact matching: Algorithms and related problems. In 40th International Symposium on Theoretical Aspects of Computer Science, 2023.
- [23] Uriel Feige. A threshold of ln n for approximating set cover. Journal of the ACM (JACM), 45(4):634–652, 1998. doi:10.1145/285055.285059.
- [24] Daniel Golovin, Anupam Gupta, Amit Kumar, and Kanat Tangwongsan. All-norms and all-l_p-norms approximation algorithms. In IARCS Annual Conference on Foundations of Software Technology and Theoretical Computer Science (2008). Schloss-Dagstuhl-Leibniz Zentrum für Informatik, 2008.
- [25] Fabrizio Grandoni, R. Ravi, Mohit Singh, and Rico Zenklusen. New approaches to multi-objective optimization. Math. Program., 146(1-2):525–554, 2014. doi:10.1007/S10107-013-0703-7.
- [26] Fabrizio Grandoni, Ramamoorthi Ravi, and Mohit Singh. Iterative rounding for multi-objective optimization problems. In European Symposium on Algorithms, pages 95–106. Springer, 2009.
- [27] Fabrizio Grandoni and Rico Zenklusen. Approximation schemes for multi-budgeted independence systems. In European Symposium on Algorithms, pages 536–548. Springer, 2010. doi:10.1007/978-3-642-15775-2_46.
- [28] Anupam Gupta, Viswanath Nagarajan, and Sahil Singla. Adaptivity gaps for stochastic probing: submodular and xos functions. In Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA ’17, pages 1688–1702, USA, 2017. Society for Industrial and Applied Mathematics. doi:10.1137/1.9781611974782.111.
- [29] G. H. Hardy, J. E. Littlewood, and G. Pólya. Inequalities. Cambridge University Press, 1934.
- [30] Oscar H. Ibarra and Chul E. Kim. Fast approximation algorithms for the knapsack and sum of subset problems. J. ACM, 22(4):463–468, October 1975. doi:10.1145/321906.321909.
- [31] Sharat Ibrahimpur and Chaitanya Swamy. Approximation algorithms for stochastic minimum-norm combinatorial optimization. In 61st IEEE Annual Symposium on Foundations of Computer Science, pages 966–977, 2020. doi:10.1109/FOCS46700.2020.00094.
- [32] Sharat Ibrahimpur and Chaitanya Swamy. Minimum-norm load balancing is (almost) as easy as minimizing makespan. In 48th International Colloquium on Automata, Languages, and Programming, volume 198 of LIPIcs, pages 81:1–81:20, 2021. doi:10.4230/LIPICS.ICALP.2021.81.
- [33] Thomas Kesselheim, Marco Molinaro, and Sahil Singla. Supermodular approximation of norms and applications. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, pages 1841–1852, 2024. doi:10.1145/3618260.3649734.
- [34] Andreas Krause and Daniel Golovin. Submodular function maximization. Tractability, 3(71-104):3, 2014.
- [35] Jon Lee, Maxim Sviridenko, and Jan Vondrák. Submodular maximization over multiple matroids via generalized exchange properties. Mathematics of Operations Research, 35(4):795–806, 2010. doi:10.1287/MOOR.1100.0463.
- [36] Jian Li and Samir Khuller. Generalized machine activation problems. In Proceedings of the Twenty-Second Annual ACM-SIAM Symposium on Discrete Algorithms, pages 80–94. SIAM, 2011. doi:10.1137/1.9781611973082.7.
- [37] André Linhares, Neil Olver, Chaitanya Swamy, and Rico Zenklusen. Approximate multi-matroid intersection via iterative refinement. Mathematical Programming, 183:397–418, 2020. doi:10.1007/S10107-020-01524-Y.
- [38] Kalen Patton, Matteo Russo, and Sahil Singla. Submodular norms with applications to online facility location and stochastic probing. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2023). Schloss-Dagstuhl-Leibniz Zentrum für Informatik, 2023.
- [39] Ram Ravi and Michel X Goemans. The constrained minimum spanning tree problem. In Algorithm Theory—SWAT’96: 5th Scandinavian Workshop on Algorithm Theory Reykjavík, Iceland, July 3–5, 1996 Proceedings 5, pages 66–75. Springer, 1996.