Max-Cut with Multiple Cardinality Constraints
Abstract
We study the classic Max-Cut problem under multiple cardinality constraints, which we refer to as the Constrained Max-Cut problem. Given a graph , a partition of the vertices into disjoint parts , and cardinality parameters , the goal is to select a set such that for each , maximizing the total weight of edges crossing (i.e., edges with exactly one endpoint in ).
By designing an approximate kernel for Constrained Max-Cut and building on the correlation rounding technique of Raghavendra and Tan (2012), we present a -approximation algorithm for the problem when . The algorithm runs in time , where and . This improves upon the -approximation of Feige and Langberg (2001) for (the special case when ), and generalizes the -approximation of Raghavendra and Tan (2012), which only applies when and does not handle multiple constraints.
We also establish that, for general values of , it is NP-hard to determine whether a feasible solution exists that cuts all edges. Finally, we present a -approximation algorithm for Max-Cut under an arbitrary matroid constraint.
Keywords and phrases:
Maxcut, Semi-definite Programming, Sum of Squares HierarchyCategory:
APPROXCopyright and License:
2012 ACM Subject Classification:
Theory of computation Design and analysis of algorithmsAcknowledgements:
This work was conducted in part while Madhusudhan Reddy Pittu was a visiting student at TTIC.Editors:
Alina Ene and Eshan ChattopadhyaySeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
Given an undirected graph on vertices and a weight function , the Max-Cut problem seeks a subset maximizing , the total weight of edges crossing the cut . Without loss of generality, we assume the weights are scaled so that the total edge weight satisfies .
Max-Cut is a fundamental problem in combinatorial optimization and approximation algorithms, with several landmark results, most notably the seminal SDP rounding algorithm by Goemans and Williamson [15], which achieves an approximation. This approximation ratio is known to be optimal under the Unique Games Conjecture (UGC) [18].
In this work, we study a variant called Constrained Max-Cut, where additional partition constraints are imposed on the solution.
Definition 1 (Constrained Max-Cut).
Given a graph , a weight function , and a set of partition constraints where and for all , the Constrained Max-Cut problem asks to find a subset such that for all , maximizing .
Several well-studied problems are special cases of . The Max-Bisection problem corresponds to and , and admits approximation factors close to – specifically, [2] (see also [14, 25, 9, 17, 11, 16, 22]). More generally, when there is a single cardinality constraint (i.e., ), the problem is known as [10]. It is also referred to as -Max-Cut in parameterized complexity, see e.g., [6, 24].
Definition 2 (Max-Cutk).
Given an undirected graph , a weight function , and an integer 111Assume that without loss of generality., the problem seeks a subset of cardinality exactly that maximizes .
For , the global correlation rounding technique of Raghavendra and Tan [22] (building on [4]) achieves an approximation. Austrin and Stankovic [3] later showed that this approximation is essentially tight for . However, when , existing results are weaker. Feige and Langberg [10] gave a -approximation for all , where is a small universal constant (). Moreover, the pipage rounding technique of Ageev and Sviridenko [1] guarantees a -approximation for all .
The special case of with and has applications in pricing in social networks [8], also referred to as influence-and-exploit [13]. In this context, consumers’s valuation depends directly on the usage of their neighbors in the network. Consequently, the seller’s optimal pricing strategy may involve offering discounts to certain influencers who hold central positions in the underlying network. Candogan et al. [8] considered a setting with two price types: full and discounted. Specifically, the objective is to maximize the total network influence, subject to the constraint of offering discounted prices to a small target set of buyers, where . Candogan et al. showed that the obtained problem can be reduced to , where .
Moreover, in certain settings where diversity among influencers is desirable, it is natural to require that the selected influencers fairly represent different groups. This requirement can be modeled as a problem with multiple capacity constraints. In most relevant cases, the number of groups is a small integer . For a comprehensive survey on fair representation for learning on graphs, see to [19].
In this paper, we also introduce and study a more general version of the problem: finding a maximum cut subject to a matroid constraint.
Definition 3 (Matroid Max-Cut).
Given an undirected graph , a weight function , and a matroid , the matroid Max-Cut problem is to maximize subject to , where is the collection of bases of .
Note that and are special cases of the matroid Max-Cut problem; we get these problems when the matroid is a partition matroid or a uniform matroid, respectively. This problem has not been explicitly studied previously. However, an algorithm by Lee, Mirrokni, Nagarajan, and Sviridenko [21] gives a -approximation to the more general problem of symmetric submodular function maximization subject to matroid base constraint.
1.1 Our Results and Techniques
Our algorithm for builds on the global correlation rounding technique introduced by Raghavendra and Tan [22], which achieves an -approximation in the regime where . We extend this approach by developing an approximate kernel and applying it in conjunction with correlation rounding, allowing us to handle the challenging case where .
This yields an approximation guarantee of for across all values of , improving upon the -approximation of [10] in the sparse regime. Formally:
Theorem 4.
For every , there is an algorithm that runs in time and obtains an -approximation to the optimal solution, where is the approximation factor of the Raghavendra–Tan algorithm.
The regime where is particularly challenging, as the correlation rounding approach of Raghavendra and Tan [22] does not extend to this setting. Our algorithm closes this gap by improving upon the -approximation of [10] in the sparse regime. For completeness, we provide a brief overview of the approach of Raghavendra and Tan in Section 1.2, and highlight the key reasons why it breaks down when .
We next address the more general setting of multiple constraints, focusing on the case .
Theorem 5.
For every , there is an algorithm that runs in time, where , and obtains an -approximation to the optimal solution of Constrained Max-Cut. In particular, when and is fixed, the running time is polynomial in .
More broadly, we study Max-Cut under an arbitrary matroid constraint , generalizing with an arbitrary number of partition constraints, especially when .
Theorem 6.
There exists a -approximation algorithm for matroid Max-Cut.
The only prior result for this setting is by Lee et al. [21], who provided a -approximation for the more general problem of symmetric submodular function maximization subject to a matroid base constraint.
Finally, we show that for general with an arbitrary number of constraints, it is NP-hard to decide whether there exists a feasible solution cutting all edges. Formally:
Theorem 7.
Given a graph , a partition of vertices into , and budget parameters , it is NP-hard to decide whether there exists a feasible solution such that .
We note that for the standard Max-Cut problem, this decision variant can be solved in polynomial time using bipartite testing.
Our Techniques.
A key technical contribution of our work is the construction of an approximate kernel for . Specifically, for any cardinality constraint , we sort the vertices by their (weighted) degrees as , and define as the top vertices. While the graph remains unchanged, we restrict our attention to solutions contained entirely within . Then, an optimal solution to over achieves a cut value that is at least a fraction of the true optimum. In other words,
| (1) |
See Theorem 12 in Section 2 for the formal statement.
This reduction is particularly useful because it allows us to focus on problem instances where . Conceptually, we can contract the vertices in into a single super vertex , and then restrict the solution to exclude . This transforms the sparse regime into one where the effective solution size is a constant fraction of the (reduced) vertex set, enabling the use of correlation rounding techniques that require .
In contrast, prior work by [24] uses kernelization to design fixed-parameter algorithms for , but their parameter is the value of the optimal solution itself, and they aim for an exact kernel. As a result, their kernel size is polynomial in , which is insufficient for our purposes. Moreover, their kernelization is sequential and adaptive, while ours is non-adaptive. Our approximate kernel also extends to with multiple constraints (), as formalized in Theorem 14.
Once we reduce to an instance with , we apply the Raghavendra–Tan algorithm [22] to obtain a subset of vertices of size , achieving a cut value that is at least an -fraction of the optimum. We then perform a random correction step: adjusting the solution by randomly adding or removing at most vertices to exactly match the required size , incurring only a negligible loss in cut value.
When , however, the rounding procedure of [22] does not directly apply. To handle the setting with multiple constraints, we introduce the notion of -block independence for SDP solutions, which generalizes the standard notion of -independence. Informally, an SDP solution is -block independent if, within each partition , the average correlation between pairs of vertices is at most .
We first show how to efficiently construct a block-independent solution. Then, by applying the rounding algorithm of [22], we obtain a subset that approximately satisfies each group constraint: for each , the size lies in the range . Finally, we apply a random correction step within each group to enforce exact feasibility, while ensuring that the cut value degrades by only a negligible amount.
1.2 Preliminaries
Our results heavily rely on the global correlation rounding technique developed in [22]. For completeness, we include the relevant definitions and theorems in this section. A quick summary of the Lasserre hierarchy is provided in Appendix A.
Naive approaches based on variants of hyperplane rounding applied to a two-round Lasserre SDP relaxation for the problem can produce subsets of expected size that achieve good approximation guarantees. However, these approaches offer no control over the concentration of around , due to potentially high correlations between the values assigned to vertices by the SDP solution.
Notation.
We use to denote a level- Lasserre pseudo-distribution, where is a distribution over partial assignments to the subset . Let denote the random variables jointly distributed according to , and the marginal variable for under . We write to denote the pseudo-probability that the assignment to lies in the event . In particular, conditional pseudo-probabilities are expressed as , which denotes the pseudo-probability that given that for some .
SDP Relaxation.
To leverage the correlations between vertices, [22] employ an -round Lasserre SDP for with a sufficiently large constant , formally described in Equation 2.
| (2) | |||||
| s.t. | |||||
Measuring Correlations.
One method to assess the correlation between two random variables and is through mutual information, defined as , where and are sampled according to the local distribution . An SDP solution is -independent if the average mutual information between uniformly random vertex pairs is at most , i.e., .
Definition 8 (-independence [22]).
An SDP solution to an -round Lasserre SDP is -independent if , where is the local distribution over . More generally, if is a distribution over , then the solution is -independent w.r.t. if . When unspecified, is assumed to be the uniform distribution over .
For many standard rounding schemes, such as halfspace rounding, the variance in the balance of the resulting cut is directly linked to the average correlation among random vertex pairs. Specifically, if the rounding scheme is applied to an -independent solution, the variance in the cut’s balance is bounded by a polynomial function of .
Obtaining Uncorrelated SDP Solutions.
If all vertices in a -round Lasserre SDP solution are highly correlated, conditioning on the value of one vertex reduces the entropy of the rest. Formally, if the solution is not -independent (i.e., ), then conditioning on a randomly chosen vertex and its value decreases the average entropy of the remaining variables by at least . Repeating this process times suffices to obtain an -independent solution. Thus, starting from a -round Lasserre SDP solution, this process results in a -round -independent solution for some .
Rounding Uncorrelated SDP Solutions.
Given an -independent SDP solution, many natural rounding schemes ensure that the balance of the output cut is concentrated around its expectation. Hence, it suffices to construct rounding schemes that preserve the expected balance. Raghavendra and Tan [22] present a simple rounding procedure that preserves the individual bias of each vertex, thereby ensuring the global balance property.
An elegant probabilistic argument from [22] shows how to convert an -round Lasserre SDP solution into an -independent -round solution, while losing only an additive factor of in the objective value (assuming the optimum is normalized to at most ).
Lemma 9 ([22]).
There exists such that .
Lemma 9 implies the existence of a such that conditioning on the joint assignment to randomly sampled vertices reduces the average mutual information between other pairs to at most .
Theorem 10 ([22]).
For every and integer , there exists an algorithm running in time that finds an -independent solution to the -round Lasserre SDP, with objective value at least , where OPT is the optimum SDP value.
Theorem 10 implies that there exists such that conditioning on vertices yields an -independent solution with probability at least . Since the sampling procedure preserves the marginal biases of the vertices, the SDP objective remains close to optimal in expectation. By Markov’s inequality, the value of the conditioned solution is at least with probability at least . Thus, there exists a small subset of vertices such that conditioning on them yields an -independent solution with near-optimal value.
Algorithm 5.3 of [22] is a rounding scheme that preserves the bias (according to the SDP solution) of every vertex while also approximately preserving the pairwise correlations up to polynomial factors. Using numerical techniques, they show that the probability of an edge being cut is at least times its contribution to the SDP objective, implying that the total cut value is at least times the SDP value.
Controlling Cut Balance.
Theorem 11 ([22]).
Given an -independent solution to two rounds of the Lasserre SDP, let denote the rounded output from Algorithm 5.3. Let be the expected balance of the cut. Then, .
By applying Chebyshev’s inequality to Theorem 11, the number of vertices in the cut lies in the range with high probability. When , we can choose small enough so that the relative deviation is within . A post-processing step can then adjust the set size to exactly (e.g., by swapping a small number of vertices), which incurs only an fractional loss in cut value. However, when , the additive error term may significantly exceed , making it difficult to ensure cardinality feasibility without substantially affecting the objective.
Notation.
For any subset and vertex , we write and . Let be a graph with non-negative edge weights . For a parameter , let denote the set of the highest-degree vertices in under weight function . If the vertex set is partitioned into disjoint groups, , then denotes the highest-degree vertices in part . When the weight function is clear from context, we abbreviate the weighted degree of a vertex as instead of .
2 Approximate Kernels for Max-Cut with Cardinality Constraints
In Section 2.1, we show that for any instance of , one can reduce the graph to a (conditioned) instance with vertices. In Section 2.2, we generalize this construction to the setting with multiple partitions. Specifically, for any instance of , we construct a conditioned instance such that for every , we have .
We use OPT to denote the optimal cut value of a given instance. For the single-group problem on a graph , we define . For the multi-group case with partition and size constraints , the optimal value is .
2.1 Approximate Kernel for
Kernel Procedure for .
We now describe the approximate kernel construction for the single-group problem.
Input: Graph , cardinality parameter , and approximation parameter .
Output: Reduced graph .
-
1.
If , return .
-
2.
Otherwise, sort the vertices of in decreasing order of weighted degree, and retain only the top vertices. Merge the rest of vertices into a super vertex , and return the resulting graph .
Note that the super vertex appears in the output graph only if .
Theorem 12.
For any instance , let be the reduced instance returned by the kernel procedure above. Then the optimal cut value of on , conditioned on not selecting the super vertex , satisfies
Proof.
Let , and recall from the notation in the preliminaries that is the set of the highest-degree vertices in . Let be an optimal solution of size in with cut value . We will construct a set of size with value close to OPT.
We iteratively transform into a set within by applying Lemma 13 up to times. At each step , we replace a vertex with a vertex such that:
Since each step increases by one, the process terminates in at most steps. Therefore,
Substituting , we get
where the final bound assumes .
Lemma 13.
Let be a subset of size such that . Then there exist vertices and such that:
Proof.
Since and , we have . Let be the vertex minimizing , the total weight of edges between and . Let be any vertex in .
We use the submodularity of the cut function:
Rearranging:
Now we bound . Since minimizes among , we have:
which implies:
Putting everything together:
completing the proof.
2.2 Approximate Kernel for Constrained Max-Cut
Kernel Procedure for Constrained Max-Cut.
We now describe the kernelization procedure for the problem with multiple vertex groups.
Input: Graph , cardinality constraints , and approximation parameter .
Output: Reduced graph .
-
1.
For each , if , retain the top vertices in by weighted degree and merge the remaining vertices into a super vertex .
-
2.
Return the resulting graph .
Note that a super vertex appears in the output graph only if . Let denote the set of all super vertices.
Theorem 14.
For any instance , let be the reduced instance returned by the kernel procedure. Then the optimal value of the reduced instance, conditioned on not selecting any super vertex, satisfies
Proof.
Let be an optimal solution to the original instance with . For each part , define as the top vertices in by weighted degree (as defined in the preliminaries).
We will transform into a solution such that for every while losing only a small fraction of the cut value. At each step , identify the smallest index for which , and apply the local exchange from Corollary 16 to swap a vertex with a vertex , yielding a new set with
For each , we perform at most such exchanges in . Hence, the total cut value at the end satisfies:
Using the inequality , we get
where the final bound assumes .
Lemma 15.
Let be a subset of size and let be a subset of size greater than such that and every vertex in has higher weighted degree than every vertex in . Then there exist and such that:
Proof.
The proof is identical to that of Lemma 13. It proceeds by selecting to minimize over and applying cut submodularity to bound the loss when replacing .
Corollary 16.
For any subset and an index such that and , there exist vertices and such that
Proof.
Using in Lemma 15, and the fact that finishes the proof.
3 Single Constraint
In this section, we describe our -approximation algorithm for , for all values of . Without loss of generality, we assume due to the symmetry of the cut function.
3.1 Algorithm
Input: Weighted graph and parameters , .
Output: A set of size .
-
1.
(Preprocessing Step) Let be the approximate kernel output by the kernel with input . Note that .
-
2.
(SDP and Conditioning)
-
(a)
Solve a -round Lasserre SDP relaxation for the problem on the graph (see Section 3.2).
-
(b)
Apply Theorem 10 with and to obtain a 2-level SDP solution that is -independent and has objective value at least , where is the optimum value of on (conditioned on not selecting the super vertex ). By Lemma 18 (2), we know that .
-
(a)
-
3.
(Rounding) Apply the rounding algorithm of Raghavendra and Tan (Algorithm 5.3 in [22]) to obtain a (random) set . Let denote the event that .
-
4.
(Correction) If event does not occur, return an arbitrary subset of size . Otherwise, adjust by randomly adding or removing vertices to produce a set of size exactly , and return .
3.2 Lasserre SDP
We now describe the SDP used in Step 2 of the algorithm above. Since a full overview of the Lasserre hierarchy is already provided in Appendix A, we only describe the relevant formulation.
After the preprocessing step, we solve the following level- Lasserre SDP relaxation for on the reduced graph , with an additional constraint ensuring the super vertex is not selected:
(3)
s.t.
(4)
3.3 Analysis
Proof of Theorem 4.
Using Lemma 18 (1), the optimal value on the kernelized instance is at least . After solving the SDP, we obtain a value at least .
The expected size of is exactly since Algorithm 5.3 from [22] preserves the bias of each vertex. Using Theorem 11, the variance of the balance is at most (Assume that the constant hidden in the -notation is for simplicity. One can absorb the constant into the in general). By Chebyshev’s inequality, the event occurs with probability at least .
The expected cut value conditioned on satisfies:
| (5) |
Let be the final corrected set. Then:
| (6) | ||||
| (7) | ||||
| (8) |
Steps 1 and 4 take and time, respectively. Steps 2 and 3, which involve solving the Lasserre SDP and rounding, dominate the runtime and require time.
Lemma 17.
Conditioned on the event and a fixed , let be the set obtained by randomly adding or deleting vertices from so that . Then:
Lemma 18.
Let denote the optimum value of on , conditioned on not selecting the super vertex .
-
1.
, where OPT is the optimum value for on .
-
2.
is at least an -fraction of the total edge weight in .
Proof.
Part (1) follows from Theorem 12. For part (2), we show that a uniformly random subset of of size cuts any edge with probability at least .
Let . Then . If edge is adjacent to , it is cut with probability . Otherwise, the cut probability is .
4 Constant Number of Constraints
In this section, we present our -approximation algorithm for , the Max-Cut problem with cardinality constraints. Our primary focus is on instances where the number of vertices to be selected from each part is relatively small, and for this reason, we assume that . (Unlike the case in , this is not without loss of generality.)
The key observation enabling this extension of [22] to multiple constraints is that the notion of -independence can be defined locally within each block. Specifically, it suffices to ensure that the average mutual information between vertex pairs within each part is small:
If this condition holds, then after rounding via Algorithm 5.3 of [22], the size of each intersection concentrates around its expectation. In particular, using Theorem 11 with as the uniform distribution over , the variance of is bounded by for every . Therefore, for an appropriate choice of , we obtain
with probability at least .
Definition 19 (-block independence).
An SDP solution to an -round Lasserre relaxation is -block independent if hold for all .
To find such solutions, we extend the conditioning technique of [22]. The following procedure begins with an -round SDP solution and returns an -round -block independent solution for .
Conditioning Procedure
-
1.
For each :
-
(a)
Sample a block index uniformly at random.
-
(b)
Sample a vertex uniformly at random.
-
(c)
Sample from its marginal distribution under the current SDP solution (conditioned on previous outcomes), and condition on this value.
-
(d)
If the resulting SDP solution is -block independent, terminate and return it.
-
(a)
Lemma 20.
For any , there exists such that
Proof.
Define the potential function:
Now, conditioned on fixed values of and , the difference in potentials is:
Taking expectation over all random choices of and gives:
| (9) |
Summing (9) over to , and noting that entropy is always non-negative, we get:
Therefore, by averaging, there exists for which the expected blockwise mutual information is at most .
Corollary 21.
If
then
Proof.
Each blockwise term is a subset of the global sum, and mutual information is non-negative.
Theorem 22.
For every and integer , there exists an algorithm running in time that finds an -block independent solution to the -round Lasserre SDP with value at least , where OPT is the optimum value of the -round SDP.
Proof.
Set . First, solve the -round Lasserre SDP relaxation (as described in Section 4.1) to obtain an initial solution.
Next, apply the conditioning procedure described above. That is, for each , sample a block index uniformly at random, then sample a vertex uniformly, sample from its marginal distribution (after the first fixings), and condition the SDP solution on that assignment. Continue this process until the resulting pseudo-distribution becomes -block independent.
We analyze this procedure by appealing to Lemma 20, which shows that:
For some . By Markov’s inequality, the probability that the total conditional mutual information (summed over all block pairs) exceeds is at most:
Thus, with probability at least , the conditioned solution is -block independent. By Corollary 21, this also implies that:
so the solution satisfies the desired independence property within each block.
Now consider the effect of the conditioning procedure on the SDP objective value. Let denote the value of the SDP after conditioning. Since conditioning preserves expectations, we have . To bound the probability that the value drops by more than , we apply Markov’s inequality to the non-negative random variable :
where the last inequality uses .
Separately, as shown earlier, the probability that the conditioned solution fails to be -block independent is at most . By a union bound, the total failure probability is
for all . Hence, there exists a choice of conditioning – i.e., some and assignment to – such that the resulting SDP solution is -block independent and has objective value at least .
This outcome can be found by brute-force search over all subsets of up to variables and their possible assignments. The overall runtime is thus
as claimed.
Proof of Theorem 5.
4.1 Algorithm
Input: Weighted graph with vertex set partitioned as , parameters for , and .
Output: A set such that for all .
-
1.
(Preprocessing Step) Let be the approximate kernel obtained via the kernel procedure with input .
-
2.
(SDP and Conditioning)
-
(a)
Solve a -round Lasserre SDP relaxation for on (see Section 4.2).
-
(b)
Apply Theorem 22 with and to obtain a 2-level SDP solution that is -block independent and has value at least . From Lemma 24, we know .
-
(a)
-
3.
(Rounding) Apply Algorithm 5.3 from [22] to obtain a (random) set . Let denote the event that for each , and define .
-
4.
(Correction) If does not occur, return an arbitrary feasible set. Otherwise, for each part , randomly add or remove vertices to ensure , and return the resulting set .
Lemma 23.
The expected value of the cut returned by Algorithm 4.1 is at least . The running time of the algorithm is where .
4.2 SDP Relaxation
We solve the following level- Lasserre SDP relaxation for on the reduced graph :
| (10) | |||||
| s.t. | |||||
4.3 Analysis
Proof of Lemma 23.
By Lemma 24, we have , and after solving the SDP and conditioning, the objective remains at least .
Since the SDP solution is -block independent, using Theorem 11 with as the uniform distribution over each , the variance of is . By Chebyshev’s inequality, the event occurs with probability at least , so the joint event occurs with probability at least .
The expected value of the cut after rounding is at least . Conditioning on , we have:
| (11) |
Let be the corrected set after adjusting to satisfy cardinality constraints exactly. Using the same argument as in Lemma 17 applied sequentially across the parts, we get:
| (12) |
The total running time is dominated by solving the SDP and brute-force conditioning, which takes . Preprocessing and postprocessing steps take time.
Lemma 24.
Let be the optimum value of on (conditioned on not picking any ).
-
1.
, where OPT is the optimum value of the instance on .
-
2.
fraction of the total edge weight in .
Proof.
Part (1) follows from Theorem 14. For (2), consider sampling , where each is a uniformly random subset of size from . Let . Since , we have .
-
1.
If an edge is adjacent to , the cut probability is at least .
- 2.
-
3.
If endpoints lie in and (), the probability that exactly one endpoint lies in is
So the expected cut value of is at least an fraction of total edge weight in .
Proof of Theorem 5.
5 Arbitrary Number of Constraints
We consider the general case of with an arbitrary number of constraints, potentially . First, we present a -approximation for the more general problem of Max-Cut under an arbitrary matroid constraint. Next, we establish an NP-hardness result for determining whether the optimal solution in a given instance of with an arbitrary number of constraints equals the total number of edges in the graph.
5.1 Approximation Algorithm
Proof of Theorem 6.
Consider the following linear program:
| (13) | ||||
| (14) | ||||
| (15) | ||||
| (16) |
where is the base polytope of the matroid . We can see that for any given , the optimal choice for is . When is integral, this function also coincides with the indicator whether edge has been cut. Since the matroid polytope is separable, we can solve the LP Equation 13 efficiently.
Now consider the following non-concave quadratic program:
| (17) | ||||
| (18) |
Observe that the function also coincides with the cut indicator function of edge when is integral. Even though we cannot solve Equation 17 efficiently, we can show that it has no integrality gap and infact that we can round any fractional solution to a solution that is integral and with value at least that of . The two crucial properties we need that are easy to see are:
-
1.
The function is convex in any direction for . Here is the indicator vector for vertex .
-
2.
The polytope is the facet of a matroid and hence solvable and integral.
Given these properties, any fraction solution can be pipage rounded (see [1] and especially section 3.2 of [7]) to an integral solution with value at least that of the fractional solution. The final observation is that for any , we have
| (19) |
from Lemma 27. This implies that the integrality gap of Equation 13 is at most and in fact provides a way to find a rounding with value at least times the LP value. Solve the LP and pipage round the solution using the quadratic objective. Note that the proof idea for Theorem 6 is essentially the same as in [1] used for the Hypergraph Max -cut with given sizes of parts problem and the pipage rounding for matroids from [7].
5.2 Hardness Result
Proof of Theorem 7.
We show a reduction from the 3D matching problem. An instance of the 3D matching problem is a tripartite graph with parts . The edges are triples . The problem is to decide if there is a subset of the edges such that every vertex is included in exactly one edge.
The reduction is as follows: For every edge , consider the star graph with four vertices with the center labeled and the leaves labeled respectively. The overall graph is simply the union of all these stars. The partition matroid consists of parts for every vertex that contains the vertices such that . We have parts similarly for elements in and . The capacity of every part is exactly .
(Completeness) If there is a collection of edges that every element of is in exactly one edge, then consider the solution where . It is easy to see that and for every .
(Soundness) Since every star graph is bipartite, any solution such that should have the center on one side and the leaves on the other side. This implies that every solution such that is of the form for some where is the collection of triples from the 3D matching instance. The partition matroid constraint that for exactly translates to being a perfect 3D matching. Since it is NP-Hard to decide if there is a perfect 3D matching, it is NP-hard to decide if there is a cut such that and when .
Remark 25.
The above theorem shows that, in general, deciding whether the Max-Cut value equals the total number of edges in the graph is solvable in polynomial time when the number of constraints is constant. Moreover, the decision problem becomes solvable in quasi-polynomial time when the number of constraints is .
References
- [1] Alexander A. Ageev and Maxim Sviridenko. Pipage rounding: A new method of constructing algorithms with proven performance guarantee. Journal of Combinatorial Optimization, 8:307–328, 2004. doi:10.1023/B:JOCO.0000038913.96607.C2.
- [2] Per Austrin, Siavosh Benabbas, and Konstantinos Georgiou. Better balance by being biased: A 0.8776-approximation for max bisection. ACM Transactions on Algorithms (TALG), 13(1):1–27, 2016. doi:10.1145/2907052.
- [3] Per Austrin and Aleksa Stanković. Global cardinality constraints make approximating some max-2-csps harder. Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, 145:24:1–24:17, 2019. doi:10.4230/LIPICS.APPROX-RANDOM.2019.24.
- [4] Boaz Barak, Prasad Raghavendra, and David Steurer. Rounding semidefinite programming hierarchies via global correlation. In Annual Symposium on Foundations of Computer Science, pages 472–481, 2011. doi:10.1109/FOCS.2011.95.
- [5] Niv Buchbinder, Moran Feldman, Joseph Naor, and Roy Schwartz. Submodular maximization with cardinality constraints. In Proceedings of the twenty-fifth annual ACM-SIAM symposium on Discrete algorithms, pages 1433–1452. SIAM, 2014. doi:10.1137/1.9781611973402.106.
- [6] Leizhen Cai. Parameterized complexity of cardinality constrained optimization problems. The Computer Journal, 51(1):102–121, 2008. doi:10.1093/COMJNL/BXM086.
- [7] Gruia Calinescu, Chandra Chekuri, Martin Pál, and Jan Vondrák. Maximizing a monotone submodular function subject to a matroid constraint. SIAM Journal on Computing, 40(6):1740–1766, 2011. doi:10.1137/080733991.
- [8] Ozan Candogan, Kostas Bimpikis, and Asuman Ozdaglar. Optimal pricing in networks with externalities. Operations Research, 60(4):883–905, 2012. doi:10.1287/OPRE.1120.1066.
- [9] Uriel Feige, Marek Karpinski, and Michael Langberg. A note on approximating max-bisection on regular graphs. Information Processing Letters, 79(4):181–188, 2001. doi:10.1016/S0020-0190(00)00189-7.
- [10] Uriel Feige and Michael Langberg. Approximation algorithms for maximization problems arising in graph partitioning. Journal of Algorithms, 41(2):174–211, 2001. doi:10.1006/JAGM.2001.1183.
- [11] Uriel Feige and Michael Langberg. The rounding technique for semidefinite programs. Journal of Algorithms, 60(1):1–23, 2006. doi:10.1016/J.JALGOR.2004.11.003.
- [12] Noah Fleming, Pravesh Kothari, and Toniann Pitassi. Semialgebraic proofs and efficient algorithm design. Foundations and Trends® in Theoretical Computer Science, 14(1-2):1–221, 2019. doi:10.1561/0400000086.
- [13] Dimitris Fotakis and Paris Siminelakis. On the efficiency of influence-and-exploit strategies for revenue maximization under positive externalities. Theoretical Computer Science, 539:68–86, 2014. doi:10.1016/J.TCS.2014.04.026.
- [14] Alan Frieze and Mark Jerrum. Improved approximation algorithms for MAX -CUT and MAX BISECTION. Algorithmica, 18(1):67–81, 1997. doi:10.1007/BF02523688.
- [15] Michel X Goemans and David P Williamson. Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming. Journal of the ACM (JACM), 42(6):1115–1145, 1995. doi:10.1145/227683.227684.
- [16] Venkatesan Guruswami, Yury Makarychev, Prasad Raghavendra, David Steurer, and Yuan Zhou. Finding almost-perfect graph bisections. In In Proceedings of Innovations in Computer Science, pages 321–337, 2011. URL: http://conference.iiis.tsinghua.edu.cn/ICS2011/content/papers/11.html.
- [17] Eran Halperin and Uri Zwick. A unified framework for obtaining improved approximation algorithms for maximum graph bisection problems. Random Structures and Algorithms, 20, May 2002. doi:10.1002/rsa.10035.
- [18] Subhash Khot, Guy Kindler, Elchanan Mossel, and Ryan O’Donnell. Optimal inapproximability results for MAX-CUT and other -variable CSPs? SIAM Journal on Computing, 37(1):319–357, 2007. doi:10.1137/S0097539705447372.
- [19] Charlotte Laclau, Christine Largeron, and Manvi Choudhary. A survey on fairness for machine learning on graphs. arXiv preprint arXiv:2205.05396, 2022. doi:10.48550/arXiv.2205.05396.
- [20] Monique Laurent. A comparison of the sherali-adams, lovász-schrijver, and lasserre relaxations for 0-1 programming. Mathematics of Operations Research, 28(3):470–496, 2003. doi:10.1287/MOOR.28.3.470.16391.
- [21] Jon Lee, Vahab S. Mirrokni, Viswanath Nagarajan, and Maxim Sviridenko. Maximizing nonmonotone submodular functions under matroid or knapsack constraints. SIAM Journal on Discrete Mathematics, 23(4):2053–2078, 2010. doi:10.1137/090750020.
- [22] Prasad Raghavendra and Ning Tan. Approximating CSPs with global cardinality constraints using SDP hierarchies. In Proceedings of the Symposium on Discrete Algorithms, pages 373–387. SIAM, 2012. doi:10.1137/1.9781611973099.33.
- [23] Thomas Rothvoß. The lasserre hierarchy in approximation algorithms. Lecture Notes for the MAPSP, pages 1–25, 2013.
- [24] Saket Saurabh and Meirav Zehavi. ()-max-cut: An -time algorithm and a polynomial kernel. Algorithmica, 80:3844–3860, 2018.
- [25] Yinyu Ye. A -approximation algorithm for max-bisection. Mathematical Programming, 90:101–111, 2001. doi:10.1007/PL00011415.
Appendix A Basics of the Lasserre SDP Hierarchy
For detailed expositions on the sum-of-squares hierarchy, we refer the reader to the excellent surveys by Laurent [20], Rothvoß [23], and Fleming et al. [12]. This section briefly summarizes key ideas drawn from these sources.
Given a binary optimization problem with a linear relaxation defined by a matrix and right-hand side , consider the feasible region .
We ask: how can we systematically strengthen this relaxation to better approximate the convex hull of integral solutions, ?222We work over rather than since this is the natural domain for problems like Max-Cut, where signs encode partition membership.
Points in this convex hull can be interpreted as distributions over the hypercube . The level- Lasserre SDP yields a pseudo-distribution , where each is a distribution over partial assignments to the subset . However, there need not exist a global distribution whose marginals agree with these . Despite this, the pseudo-distribution satisfies the following key properties:
-
1.
Marginal Consistency: The pseudo-distributions are consistent across overlapping subsets. That is, for any subsets with and any assignment , we have:
(20) where denotes the extension of to using , and similarly for on .
-
2.
Conditioning: The SDP solution supports conditioning on the value of any variable . Given a level- pseudo-distribution and variable , there exist level- pseudo-distributions , and a weight such that for all with and ,
(21) where is nonzero only if and is nonzero only if .
While these properties are also satisfied by weaker hierarchies like Sherali–Adams, the Lasserre hierarchy is uniquely characterized by an additional sum-of-squares condition: for every polynomial of degree at most , the pseudo-expectation of its square is non-negative:
| (22) |
Assuming polynomials are multilinear (since we evaluate over the hypercube), any such polynomial can be written as , where and . Then the pseudo-expectation becomes:
Moreover, to incorporate the linear constraints , the Lasserre relaxation requires that:
| (23) |
A level- pseudo-distribution satisfying (22) and (23) can be found by solving a semidefinite program, as described below.
A.1 SDP Formulation
Definition 26 (Lasserre Hierarchy [23]).
Let . The level- Lasserre relaxation, denoted , consists of vectors satisfying:
Here, is the moment matrix, and is the moment matrix of slacks. The projection onto the original variables is denoted by .
This is a valid relaxation: for any integral solution , the assignment for all yields a feasible point in .
Each variable represents the pseudo-moment corresponding to all variables in being assigned . From these, one can recover the pseudo-distribution via Möbius inversion:
| (24) |
where denotes the partial assignment that sets variables in to and the remaining in to .
Appendix B Omitted proofs
Lemma 27.
For any , .
Proof.
We consider the following cases,
-
1.
When , the required inequality to prove is
The left most inequality is equivalent to which is trivially true. The right most inequality is equivalent to which is true because .
-
2.
When , substituting , the required inequality to prove is
The conditions on are that they are in the range and that . This is exactly the case above.
Proof of Lemma 17.
Suppose , then is obtained by adding a uniformly random set of vertices from to . The value is equal to if and otherwise. In the first case, the probability of an element in to be added to is at most . In the second case, it is . For it to be at most , it suffices to have , which is true because in fact, we can show that
The leftmost inequality is equivalent to which is true because . Using Lemma 28, we can imply that .
If , then is obtained by removing a uniformly random subset of size . The probability of an element in to be removed is . Hence, it suffices to have , which is true because in fact, . The rightmost inequality is equivalent to . This is true because . Using Lemma 28 on , we get .
Lemma 28 (Lemma 2.2 in [5]).
For any set , if is a random set such that each element in is included in (not necessarily independently) with probability at most , then
| (25) |
This fact is generally true for any non-negative submodular function.
