Approximating Maximum Cut on Interval Graphs and Split Graphs Beyond Goemans-Williamson
Abstract
We present a polynomial-time -approximation algorithm for the Maximum Cut problem on interval graphs and split graphs, where is the approximation guarantee of the Goemans-Williamson algorithm and is a fixed constant. To attain this, we give an improved analysis of a slight modification of the Goemans-Williamson algorithm for graphs in which triangles can be packed into a constant fraction of their edges. We then pair this analysis with structural results showing that both interval graphs and split graphs either have such a triangle packing or have maximum cut close to their number of edges. We also show that, subject to the Small Set Expansion Hypothesis, there exists a constant such that there is no polyomial-time -approximation for Maximum Cut on split graphs.
Keywords and phrases:
Maximum cut, graph theory, interval graphs, split graphsCategory:
APPROXFunding:
Jungho Ahn: Supported by Leverhulme Trust Research Project Grant RPG-2024-182.Copyright and License:
2012 ACM Subject Classification:
Theory of computation Approximation algorithms analysisEditors:
Alina Ene and Eshan ChattopadhyaySeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
Given a graph , the Maximum Cut problem asks for a subset of vertices that maximizes the number of edges with exactly one endpoint contained in . In this paper, we study the approximability of the Maximum Cut problem on interval graphs and split graphs.
A graph is interval if there exists a collection of intervals on the real line such that if and only if . See Figure 1 for an example. Interval graphs are used in the field of biology, where they model natural phenomena such as DNA and food webs [13]. They have also been used in the study of register allocation, where vertices correspond to variables and intervals correspond to “live ranges” [11]. Finally, they have numerous desirable theoretical properties and have arisen as a natural class of graphs to design algorithms for. For example, it is shown in [5] that a certain variant of the graph homomorphism problem is polynomial-time solvable if and only if the label graph is interval.
A graph is split if there exists a partition of such that , the graph induced by , is a clique and , the graph induced by , is independent. See Figure 2 for an example. Split graphs and interval graphs are both important subclasses of chordal graphs, which themselves are a subclass of perfect graphs. In particular, split graphs are often the “simplest” subclass of perfect graphs in which problems are difficult to approximate. Thus, considering split graphs is a natural first step when attempting to characterize a problem on chordal or perfect graphs. In this paper, we show that, assuming the Small Set Expansion Hypothesis, there is some constant such that there is no polynomial-time -approximation for Maximum Cut on split graphs. To our knowledge, this is the first known hardness of approximation result for Maximum Cut on perfect graphs.
Maximum Cut is one of the original 21 problems shown to be NP-Complete by Karp [14] and has long been a staple problem among algorithm researchers. The seminal work of Goemans and Williamson shows that there is a polynomial-time -approximation algorithm for Maximum Cut on all graphs [8]. The optimality of this result was open until 2007, when Khot et al. showed that, subject to the Unique Games Conjecture, there is no polynomial-time approximation algorithm for Maximum Cut with an approximation ratio better than [15].
Remarkably, there is relatively little known about Maximum Cut when the input graph is restricted to graphs from some structured class. Even for subclasses of perfect graphs, where problems such as Independent Set and Chromatic number admit polynomial-time algorithms based on semidefinite programming, Maximum Cut, another flagship application of semidefinite programming, remains mostly unexplored. In particular the extremely well-structured classes of interval graphs and split graphs, two important subclasses of perfect graphs, had no known approximation algorithm with a ratio better than prior to this work. Our main results provide the first improved approximation for these classes of graphs since the work of Goemans and Williamson.
Theorem 1.
There is a polynomial-time -approximation algorithm for Maximum Cut on interval graphs.
Theorem 2.
There is a polynomial-time -approximation algorithm for Maximum Cut on split graphs.
On the hardness side, it was shown by Bodlander and Jansen in 2000 that Maximum Cut on split graphs is NP-Complete [3]. Maximum Cut on interval graphs has seen further attention lately - it was shown only recently that Maximum Cut on interval graphs is NP-Complete [1]. Further work refined this hardness for interval graphs which have at most interval lengths [4], and later at most interval lengths [2]. Hardness for unit interval graphs - interval graphs with only interval length - remains open. We remark that none of these hardness results imply any hardness of approximation. Our final main result is that, subject to the Small Set Expansion Hypothesis of [17], Maximum Cut on split graphs is hard to approximate to some constant factor.
Theorem 3.
There exists a constant such that there is no polynomial-time -approximation algorithm for Maximum Cut on split graphs, subject to the Small Set Expansion Hypothesis under randomized reductions.
| lower bound | upper bound | |
|---|---|---|
| General graphs | [8] | [15] |
| Degree graphs | [12] | [19] |
| Interval graphs | NP-Complete [1] | |
| Split graphs | ||
| Planar graphs | 1 [10] | |
| Line graphs | 1 [9] |
1.1 Our Techniques
1.1.1 Analysis of the Perturbed Goemans-Williamson Algorithm
Our starting point is the Goemans-Williamson algorithm [8]. This algorithm first solves the following semidefinite program (SDP)
That is, the algorithm maps each vertex to a unit vector in a way that maximizes the sum of , where is the angle between and . Next, the algorithm samples a random Gaussian vector , creates the set , and returns the cut defined by . This step is equivalent to sampling a random hyperplane and taking to be the vertices on one side of this hyperplane. It is a straightforward calculation to see that the probability of an edge being cut by is equal to . Thus, the approximation guarantee of the Goemans-Williamson algorithm is equal to
Define to be the “critical angle” at which this ratio is minimized. We plot the performance guarantee over all angles in Figure 3. Any edge with has approximation ratio strictly better than . Intuitively, if at least a constant fraction of edges have angle bounded away from , we should expect the Goemans-Williamson algorithm to achieve a better approximation ratio. Unfortunately, this is not exactly true.
Consider the graph , the path on vertices. Suppose the SDP is solved suboptimally, so that the first edge has angle and the second edge has angle . Even though half of the edges have angle bounded away from , the expected size of the cut from rounding is still only times the optimal value of the SDP. To handle situations like this, we introduce a “perturbed” version of the Goemans-Williamson algorithm. In this perturbed algorithm, any vertex with for some small will instead be included in with probability . The motivation behind this perturbation is to consider what happens when every is “moved” by a small amount. If and are close together, then they are likely to move further apart after this perturbation. Inversely, if and are far apart, they are likely to move closer together after this perturbation.
Indeed, in Lemma 17, we show that the perturbed algorithm is more likely to cut any edge with , and less likely to cut any edge with . Moreover, the further away is from , the more the results of the perturbed algorithm vary from the unperturbed algorithm. Thus, if there are many edges with angle near , but few edges with angle near , the perturbed algorithm will achieve an approximation ratio above , see Lemma 19. On the other hand, if there are many edges with angle near , then the Goemans-Williamson algorithm itself will achieve an approximation ratio above , see Lemma 15. Thus, if we take the maximum result of the perturbed and unperturbed Goemans-Williamson algorithm, we can ignore the case of having many edges which contribute little to the optimal. That is, as shown in Lemma 21, we can indeed assume that if at least a constant fraction of edges have angle bounded away from , the Goemans-Williamson algorithm will achieve an approximation ratio above .
Consider a single triangle . A straightforward analysis shows that
regardless of the values of . Thus, there is some edge with . That is, for every triangle in our graph, we can expect at least one of its edges to have an angle bounded away from . This leads us into Section 2, where we show that every interval graph and split graph either have a large edge-disjoint triangle packing, or have a cut with at least edges.
1.1.2 Finding an Edge-Disjoint Triangle Packing
Suppose is a split graph. If at least of its edges are contained in the clique , then we can pack triangles almost perfectly into those edges. Otherwise, we have that at least edges are crossing between and , and so we have found a cut with edges, where is the size of a maximum cut of . Thus, we find ourselves in a “win-win” situation, where either has an edge-disjoint triangle packing on a constant fraction of its edges, or is nearly bipartite, and we can find a large cut. This analysis leads directly to an improved approximation for Maximum Cut on split graphs.
It turns out that the same “win-win” structural result also holds for interval graphs, which we prove in Theorem 5. To show this, we employ a marking scheme, where each vertex is classified as either “small” or “large” based on how it interacts with the rest of the graph. We first show in Lemma 8 that the number of edges between large vertices is low compared to . The situation for small vertices is more complicated. We show in Lemmas 9 and 10 that one of the following conditions must always hold:
-
1.
has a bridge; or
-
2.
there is a clique in such that the sum of degrees is not much more than ; or
-
3.
the number of edges between small vertices is low compared to .
If Condition 1 holds, we can essentially delete the bridge; because adding a bridge to a bipartite graph does not create an odd cycle, we can always add back in the bridge after finding a cut in the rest of the graph. If Condition 2 holds and , then we can pack edge-disjoint triangles into and delete from the graph. Because the sum of degrees is not much more than , we have packed triangles into at least a constant fraction of the deleted edges. The case of is more tricky, as one cannot pack triangles into one or two vertices, and must be handled separately. If Condition 3 holds, then most of the edges in go between small and large vertices. If we have already packed triangles into a constant fraction of edges via Condition 2, then we are done. Otherwise, “most” edges are remaining in the graph, and we have found a cut on “most” of these remaining edges. For the right definition of “most,” this is a large cut of at least edges.
1.1.3 Hardness of Approximating Maximum Cut on Split Graphs
To show that Maximum Cut is hard to approximate on split graphs, we start from the following hardness result that follows using standard techniques from [18]. Assuming the Small Set Expansion Hypothesis holds under randomized reductions, for any sufficiently small , there is no polynomial time algorithm that, given a graph , can distinguish between the following two cases.
-
1.
There exists a cut with such that .
-
2.
For all cuts with , either or .
That is, in Case 1, there is a cut with nearly half the vertices that cuts very few edges. In Case 2, all cuts with at least a constant fraction of vertices must cut many more edges. We transform into a split graph by turning into a clique and setting , setting , and adding an edge from to each endpoint of . Any cut of can have at most edges from those edges contained in , and at most edges from those going between and . In Case 1, there exists a cut that does cut nearly edges. In Case 2, any cut of either cuts at most edges from those edges contained in , or at most from those edges crossing between and . After ensuring that , this shows that Maximum Cut is hard to approximate on split graphs.
1.2 Preliminaries
1.2.1 Graphs
We consider only finite graphs in this paper. Apart from Section 4, all considered graphs will also be simple and unweighted. When the subject graph is clear from context, we will use to refer to the vertex set of , to refer to the edge set of , and to refer to the number of vertices in . Let be a subset of vertices of . We define as the subgraph of induced by and as the set of edges in this subgraph. We define to be the cut induced by and to be the maximum cut size of . We will often omit and write only when the graph is clear from context. Let be any vertex of . We define as the set of edges adjacent to , as the set of neighbors of , and as the degree of . We define a triangle of as a set of edges forming a triangle. We say that has a triangle packing of size if there exist edge-disjoint triangles .
1.2.2 Gaussians
We define as the Gaussian distribution with mean and variance . Moreover, we define as the -dimensional Guassian distribution, where a sampled vector has for each and is independent of for . We make use of the fact that sampling a vector in this way is equivalent to sampling a random direction; that is, after normalizing, becomes a uniformly random unit vector in -dimensional space. In particular, this means that is symmetric up to rotations, which we will exploit frequently. We will also make use of the following lemma, which says that is roughly equivalent to a uniform distribution close to .
Lemma 4.
For a randomly sampled , we have that
for all .
The proof of Lemma 4 follows from direct calculation and is not instructive, so we omit it.
2 Triangle Packing and Maximum Cut Tradeoff
This section is devoted to proving the following structural result.
Theorem 5.
If is an interval graph, then either has a triangle packing of size or has a cut of size at least that can be found in polynomial time.
2.1 Warmup: Tradeoff for Split Graphs
As a warmup, we prove the following structural result that is essentially equivalent to Theorem 5, except it is for split graphs instead of interval graphs.
Theorem 6.
If is a split graph, then either has a triangle packing of size or has a cut of size at least that can be found in polynomial time.
Before we prove this, we need the following helpful lemma.
Lemma 7.
The complete graph on vertices has an edge-disjoint triangle packing of size .
Edge-disjoint triangle packings in complete graphs have been studied in the literature before, with [6] giving an optimal bound. Lemma 7 is not very close to the optimal bound, but it is sufficient for our purposes and substantially easier to prove.
Proof.
Label the vertices of with numbers . Label each triangle with . Fix any edge . Then we have that, for each possible triangle label, there is only one triangle involving with that label. So we may take all the triangles of any specific label and find an edge-disjoint triangle packing. By the pigeon hole principle, some label has at least triangles. Therefore, has a triangle packing of size , as wanted.
Theorem 6.
Write where and are the clique and independent portions of , respectively. Then we can partition . If , then we are done, as we have constructed a cut of size at least in polynomial time. Otherwise, we have that . Now, recalling that is the complete graph on vertices, we use Lemma 7 to find that (and thus ) has an edge-disjoint triangle packing of size at least .
2.2 Finding a Cut
The proof of Theorem 5 maintains a similar flavor as that of Theorem 6. While the proof of Theorem 6 either finds a large cut or one large clique to pack triangles into, we need to repeatedly apply Lemma 7 during the proof of Theorem 5, as there may be no single large-enough clique. That is, at each stage, we either identify a clique that we may pack triangles into and remove while being careful to bound the number of non-clique edges we remove, or we conclude there is a large cut.
Given an interval graph , we proceed by partitioning the vertices into “small” and “large” vertices . At each step, we will either certify that the implied cut is large , or find a way to expand a triangle packing on edges in .
For any , define to be the “bag” of vertices whose intervals occupy position . For a vertex , define to be the size of the smallest bag in ’s interval. Let be a constant we will decide the value of later. We say that is “small” if and “large” otherwise. In other words, if at least a constant fraction of ’s edges “come from” , then is small. We define as the set of small vertices and as the set of large vertices.
Lemma 8.
.
Proof.
Define an ordering on by if . If we say if the leftmost point of is to the left of the leftmost point of . For with and equal leftmost point, we break ties arbitrarily to extend into a total ordering.
Fix any . We will bound the number of edges with as a function of . Consider any such that does not intersect either the leftmost point or rightmost point of . Then we must have that and thus . Combined with the fact that the leftmost bag of is to the right of the leftmost bag of this implies that . Thus, we need only consider such that intersects either the leftmost or rightmost bag of .
Let be the set of such that contains the leftmost point of . Let be the first of these neighbors sorted by increasing leftmost point. Similarly, let be the last of these neighbors sorted by increasing rightmost point. Now notice that all have and thus . This is because every point of either intersects all of or all of , which have size at least assuming is non-empty. See Figure 4 for an illustration. Thus, the number of vertices such that is at most . We can similarly bound the number of vertices such that contains the rightmost point of and by .
Putting these bounds together with the fact that , we can bound the total number of edges with by . Thus, we find that .
Unlike with large vertices, we cannot unconditionally bound . We instead introduce a condition that, if unsatisfied, will allow us to make progress towards a triangle packing.
Lemma 9.
If, for some , all have , then .
Proof.
Fix any non-isolated . Let denote the leftmost point of . Let denote the set of small neighbors of whose intervals contain . Note that . By the definition of , we can bound . Additionally, we have that and thus . Now we can bound by iterating over all small vertices
2.3 Building a Triangle Packing
Lemma 10.
If, for some and , , then either
-
1.
we can pack at least edge-disjoint triangles into or
-
2.
has a bridge.
Proof.
Suppose . Then by Lemma 7, we can pack at least edge-disjoint triangles into . By the definition of , we have that
This fulfills Condition 1.
Now suppose . Label . If , then is a bridge of and thus Condition 2 is fulfilled. Otherwise, let . We can pack a single triangle into the edges . Further, we have that , and so . This fulfills Condition 1.
This leads us to Algorithm 1. We will iteratively find a clique to pack triangles into, delete the clique, and continue. If we reach a point where we have deleted at least edges, then we can finish as we have packed triangles into a constant fraction of the edges. Otherwise, if we run out of cliques to pack triangles into, we certify that we have an almost-complete cut on the remaining graph, which still contains most of the original edges of . Finally, whenever we identify a bridge, we can simply delete it from the graph and add it to our final cut, as bridges can always be added to any cut.
Note that Algorithm 1 runs in polynomial time. There are at most iterations of the while loop, as each iteration either returns or removes at least one edge.
Lemma 11.
If returns the set from within the while loop, then and is a valid cut.
Proof.
Suppose that Algorithm 1 returns on the th iteration of the while loop. By Lemmas 8 and 9, we have that edges. Note that and due to the while loop condition, we have that . Thus, .
To see that is a valid cut, note that adding a bridge to a bipartite graph cannot introduce an odd cycle. Thus, we can iteratively add each edge of to without invalidating our cut.
Lemma 12.
If exits the while loop, then has an edge-disjoint triangle packing of size at least .
Proof.
At the conclusion of the while loop, we have that . Due to Lemma 10, each time we expand by edges, we can pack an additional edge-disjoint triangles into . By iterating this process, there exists an edge-disjoint triangle packing of size at least in .
2.4 No Tradeoff for Chordal Graphs
In this subsection, we will show that there is no equivalent tradeoff for chordal graphs as there are for interval graphs and split graphs. Thus, improving upon for chordal graphs will likely require new algorithmic insights.
Theorem 13.
For all , there exists a chordal graph such that and has no edge-disjoint triangle packing of size .
Proof.
Suppose we are given a chordal graph with and no triangle packing of size . We will show later how to obtain such a . Construct by setting and . That is, for each edge , we attach one fresh vertex to each endpoint of the edge, and create a “universal” vertex connected to all vertices in the graph.
We first note that is chordal, as it is obtained from a chordal graph by iteratively adding simplicial vertices. By inspection, any triangle must contain at least one edge from . Thus, the maximum number of edge-disjoint triangles in is at most . For each edge , contains a 5-cycle , and so contains edge-disjoint 5-cycles. Thus, we have that . Note that , so .
It remains to show that a chordal graph with the desired properties exists. We give the following interval graph construction for . Select large enough. Create vertices in a “segment-tree” pattern as follows. In the first layer, create one vertex with interval . In the second layer, create two vertices with intervals and respectively. In the third layer, create four vertices with intervals and respectively. Then iterate this process for total layers, see Figure 5 for an illustration.
We have that and by counting only the edges adjacent to the bottom layer. Thus, for sufficiently large, we have that .
To show that has no edge-disjoint triangle packing of size , it is sufficient to show that , as each triangle results in at least one un-cuttable edge. Consider the cut of obtained by taking the bottom layers as one side of the cut. The number of edges in this cut is . Note also that . Therefore, we have that . As tends to infinity, the term tends to . So, by setting to be a sufficiently large constant based on and letting be sufficiently large based on , we have that and so has no edge-disjoint triangle packing of size . Thus, fulfills all the conditions we needed to produce .
3 Analysis of the Perturbed Goemans-Williamson Algorithm
Our second main contribution is an improved approximation for Maximum Cut in graphs with large triangle packings. Given a fixed triangle , it is impossible for all angles to be equal or very close to the critical angle . However, it is possible for to achieve the critical angle and have . In this case, despite the entire graph being a single triangle, the Goemans-Williamson rounding algorithm will not perform beyond its worst-case guarantee. This is because the contribution of the edge to the objective function is , and so rounding it “better” does not actually increase our expected value.
To grapple with this issue, we introduce the “Perturbed Goemans-Williamson Algorithm.” Intuitively speaking, this algorithm randomly “perturbs” each vector slightly. We will see that edges with near-zero angle stand to gain much more in this perturbation process than any other edges have to lose besides those with an angle of nearly . Thus, any SDP solution with many near-zero angle edges and few near- angle edges can be rounded with a guarantee better than .
Before presenting the algorithm, we must first define the semidefinite program from which we will round a solution.
| maximize: | (SDP-GW) | |||||
| subject to: | ||||||
Note that Algorithm 2 runs in polynomial time, as solving SDPs and sampling from a Gaussian distribution can be made to run in polynomial time.
Theorem 14.
If has an edge-disjoint triangle packing of size at least , then
Theorem 14, when combined with Theorem 5 and Theorem 6, immediately shows that there is a polynomial-time -approximation for Maximum Cut on interval graphs and a polynomial-time -approximation for Maximum Cut on split graphs.
Note that is the result of running the original Goemans-Williamson algorithm, so the result of Algorithm 2 is immediately at least .
For an edge , let and be the random variables indicating that is cut by and , respectively. Define as the angle between and . For , let
be the set of edges with angle “close to” , where is a parameter of Algorithm 2. We first deal with the case where there are many edges in .
Lemma 15.
If and , then .
Proof.
Let equal the value of SDP-GW at optimal solution . Consider any . Recall that the contribution of to is . Also, by simple calculation, we have that . We calculate
The first inequality is by calculating for . The second inequality follows from the fact that is increasing on . The lemma now follows from the fact that .
Now we show that for edges not in , is not much worse than . For technical reasons, our lemma statement also excludes edges in . We will see later that for all edges with angle at most , is no worse than . Additionally, for edges with angle very close to , is substantially better than .
Lemma 16.
For all , we have that .
Proof.
For any vertex , let be the random variable indicating that is randomly selected. Fix any . We first note that and .
We now turn our attention to . Note that and lie on a single plane through the origin, so by symmetry of , we can assume without loss of generality that and for . Label the random variable . By symmetry of output, we can assume without loss of generality that . Note that these assumptions imply that is independent of the value of .
Suppose that . Then we have that , and so . That is, the sign of is entirely determined by the sign of . This implies that, when , . Thus, by independence of and the value of ,
We then apply Lemma 4 to bound
Let . We wish to bound . To this end, we manipulate probabilities
Now we find
To bound the numerator of the error term, first consider . Recall that and . Thus, assuming holds, in order for to hold, we must have that . Using Lemma 4, we can bound
and using Lemma 4 again,
Applying this with yet another application of Lemma 4 gives
This yields the desired result, that .
Putting it all together, we find that
and using Lemma 4,
Noting that because completes the proof of the lemma.
Lemma 16 allows us to bound the perturbation loss on all angles sufficiently bounded away from and . However, we will not be able to guarantee that the angles we consider are sufficiently bounded away from to properly utilize Lemma 16. Thus, we strengthen Lemma 16 to show that perturbation does not cause any loss on angles below .
Lemma 17.
For all with , we have that .
Proof.
Take any such that . As in the proof of Lemma 16, restrict to two dimensions, rotate, and reflect so we may assume for , and . Due to our assumption on , we have that . As earlier, define the event for . As in the proof of Lemma 16, our main task is to bound . However this time, we will show .
The event is equivalent to . Since , if happens, then we must have , so
We have that
Recall that and , so . Thus, the event contains a larger portion of the negative space than the positive space and
Recalling that , and and are independent, we can now calculate
and by symmetry, . A similar computation as that in the proof of Lemma 16 completes the proof.
Lemma 18.
For all , we have that .
Proof.
Take any and calculate, using Lemma 4,
Now notice that if , then . Thus, if a substantial fraction of has very small angle, then will be noticeably larger than . To this end, define
Note that for each , we have .
Lemma 19.
If and , then
Proof.
We have shown that if either or contain a constant fraction of , then we can improve upon Goemans-Williamson, so we may assume from now on that both of these sets have negligible size. Now we will utilize the fact that has an edge-disjoint triangle packing of size to show that there are many edges not near the critical angle . Define the set
Lemma 20.
.
Proof.
Fix any triangle . We can bound the sum of angles , and so at least one has . Thus, in any triangle , we have that is non-empty. Because has edge-disjoint triangles, there are at least edges in . Subtracting out those edges in completes the proof.
We note that , which motivates the following lemma.
Lemma 21.
If and , then
Proof.
The assumptions on and imply, through Lemma 20, that . We calculate
Proof.
4 Hardness of Approximating Maximum Cut on Split Graphs
In this section, we show that, subject to the Small Set Expansion Hypothesis, Maximum Cut on split graphs is hard to approximate to some constant. Our starting point is hardness of finding a small balanced cut on weighted graphs. Given a graph with weights for , we define
as the normalized set size of . Here, is defined as the weighted degree of . In an unweighted regular graph, we have that . Also define
as the edge expansion of . In an unweighted -regular graph, we have that .
Lemma 22 (Corollary 3.6 of [18]).
There is a constant such that for any sufficiently small , it is SSE-hard to distinguish between the following two cases for a weighted graph :
Yes: There exists such that and .
No: Every with satisfies .
To utilize this hardness, we first create an unweighted instance as follows.
Lemma 23.
There is a constant such that for any sufficiently small , it is SSE-hard to distinguish between the following two cases for an unweighted, non-simple graph with :
Yes: There exists such that and .
No: Every with satisfies .
Proof.
We begin with a gap instance of the form in Lemma 22. We will create a weighted instance with the same vertex and edge set such that each edge has weight in . Duplicating edges according to their weight will then yield the desired statement for unweighted graphs.
First, we may assume without loss of generality that the edges in are scaled such that the sum of degrees is . This is because multiplying the weight of all edges by an equal constant factor does not change the results of or . Now produce the weights by setting . Note in particular that this implies for all . Define , and analogously to , and , except for weights . Note that for all and so for all .
Suppose that is in the Yes case of Lemma 22, and let have and . Then we calculate
Thus, for sufficiently large , we have that . By a similar calculation, we have that and so . We now aim to show that . We calculate
Here, the second inequality uses the fact that and so . Thus, for sufficiently large , we have that . This establishes that the Yes case of maps to the Yes case of .
Suppose that is in the No case of Lemma 22. Then any with either has or . Take any . If , then . So consider the case in which . Then we calculate
If we set , then for sufficiently large , . This establishes that the No case of maps to the No case of .
Our final reduction to Maximum Cut requires our graph to be regular and , so before proceeding, we must first take a standard sparsification step.
Lemma 24.
There is a constant such that for any sufficiently small , it is SSE-hard under randomized reductions to distinguish between the following two cases for an unweighted, non-simple graph with and maximum degree at most , where is the minimum degree:
Yes: There exists such that and .
No: Every with satisfies .
Proof.
We begin with a gap instance of the form in Lemma 23. We randomly create as follows. For each vertex , create copies and add them to . That is, we set Let map copies to their original vertex . For each edge and copies , let for a parameter we will adjust later. Now add copies of to and randomly add a final edge with probability . Note that the expected number of edges is equal to .
Let us consider the properties of . We have that and
Let be the random variable denoting the number of edges in “produced” from . By a standard application of Chernoff bounds and the union bound over and , we can show, for any constant , that
-
1.
for all and
-
2.
for all
with high probability, assuming . In particular, item 1 implies that for all , where is the degree of in . Suppose from now on that items 1 and 2 are both true. Now suppose is a Yes instance of Lemma 23. That is, there is a such that and . Let , and be defined for as is defined for . Then we calculate
Applying the same calculation for yields the upper bound
For sufficiently small , this implies that . Let be defined for as for . We calculate
For sufficiently small , we get the upper bound . This establishes that the Yes case of maps to the Yes case of with high probability.
Now suppose that is a No instance of Lemma 23. Consider any with . Let be a vector defined by . That is, indicates the proportion of each original vertex selected by . We calculate
Let be a random variable sampled by including with probability . Then we have that
We can apply Chernoff bounds on to find that with high probability. We additionally calculate
By setting sufficiently small and applying Chernoff bounds, this implies that with probability at least . Thus, by the probabilistic method, there exists some such that and . Applying Lemma 23, we find that
So, set . This establishes that the No case of maps to the No case of with high probability.
We now convert the language of and to the simpler language of cardinalities of sets.
Lemma 25.
There exists a constant such that for all sufficiently small , it is SSE-hard under randomized reductions to distinguish between the following cases for an unweighted graph with :
Yes: There exists a cut with such that .
No: For all cuts with , either or .
Proof.
We begin with a gap instance of the form in Lemma 24. We will not need to modify this instance to produce our desired result. Suppose that is a Yes instance of Lemma 24. Then there is some with and . Using that every vertex has degree , where is the minimum degree of , we can bound
Thus, . For sufficiently small , we can then bound below . Because , we also bound . If , then swap with , noting that . This fulfills the condition. Similarly, we can bound
With the bound of from Lemma 24, this implies that
for sufficiently small . Thus, being a Yes instance of Lemma 24 implies the Yes conditions of this lemma.
Suppose that is a No instance of Lemma 24. Consider any . Suppose first that . Then we have that
and so . For sufficiently small, this implies . Suppose otherwise, that . Then we have that
and so . Setting finishes the proof.
We now reduce from gap instances of the type in Lemma 25 to show hardness for Maximum Cut on split graphs.
Theorem 3. [Restated, see original statement.]
There exists a constant such that there is no polynomial-time -approximation algorithm for Maximum Cut on split graphs, subject to the Small Set Expansion Hypothesis under randomized reductions.
Proof.
We reduce from a graph of the form in Lemma 25 to a (simple) split graph as follows. Let the clique portion of be , and let the independent portion of be . We define the edge set of as . That is, we place a copy of in the clique portion of , and place one vertex for each edge of in the independent set portion, each connected to its two endpoints in .
Note that in a split graph, the maximum cut is defined solely by its intersection with the clique portion of the graph, as the decision for vertices in the independent set portion can be made greedily. Define for to be the edges in the maximum cut of defined by , breaking ties arbitrarily.
Suppose that is a Yes instance as defined in Lemma 25. Then by direct counting, we have that, in ,
Define . Suppose instead that is a No instance and consider any with . We will show that, with the right choice of and , for some constant . If , then
Because , and for sufficiently small , this implies that for some constant when and are sufficiently small. Otherwise, if , then we must have that . Then
As above, because for sufficiently small , we have that for some constant when and are sufficiently small. Thus, in the No case, we have that . Therefore, it is SSE-hard to distinguish between and . This implies SSE-hardness of approximating Maximum Cut to a factor of on split graphs.
It is critical that Lemma 25 allows for non-simple graphs. If were simple, then our reduction in the proof of Theorem 3 results in the relation , where is the complement of . This is proved explicitly in [3]. Then, if , the value is a good estimate for . If , then . In this case, is a dense graph and so there is a PTAS for [7, 16]. In particular, these cases imply that there is a PTAS for whenever is a 2-split graph (i.e., all vertices in have degree ) without any “duplicated” vertices of that have the same neighborhood.
References
- [1] Ranendu Adhikary, Kaustav Bose, Satwik Mukherjee, and Bodhayan Roy. Complexity of maximum cut on interval graphs. Discrete & Computational Geometry, 70(2):307–322, 2023. doi:10.1007/S00454-022-00472-Y.
- [2] Alexey Barsukov and Bodhayan Roy. Maximum cut on interval graphs of interval count two is np-complete. arXiv preprint, 2022. arXiv:2203.06630.
- [3] Hans L Bodlaender and Klaus Jansen. On the complexity of the maximum cut problem. Nordic Journal of Computing, 7(1):14–31, 2000.
- [4] Celina MH de Figueiredo, Alexsander A de Melo, Fabiano S Oliveira, and Ana Silva. Maximum cut on interval graphs of interval count four is np-complete. Discrete & Computational Geometry, 71(3):893–917, 2024. doi:10.1007/S00454-023-00508-X.
- [5] Tomas Feder and Pavol Hell. List homomorphisms to reflexive graphs. Journal of Combinatorial Theory, Series B, 72(2):236–250, 1998. doi:10.1006/JCTB.1997.1812.
- [6] Tomás Feder and Carlos Subi. Packing edge-disjoint triangles in given graphs. In Electronic Colloquium on Computational Complexity (ECCC), volume 19, page 13, 2012.
- [7] W Fernandez De La Vega. Max-cut has a randomized approximation scheme in dense graphs. Random Structures & Algorithms, 8(3):187–198, 1996. doi:10.1002/(SICI)1098-2418(199605)8:3\%3C187::AID-RSA3\%3E3.0.CO;2-U.
- [8] Michel X Goemans and David P Williamson. Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming. Journal of the ACM (JACM), 42(6):1115–1145, 1995. doi:10.1145/227683.227684.
- [9] Venkatesan Guruswami. Maximum cut on line and total graphs. Discrete applied mathematics, 92(2-3):217–221, 1999. doi:10.1016/S0166-218X(99)00056-6.
- [10] Frank Hadlock. Finding a maximum cut of a planar graph in polynomial time. SIAM Journal on Computing, 4(3):221–225, 1975. doi:10.1137/0204019.
- [11] Laurie J Hendren, Guang R Gao, Erik R Altman, and Chandrika Mukerji. A register allocation framework based on hierarchical cyclic interval graphs. In Compiler Construction: 4th International Conference, CC’92 Paderborn, FRG, October 5–7, 1992 Proceedings 4, pages 176–191. Springer, 1992. doi:10.1007/3-540-55984-1_17.
- [12] Jun-Ting Hsieh and Pravesh K Kothari. Approximating max-cut on bounded degree graphs: Tighter analysis of the fkl algorithm. arXiv preprint arXiv:2206.09204, 2022. doi:10.48550/arXiv.2206.09204.
- [13] John R. Jungck and Rama Viswanathan. Chapter 1 - graph theory for systems biology: Interval graphs, motifs, and pattern recognition. In Raina S. Robeva, editor, Algebraic and Discrete Mathematical Methods for Modern Biology, pages 1–27. Academic Press, Boston, 2015. doi:10.1016/B978-0-12-801213-0.00001-0.
- [14] Richard M. Karp. Reducibility among combinatorial problems. In Complexity of Computer Computations, pages 85–103, 1972. doi:10.1007/978-1-4684-2001-2_9.
- [15] Subhash Khot, Guy Kindler, Elchanan Mossel, and Ryan O’Donnell. Optimal inapproximability results for max-cut and other 2-variable csps? SIAM Journal on Computing, 37(1):319–357, 2007. doi:10.1137/S0097539705447372.
- [16] Claire Mathieu and Warren Schudy. Yet another algorithm for dense max cut: go greedy. In SODA, volume 8, pages 176–182, 2008. URL: http://dl.acm.org/citation.cfm?id=1347082.1347102.
- [17] Prasad Raghavendra and David Steurer. Graph expansion and the unique games conjecture. In Proceedings of the forty-second ACM symposium on Theory of computing, pages 755–764, 2010. doi:10.1145/1806689.1806792.
- [18] Prasad Raghavendra, David Steurer, and Madhur Tulsiani. Reductions between expansion problems. In 2012 IEEE 27th Conference on Computational Complexity, pages 64–73. IEEE, 2012. doi:10.1109/CCC.2012.43.
- [19] Luca Trevisan. Non-approximability results for optimization problems on bounded degree instances. In Proceedings of the thirty-third annual ACM symposium on Theory of computing, pages 453–461, 2001. doi:10.1145/380752.380839.
