On Average Baby PIH and Its Applications
Abstract
The Parameterized Inapproximability Hypothesis (PIH) asserts that no FPT algorithm can decide whether a given CSP instance parameterized by the number of variables is satisfiable, or at most a constant fraction of its constraints can be satisfied simultaneously. In a recent breakthrough, Guruswami, Lin, Ren, Sun, and Wu (STOC 2024) proved the PIH under the Exponential Time Hypothesis (ETH). However, it remains a major open problem whether the PIH can be established assuming only . Towards this goal, Guruswami, Ren, and Sandeep (CCC 2024) showed a weaker version of the PIH called the Baby PIH under . In addition, they proposed one more intermediate assumption known as the Average Baby PIH, which might lead to further progress on the PIH. As the main contribution of this paper, we prove that the Average Baby PIH holds assuming .
Given a CSP instance where the number of its variables is the parameter, the Average Baby PIH states that no FPT algorithm can decide whether (i) it is satisfiable or (ii) any multi-assignment that satisfies all constraints must assign each variable more than values on average for any fixed constant . So there is a gap between (i) and (ii) on the average number of values that are assigned to a variable, i.e., vs. . If this gap occurs in each variable instead of on average, we get the original Baby PIH. So central to our paper is an FPT self-reduction for CSP instances that turns the above gap for each variable into a gap on average. By the known -hardness for the Baby PIH, this proves that the Average Baby PIH holds under .
As applications, we obtain (i) for the first time, the -hardness of constant approximating -ExactCover, and (ii) a tight relationship between running time lower bounds in the Average Baby PIH and approximating the parameterized Nearest Codeword Problem (-NCP).
Keywords and phrases:
Average Baby PIH, Parameterized Inapproximability, Constraint Satisfaction Problem, Exact Set Cover, W[1]-hardnessCopyright and License:
![[Uncaptioned image]](x1.png)
2012 ACM Subject Classification:
Theory of computation Problems, reductions and completeness ; Theory of computation Parameterized complexity and exact algorithmsAcknowledgements:
The authors want to thank Guohang Liu, Mingjun Liu, Yangluo Zheng for discussions in the early stage of this work. The comments and suggestions from anonymous reviewers also help to improve the paper significantly. In particular, Theorem 2 is due to one of them.Funding:
Yuwei Liu and Yijia Chen are supported by the National Natural Science Foundation of China (Project 62372291).Editors:
Olaf Beyersdorff, Michał Pilipczuk, Elaine Pimentel, and Nguyễn Kim ThắngSeries and Publisher:

1 Introduction
In classical complexity theory, the PCP theorem [3, 2, 8] serves as an essential tool for proving most of the existing results in the hardness of approximation. As a variant, the Multi-Assignment PCP theorem [1, Lemma 11] states that, for any constant and , it is even NP-hard to decide whether a CSP instance is satisfiable, or any multi-assignment (see Definition 7.) that assigns each variable no more than values cannot satisfy a -fraction of constraints. Among others, the Multi-Assignment PCP theorem was used to show the NP-hardness of approximating SetCover [1]. It turns out that the Multi-Assignment PCP theorem is a simple consequence of the PCP theorem by a straightforward probabilistic argument. Nevertheless, Barto and Kozik [4] gave a direct and purely combinatorial proof for the simple case of . Observe that it means that the CSP instances under consideration are either satisfiable or cannot be satisfied by a desired multi-assignment. This restricted version of the Multi-Assignment PCP theorem is termed the Baby PCP Theorem in [4].
As an analog of the PCP theorem in parameterized complexity theory, the Parameterized Inapproximability Hypothesis [21], PIH for short, is an important assumption from which we can prove many FPT inapproximability results, including the inapproximability of -Clique, -ExactCover [13], and Direct Odd Cycle Transversal [21], etc. It claims that for some constant , no -time (i.e., FPT) algorithm can distinguish a satisfiable 2CSP instance with variables from one where less than -fraction of constraints can be satisfied simultaneously [21]. Unlike the PCP theorem, the PIH is still a major open problem in parameterized complexity. The current state of the art is that the PIH holds under the Exponential Time Hypothesis (ETH) [12, 11], and a proof of the PIH under the minimum assumption remains elusive. Toward this goal, studying some consequences of the PIH and proving them under might provide new insights and valuable lessons.
Recently, Guruswami, Ren, and Sandeep [13] proved a parameterized version of the Baby PCP Theorem, appropriately coined as the Baby PIH, under . The Baby PIH states that for any constant , no FPT algorithm can distinguish a satisfiable 2CSP instance from one with no satisfying multi-assignment which assigns each variable no more than values. Just like the relationship between the PCP theorem and the Baby PCP theorem, the Baby PIH is a direct consequence of the PIH. As a next step, a further complexity assumption is suggested, i.e., the Average Baby PIH [13, Conjecture 3], which seems to be sandwiched between the PIH and the Baby PIH. It postulates the -hardness of the problem known as Avg--Gap- (see Definition 8) which asks for distinguishing a satisfiable 2CSP instance from one without satisfying multi-assignment which assigns each variable no more than values on average. The authors of [13] also demonstrated the difference between the Baby PIH and the Average Baby PIH. In fact, for all and , they constructed 2CSP instances with variable set that cannot be satisfied by any multi-assignment assigning each variable in no more than values, but can be satisfied by a multi-assignment that assigns in total values to all the variables in , that is, every variable is assigned values on average. Compared to proving -hardness111Strictly speaking, the PIH is not a computational problem and we cannot directly define its hardness. The formal statement should be “proving -hardness of the problem described in the PIH”, and we use “-hardness for the PIH” for short in the introduction. of the PIH, it is apparently easier to show the -hardness of the Average Baby PIH, and studying the Average Baby PIH might bring us further closer to a proof of the -hardness for the PIH. Moreover, the Average Baby PIH is already sufficient for proving some non-trivial inapproximability results such as -ExactCover [13].
1.1 Main Results
Let be a 2CSP instance with a set of variables, an alphabet , and a set of constraints. A multi-assignment relaxes the standard notion of assignments by assigning each variable a set of values in , i.e., . Thereby, is said to be satisfied by if for every constraint , one can pick for each variable of a value from the set assigned to this variable to satisfy . We say that assigns values to in total, or equivalently, each variable in is assigned values on average (see Definition 7). Our main result is the following theorem stating that the Average Baby PIH holds under .
Theorem 1 (Informal, see Theorem 16).
Assume . Then for any constant , given a 2CSP instance parameterized by , no FPT time algorithm can distinguish between:
-
is satisfiable.
-
Any multi-assignment assigning no more than values to does not satisfy .
Clearly any standard assignment can be identified with a multi-assignment that assigns each variable a set of a single value, i.e., . Hence, there is a constant gap in Theorem 1 between YES and NO instances on the average number of values assigned to each variable, which gives us the aforementioned Avg--Gap- problem. On the other hand, the constant gap for the PIH is on the fraction of constraints that can be satisfied by an assignment. That is, a YES instance is a 2CSP instance whose all constraints can be satisfied by an assignment, while any assignment can only satisfy at most a constant fraction of constraints in a NO instance. So, perhaps surprisingly, the difference between the Average Baby PIH and the PIH can be pinpointed within the Avg--Gap- problem precisely in terms of whether a given instance contains a “dense” or “sparse” set of constraints.222This is pointed out by an anonymous reviewer. More precisely:
Theorem 2.
-
Under , the Average Baby PIH holds for Avg--Gap- instances with .
-
If the Average Baby PIH holds for Avg--Gap- instances with , then the PIH holds as well.
As a first application of the Average Baby PIH under , using a reduction in [13], we obtain the -hardness of constant approximating the -ExactCover problem (see Definition 9), improving its previous approxmation lower bound under a stronger assumption, i.e., the Gap-ETH [23].
Theorem 3 (Theorem 27 restated).
For any constant , -approximating -ExactCover is -hard.
We remark that the -hardness of approximating -ExactCover has been a long-standing open problem in parameterized complexity. Although the , , ETH-hardness of approximating the -SetCover problem has been established in [6, 15, 18, 20], as a special case of -SetCover, the hardness of approximating -ExactCover was only known under the PIH [23] prior to our work.
The second application is a close relationship between running time lower bounds for constant approximating the parameterized Nearest Codeword Problem -Gap--NCPp (see Definition 10) and Avg--Gap-. Its proof is a straightforward composition of two known reductions in [23, 13] .
Theorem 4.
For any prime , computable function , and constant , if no -time algorithm can decide Avg--Gap-2CSP with variables for any computable function , then -Gap--NCPp cannot be solved in time for any computable function .
Proof Sketch.
Theorem 4 follows from the gap-preserving reduction [13] from Avg--Gap-2CSP to -ExactCover (see also Appendix A), and the gap-preserving reduction from -ExactCover to -Gap--NCPp [23, Theorem 28]. Note that in both reductions the parameter is preserved.
1.2 Technical Overview: Local-to-Global Reduction For 2CSP
To prove Theorem 1, we show that the -hardness for the Baby PIH implies the -hardness for the Average Baby PIH. Here, the “-hardness for the Baby PIH” means that, for all , it’s -hard to decide whether (i) a 2CSP instance is satisfiable, or (ii) it cannot be satisfied by any multi-assignment assigning each variable no more than values. Thus there is a constant gap between (i) and (ii) in the number of values assigned to each variable. Thereby, the gap is “local.” As already mentioned, for the Average Baby PIH, the gap is on the average number of values, or equivalently, the total number of values assigned to all the variables. Hence, the gap is “global.” Our reduction from the Baby PIH to the Average Baby PIH is thus said to be “local-to-global.”
Technically, our reduction relies on a simple but crucial property of high-distance error-correcting codes (ECC) shown in [16, 20]. In particular, we need an ECC with relative distance such that any two distinct codewords can agree on at most entries. So if we have a set of codewords and an -fraction of entries (denoted by ) with such that for each , we can find distinct that agree on their -th entry, then the size of must be large. The lower bound of is called the collision number of , denote by . A simple counting argument in [20] shows that .
Now given a 2CSP instance , let , parameter , and . We fix some ECC with very large relative distance (e.g., a Reed-Solomon code) for prime and . Then we have . We construct a new 2CSP instance as:
-
, .
-
Each takes value from the (encoding of) satisfying assignments of , i.e., . Each takes value from .
-
For each and , there is a constraint that checks whether ’s assigned value and ’s assigned value satisfy and .
See Figure 1 for an illustration. Finally, we duplicate and each an appropriate number of times to make them equal in size, finishing our reduction.
In general, an assignment to should correspond to a satisfying multi-assignment to the original instance , and the value assigned to each is a “guess” of the -th entry of the encoding of every . It is easy to see that if the input instance is satisfiable, then so does the new 2CSP instance , since for each corresponding to , we can simply assign it the value , where is a satisfying assignment for . At the same time, the value assigned to each is .
For soundness, suppose has no satisfying multi-assignment assigning at most values to each variable, we need to argue that has no satisfying multi-assignment that assigns values in total. To that end, we exploit the collision number of the code . Fix any satisfying multi-assignment to . Recall that each variable is assigned some value from a satisfying partial assignment to . Then, naturally gives a satisfying multi-assignment to , which implies that there exists a variable such that more than different values are assigned by to , for which the corresponding constraint contains .
Now we have two possible cases for . In the first case, for -fraction of variables in , the multi-assignment assigns each of them more than values. We are done, since this implies that the total number of assigned values by is more than . For the second case, there exists an -fraction of variables in , each of which is assigned by at most values. Given such a variable , we have more than different codewords assigned to in ’s position which has at most possible values in the -th entry. This entails the existence of two different codewords with the same -th entry. Since there are such entries, the assignment to must contain at least different codewords. Each assignment to contributes two codewords, so the total number of assigned values by is at least . In summary, both cases guarantee a constant gap on either or , showing that any satisfying multi-assignment to must assign values to in total.
More details are referred to Section 3.
1.3 Discussions
For minimization problems, the technique of constructing two parts of variables and arguing that at least one part has a large gap seems quite general, as exhibited by the previous works showing the -hardness of approximating -SetCover [6, 20] and -NCP [17]. The use of the collision number of error-correcting codes in the context of parameterized inapproximability was first introduced in [16] and further developed in [20, 17]. We ask whether these techniques can be unified.
Question 1.
Is there a general framework for proving parameterized inapproximability of minimization problems?
We also suggest two new variants of the Average Baby PIH, which might serve as a next step towards proving the PIH under . On closer inspection of our construction, particularly Case 1 in the proof of Lemma 20, it only guarantees the total number of values assigned to (i.e., ) is large. This could happen if an -fraction of ’s in is assigned super-constant number of values. We ask if a larger gap can be achieved. So the first one asks whether the Average Baby PIH can be strengthened by requiring a constant fraction of variables to be assigned multiple values.
Question 2.
Under , can we prove that for any constant , there exists a constant such that no FPT algorithm can decide whether a 2CSP instance is satisfiable, or any satisfying multi-assignment must have at least a -fraction of variables assigned values?
The second variant is already contained in Theorem 2, thus equivalent to PIH.
Question 3.
Under , can we prove that the Average PIH holds even for Avg--Gap- instances with the number of constraints being linear in the number of variables?
It is also interesting to consider whether the inapproximability factor in the Average Baby PIH can be improved to , since this would directly lead to better lower bounds for approximating -ExactCover. The current obstacle is that, although the running time of our reduction does not depend on the approximation factor, our reduction relies on the gap created in the Baby PIH [13]. In order to achieve an -gap in the Baby PIH, the reduction in [13] runs in time , consequently, the existing FPT reduction cannot create a super-constant gap .
Question 4.
Under , can we prove that the Average PIH (Theorem 1) holds for inapproximability factor , hence giving better inapproximability result for -ExactCover?
1.4 Organization
In Section 2, we introduce the main computational problems and complexity assumptions studied in this paper. As the central contribution, Section 3 explains our reduction from the Baby PIH to the Average Baby PIH. This, in fact, establishes the Average Baby PIH under . In Appendix A, we present a reduction from Avg--Gap- to the constant approximation of -ExactCover, which slightly differs from the construction in [13].
2 Preliminaries
For a positive integer , we use to denote the set . is the disjoint union of sets , where we tacitly assume that are pairwise disjoint. We use (without subscript) to denote the logarithm number with base . For any prime number , we write for the (unique) finite field of size . The asymptotic notations, i.e., , and , are used following the general convention. The reader is assumed to be familiar with basic notions in parameterized complexity theory, in particular FPT and . Otherwise, the standard references are, e.g., [10, 9, 7].
2.1 Problems
Definition 5 (Parameterized 2CSP).
A instance is defined as a triple where:
-
is a set of variable.
-
, where each contains values that the variable can be assigned. Often, we assume that there exists an such that for all .
-
, where each for some , and is a subset of .
The problem is to decide whether there exists an assignment that satisfies:
-
For all , .
-
For all , .
The parameter for this problem is , the number of variables. Each pair of variables has at most one constraint, so . Without loss of generality, each variable is related to some constraint in . The size of instance is defined as , where the size of each is defined as .
The approximation of parameterized 2CSP refers to the following problem.
Definition 6 (-Gap-2CSP).
Given a 2CSP instance with parameter , we want to distinguish between:
-
is satisfiable;
-
any assignment can satisfy at most an -fraction of constraints in .
As already mentioned, the notion of multi-assignment extends the usual assignment in such a way that each variable can be assigned multiple values.
Definition 7 (Multi-assignment).
A multi-assignment of a 2CSP instance is a function , 333Here we use to denote the power set of . such that for all we have . Furthermore, we say that satisfies if:
-
For all , there exist and with .
The individual size of is defined as , and the total size of is .
Let . We say that a 2CSP instance is -list satisfiable if there exists a multi-assignment with individual size no more than which satisfies , and is -average list satisfiable if there exists a multi-assignment with total size no more than which satisfies .
Definition 8 (Avg--Gap-).
Given a instance , the goal is to distinguish between the following two cases:
-
is satisfiable.
-
is not -average list satisfiable.
We also consider the -ExactCover problem (aka, the -UniqueSetCover problem) and the -NCP problem (aka, -MLD, for the parameterized Maximum Likelihood Decoding problem) as defined below.
Definition 9 (-ExactCover).
Given a set (which we call universe) and a collection of ’s subsets , the goal is to distinguish between the following two cases:
-
there exist at most disjoint sets in that form a partition of ,
-
or is not the union of any sets in .
Definition 10 (-NCP).
For prime , integer , given a (multi-)set of vectors in , and a target vector , the -NCPp problem asks for distinguishing between:
-
the Hamming distance between and the vector space spanned by is at most ,
-
or the Hamming distance between and the vector space spanned by is at least .
2.2 Hypotheses
Hypothesis 11 (PIH [21]).
For every constant , there is no FPT algorithm solving -Gap-2CSP.
The Baby PIH, a hypothesis implied by PIH, asserts the hardness of approximating individual size of a satisfying multi-assignment. Formally,
Hypothesis 12 (Baby PIH [13]).
For any constant , no FPT algorithm can on input a 2CSP instance, distinguish whether it is satisfiable, or cannot be satisfied by any multi-assignment with individual size at most .
We emphasize that the Baby PIH is a hardness hypothesis with a local condition, i.e., the individual size of satisfying assignments. It is shown that the standard assumption implies the Baby PIH:
Theorem 13 ([13]).
The Baby PIH holds under .
In contrast, the Average Baby PIH is defined on a global condition concerning the total size of satisfying assignments. The precise statement of this complexity assumption contains a technical property on the “shape” of the constraints in a 2CSP instance.
Definition 14 (Rectangular relation).
A 2CSP instance is said to have rectangular relations if for each , there exist a set and mappings , such that iff . We call the underlying set of .
Some explanation for “rectangular” might be in order. Recall that a subset is a (combinatorial) rectangle if and only if there exist such that . It is easy to verify that is rectangular if and only if is the union of a set of pairwise disjoint rectangles.
Hypothesis 15 (Average Baby PIH).
For any constant , there exists no FPT algorithm solving the Avg--Gap-2CSP problem, even when the instance contains only rectangular relations.
3 Average Baby PIH from Baby PIH
In this section, we show that the Average Baby PIH even for instances with only rectangular relations, is implied by the Baby PIH.
3.1 Proofs of Main Results
We employ a local-to-global reduction developed in [17] to amplify the local gap for one variable (Theorem 13) into a global gap for all variables, thus proving the Average Baby PIH from the Baby PIH.
Theorem 16.
Under , for any constant , no FPT algorithm can distinguish a given CSP instance with rectangular relation is satisfiable, or cannot be satisfied by any multi-assignment with total size no more than .
To show Theorem 16, we first introduce some tools from coding theory. The collision number of an error-correcting code characterizes the number of codewords needed to find “collision” on a constant fraction of coordinates. We use the definition in [17]:
Definition 17 (-Collision Number).
Let and with . For every we say that and collide on position if . Furthermore, a subset collides on position if there exist distinct with . We define the collision set of as
Observe that if , then .
Now for every and the -collision number of , denoted by , is the maximum such that for all we have
For Reed-Solomon codes, we have the following lower bounds on their collision number.
Theorem 18 (Theorem 10 in [20], see also [16]).
For any , any Reed-Solomon code with sufficiently large , .
Our proof of Theorem 16 consists of two reductions. The first one (Lemma 20) reduces CSP instances from the Baby PIH, i.e., with a “local” gap as explained in Section 1.2, to a new instance whose constraints are between two disjoint groups of variables. The new instance has a different “global” gap on each group of variables. As the sizes of the two groups might not be balanced, we do not necessarily have a “global” gap on all the variables. But this is easily remedied by the second reduction (Lemma 22) which makes an appropriate number of copies of the two groups.
Definition 19 (-Average Multi-Assignment).
Let be a bipartite CSP instance, in particular and every has and . Then for an -average multi-assignment of is a multi-assignment such that
and |
That is, the total size of restricted to is at most , and the total size of restricted to is at most (cf. Definition 7). We say is -average list satisfiable if there is an -average multi-assignment which satisfies .
Lemma 20.
There is an algorithm which on input a CSP instance , , and computes a bipartite CSP instance with the following properties.
- Completeness.
-
If is satisfiable, then so is ,
- Soundness.
-
For every if is not -list satisfiable, then is not -average list satisfiable for every with
- Rectangularity.
-
All constraints in are rectangular.
In addition, there exists a computable function upper bounding the running time of as
(1) |
And the number of variables and the number of constraints in can also be upper bounded by .
Proof.
For the given CSP instance we let
and |
Thereby we fix some enumerations of the variables in and the constraints in as
and |
Let be a Reed-Solomon code with
and |
Clearly , and therefore we can assume without loss of generality
Moreover, we only consider the case that
i.e., is sufficiently larger than and .444 Otherwise, the original instance can be solved in time of the form (1), and we can then output some predetermined depending on whether is satisfiable. Hence we can invoke Theorem 18 on , , , and to obtain
(2) |
where the second inequality is by our choice of .
Now the algorithm constructs the following bipartite CSP instance .
- Variables.
-
with
and - Alphabets.
-
where:
-
For every the alphabet of the variable is
(3) That is, contains all the (partial) satisfying assignments of encoded by as pairs of vectors in . (Recall .)
-
For every we have . Since , we have .
-
- Constraints.
-
Let and . Then for every we have a constraint between the variable and which checks whether is assigned to and to such that
and (4) Consequently (4) implies that the constraint is rectangular.555To see this, we take and . Then equation (4) is precisely as in Definition 14. Moreover, the number of constraints in is
The completeness of our reduction is straightforward. So we turn to the soundness. In particular, we assume that the given CSP instance is not -list satisfiable. Furthermore, let be a satisfying multi-assignment for . We need to show that, for any if there is a satisfying -average multi-assignment , then
(5) |
To that end, let
(6) |
That is, is the set of all codewords in that uses for the variables in .
Claim 21.
Let with . Then collides on position .
Proof of Claim 21.
Let be fixed with .
Consider an arbitrary constraint (i.e., ). Since is a satisfying multi-assignment for , there exist
and |
such that and satisfy the constraint between and in . By and (3) there are with and such that
(7) |
Then we say that is -suitable for with respect to , and similarly is -suitable for with respect to .
In addition, by (4)
and | (8) |
Now we define a multi-assignment for the original instance as follows. For every let
(9) |
(Recall that we have fixed an and hence .) Since every variable must appear in at least one constraint (cf. Definition 5), it is easy to see that is a satisfying multi-assignment for by (7).
As is not -list satisfiable, there exists an (i.e., ) with
We have assumed that
so by (9) there is an such that
Hence there are with and such that
-
is -suitable for with respect to ,
-
and is -suitable for with respect to .
Then (8) implies that
In other words, and collide on position . Clearly , so this finishes the proof of the claim.
Let
and |
Now we distinguish two cases.
- 1.
-
2.
There are at most fraction of with . Or equivalently, there are at least fraction of with . Then
So both cases lead to (5) as desired.
With some proper replication, the unbalanced -gap can be turned into a balanced one, and yield the desired -average list unsatisfiability.
Lemma 22.
For any bipartite CSP instance and we can compute in polynomial time a CSP instance with
such that
- Completeness.
-
If is satisfiable, then so is ,
- Soundness.
-
Let . If is not -average list satisfiable for every with , then is not -average list satisfiable. Or equivalently, if is -average list satisfiable, then for some with the bipartite is -average list satisfiable.
Furthermore, if is rectangular, then so is .
Proof.
Let
and |
The desired is constructed as below.
- Variables.
-
consists of copies of and copies of , i.e., where
and Note, , therefore contains many variables.
- Alphabets.
-
where:
-
For every and , let . Recall that is the alphabet for the variable in the original CSP instance .
-
Similarly, for every and let .
-
- Constraints.
-
For every constraint with and , , and we have a constraint
That is, is a copy of where the variable is replaced by its -th copy and by its -th copy . It immediately implies that if is rectangular, then is rectangular too.
Again the completeness is immediate. Towards the soundness, let be a satisfying -average multi-assignment for . In particular,
We set
and | (10) |
It follows that
Note that
(11) |
Therefore,
(by ) | ||||
(by (10)) | ||||
(by (11)) |
Hence, there exists an such that
by . Arguing similarly for we get an such that
Finally we define a multi-assignment for the original instance by
By the above argument, is -average. Moreover, it satisfies , since satisfies .
Putting all pieces together, we have Theorem 16.
Proof of Theorem 16.
We give an FPT reduction from instances in the Baby PIH (Theorem 13) to Avg--Gap-. Then, since the Baby PIH holds under , we deduce that the Average Baby PIH also holds under .
For any CSP instance , we can construct a bipartite CSP instance by Lemma 20, and then construct an Avg--Gap- instance from by Lemma 22. Trivially, is satisfiable when is satisfiable. When is not -list satisfiable, is not -average list satisfiable for all constants with , and thus is not -average list satisfiable. Furthermore, has rectangular relations because has rectangular relations.
Moreover, the running time of this reduction can be bounded by
for a computable function , and
as well, so the reduction is an FPT reduction.
3.2 Average Baby PIH on Dense and Sparse Instances
In this section we prove Theorem 2, which is divided into two separate lemmas to help with readability. For our purposes, a 2CSP instance is dense if ; or it is sparse, if .
Lemma 23.
Under , the Average Baby PIH holds for all Avg--Gap- instances that are dense.
Proof.
The reduction in the proof of Lemma 20 yields complete bipartite 2CSP instances , i.e., for each and , there exists a constraint . Then the reduction in the proof of Lemma 22 makes copies of and copies of , while keeping the constraints in each pair of copies. So in the final instance from the proof of Theorem 16, the number of constraints is
Now consider any function . We produce a new instance by simply copying for times, where is chosen as the minimum number satisfying
Note that there is no constraint between different copies. Then, the new parameter is , and
It’s clear that this reduction runs in FPT time. Also, if is satisfiable, then is satisfiable. If any satisfying multi-assignment to must have total size more than , then any satisfying multi-assignment to must assign each copy of more than values, so in total more than values, preserving the gap.
Lemma 24.
Let . If there exists a constant such that no FPT algorithm can solve Avg--Gap- on instance with , i.e., is sparse, then the PIH holds.
Proof Sketch.
Let be a NO instance of Avg--Gap-. For any (standard) assignment , assume that violates constraints, then one can simply add at most values to and obtain a satisfying multi-assignment with total size . Since is a NO instance, we have . Thus,
In other words, any assignment to must violate a constant fraction of the constraints in . This gives a reduction from Avg--Gap- to PIH.
References
- [1] Noga Alon, Dana Moshkovitz, and Shmuel Safra. Algorithmic construction of sets for k-restrictions. ACM Trans. Algorithms, 2(2):153–177, 2006. doi:10.1145/1150334.1150336.
- [2] Sanjeev Arora, Carsten Lund, Rajeev Motwani, Madhu Sudan, and Mario Szegedy. Proof verification and the hardness of approximation problems. J. ACM, 45(3):501–555, 1998. doi:10.1145/278298.278306.
- [3] Sanjeev Arora and Shmuel Safra. Probabilistic checking of proofs: A new characterization of NP. J. ACM, 45(1):70–122, 1998. doi:10.1145/273865.273901.
- [4] Libor Barto and Marcin Kozik. Combinatorial gap theorem and reductions between promise CSPs. In Joseph (Seffi) Naor and Niv Buchbinder, editors, Proceedings of the 2022 ACM-SIAM Symposium on Discrete Algorithms, SODA 2022, Virtual Conference / Alexandria, VA, USA, January 9 - 12, 2022, pages 1204–1220. SIAM, 2022. doi:10.1137/1.9781611977073.50.
- [5] Yijia Chen, Yi Feng, Bundit Laekhanukit, and Yanlin Liu. Simple combinatorial construction of the k-lower bound for approximating the parameterized k-Clique. In 2025 Symposium on Simplicity in Algorithms (SOSA), pages 263–280. Society for Industrial and Applied Mathematics, 2025. doi:10.1137/1.9781611978315.21.
- [6] Yijia Chen and Bingkai Lin. The constant inapproximability of the parameterized Dominating Set problem. SIAM J. Comput., 48(2):513–533, 2019. doi:10.1137/17M1127211.
- [7] Marek Cygan, Fedor V. Fomin, Lukasz Kowalik, Daniel Lokshtanov, Dániel Marx, Marcin Pilipczuk, Michal Pilipczuk, and Saket Saurabh. Parameterized Algorithms. Springer, 2015. doi:10.1007/978-3-319-21275-3.
- [8] Irit Dinur. The PCP theorem by gap amplification. J. ACM, 54(3):12, 2007. doi:10.1145/1236457.1236459.
- [9] Rodney G. Downey and Michael R. Fellows. Fundamentals of Parameterized Complexity. Texts in Computer Science. Springer, 2013. doi:10.1007/978-1-4471-5559-1.
- [10] Jörg Flum and Martin Grohe. Parameterized Complexity Theory. Texts in Theoretical Computer Science. An EATCS Series. Springer, 2006. doi:10.1007/3-540-29953-X.
- [11] Venkatesan Guruswami, Bingkai Lin, Xuandi Ren, Yican Sun, and Kewen Wu. Almost optimal time lower bound for approximating parameterized Clique, CSP, and more, under ETH. CoRR, abs/2404.08870, 2024. doi:10.48550/arXiv.2404.08870.
- [12] Venkatesan Guruswami, Bingkai Lin, Xuandi Ren, Yican Sun, and Kewen Wu. Parameterized Inapproximability Hypothesis under Exponential Time Hypothesis. In Bojan Mohar, Igor Shinkar, and Ryan O’Donnell, editors, Proceedings of the 56th Annual ACM Symposium on Theory of Computing, STOC 2024, Vancouver, BC, Canada, June 24-28, 2024, pages 24–35. ACM, 2024. doi:10.1145/3618260.3649771.
- [13] Venkatesan Guruswami, Xuandi Ren, and Sai Sandeep. Baby PIH: parameterized inapproximability of min CSP. In Rahul Santhanam, editor, 39th Computational Complexity Conference, CCC 2024, July 22-25, 2024, Ann Arbor, MI, USA, volume 300 of LIPIcs, pages 27:1–27:17. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024. doi:10.4230/LIPICS.CCC.2024.27.
- [14] Karthik C. S. and Subhash Khot. Almost polynomial factor inapproximability for parameterized k-Clique. In Shachar Lovett, editor, 37th Computational Complexity Conference, CCC 2022, July 20-23, 2022, Philadelphia, PA, USA, volume 234 of LIPIcs, pages 6:1–6:21. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2022. doi:10.4230/LIPICS.CCC.2022.6.
- [15] Karthik C. S., Bundit Laekhanukit, and Pasin Manurangsi. On the parameterized complexity of approximating Dominating Set. J. ACM, 66(5):33:1–33:38, 2019. doi:10.1145/3325116.
- [16] Karthik C. S. and Inbal Livni Navon. On hardness of approximation of parameterized Set Cover and Label Cover: Threshold graphs from error correcting codes. In Hung Viet Le and Valerie King, editors, 4th Symposium on Simplicity in Algorithms, SOSA 2021, Virtual Conference, January 11-12, 2021, pages 210–223. SIAM, 2021. doi:10.1137/1.9781611976496.24.
- [17] Shuangle Li, Bingkai Lin, and Yuwei Liu. Improved lower bounds for approximating parameterized nearest codeword and related problems under ETH. In Karl Bringmann, Martin Grohe, Gabriele Puppis, and Ola Svensson, editors, 51st International Colloquium on Automata, Languages, and Programming, ICALP 2024, July 8-12, 2024, Tallinn, Estonia, volume 297 of LIPIcs, pages 107:1–107:20. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024. doi:10.4230/LIPICS.ICALP.2024.107.
- [18] Bingkai Lin. A simple gap-producing reduction for the parameterized Set Cover problem. In Christel Baier, Ioannis Chatzigiannakis, Paola Flocchini, and Stefano Leonardi, editors, 46th International Colloquium on Automata, Languages, and Programming, ICALP 2019, July 9-12, 2019, Patras, Greece, volume 132 of LIPIcs, pages 81:1–81:15. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPICS.ICALP.2019.81.
- [19] Bingkai Lin. Constant approximating k-Clique is W[1]-hard. In Samir Khuller and Virginia Vassilevska Williams, editors, STOC ’21: 53rd Annual ACM SIGACT Symposium on Theory of Computing, Virtual Event, Italy, June 21-25, 2021, pages 1749–1756. ACM, 2021. doi:10.1145/3406325.3451016.
- [20] Bingkai Lin, Xuandi Ren, Yican Sun, and Xiuhan Wang. Constant approximating parameterized k-SetCover is W[2]-hard. In Nikhil Bansal and Viswanath Nagarajan, editors, Proceedings of the 2023 ACM-SIAM Symposium on Discrete Algorithms, SODA 2023, Florence, Italy, January 22-25, 2023, pages 3305–3316. SIAM, 2023. doi:10.1137/1.9781611977554.CH126.
- [21] Daniel Lokshtanov, M. S. Ramanujan, Saket Saurabh, and Meirav Zehavi. Parameterized complexity and approximability of Directed Odd Cycle Transversal. In Shuchi Chawla, editor, Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, SODA 2020, Salt Lake City, UT, USA, January 5-8, 2020, pages 2181–2200. SIAM, 2020. doi:10.1137/1.9781611975994.134.
- [22] Carsten Lund and Mihalis Yannakakis. On the hardness of approximating minimization problems. J. ACM, 41(5):960–981, 1994. doi:10.1145/185675.306789.
- [23] Pasin Manurangsi. Tight running time lower bounds for strong inapproximability of maximum k-Coverage, Unique Set Cover and related problems (via t-wise agreement testing theorem). In Shuchi Chawla, editor, Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, SODA 2020, Salt Lake City, UT, USA, January 5-8, 2020, pages 62–81. SIAM, 2020. doi:10.1137/1.9781611975994.5.
Appendix A From Average Baby PIH to Inapproximability of -ExactCover
We present a proof relying on a construction that slightly differs from the one in [13]. Their proof makes use of the -set gadget [22, 1] that was previously used to show the hardness of approximating SetCover problem. On the other hand, our proof develops a novel composition of the collision number of ECCs (recall Definition 17) with the following well-known combinatorial object.
Definition 25 (Hypercube Partition System).
Let be two sets. Then the -hypercube partition system is defined by
-
the universe , and
-
a collection of subsets where each .
Theorem 26 (cf. Theorem 21 in [13]).
Assume that the Average Baby PIH holds on all 2CSP instances with rectangular relations. Then -ExactCover cannot be approximated in FPT time within any constant factor. More precisely, for every constant no FPT algorithm, on a given -SetCover instance with size and , can distinguish between the following two cases:
-
We can choose disjoint sets in whose union is .
-
is not the union of any sets in .
Proof.
Let be an Avg--Gap- instance with rectangular relations. We set . Moreover, for each rectangular constraint we use to denote the underlying set and the associated mappings as in Definition 14. That is, for every , it holds that if and only if . Then we set
(12) |
Clearly, we can assume without loss of generality
Now we reduce to a -ExactCover instance. To that end, we choose a further alphabet whose size is a prime and satisfies
Moreover,let
This leads to the following code with very large distance (here we simply use Reed-Solomon code again)
Plugging
in Theorem 18 we conclude that the -collision number of Enc is
Observe that (12) implies that every tuple in can be identified with a string in , i.e., the domain of Enc.
Then, for each variable and every its possible value , we define a set as follows. For each constraint with associated set and mappings , and for each , we construct a -hypercube partition system
(13) |
Then for each we add to and similarly to . Finally, let the universe be
For the completeness, let be a satisfying assignment of , it is routine to check that is a partition of .
For the soundness, assume that every satisfying multi-assignment of has total size at least (cf. Definition 7). Let be a cover of . Consider the multi-assignment that maps every variable to . If this multi-assignment satisfies , the our assumption implies . Otherwise, assume that there exists some constraint which is not satisfied. Note that the above multi-assignment assigns to and to . Since is not satisfied, for all we have . However, for each , since is covered by , there must exist and with . Therefore, the set collides on all coordinates , hence it must have size at least . We deduce
Finally, in each hypercube partition system (13) it holds that
and there are at most such systems. The size of the universe is thus at most for some appropriate computable function , while the parameter of the -ExactCover instance remains . It follows easily that the running time of this reduction is FPT.
Combining Theorem 26 and Theorem 16, we obtain:
Theorem 27.
For any constant , -approximating -ExactCover is -hard.