Redundancy Rules for MaxSAT
Abstract
The concept of redundancy in SAT leads to more expressive and powerful proof search techniques, e.g., able to express various inprocessing techniques, and originates interesting hierarchies of proof systems [Heule et.al’20, Buss-Thapen’19]. Redundancy has also been integrated in MaxSAT [Ihalainen et.al’22, Berg et.al’23, Bonacina et.al’24].
In this paper, we define a structured hierarchy of redundancy proof systems for MaxSAT, with the goal of studying its proof complexity. We obtain MaxSAT variants of proof systems such as , , , and others, previously defined for SAT.
All our rules are polynomially checkable, unlike [Ihalainen et.al’22]. Moreover, they are simpler and weaker than [Berg et.al’23], and possibly amenable to lower bounds. This work also complements the approach of [Bonacina et.al’24]. Their proof systems use different rule sets for soft and hard clauses, while here we propose a system using only hard clauses and blocking variables. This is easier to integrate with current solvers and proof checkers.
We discuss the strength of the systems introduced, we show some limitations of them, and we give a short proof that any assignment for the weak pigeonhole principle falsifies at least clauses.
Keywords and phrases:
MaxSAT, Redundancy Rules, Pigeonhole PrincipleFunding:
Ilario Bonacina: The author was supported by grant PID2022-138506NB-C22 (PROOFS BEYOND) funded by AEI.Copyright and License:
2012 ACM Subject Classification:
Theory of computation Proof complexityAcknowledgements:
The authors would like to thank the Simons Institute for the Theory of Computing: part of this work has been done during the “Extended Reunion: Satisfiability” program (Spring 2023). Another part of this work has been done during the Oberwolfach workshop 2413 “Proof Complexity and Beyond” and during the 2023 Workshop on Proof Theory and its Applications organized by the Proof Society.Event:
28th International Conference on Theory and Applications of Satisfiability Testing (SAT 2025)Editors:
Jeremias Berg and Jakob NordströmSeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
This paper investigates new proof systems for MaxSAT that incorporate redundancy inferences tailored to work for MaxSAT. Redundancy inferences were introduced as extensions to SAT solvers to allow non-implicational inferences that preserve satisfiability and non-satisfiability. For resolution and SAT solvers, the first redundancy inferences were based on blocked clauses (BC) [25] and Resolution Asymmetric Tautology (RAT) [22, 15]. Other work on redundancy reasoning includes [14, 16, 20, 23, 17]; and, of particular relevance to the present paper, are the work of Heule, Kiesl, and Biere [19], and the work of Buss and Thapen [9]. Redundancy inferences formalize “without loss of generality” reasoning [30] and can substantially strengthen resolution and, in some cases, the effectiveness of SAT solvers for hard problems such as the pigeonhole principle (PHP) [20]. Indeed, in their full generality, redundancy inferences allow resolution to polynomially simulate extended resolution.
MaxSAT is a generalization of SAT; it is the problem of determining a truth assignment for a CNF formula that minimizes the number of falsified clauses. Although the MaxSAT problem is inherently more difficult than SAT, in some cases MaxSAT can be adapted to be more efficient in practice than CDCL solvers for hard problems such as PHP [7]. There are several approaches to MaxSAT solvers, including MaxSAT resolution [8, 26], core-guided MaxSAT [12, 1, 28, 27, 29], and maximum-hitting-set MaxSAT [2, 11, 31] ; the present paper discusses only MaxSAT resolution. The MaxSAT resolution proof system was first defined by Larrosa and Heras [26] and proved completed by Bonet, Levy and Manya [8].
We define cost preserving redundancy rules for MaxSAT mirroring the redundancy rules for SAT, called “”, “”, “”, “”, and “” (see Definition 3.2). The strongest of these is “” based on the substitution redundancy () [9]. All five of these new inferences are sound for MaxSAT reasoning (Theorem 4.2). Furthermore, we prove that , and are complete for MaxSAT (Theorem 4.3). On the other hand, we prove that and are incomplete for MaxSAT (Corollary 5.2). We illustrate the power of by giving polynomial size proofs of the cost of the blocking-variable version of weak pigeonhole principle for arbitrary numbers of pigeons and holes (Theorem 6.3).
Ours is not the first paper bringing redundancy reasoning to the context of optimization. For instance, the work of Ihalainen, Berg and Järvisalo [21], building on [4], introduced versions of redundancy inferences that work with MaxSAT. In contrast to the system CPR of [21], all of our “cost-” inferences are polynomial-checkable for validity, and thus all give traditional Cook-Reckhow proof systems for MaxSAT.
Another system with redundancy rules for certifying unsatisfiability and optimization is veriPB [5]. Being rooted in cutting planes, veriPB is particularly apt at certifying optimality, and it can log the reasoning of MaxSAT solver strategies that are way out of reach of MaxSAT resolution [3]. The propositional fragment of is as strong as Extended Resolution and DRAT, and plausibly even stronger [24]. In contrast, our systems are plausibly weaker and simpler, but yet strong enough to prove efficiently interesting formulas. Still our systems might be amenable to proving lower bounds. Indeed, our explicit goal is to study the proof complexity of redundancy rules for MaxSAT in a similar vein of what [9] does for SAT, something that is beyond the scope of , focused on proof logging actual solvers. As a starting point, we show that “cost-BC” and “cost-LPR” are not complete (Corollary 5.2) and we show width lower bounds on “cost-SPR” and an analogue of a width lower bound for “cost-SR” (Corollary 5.3). Finally there is another work that tries to modify the redundancy rules to make them compatible with the MaxSAT model of soft and hard clauses [6]. That work has a similar spirit to this one, but departs in the redundancy test, which checks for inclusion rather than for reverse unit propagation. Such choice makes the inference weaker but preserves the number of unsatisfied clauses, hence the rule applies directly to soft clauses.
Before proceeding with the description and analysis of our systems, we should highlight two aspects of redundancy inference that are somehow orthogonal to the choice of the concrete inference rule: clause deletion and new variable introduction. The applicability of a redundancy rule depends on the clauses present in the database, and it seems that allowing deletion of past clauses makes the system stronger, and indeed collapses together the power of several types of inference rules (see [9]). Likewise the possibility of introducing new variables in redundant clauses makes all such systems as powerful as extended resolution [25]. Coherently with the stated goal of having systems that are simple and amenable to proof complexity analysis, in this paper we do not allow neither clause deletion nor new variables.
Structure of the paper
Section 2 contains all the necessary preliminaries, including notation on MaxSAT and the blocking variables encoding of MaxSAT instances (blocking variables are also used by [21]). Section 3 introduces the redundancy rules for MaxSAT, proves their basic properties, and defines calculi based on those rules. Section 4 shows their soundness, and their completeness. Section 5 shows the incompleteness of / and some limitations of and even . Section 6 gives examples of applications of the redundancy rules, including a polynomial size proof of the optimal cost of the weak Pigeonhole Principle and a general result about the polynomial size provability of minimally unsatisfiable formulas. Section 7 gives some concluding remarks.
2 Preliminaries
For a natural number , let be the set . Sets and multi-sets are denoted with capital Roman or Greek letters.
Propositional logic notation
A Boolean variable takes values in . A literal is either a variable or its negation . A clause is a finite disjunction of literals, i.e., . The empty clause is . A formula in Conjunctive Normal Form (CNF) is a conjunction of clauses . We identify a CNF with the multiset of its clauses, and denote as the number of its clauses (counted with multiplicity). We denote as the set of variables in .
Substitutions and assignments
A substitution for a set of variables is a function so that is either , or some literal defined on . For convenience, we extend a substitution to constants and literals, setting , , and for any variable . The composition of two substitutions is the substitution , where for . A substitution is an assignment when for any . The domain of an assignment is , and is a total assignment over if is its domain, i.e., maps all variables in to Boolean values. Given a clause and a substitution , the clause restricted by , is simplified using the usual logic rules, i.e., , , and . Another notation for is . If or is tautological we say that , i.e., satisfies .
The restriction of a CNF formula by , denoted as , is the conjunction of all clauses where and . The CNF is also a multiset. We say that satisfies () if for every , , i.e., . We say that if for every substitution , if then .
We identify a literal with the substitution that assigns to and leaves all other variables unassigned. Hence we use notations like . Likewise, given a clauses we denote as the assignment that maps all literals in to false, and we use the notation .
Unit propagation
A unit clause is a clause of just one literal. Unit propagation works as follows. Start with a CNF : if has no unit clauses, the process ends, otherwise pick some unit clause in arbitrarily, remove from all clauses containing and remove literal from all clauses containing it. Keep repeating until has no more unit clauses. Regardless of the choice of the unit clause, the process always ends with the same formula.
We say that when the application of unit propagation to the formula produces the empty clause. For two CNF formulas we say that if for every , . Clearly, if then , and if , then . It is important to stress that the relation is efficiently checkable.
Observation 2.1 ([9, Fact 1.3]).
Let be a substitution and be CNF formulas, if , then .
Resolution
Resolution is a well-studied propositional deduction system with two inference rules: (i) from a clause we can deduce any s.t. ; (ii) from clauses and we can deduce . A resolution proof from a set of clauses is a sequence of clauses where each is either already in or is deduced from earlier clauses in the sequence using one of the two inference rules. Resolution is complete, thus deciding whether a clause can be deduced from is the same as deciding whether .
MaxSAT
Given a CNF formula , MaxSAT asks to find the maximum number of clauses in which can be simultaneously satisfied. In applications, it is useful to consider a generalization for which we divide the clauses into hard or soft (partial MaxSAT). Hard clauses must be satisfied, while soft clauses can be falsified with a cost. Consider where is the multiset of hard clauses and is the multiset of soft ones. In this model, MaxSAT asks to find the maximum number of clauses in that can be simultaneously satisfied by an assignment that satisfies all clauses in . Observe that the optimization problem is not well defined if is not satisfiable.111An even more general version is weighted MaxSAT, where we would consider weighted set of clauses , i.e., each clauses has an associated weight where . In this model the goal is to minimize the weight of the falsified clauses. The role of the weight is to model hard clauses. In this paper we do not focus on this model. It is not relevant whether is a set or a multiset. In , on the other hand, the multiplicity of soft clauses must be accounted for.
Proof systems for MaxSAT aim to show lower bounds on the cost of (partial) MaxSAT instances, one such system is MaxSAT resolution.
MaxSAT with blocking variables
Without loss of generality we can assume that all soft clauses in a MaxSAT instance are unit clauses; indeed, using a new variable , a soft clause can be replaced with a hard clause and a soft clause , without affecting the cost. The variable is usually called a blocking variable. This appears in [13], but it might have been used even earlier.
Definition 2.2.
Let with soft clauses . The blocking variables formulation of is where
-
,
-
,
and are new variables (blocking variables) not appearing in . We say that is a MaxSAT instance encoded with blocking variables, when it is given as a set of hard clauses of the form as in above. The soft clauses, then, are implicit.
Observation 2.3.
Let be a MaxSAT instance and be the blocking variables formulation of . Any assignment that satisfies and falsifies clauses in can be extended to an assignment that satisfies and sets blocking variables to true. Vice versa, any assignment that satisfies and sets blocking variables to true satisfies too and falsifies at most clauses in .
Because of ˜2.3, for the rest of this work we consider to be a MaxSAT instance encoded with blocking variables usually named . The goal is to satisfy while setting to true the least number of blocking variables. More formally, given a total assignment for , we define
and the goal is to find the value of . Notice that, the notation is defined even for assignments not satisfying .
3 Redundancy rules for MaxSAT
In the context of SAT, a clause is redundant w.r.t. a CNF instance if and are equisatisfiable, that is either they both are satisfiable or both unsatisfiable [25]. The natural adaptation of this notion to MaxSAT is a clause that does not affect the cost of .
Definition 3.1 (redundant clause, [21]).
A clause is redundant w.r.t. a MaxSAT instance when
| (1) |
Clauses that logically follow from are obviously redundant, but there may be other useful clauses that do not follow logically, and yet do not increase the cost if added.
The condition in eq. (1) is not polynomially checkable (unless, say ). Therefore, we consider efficiently certifiable notions of redundancy, i.e., ways to add redundant clauses (in the sense of eq. (1)) while certifying efficiently their redundancy. This is done showing how to extend in a systematic way the notions of efficiently certifiable redundancy already studied in the context of SAT (BC, RAT, LPR, SPR, PR, SR) [19, 9] to the context of MaxSAT. This is an alternative to the approach of [21]. This definition could also be seen as a very special case of veriPB [5] (see Section 3.1 for more details on the connections with veriPB).
Definition 3.2.
A clause is cost substitution redundant () w.r.t. to if there exists a substitution such that
-
1.
(redundancy)
-
2.
for all total assignments , (cost).
If the substitution has some additional structure, we have the following redundancy rules listed in decreasing order of generality:
- Cost propagation redundant ()
-
if is a partial assigment.
- Cost subset propagation redundant ()
-
if is a partial assigment with the same domain as . In other words, flips some variables in .
- Cost literal propagation redundant ()
-
if is a partial assignment with the same domain as , but differs from on exactly one variable.
- Cost blocked clause ()
-
if is a partial assignment with the same domain as , which differs from on exactly one variable , and moreover, for every clause containing the variable , .222The definition of blocked clause is written to match the previous definitions. In this case, the redundancy condition is always satisfied.
Item 1 in Definition 3.2 claims that adding does not make unsatisfiable, unless it was already the case. Together with Item 2, it ensures that any assignment that falsifies the new clause can be patched with a substitution so that is satisfied without increasing the minimum cost (see Lemma 3.4).
Indeed, Item 1 in Definition 3.2 and its variants correspond to the redundancy rules of proof systems , , , and from [9], adapted here to consider cost. Since is the same as (see [9, Theorem 1.10]), the notion of cost literal propagation redundancy could as well be called cost- redundancy.
Remark 3.3.
It is important to compare Definition 3.2 with [21, Definition 2]. Redundancy conditions are very similar and the main differences are in the cost conditions. Let us compare their CPR rule with our . In , the cost condition requires the witness to be at least as good as all possible extensions of , while in CPR the requirement is enforced only on those extensions of that satisfy . The latter more permissive condition allows to derive more clauses, but it is unlikely to be polynomially checkable, while the condition in is polynomially checkable (see Lemma 3.5). In [21], the authors also define two polynomially checkable rules where the cost condition is not present, but implicitly enforced via restrictions on the type of assignments used. Those rules are special cases of and respectively.
Lemma 3.4.
If is w.r.t. to , then is redundant w.r.t. .
Proof.
It is enough to show that . Let . To show that adding to does not increase the cost, consider an optimal total assignment that satisfies and sets to true exactly blocking variables. If we already have that and . Otherwise, extends and, by assumption, there is a substitution such that . To show that , it remains to show that indeed . By assumption,
and, since and extends , then too. Equivalently, .
Both Item 1 and Item 2 of Definition 3.2 are stronger than what is actually needed for Lemma 3.4 to hold. Indeed, for Item 1, it would be enough that , and, for Item 2, it would be sufficient to check it for any such that . Unfortunately, these latter versions of Item 1 and Item 2 are in general not polynomially checkable. Instead, our conditions are checkable in polynomial time.
Lemma 3.5.
Let be a MaxSAT instance, a clause and a substitution. There is a polynomial time algorithm to decide whether is w.r.t. , given the substitution .
Proof (sketch).
The redundancy condition in Definition 3.2 is polynomially checkable, since it is a unit propagation. To check the cost condition we need to check whether
| (2) |
is at most . This amounts to maximizing the function
| (3) |
over all assignments on variables in . Observe that eq. (3) is a function of the form , for suitable constants and s. Therefore its maximum can be easily determined in polynomial time and corresponds to assigning each to when and to otherwise.
Lemma 3.5 allows the definition of proof systems that extend resolution using cost redundancy rules in the sense of Cook-Reckhow [10].
Definition 3.6 ( calculus).
The calculus is a proof system for MaxSAT. A derivation of a clause from a MaxSAT instance (encoded with blocking variables) is a sequence of clauses where, , each is either already in or is deduced from earlier clauses in the sequence using a resolution rule, or is w.r.t. to with .333We only consider the case where no new variables are added via rules. To be coherent with [9], should be , following the notational convention of adding an exponent with “” to denote , and when the systems are not allowed to introduce new variables. We ignore that convention to ease an already rich notation. The length of such derivation is , i.e., the number of derived clauses. To check the validity of derivations in polynomial time, any application of the rule comes accompanied by the corresponding substitution that witnesses its soundness.
If the goal is to certify that , we can accomplish this deriving distinct unit clauses of the form (see Theorem 4.2). If the goal is to certify that , we can accomplish this deriving distinct unit clauses of the form together with the unit clauses (see Theorem 4.2).
In a similar fashion we define , , , and calculi. A remarkable aspect of these calculi is that a proof must somehow identify the blocking variables to be set to true. When there are multiple optimal solutions, it is quite possible that none of the follows logically from . Nevertheless, the redundancy rules, often used to model “without loss of generality” reasoning, can reduce the solution space.
3.1 Simulation by veriPB
We show that and its subsystems are p-simulated by . This is actually not surprising since proof steps are based on cutting planes, which in turn easily simulates resolution. Furthermore includes the SAT redundancy rules equipped with some criteria of cost preservation. can argue within the proof whether some substitution is cost preserving. If can do that for the substitutions that respect the second condition in Definition 3.2, then it p-simulates .
Since is a complex system, we describe the minimal fragment needed to simulate . Consider a CNF formula and some objective linear function to be minimized. A derivation starts with encoded as linear inequalities. At each step a linear inequality is derived from the set of previously derived inequalities in one of the following ways:
-
1.
is derived from using some cutting planes rule;
-
2.
there exists a substitution so that ,where the symbol means that there is a cutting planes proof that witnesses the implication.
In Item 2 the cutting plane proof is included in the proof to make it polynomially verifiable, unless it just consists in unit propagation steps or certain uses of Boolean axioms. In the latter cases the verifier can efficiently recover such derivation steps on its own.
The resolution inference rule of can be simulated using 1. The redundacy inference rule of is simulated by 2. The redundancy condition is syntactically the same, while the cost condition is argued via a cutting planes proof. The latter is the only non-trivial part. This argument gives a proof sketch of the following proposition.
Proposition 3.7.
Let be a MaxSAT instance over variables with blocking variables . If there is a step derivation of a set of clauses from it, then there is a proof from and objective of some in steps.
4 Soundness and completeness
The calculi // are sound and complete. Before proving the soundness we show an auxiliary lemma, that shows that when the calculus certifies the lower bound for the optimal values, it can also certify the optimality.
Lemma 4.1.
Let be a MaxSAT instance encoded with blocking variables , of , and suppose contains the unit clauses . Then can prove for each in steps.
Proof.
Let be a total assignment satisfying with , that is maps all the s to and the other blocking variables to . We derive all the clauses
with using the rule. For all clauses , the substitution witnessing the validity of the rule is always . The redundancy condition from Definition 3.2 is trivially true since union an arbitrary set of s is mapped to under . The cost condition is true because for every , and .
To conclude, by resolution, derive from and the unit clauses .
Theorem 4.2 (soundness of ).
Let be a MaxSAT instance encoded with blocking variables. If there is a proof of distinct blocking variables, then . If there is a proof of distinct blocking variables and all the unit clauses for , then .
Proof.
Let be the blocking variables of . Let the set of clauses in plus all the clauses derived in the proof of . That is also contains contains distinct unit clauses , hence . By Lemma 3.4, the cost is preserved along proof steps, therefore . In the case where we have all the s then and therefore .
As an immediate consequence of Theorem 4.2, also all the , , calculi are sound. Moreover, we can always prove the optimal lower bound in calculus.
Theorem 4.3 (completeness of ).
Let be a MaxSAT instance encoded with blocking variables, of . There is a derivation of the unit clauses for some distinct blocking literals and all the for .
Proof.
Let be the blocking variables of . Take to be an optimal assignment, that is , , and for every total assignment that satisfies , . Without loss of generality we can assume sets variables to and the remaining s to .
Given any assignment , let be the largest clause falsified by . Let be the set of all clauses where is a total assignment that satisfies is different from . We want to derive from , essentially forbidding any satisfying assignment except for .
We can add all clauses in one by one by the rule. Indeed, for any clause and any , the clause is w.r.t. , with as the witnessing assignment. The redundancy condition
holds because so the RHS is just true. The cost condition holds by optimality of . In the end, the only assignment that satisfies is . By the completeness resolution we can prove all its literals, in particular literals for and literals for .
5 Incompleteness and Width Lower bounds
Theorem 4.3 shows the completeness of , hence for and . It is not a coincidence that the proof does not apply to , indeed the latter is not complete. We will see that for some redundant clause derived according to Definition 3.2, the number of values that partial assignment flips with respect to any may have to be large. In a proof this number is always at most one. For a redundant clause derived in some subsystem of via a witness substitution we define
where is the number of different bits between two total assignments. The following result shows sufficient conditions for to be large.
Theorem 5.1.
Let be a MaxSAT instance encoded with blocking variables, of , and let be the set of optimal total assignments for , i.e., when and . If is such that
-
1.
all pairs of assignments in have Hamming distance at least , and
-
2.
for every blocking variable there are s.t. and ,
then to derive any blocking literal , must derive a redundant clause with , where is the witnessing substitution for .
Proof.
Consider a derivation from as a sequence where each with either derived by resolution from clauses in , or is w.r.t. . For , let be the number of the optimal assignments for .
At the beginning by construction. If at some point contains some clause , then the value must be strictly smaller than because contains at least some assignment with (by assumption 2).
Let be the first step where drops below . The clause introduced at that moment must be w.r.t. , because the resolution steps do not change the set of optimal assignments. Let then be the witnessing substitution used to derive , we have
| (4) |
Since dropped below , clause must be incompatible with some , that is . Therefore, by cost preservation,
| (5) |
Since was the first moment when we have that and therefore . In particular, . By eq. (5) then it must be . But then, by assumption 1, and have Hamming distance at least , which implies .
We see an example application of this result to , where the number of values allowed to be flipped by the rule is at most .
The formula has cost and its optimal assignments to variables are . These assignments fulfil the premises of Theorem 5.1, with Hamming distance . Therefore cannot prove the cost of to be , and hence is incomplete.
Corollary 5.2.
Proof systems , are incomplete.
Corollary 5.3.
There is a formula family with variables, clauses and where, in order to prove , any proof derives a redundant clause with , where is the witnessing substitution for .
Proof.
We define on variables , and variables . The formula contains hard clauses to encode the constraint , and constraints , for . Furthermore has the soft clauses, encoded as hard clauses with blocking variables, and for .
To satisfy the formula an assignment must either set all ’s to true and ’s to false, or vice versa. Both such assignments set to true of the blocking variables, and no blocking variable is fixed to a constant value. Therefore the claim follows from Theorem 5.1.
Corollary 5.4.
Any proof for a formula respecting the hypothesis of Theorem 5.1 requires some redundant clauses of width at least .
Proof.
Any redundant clause derived in must have , and by Theorem 5.1, . For instance, the formula from Corollary 5.3 requires refutations of width .
6 Short proofs using redundancy rules
We show applications demonstarting the power of the redundancy rules on notable families of CNF formulas. In Section 6.1 we consider minimally unsatisfiable formulas, while in Section 6.2 we consider the weak Pigeonhole Principle.
Remark 6.1.
Due to Theorem 4.2 and Theorem 4.3, we refer to a (resp. , ) derivation from of for some distinct blocking literals and all the for , as a proof of in (resp. , ).
6.1 Short proofs of minimally unsatisfiable formulas
Recall the definition of from [18] (see also [9, Definition 1.16]). A calculus refutation of a CNF formula is a sequence of clauses where , and each is either a clause in , or derived by resolution, or is w.r.t. , that is satisfies
that is, Item 1 of Definition 3.2 for a which is a partial assignment. A refutation is a derivation of . The size of a refutation is the number of clauses in it.
An unsatisfiable set of clauses is minimally unsatisfiable if no proper subset of is unsatisfiable.
Theorem 6.2.
If a minimally unsatisfiable CNF formula has a refutation of size , then there is a proof of of at most many clauses.
Proof.
Let and be the corresponding MaxSAT instance. Let be a refutation of , so . First we show that
with , is a valid derivation of from . In particular, assuming we already derived the first steps of , we show how to derive .
When , the clause is the weakening of some clause in . If was derived using a resolution rule on some premises in , then can be derived in the same way from the corresponding premises in . The remaining case is when is w.r.t. . Let be the assignment that witnesses it. This assignment only maps variables from the original formula , so we extend it to , and then use to witness that indeed is w.r.t. . For the cost condition in Definition 3.2, just observe that any extension of has cost . For the redundancy condition, observe that, by construction, , , and .
The last clause of is . Let be an optimal assignment of . Since is minimally unsatisfiable, . W.l.o.g. assume sets and all for .
Now, for each , the clause is w.r.t. , using itself as the witnessing assignment: redundancy holds since satisfies every clause in and all clauses . The cost condition follow since for any and .
In the end we use steps to derive from and , and to derive in calculus all the units via Lemma 4.1.
Theorem 6.2 shows that the propositional refutations for the minimally unsatisfiable formulas in [9] translate immediately to certificates in the MaxSAT. In particular, as a corollary of Theorem 6.2, we have that proves in polynomial size that
-
the Bit-Pigeonhole Principle [9, Theorem 4.4],
-
the Parity Principle [9, Theorem 4.6],
-
the Tseitin Principle on a connected graph [9, Theorem 4.10],
have all cost 1, since they are all minimally unsatisfiable. In MaxSAT resolution that would require exponentially long derivations.
6.2 Short proofs of the minimum cost of
Let . The pigeonhole principle from pigeons to holes, with blocking variables, has the following formulation, that we call : the totality clauses for , and the injectivity clauses for and . We use as an alias of the variable , given that .
Theorem 6.3.
proves in polynomial size.
This is the main result of the section. Before proving it we show two useful lemmas. The first lemma is used to “clean up” the set of clauses during a derivation. For each new step in a calculus derivation the redundancy condition must be checked against an ever increasing set of clauses. It turns out that some already derived clauses can be completely ignored for the rest of the derivation under some technical conditions. This makes up for the lack of a deletion rule, that we do not have, and in the context of SAT seems to give more power to the systems [9].
Lemma 6.4.
Let and be two sets of clauses. Any derivation from is also a valid derivation from if either of the two cases applies
-
1.
Variables in do not occur in .
-
2.
For every clause there is a clause so that .
Proof.
The cost condition does not depend on the set of clauses, therefore we only need to check the validity of the redundancy condition. In the first case, the redundancy condition applies because the clauses of are unaffected by the substitutions involved.
For the second case, consider the derivation of a clause witnessed by . The clauses in and are subsumed by clauses in and , respectively. Hence
which implies the validity of the redundancy condition.
The second lemma is used as a general condition to enforce clauses to be .
Lemma 6.5.
Let be a clause and a set of clauses. If there exists a permutation such that
-
1.
maps the set of blocking variables to itself,
-
2.
the substitution satisfies , and ,
then is w.r.t. . Notice that the second condition in item (2) is automatically satisfied if is a symmetry of , i.e. .
Proof.
Proof of Theorem 6.3.
The proof is by induction. The goal is to reduce the formula to pigeons and holes. First we do some preprocessing: from we derive a slightly more structured formula . Then we show how to derive in a polynomial number of steps. The results follows because after such derivations we obtain the formula that contains the clauses . Moreover, we also derive along the way.
We derive from using the rules of calculus. We divide the argument into several steps, but first we show how to derive from .
Preprocessing 1.
“Make full-fledged extension variables”. 444This is true in general: if a MaxSAT instance contains a clause then it is possible to make the blocking variable a full-fledged extension variable () by . Turn all the variables into full-fledged extension variables that satisfy , by adding the clauses
one by one in .
We need to derive clause for every . Assume that we already got , we derive as a clause w.r.t. . The witnessing assignment is . Since , the cost condition is satisfied. The redundancy condition follows from
Indeed, on clauses of that do not contain the variable , the assignments and behave identically, while all the clauses containing are satisfied by . Repeat the previous argument to get all the clauses . The current database of clauses is .
Preprocessing 2.
“Enforce injectivity”. Optimal assignments for can have unassigned pigeons or have collisions between pigeons. It is more convenient to avoid collisions and just focus on assignments that are partial matching. A moment’s thought suffices to realize that such restriction does not change the optimal cost but simplifies the solution space. We enforce collisions to never occur by deriving all the unit clauses by .
These clauses can be derived in any particular order: to show that is w.r.t. and the previously derived we pick one of the two pigeons involved (say ) and use as the witnessing assignment. The cost is not increased, and to check the redundancy condition observe that satisfies all the clauses that touches, so on the right side of the redundancy condition has a subset of with no occurrences , while the left side has the same set of clauses, but restricted with .
Now that we have all clauses we resolve them with the corresponding clauses to get the set of clauses
for all holes and pair of pigeons and .
We do not need variables anymore. By one application of Lemma 6.4, from now on we can ignore all clauses . By another application, we can also ignore the clauses . We will do induction on the current database of clauses.
For clarity we list all its clauses again.
Formula for (totality 1), for and (totality 2), for and (injectivity).
The core idea of the induction is that if a pigeon flies to a hole, we can assume without loss of generality that it is pigeon that flies into hole .
Step 0.
“If some pigeon flies, we can assume it is pigeon who flies”. We want to derive, in this order, the set of clauses
from , to claim that if some pigeon is mapped, then pigeon is mapped too. For each we apply Lemma 6.5 using as the witnessing permutation , the permutation that swaps pigeons and .
Namely, , , , , and is the identity on all other variables, therefore satisfies the first requirement for the lemma. Likewise , and we need to check that
By symmetry , and for , , hence the inclusion is true. The current database of clauses is .
Step 1.
“If pigeon flies to some hole, we can assume it flies to hole ”. Using inferences, we derive from , in this order, the clauses
expressing that if pigeon flies to some hole, this hole is the last one.
For each we apply Lemma 6.5 with the witnessing permutation swapping holes and .
Namely and , and is the identity on all other variables. By construction satisfies the first requirement for the lemma, and likewise , and, again, we need to check
By symmetry , and for , , hence the inclusion is true. The current database of clauses is .
Step 2.
“Obtain for every via resolution”. Resolve the clause (totality 1) with , the resulting clause with all clauses from step 2, to get . Then resolve again with the injectivity clause , then the result with clause (from step 1), and again this latter result with clause (totality 2). The final result is .
The clauses subsume the clauses in of the form and all the intermediate clauses from the previous resolution steps. Therefore we use Lemma 6.4 to be able to ignore the subsumed clauses.
The current database of clauses is is equal to
Step 3.
“Assign pigeon to hole : derive unit clauses and ”. The goal is to enforce pigeon to be mapped to hole , by deriving the clause using the rule. Then we get immediately by resolving with (totality 2).
The unit clause is w.r.t. , using partial assignment as witness.
Clearly satisfies the cost condition. To see that the redundancy condition holds as well, we need to show that for all in that contain , but the only such clause that remains in is , which is satisfied by . The current database of clauses is .
Step 4.
“Derive by ”. We can derive them in any order using as witnessing substitution of the rule the assignment setting , , and .
The cost condition is immediate, and the redundancy condition follows from the fact that .
Step 5.
“Reduction to pigeons and holes”. First we derive by unit propagation all the the totality clauses of . That is, we remove the hole from the totality axioms of the pigeons in the current database. Now, the current database is , the unit clauses , , for and for , and clauses that are subsumed by one of these unit clauses. Therefore by Lemma 6.4 we can ignore all the unit clauses and all the clauses subsumed by them. That is we can carry on the derivation using only .
Thus steps (1)–(6) are repeated times, up to derive .
The unit clauses derived in the whole process include
-
(totality clauses in ).
-
(derived at each step of the induction),
-
for all and (derived at the preprocessing).
Therefore .
7 Conclusions and open problems
We proposed a way to extend redundancy rules, originally introduced for SAT, into polynomially verifiable rules for MaxSAT. We defined sound and complete calculi based on those rules and we showed the strength of some of the calculi giving short derivations of notable principles and we showed the incompleteness of the weaker ones and width lower bounds for the stronger ones. We conclude this article with a list of open problems:
- 1.
-
2.
Does simulate MaxSAT Resolution? That is, if we have a MaxSAT instance with blocking variables and MaxSAT Resolution proves in size that , is there a proof of in of size ?
-
3.
We proved a width lower bound for and an analogue of a width lower bound for on formulas with optimal assignments far from each other in the Hamming distance. We reiterate the open problem of proving size lower bounds for and stronger systems.
References
- [1] Carlos Ansótegui, Maria Luisa Bonet, and Jordi Levy. SAT-based MaxSAT algorithms. Artif. Intell., 196:77–105, 2013. doi:10.1016/J.ARTINT.2013.01.002.
- [2] Fahiem Bacchus, Matti Järvisalo, and Ruben Martins. Maximum satisfiabiliy. In Armin Biere, Marijn Heule, Hans van Maaren, and Toby Walsh, editors, Handbook of Satisfiability - Second Edition, volume 336 of Frontiers in Artificial Intelligence and Applications, pages 929–991. IOS Press, 2021. doi:10.3233/FAIA201008.
- [3] Jeremias Berg, Bart Bogaerts, Jakob Nordström, Andy Oertel, and Dieter Vandesande. Certified core-guided MaxSAT solving. In Proceedings of the 29th International Conference on Automated Deduction (CADE), volume 14132, pages 1–22, 2023. doi:10.1007/978-3-031-38499-8_1.
- [4] Jeremias Berg and Matti Järvisalo. Unifying reasoning and core-guided search for maximum satisfiability. In Proceedings of the 16th European Conference on Logics in Artificial Intelligence (JELIA), volume 11468, pages 287–303, 2019. doi:10.1007/978-3-030-19570-0_19.
- [5] Bart Bogaerts, Stephan Gocht, Ciaran McCreesh, and Jakob Nordström. Certified dominance and symmetry breaking for combinatorial optimisation. Journal of Artificial Intelligence Research, 77:1539–1589, 2023. doi:10.1613/JAIR.1.14296.
- [6] Ilario Bonacina, Maria Luisa Bonet, and Massimo Lauria. MaxSAT Resolution with Inclusion Redundancy. In Proceedings of the 27th International Conference on Theory and Applications of Satisfiability Testing (SAT), volume 305, pages 7:1–7:15, 2024. doi:10.4230/LIPIcs.SAT.2024.7.
- [7] Maria Luisa Bonet, Sam Buss, Alexey Ignatiev, João Marques-Silva, and António Morgado. MaxSAT resolution with the dual rail encoding. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, (AAAI), the 30th innovative Applications of Artificial Intelligence (IAAI), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI), pages 6565–6572, 2018. doi:10.1609/AAAI.V32I1.12204.
- [8] Maria Luisa Bonet, Jordi Levy, and Felip Manyà. Resolution for Max-SAT. Artificial Intelligence, 171(8-9):606–618, 2007. doi:10.1016/J.ARTINT.2007.03.001.
- [9] Sam Buss and Neil Thapen. DRAT and propagation redundancy proofs without new variables. Logical Methods in Computer Science, 17(2), 2021. URL: https://lmcs.episciences.org/7400.
- [10] Stephen A. Cook and Robert A. Reckhow. The relative efficiency of propositional proof systems. Journal of Symbolic Logic, 44(1):36–50, 1979. doi:10.2307/2273702.
- [11] Jessica Davies and Fahiem Bacchus. Solving MAXSAT by solving a sequence of simpler SAT instances. In Proceedings of the 17th International Conference on Principles and Practice of Constraint Programming (CP), pages 225–239, 2011. doi:10.1007/978-3-642-23786-7_19.
- [12] Zhaohui Fu and Sharad Malik. On solving the partial MAX-SAT problem. In Proceedings of the 9th International Conference on Theory and Applications of Satisfiability Testing (SAT), pages 252–265, 2006. doi:10.1007/11814948_25.
- [13] Michel X. Goemans and David P. Williamson. New -approximation algorithms for the maximum satisfiability problem. SIAM Journal on Discrete Mathematics, 7(4):656–666, 1994. doi:10.1137/S0895480192243516.
- [14] Marijn Heule, Warren A. Hunt Jr., and Nathan Wetzler. Trimming while checking clausal proofs. In Proceedings of the conference on Formal Methods in Computer-Aided Design (FMCAD), pages 181–188, 2013. URL: https://ieeexplore.ieee.org/document/6679408/.
- [15] Marijn Heule, Warren A. Hunt Jr., and Nathan Wetzler. Verifying refutations with extended resolution. In Proceedings of the 24th International Conference on Automated Deduction (CADE), volume 7898, pages 345–359, 2013. doi:10.1007/978-3-642-38574-2_24.
- [16] Marijn Heule, Warren A. Hunt Jr., and Nathan Wetzler. Expressing symmetry breaking in DRAT proofs. In Proceedings of the 25th International Conference on Automated Deduction (CADE), volume 9195, pages 591–606, 2015. doi:10.1007/978-3-319-21401-6_40.
- [17] Marijn J. H. Heule and Armin Biere. What a difference a variable makes. In Proceedings of the 24th International Conference on Tools and Algorithms for the Construction and Analysis of Systems (TACAS), Held as Part of the European Joint Conferences on Theory and Practice of Software (ETAPS), volume 10806, pages 75–92, 2018. doi:10.1007/978-3-319-89963-3_5.
- [18] Marijn J. H. Heule, Benjamin Kiesl, and Armin Biere. Short proofs without new variables. In Proceedings of the 26th International Conference on Automated Deduction (CADE), volume 10395, pages 130–147, 2017. doi:10.1007/978-3-319-63046-5_9.
- [19] Marijn J. H. Heule, Benjamin Kiesl, and Armin Biere. Strong extension-free proof systems. Journal of Automated Reasoning, 64(3):533–554, 2019. Extended version of [18]. doi:10.1007/s10817-019-09516-0.
- [20] Marijn J. H. Heule, Benjamin Kiesl, Martina Seidl, and Armin Biere. Pruning through satisfaction. In Proceedings of the Hardware and Software: Verification and Testing - 13th International Haifa Verification Conference (HVC), volume 10629, pages 179–194, 2017. doi:10.1007/978-3-319-70389-3_12.
- [21] Hannes Ihalainen, Jeremias Berg, and Matti Järvisalo. Clause redundancy and preprocessing in maximum satisfiability. In Proceedings of the 11th International Joint Conference on Automated Reasoning (IJCAR), volume 13385, pages 75–94, 2022. doi:10.1007/978-3-031-10769-6_6.
- [22] Matti Järvisalo, Marijn Heule, and Armin Biere. Inprocessing rules. In Proceedings of the 6th International Joint Conference on Automated Reasoning (IJCAR), volume 7364, pages 355–370, 2012. doi:10.1007/978-3-642-31365-3_28.
- [23] Benjamin Kiesl, Adrián Rebola-Pardo, and Marijn J. H. Heule. Extended resolution simulates DRAT. In Proceedings of the 9th International Joint Conference on Automated Reasoning (IJCAR), Held as Part of the Federated Logic Conference, FloC 2018, volume 10900, pages 516–531, 2018. doi:10.1007/978-3-319-94205-6_34.
- [24] Leszek Aleksander Kołodziejczyk and Neil Thapen. The Strength of the Dominance Rule. In Proceedings of the 27th International Conference on Theory and Applications of Satisfiability Testing (SAT), volume 305, pages 20:1–20:22, 2024. doi:10.4230/LIPIcs.SAT.2024.20.
- [25] Oliver Kullmann. On a generalization of extended resolution. Discrete Applied Mathematics, 96-97:149–176, 1999. doi:10.1016/S0166-218X(99)00037-2.
- [26] Javier Larrosa and Federico Heras. Resolution in max-sat and its relation to local consistency in weighted csps. In Proceedings of the 19h International Joint Conference on Artificial Intelligence (IJCAI), pages 193–198, 2005. URL: http://ijcai.org/Proceedings/05/Papers/0360.pdf.
- [27] António Morgado, Carmine Dodaro, and João Marques-Silva. Core-guided MaxSAT with soft cardinality constraints. In Proceedings of the 20th International Conference on Principles and Practice of Constraint Programming (CP), pages 564–573, 2014. doi:10.1007/978-3-319-10428-7_41.
- [28] António Morgado, Federico Heras, Mark H. Liffiton, Jordi Planes, and João Marques-Silva. Iterative and core-guided MaxSAT solving: A survey and assessment. Constraints An Int. J., 18(4):478–534, 2013. doi:10.1007/S10601-013-9146-2.
- [29] Nina Narodytska and Fahiem Bacchus. Maximum satisfiability using core-guided MaxSAT resolution. In Proceedings of the 28th Conference on Artificial Intelligence (AAAI), pages 2717–2723, 2014. doi:10.1609/AAAI.V28I1.9124.
- [30] Adrián Rebola-Pardo and Martin Suda. A theory of satisfiability-preserving proofs in SAT solving. In Proceedings of the 22nd International Conference on Logic for Programming, Artificial Intelligence and Reasoning (LPAR), volume 57, pages 583–603, 2018. doi:10.29007/TC7Q.
- [31] Paul Saikko, Jeremias Berg, and Matti Järvisalo. LMHS: a SAT-IP hybrid MaxSAT solver. In Proceedings of the 19th International Conference on Theory and Applications of Satisfiability Testing (SAT), pages 539–546, 2016. doi:10.1007/978-3-319-40970-2_34.
