Partial Minimum Branching Program Size Problem Is ETH-Hard
Abstract
We show that assuming the Exponential Time Hypothesis, the Partial Minimum Branching Program Size Problem () requires superpolynomial time. This result also applies to the partial minimization problems for many interesting subclasses of branching programs, such as read- branching programs and s.
Combining these results with the recent unconditional lower bounds for MCSP [6], we obtain an unconditional superpolynomial lower bound on the size of Read-Once Nondeterministic Branching Programs () computing the total versions of the minimum , read--, and size problems.
Additionally we show that it is -hard to check whether a given computing a partial Boolean function can be compressed to a of a given size.
Keywords and phrases:
MCSP, branching programs, meta-complexity, lower boundsCopyright and License:
2012 ACM Subject Classification:
Theory of computation Circuit complexityAcknowledgements:
The authors are grateful to Rahul Santhanam for suggesting working on connections between meta-complexity and branching programs, and to Mark Bun and anonymous reviewers for very helpful feedback on the preliminary versions of this paper.Funding:
Artur Riazanov was supported by the Swiss State Secretariat for Education, Research and Innovation (SERI) under contract number MB22.00026. Ludmila Glinskih was supported by a Sloan Research Fellowship to Mark Bun. Most of this work was done while Ludmila was at Boston University, and in part while she was visiting the Simons Institute for the Theory of Computing.Editor:
Raghu MekaSeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
The Minimum Circuit Size Problem (MCSP) has been a central problem of study in computational complexity in recent years. Given the truth-table of a Boolean function, and a size parameter , MCSP asks one to determine whether the function can be computed by a Boolean circuit of size . As the input length of MCSP is exponential in the input length of the function defined by the given truth table, it is clear that MCSP is contained in . Although it is not known whether MCSP is -complete, it is widely believed to be hard to compute. Starting from the work of Kabanets and Cai [11] there has been mounting evidence that [20]. For example, implies that widely believed cryptographic assumptions do not hold. On the other hand, there are complexity-theoretic barriers to proving -hardness of MCSP [15, 21].
Even before the work of Kabanets and Cai [11], variations of MCSP were studied in the context where the function is given not as a truth table, but represented in some computational model. One family of models that was extensively studied in this context is branching programs. The line of work [2, 17, 23, 22, 1] showed that branching program minimization is -hard for the natural cases of succinct representation of a function: either the function itself is represented as a branching program as in [2, 22, 1], or an input is a partial function with few fixed values as in [23]. The results, however, are limited to weak models of branching programs such as s or read-once branching programs. To our knowledge, no hardness results are known even for minimizing read-twice branching programs.
The potential cross-pollination between these two research areas motivates us to study Minimum Branching Program Size Problem (MBPSP): given a function represented as a truth table and a size parameter determine whether exists a branching program of size at most computing the given function.
On the meta-complexity side, one of the approaches to prove -hardness of MCSP is to show lower bounds for -MCSP problems, in which we ask whether an input Boolean function can be computed by a circuit from the class , where is a weaker class than general Boolean circuits. There are already results showing -hardness of -MCSP [14], -MCSP [8], and the hardness of DeMorgan Formulas-MCSP (MFSP) assuming the Exponential Time Hypothesis (ETH) [10]. All of these reductions share a common structure.
-
First they show hardness of the Partial -MCSP (-) problem in which the input is the truth table of a partial function.
-
Then they show a reduction from the - problem to the -MCSP problem.
Hardness results for Partial MCSP () are already known. In [9] Ilango showed that requires superpolynomial time to compute111The same hardness result holds for Partial MFSP ()., assuming the Exponential Time Hypothesis. Recently, Hirahara [7] showed the -hardness of under randomized reductions. So the only missing component for showing hardness of MCSP is to show an efficient reduction from to MCSP.
The study of hardness of MBPSP under the ETH continues the line of work above. Expressive power of branching programs is situated between that of Boolean formulas and Boolean circuits [18, 5]. On the other hand our understanding of branching programs seems to be slightly better than our understanding of circuits. One evidence for this is that the strongest lower bound on the size for an explicit function is [16] which is much stronger than lower bound for Boolean circuits [3]. Showing the ETH-hardness of MBPSP would improve the recent ETH-hardness result for MFSP [10] which is the frontier hardness result for total circuit minimization problems and move us one step closer to the ETH-hardness of MCSP itself. Our main result is completing the first step of the common recipe for establishing the hardness of -MCSP by showing hardness of (that we call for short).
Theorem 1.
Every deterministic algorithm computing on inputs of length requires time assuming the Exponential Time Hypothesis.
Note that the ETH-hardness of or does not imply the ETH-hardness of . Although on surface it seems that MBPSP () is just a special case of MCSP (), the hardness of the latter does not necessarily imply the hardness of the former. For example, in some special settings of parameters, an explicit construction of a many-one reduction between MCSP and MBPSP yields breakthrough circuit or lower bounds.222Suppose that is such a reduction from (the language of the truth-tables computable by circuits with at most gates) to . Then taking any function requiring circuits of size larger than [3, 12] and applying to it yields lower bound for general branching programs superior to the current state of the art. In a similar way, we get superlinear circuit lower bounds from any reduction from to . On the other hand, ruling out more natural parameter configurations does not seem as trivial: indeed, our main result shows that the identity function is a reduction on a certain subset of the instances.
On the branching program minimization side, the consequences are twofold. First, our result implies that all partial branching program minimization problems require time under ETH for any representation of a function that is at least as efficient as truth table.333We note that results in this area often establish -hardness and even hardness of size approximation for the case of concise representation, while our result only establishes ETH-hardness of the exact version of the problem. Second, we show that our result implies -hardness of minimization of a branching program computing a partial function that itself is represented as a branching program (see Section 1.2.3 for details).
In particular, our result implies that the problem of finding the smallest (a subclass of branching programs) that computes the given partial function is ETH-hard. But surprisingly, there is a polynomial-time algorithm that finds the smallest computing a total function [4]. This means that assuming ETH holds, the potential reduction from to MBPSP must fail on s.
1.1 Techniques
To show Theorem 1 we adapt the proof of a similar result of Ilango [9] for the hardness of and problems.
Theorem 2 ([9]).
Every deterministic algorithm computing on inputs of length requires time assuming the Exponential Time Hypothesis holds. The same holds for .
Ilango’s proof of Theorem 2 is a reduction from the -Bipartite Permutation Independent Set Problem (). For now we can think of it as an independent set problem on some special set of graphs. This is an -hard graph problem introduced in [13], where the authors show that any deterministic Turing machine requires -time444The input to this problem has bit size. to solve it if the Exponential Time Hypothesis holds. For an input graph with vertices labeled by elements of Ilango constructs a truth-table of a -variable partial Boolean function , such that
-
is a Yes-instance of iff can be computed by a DeMorgan formula with at most gates.
The key point here is that if we constrain a circuit or a formula to have at most (inner) gates, its structure becomes simple enough for a very precise analysis. [9] characterizes the structure of such formulas that compute which would yield the correspondence between such formulas with independent sets in .
Although branching programs are situated between formulas and circuits in computational expressivity [5], Theorem 2 does not apply to , since with a sharp constraint on the size, branching programs may become stronger than circuits. We observe in Proposition 7 that this is indeed the case, showing that branching programs are stronger than circuits if we constrain the size parameter of both models to “minimal viable” size: inner nodes for a DeMorgan circuit and non-sink nodes for a for the case of a Boolean function on input bits.
Nevertheless, we show that the reduction to from Theorem 2 does yield ETH-hardness for . Indeed, in our proof of Theorem 1 we only change the part , keeping the construction of . The key part of the proof is to understand what structure a branching program has if it has at most (non-sink) nodes. As is sensitive in every input variable, we need to understand a structure of a branching program that has unique variable labels. We call such programs once-appearance branching programs (). We show that can be computed by a once-appearance branching program iff is a Yes-instance of .
1.2 Corollaries
1.2.1 Weaker Classes
is not the only model that collapses to when constrained to have at most non-sink nodes. The same also holds for , and - for every positive integer . From this we get a family of hardness of minimization results for restricted classes of branching programs.
Corollary 3.
Every deterministic algorithm computing - for 555Or any other class of branching programs that collapses to on the -node programs. on inputs of length requires time assuming the Exponential Time Hypothesis.
1.2.2 Unconditional Results for a Weaker Model
In the recent work [6] the authors employed the framework of Ilango from [9] to show an unconditional lower bound on the size of the read-once nondeterministic branching program () computing the total version of MCSP. The proof consists of three components.
-
1.
Show that unconditionally cannot be computed by a of smaller than size.
-
2.
Show that Ilango’s reduction from to can be efficiently encoded using . This shows that if on inputs of length could be computed by a of size , then could be computed by a of size .
-
3.
Show that MCSP and have the same complexity up to a linear multiplicative factor. This implies that complexity of MCSP is at least .
Since the proof of Theorem 1 follows similar steps as the proof of the Theorem 2, and analogs of the third property from [6] also hold for , -, and , we get similar unconditional lower bounds for total minimization problems in all these models.
Corollary 4.
Every computing -MCSP for on inputs of length requires size at least .
1.2.3 -hardness of Compressing BPs
Additionally, we show that given a graph we can construct a branching program of polynomial size that exactly computes the partial function . Namely it outputs values on inputs on which is defined and on all other inputs. Then can be compressed to an which agrees with on its domain iff is a Yes-instance of . Combining this result with a folklore result of -hardness of compressing a branching program, we get the following result on the hardness of compressing branching programs.
Corollary 5.
The language of pairs where is a branching program666The same is true for all sub-classes of branching programs that simulate -. over the alphabet and is in integer, such that an extension of the partial function described by can be computed by a of size at most , is and -hard.
1.3 Organization
The structure of the paper is as follows:
2 Preliminaries
Throughout this work we denote by the set of elements . Let and be partial assignments. If the supports of and do not intersect, denote by the partial assignment that coincides with on the support of and with on the support of . For a graph we denote its set of vertices by and its set of edges by . Throughout the paper a circuit is a DeMorgan circuit (, and gates) with gates of arity 2. A formula is a tree-like DeMorgan circuit. A read-once formula is a formula such that each input bit feeds into exactly one gate. E.g. is a read-once formula, and is not. Since every DeMorgan formulas can be transformed in such a way that all -gates are adjacent to the inputs, we identify DeMorgan formulas with trees with literals in the leaves and / in the inner nodes.
A Branching Program () is a form of representation of functions. A -ary function777We will instantiate this definition for and , where , is represented by a directed acyclic graph with exactly one source and two sinks. Each non-sink node is labeled with a variable; every internal node has exactly outgoing edges: labeled with elements of . One of the sinks is labeled with and the other is labeled with . We say that a node queries if it is labeled with a variable . The size of a branching program is the number of nodes in it.
The value of the function for given values of the variables is evaluated as follows: we start in the source and then follow the edge labeled with the value of the variable in the source, we repeat this process until the current node is a sink. The label of this sink is the value of the function. We refer to a path that terminates in a -labeled sink as -path.
A Non-deterministic Branching Program () is a Branching Program over the alphabet with additional non-labeled guessing nodes with two non-labeled outgoing edges. The value computed by a Non-deterministic Branching Program is if there exists at least one -path that agrees with the given assignment.
A Branching Program (deterministic or not) is called read-once (denoted as in the deterministic case and in the non-deterministic case) if any path from the source to a sink contains each variable label at most once.
For a partial assignment and a branching program with input bits over the alphabet we write for a branching program obtained by removing every edge of form and contracting every edge of form if . It is easy to see that if computes the function , then computes the function .
The Minimum Branching Program Size Problem () gets as input the truth-table of a Boolean function of length and a size parameter , and outputs if there exists a branching program of size at most that computes . We use notation .
The Partial Minimum Branching Program Size Problem () gets as input the partial truth-table of a function of length and a size parameter , and outputs if there exists a substitution of each in the truth-table to a , transforming to a Boolean function , such that . We use notation .
2.1 Once-Appearance Branching Programs
A once-appearance branching program () is a branching program with distinct node labels. Hence, for a function depending on variables such representation contains at most non-sink nodes. For we identify the non-sink nodes with their labels, and identify the edges with equalities of form for a variable and from the alphabet.
It is easy to see, that s are strictly stronger than read-once DeMorgan formulas.888Notice that for functions sensitive to all variables, read-once DeMorgan formulas are equivalent to general DeMorgan circuits with inner gates. Thus Proposition 6 and Proposition 7 hold for such circuits as well.
Proposition 6.
If is computable by a read-once DeMorgan formula, then it is computable by an .
Proof.
We first observe that if and are computable by s, then so are and whenever and have disjoint set of variables. Indeed for is constructed by redirecting all edges going to the -in the computing to the source of the computing , and by redirecting all edges to the -sink in to the -sink of . For we redirect all edges to the -sink of to the source of the for , and all edges to the -sink to the -sink of . Now to finish the proof observe that all literals are trivially computable by an .
Proposition 7.
Let . Then is computable by an but is is not computable by a read-once DeMorgan formula.
Proof.
The for is a decision tree with the root , the edge going to , the edge going to , the edge , going to the -sink for .
Consider a read-once formula computing . Then computes a function , so must have a leaf . But computes a function , so must have a leaf . That contradicts the read-once property of . We note that this example can be extended to a function that is sensitive in an arbitrary number of variables, say for .
2.2 Bipartite Permutation Independent Set Problem
In [9] Ilango showed that Partial MCSP is ETH-hard. The main idea of the proof is to show a reduction from the -Bipartite Permutation Independent Set problem that was previously shown to be ETH-hard in [13]. In this section and later we assume that is even.
In -Bipartite Permutation Independent Set problem () we are given the adjacency matrix of an undirected graph over the vertex set where every edge is between the sets of vertices and . A graph is a Yes-instance iff it contains an independent set of size such that the coordinates of vertices in define a permutation of , i.e. .
Note that a permutation of corresponding to a Yes-instance of can be viewed as two permutations on disjoint -element subsets. One permutation is defined by vertices from (corresponding to a permutation of elements ), and another by vertices from (corresponding to a permutation of elements ). We use each of these interpretations interchangeably.
Theorem 8 ([13]).
-BPIS cannot be solved in deterministic time unless ETH fails.
Ilango in [9] gave a -time reduction from -BPIS to MCSP*, hence, showing that MCSP* cannot be solved in deterministic time unless ETH fails.
3 Main Theorem
In this section we show that the reduction Ilango constructed from ETH-hard problem to a formula minimization problem holds for the minimization problem as well. More precisely we show:
Theorem 9.
Suppose that is in where , then the Exponential Time Hypothesis is false.
Let us recall the reduction from [9]999We ignore all the edges except for ones between the sets and , the second argument is represented in binary form.. Let be an instance of . Then the reduction outputs the pair where is the truth table of a partial function defined as
| (1) | ||||
| (2) | ||||
| (3) | ||||
| (4) | ||||
| (5) | ||||
| (6) | ||||
| (7) | ||||
| otherwise | (8) |
where , , , and is the vector with 1 in the th entry and zeroes in all the others, and is such that . In the rest of the proof, we use as a vector with the th coordinate equal to and all the rest equal to zero with the appropriate dimension.
The intuition behind this reduction is that for every read-once formula, computing this function has to group triples of bits and such that , and encodes that an element maps to an element in the permutation. The cases (1)-(4) ensure that the read-once formula encodes some permutation on . The cases (5) and (6) ensure that the permutation maps the first elements into the first elements and the last elements into the last elements. And, finally, the case (7) enforces that for every edge in the original graph, the permutation defined by the read-once formula contains at most one endpoint of this edge. Note, that every endpoint or vertex chosen in the independent set encodes a mapping of one element of to one element of in the permutation.
As was shown in [9], this reduction maps Yes-instances of BPIS to Yes-instances of where is the number of input bits of :
Theorem 10 ([9]).
For the reduction as above iff .
To prove Theorem 10, Ilango shows that extends to a total function computable by a read-once monotone formula iff is a Yes-instance of . In our work we show that a similar statement holds for once-appearance branching programs ().
Theorem 11.
For the reduction as above iff can be computed by .
The plan of the proof is the following:
-
In Section 4 we show that the topological order of an computing defines a unique permutation .
-
In Section 5 we show that for as above is an independent set in , thus showing the “only if” direction of the equivalence.
-
Since by Proposition 6 that losslessly simulates read-once formulas, we conclude that Theorem 10 implies the “if” direction of the equivalence.
4 Structure of Computing
For the first step of our plan it is convenient to simplify the definition of . Let be equal to whenever one of (1)-(4) is satisfied, and otherwise. Here we may choose an arbitrary as in this domain does not depend on it.
We usually shorthand as . In Section 4 we identify the structure of an computing . Let us start by constructing an example of such an . As we mentioned before, whenever a function is computable by a read-once formula, it is computable by . Ilango [9] shows that all read-once formulas computing have form
| (9) |
for a permutation . One way to translate this into an is to chain together three-node s each of them computing one disjunct of (9). A topological order of such a is a sequence of triplets of with variables in triplets permuted in some order. We claim that topological order of every computing has this form101010Notice that we do not show that the in question has the exact form as the one constructed according to (9).. We prove this claim in Lemma 12.
We then say that an computing defines a permutation mapping iff for some . We refer to the triplets of consecutive nodes of form as -triplets or -blocks.
In this section we prove the following lemma.
Lemma 12.
Suppose is an that computes . Then let be labels of vertices of in a topological order. Then for every we have for some .
The proof of Lemma 12 is surprisingly delicate. The starting point is the characterization of the structure of s computing the subfunctions of corresponding to the conditions (1)-(4). The main tool in the proofs of these characterizations is the reduction to query complexity, e.g. has maximum possible query complexity , which enforces a very tight constraint on the structure of the computing it. A potential easy proof of Lemma 12 could have followed a logic similar to the proof above if had maximal query complexity . This, however, is not the case, query complexity111111Query , if or , then we may return any value, as none of the conditions (1)-(4) hold. Otherwise, for every possible combination of values we have only two possible conditions among (1)-(4) that might hold. If we return right away, if and , it suffices to query and to get the answer, if and , it suffices to query and to get the answer. of is at most , so such a simple argument is unfortunately ruled out.
4.1 Structure of Conditioned on (1)–(4)
We start by characterizing the structure of that computes .
Lemma 13.
The structure of an computing the function is the following: all nodes except for the -sink form a single -edge path where for some permutation . For each , the edge goes to , the edge goes to , the edge goes to (for it goes to the -sink), and the edge goes to the -sink.
Proof.
The query complexity of the function is , hence there should exist at least one path of length in this branching program. Next, consider the first two variables queried in this path. Assume they do not correspond to and with the same label. Without loss of generality assume the first node is labeled by and the second node is labeled by (observe that the cases of the first two labels being and or and are identical, since the variables in the pairs are symmetric). Now, both edges outgoing from should be parts of the paths that lead to . As for both values of and exists a -assignment that is zero in all variables other than . Hence, if after querying we do not query , we get a BP that incorrectly computes the function. As we consider the first two nodes of the longest path, we get that both edges from should directly go to the node labeled by . But it turns the to be insensitive to a change in the value of that contradicts the sensitivity of function. Hence the edge must lead to the node labeled with . Then it remains to observe that after applying the substitutions or the query complexity of the function decreases by exactly . Note that paths corresponding to these two substitutions must lead to the third node in the branching program . This node then must compute , so the rest of the has the desired structure by the simple induction on . Next, we summarize the structures for the conditions (2)–(4).
Lemma 14.
The following structural properties hold.
- (2)
-
An computing the function has the following property. Every node of this reachable from the source computes the function where is the set of indices of all -variables that do not precede in the topological order.
- (3)
-
An computing the function is a Hamiltonian path terminating in the -sink with edges of the form , going to the next node in the path, and the edges , going to the -sink.
- (4)
-
In an computing the function the -sink is not reachable from the source.
Proof.
The structural property for (4) is obvious: if there is a path from the source to the -sink, the function is not identically zero.
The structural property for (3) is also easy to see. The query complexity of is , so the must have a -edge path. The edges in these path are the -edges and the path terminates in the -sink. Suppose an edge of the form or does not go to the -sink, say it goes from a node to a node , then the path from the source to according to the Hamiltonian path, then following the edge and then continuing to the -sink by the Hamiltonian path contradicts the fact that the computes the disjunction of all variables.
Now let us show the structural property for (2). The source computes the function by definition. For any node computing for a set we have that the edge goes to the node computing . For any node that is reachable from the source its edges go to the nodes that compute the same function since otherwise the function computed by the source depends on . Then take any node . Since the function depends on all bits of , this node is reachable from the source. Hence, by the observations above, it computes a function of the form with .
Observe that every path from the source to must visit all nodes from that precede in the topological order. Indeed, if there is a path from the source to that does not visit that precedes in the topological order, then the returns given the assignment , which is a contradiction. Hence, if the node computes and the node computes , where precedes in the topological order, then . Moreover and so . Therefore, the sets form a chain by inclusion and are all different, which implies the property given that does not contain indices of nodes that precede in the topological order.
4.2 Proof of Lemma 12
We derive Lemma 12 from the following two lemmas.
Lemma 15.
Let be an computing . Then for every topological ordering of the nodes of there exists a mapping such that neighbors in the topological order. Additionally, the following property holds
-
if precedes and in the topological order, then the edge does not go to 1-sink.
Lemma 16.
Suppose an computes . Suppose are the labels of the nodes in some topological ordering of . Then for every and every we have .
Proof of Lemma 12.
Let denote the topological order of the labels of the nodes in . Consider a modified version of that we denote such that whenever possible or . Observe that this modification never introduces additional violations of injectivity since cannot neighbor a node between and in the topological order for .
Now suppose for some . WLOG assume that and . Then, by construction of and there exists such that . Notice that must be a single element of that neighbors and , thus and . Then we get a contradiction with the statement of Lemma 16. Hence is an injective mapping. As is an injective mapping that maps each element of a set of the size into an element of a set of the size , we get that is a bijection.
By Lemma 13 has a subsequence of form where for a permutation . Observe that must contain for , otherwise , contradicting Lemma 15 as does not neighbor .
If contains more than one element of then by surjectivity of . Moreover, and either , or , for some . Then by Lemma 13 we get . Thus does not belong to the image of , which is a contradiction. Therefore, . Iteratively applying this procedure to and restricted to for concludes the proof of the theorem.
4.3 Proof of Lemma 15
Let for be the diagram obtained from by contracting all edges of form and removing all edges of form . Then the diagram computes . Then from Lemma 13 and (1) we get that consists of a path of length that is split into consequent blocks with two nodes in each, where the labels of the nodes in each block are for .
Consider an arbitrary block in and let and be the nodes of corresponding to the nodes in the block, such that precedes in the topological order of .
4.3.1 Case 1: and are labeled with and respectively
must still be present in and be reachable from the source (say, with the assignment ). Then consider edges in labeled with and : and , where is the first node in the subsequent block of , these are uniquely defined since is an . We claim that at least one of these edges does not appear in i.e. it is present in due to a contraction.
Assume towards a contradiction that both of the edges are present in . Then let us trace a path from the source of to the -sink according to the assignment . This path must contain and , as -node, corresponding to should be queried on substitution , and then, after assigning and this path leads to the node . By Lemma 14 for (2) the node in and our choice of , computes . Since precedes in the topological order, . Hence, we get which with value leads to a contradiction.
Now, since at least one of the edges and does not appear in , but appears in , the node in is adjacent to a node labeled with an element of . Since and belong to the Hamiltonian path in , must neighbor in the topological order. Then define where is the label of . Notice that here the node never precedes both and , so the property is satisfied.
4.3.2 Case 2: and are labeled with and respectively
First, suppose and are the first two nodes in . By Lemma 13 the edge in goes to the first node of the next -pair, let us denote it by . The edge goes to the -node in . Now, assume, towards a contradiction, that the -edge in goes to as well. Then, a path in corresponding to an assignment skips the node , as is the source of this , and is going after in the topological order. Similarly to the proof of Case 1, we get a contradiction with the fact that the subdiagram of rooted in computes the function with , which is on , hence , whereas .
Now, assume there is at least one other -pair queried before and . Let and be the pair that precedes and in the topological order of . Note, that in both edges and go to . And edge in goes to a node (skipping ), which is the first node in the next -block.
Claim 17.
One of the edges among and differ in and (that is, the edge with this label connects different pairs of nodes).
Proof.
The proof of this claim repeats a similar argument in the Case 1. Assume these two edges in are the same as in . Consider a substitution . As computes the disjunction of , a path corresponding to in should query all bits of preceding in the topological order. Hence, goes through , and to . Now, as skips and goes to , which by Lemma 14 computes in . Analogously to the Case 1, , so we get that , which contradicts . Thus, one of the edges or in leads to a different node compared to .
4.3.3 Case 2a: the edge goes to
By Claim 17 does not go to . But in this case it goes to a node labeled by an -variable in . Then as queries all bits in on the assignment , we get that goes before in the topological order. So we get that neighbors an -node in the topological order.
4.3.4 Case 2b: the edge does not go to
Let be the endpoint of in (notice that can not go to since then it would go there in as well). Consider the path from to in according to the assignment , by the construction of all nodes on this path are labeled with -variables. Let be the last node on this path before . We define . Now it remains to show the following:
-
1.
is the immediate predecessor of in the topological order.
- 2.
Let us first prove Item 1. Consider the program . By Lemma 14 for (3), consists of a single Hamiltonian path of -edges. Then, consider the edge in . This edge goes to , so is present in , so in topological order there are no -node or -node between and . Given that precedes in the topological order, is the immediate predecessor of .
Now let us prove Item 2. Consider an assignment , and . Let us trace the path corresponding to this assignment from the root of to . Notice that is reachable by this path since it is reachable in by the path according to , . After let us follow the edge to and then to following the -edges. Suppose that goes to the -sink. Let be the path from the root to the -sink that we have traced. Then consider the program . Let be the result of contraction of the edges of form in . exists in , hence the -sink is reachable in this program. This contradicts the fact that is identically zero.
4.4 Proof of Lemma 16
We are going to gradually derive the structure of the edges going out of the nodes .
By (3) we have , so by Lemma 14 the edges of from and must either go to a variable in or to the next variable in in the topological order. Hence leads to , leads to .
By (1) we have , the diagram has structure defined by Lemma 13. Then leads to and leads to and leads either to or . Now observe that the edge goes to the -sink by the structure of the diagram . If it does go to the assignment , and is evaluated to by , which is a contradiction with (3). See Figure 1 for the edges we identified so far.
Now consider an edge . By Lemma 14 for (3) it must go to the -sink in the diagram , so in it goes either to or to the -sink. Recall that by (2) we have . Suppose goes to for . Then consider the assignment , , . A path , corresponding to it passes through the node by the structure of given by Lemma 14. Hence, then leads to and to , which computes for some . Thus , yet . Hence goes to the -sink.
Now we can pinpoint the endpoint of the edge . Suppose it goes to . Then we get the contradiction with the assignment , and . since we go from to and then to the -sink, yet .
Then consider the edge , we claim that it goes to . Otherwise consider the assignment , and , since the path of this assignment passes through , then goes to and then to a node strictly below that by Lemma 14 computes for . Thus we get a contradiction with .
Then we get the final contradiction as follows: depends on , but does not, in other words , yet .
5 Structure of Computing : proof of Theorem 11
Lemma 18.
For every computing , the permutation defined by maps elements of into , and elements of into elements of .
Lemma 19.
For every computing , the permutation defined by corresponds to an independent set in , that is is independent in .
We start by deriving the theorem from the lemmas.
Proof of Theorem 11.
From Lemma 18 and Lemma 19 we get that an computing defines a permutation on which maps elements of the first/second half of to the elements of the first/second half of respectively, and for every edge it contains at most one of the vertices of a BPIS instance. Hence, the permutation defined by any computing is a correct certificate for a instance. Thus an computing exists iff there is a correct permutation defining independent set in an input instance.
5.1 Proofs of Lemma 18 and Lemma 19
Both of the lemmas follow from a helper lemma about the structure of for :
Lemma 20.
Let be an computing . Then, if a path in goes to the -sink from one of the nodes in a triplet , then sets at least two variables among to .
We defer the proof of Lemma 20 to Section 5.2. Now we are ready to prove the lemmas.
Proof of Lemma 18.
First, as is a restricted version of , we get that is well-defined by Lemma 12. Moreover, we know that the topological order of has the form , where for every there exist , such that , where .
Let form a consequent -triplet in the , with . Assume towards a contradiction that and are from different halves of the set . WLOG let and .
Consider a substitution , in which . Consider a path consistent with . By property (7) it terminates in the -sink. Now, consider the last -triplet, that goes through before getting to the -sink. This could either be -triplet, or , such that . In both cases, at most one of the variables in such -triplets is set to . Hence, we get a contradiction with Lemma 20. The proof for the -triplets in the last -elements of goes similarly.
Proof of Lemma 19.
Consider an edge connecting with in . Assume towards a contradiction that maps an element to , and to . Then, there are two triplets , and such that within each -triplet the elements are layered consequently in . Consider a substitution , which sets , all other bits of to , , all other bits of to , and all bits of to . , hence there is a path in consistent with which goes from the source to 1-sink. Consider a triplet , which is queried the last on the path . Similarly to the previous case we consider two cases, if this triplet is (or , proof for which is analogous) or not. In the latter case, by our assumptions on the structure of , all , , , and .
If and , then, by our choice of and this triplet, we get that goes to 1-sink after substituting . Which is a contradiction with Lemma 20.
Finally, assume either and . Then there is a path in from the nodes directly to 1-sink which is consistent with a substitution . Thus, again, leads to a contradiction with Lemma 20.
5.2 Proof of Lemma 20
Assume the opposite: there exist paths in , which go to the -sink immediately after substituting exactly one -value to variables of an -triplet. Consider a path among these paths, which goes to the -sink from the earliest node according to the topological order of . Let us denote the -triplet, from which goes to the -sink as , and let be the last non-sink node in .
Claim 21.
assigns to .
Proof.
Assume the opposite: the last edge in going to the -sink is .
Consider a case . We want to show that -edge cannot lead to the 1-sink. To see this observe that is sensitive in all -variables. Hence, in the a path consistent with setting goes through , and goes to the -sink. This contradicts our assumption that the edge goes to .
Now consider a case when . Consider a path in the such that for every and if precedes in the topological order, and and otherwise. By (1) we have , so every such ends in the -sink. This contradicts the fact that the edge goes to the -sink, since at least one of such paths contains this edge. The claim then follows.
In the remainder of the proof, we show that cannot be the only -value assigned by to the variables in the last triplet . Consider the case when or and the value of is assigned to by . Consider an corresponding to a partial substitution . We know the structure of from Lemma 13. For each -block there is a path to the -sink iff we set both . But that contradicts the existence of a path of a form or from the triplet to the -sink. Hence, could not be or variable.
Consider the only remaining case, when , and sets . In the following two claims we show that should ask either or prior to .
Claim 22.
Either or precedes in the topological order.
Proof.
Assume the opposite, is the first node in the triplet. By Lemma 15 goes either right before or right after in the topological order. Hence, we get that nodes in this triplet appear in the order in the topological order. But then, by the property in Lemma 15, if is the first node of the triplet then the edge does not go to the -sink, contradicting the properties of .
Claim 23.
Path queries either a value of or before querying .
Proof.
By Claim 22 we know is not the first node of the triplet in the topological order. Now assume that skips querying -nodes preceding in this triplet and goes directly to . Let be the last node prior to in . By our assumption belongs to an -triplet different from .
First let us rule out the case or . If assigns to , then the edge in goes to the endpoint of the edge in , which contradicts Lemma 13, since this edge must go to the first variable of the subsequent -block, which precedes in the topological order of . If assigns to , then in the edge goes to the -sink (as it does not go to the node in the same -block), hence in the edge must go to the -sink, which implies that the function computed by does not depend on , which is a contradiction, since does depend on .
Hence . Let us now show that assigns to . Assume the opposite: is set to in . Let us then trace the path corresponding to the assignment in until we reach the node . This must happen by Lemma 14 for (3). Now by the Lemma 13 and property (1) the edge must go to the first node of the subsequent -block, which is a contradiction with the assumption that ends in . Therefore assigns .
Consider an assignment of , . By (3) we have . But by our assumptions an edge does not go to the -sink immediately, it goes to first. A node should also be reachable by a substitution which assigns , , as is sensitive to the value of by (3). But . Hence, we get a contradiction with the fact that paths corresponding to assignments and meet in the node , and follow the same path in , but one must end in the -sink, and another in the -sink. Therefore, either or node is not queried by .
Now, let be the immediate predecessor of in . By Claim 23 we know that either . First, consider the case of . Consider a substitution of and . We know that the queries all values of . Hence, contains a path , which goes to the -sink. But, as the assignment does not satisfy , and is consistent with the substitution we get a contradiction (since the path passes through ).
Finally, to get a contradiction with the case when we show the following two claims.
Claim 24.
If , then goes before and in the topological order.
Proof.
Assume the opposite. Let go right after in the topological order. So we get that the order of nodes in this triplet is . First, note that edge goes to as queries all s and s sequentially, and every edge labeled with -edge in goes to either the next node or to the -sink. We also know that -edge goes to the -sink. Now note that that a path consistent with a substitution defined as should reach node . But then this path will follow edge to and edge to sink, though . A contradiction.
Claim 25.
If is the first node in a -triplet then path goes through .
Proof.
Consider all possible edges which get to this triplet. Let be the triplet right before this one. From the structure of we know that all edges from and either go to a node within this triplet, or the -sink, or to . We also know that cannot skip , as -node is reachable by a path consistent with a substitution . Hence if skips then would be insensitive to .
It now remains to show that cannot skip either. In every edge should go to -sink. Hence edge either goes to -node or to -sink. This finishes the proof.
From these two claims, we get that should necessarily go through . Hence, by considering we get that a path consistent should either go to a node in the next triplet, or go to the -sink, in case it is the last triplet of . Which contradicts the fact that after substituting and goes to the -sink. This finishes the proof of Lemma 20.
6 Corollaries
Multiple corollaries follow from Theorem 11. For all classes of branching programs that degenerate to an if constrained on having at most nodes on an -bit input, we get that the corresponding minimization problem is ETH-hard. In particular it holds for general branching programs (), ordered binary decision diagrams () and read- branching programs for any (-). For definitions of the latter two classes we refer the reader to [27]. Formally we have
Corollary 26.
Let be a circuit class such that for every that depends on all its variables there exists with computing if and only if there exists computing .
Then, assuming ETH holds we have that requires time .
It is easy to see that for this statement holds, so in particular Theorem 1 is implied.
Proof of Corollary 26.
Let be the reduction given by Theorem 11. Observe that every partial function in the image of can only be extended to a function that depends on all of its variables.
Hence by the assumption on , iff is computable by an . Therefore, Theorem 11 and Theorem 8 imply the statement.
Following on the results from [6], in which the authors show that is unconditionally hard for s by showing that reduction of Ilango is computable by a s, we get that similar lower bound holds for all total minimization problems for branching programs.
Corollary 27.
For as in Corollary 26, the size of the smallest that computes is , where is the input size of .
Proof.
By [6, Theorem 6] we have that the sizes of s computing - and -MCSP are linearly related (the theorem is only stated for vanilla MCSP, but the proof does not use this). By [6, Threorem 4], the reduction can be implemented in , so by Theorem 11 size of - with input size is lower bounded by the size of which by [6, Theorem 3] is .
Another corollary follows from the fact that always outputs a partial functions that can be encoded as a -.
Corollary 28.
It is -hard to check, whether any extension of a partial function expressed by a - over the alphabet can be computed with an .
Proof.
Since is an -complete problem, it suffices to check that for every the function can be represented as a - of polynomial size. First, notice that a - can distinguish between the cases (1)-(6) and the case that either (7) is true or the function equals to : we read variables in the order and for each of the patterns among (1)-(6) we maintain if the current prefixes of , , and coincide with the prefix of the pattern. Thus, at each level of our - we have states. For each of the states at the last level, we choose one of the patterns which is satisfied arbitrary, so we can merge the states into 7: 6 corresponding to the conditions (1)-(6) and one to the case where none of the conditions hold. For each of the cases (1)-(6), the restricted function can easily be computed by a -.
Now it remains to distinguish the case (7) from a star with a -. In order to do this we split the input into chunks of length and read then sequentially. For each chunk we maintain the set of possible values among that the chunk might have given the current prefix. Observe, that the number of potential states is linear. After we read all the chunks we have states at the last level, so we know the only tuple that may satisfy the predicate in (7). Since the edges are fixed, we can unite each of the states at the last level with the -sink or the -sink.
Finally, to show the -hardness of compressing s we use the following fact.
Proposition 29 ([24]).
Let -SAT be a language of satisfiable - formulas in which every variable occurs in at most 4 clauses. -SAT is -complete.
Corollary 30.
It is -hard to check whether an input - for a total function can be compressed to a of a constant size.
Proof.
By Proposition 29 we get that is -hard. Now, we construct a reduction from to the partial - minimization as follows. For a - formula in which every variable appears at most times we construct a branching program by consequently adding nodes of each clause. Denote the first clause of as , and let be the variables in it. We add nodes labeled by , , and to . We set the node labeled by as a source of . Then, if was negated in we direct a 1-edge from the -node to the node labeled by , otherwise we direct a 0-edge from to the -node. Then we similarly define one of the edges of to . Finally, if was negated in we direct its -edge to the 1-sink. Then we add nodes from the next clause of , and direct all missing edges for nodes , , and to the first node corresponding to the next node. We repeat this procedure for all clauses, redirecting all unassigned edges for nodes corresponding to variables in the last clause of to the -sink.
Now, for any substitution of values of the variables outputs if was not satisfied by , and otherwise. But note, that if was an unsatisfiable formula than would had to output on all inputs. Hence, it could be compressed to a branching program of size 0 (as we count the number of all nodes other than sinks). On the other hand, if is satisfiable and has at least one unsatisfying assignment has to contain at least one node other than a sink. To account for the fact that the size of corresponding to that is satisfiable by any assignment is also (it is just a 0-sink), we change our reduction to first substitute all 0 assignment to . If we set to be a for an identity function on one bit which has the size . Otherwise, construct as described before. Additionally, as for any -SAT instance in which every variable appears in at most 4 clauses we can easily construct an equivalent - of polynomial size. we get that checking whether this BP can compressed to a of size allows us to check whether the formula is unsatisfiable.
7 Future Work
The main future direction is to strengthen our results to the total version of MBPSP, and to show -hardness of MBPSP (or at least ). We discuss a couple of barriers for improving results in both of these directions as well as additional open questions motivated by our work.
7.1 Show polynomial-time oracle reduction from to MBPSP
As noted in [10] such reductions are very ad hoc for each specific model. Indeed, the reduction that Ilango showed for the Minimum DeMorgan Formula Size Problem in [10] does not work for branching programs as it relies on the tree-like structure of formulas. The ability of branching programs to reuse the same computation in different branches makes it harder to come up with a suitable reduction. For now we do not even know such a reduction for s but the natural next step would be to reduce - to -MCSP.
7.2 Prove a stronger conditional lower bound
It would be interesting to find a lower bound stronger than for the time complexity of (or ). The obstacle here is that we actually show the ETH-hardness of , for which this time complexity is actually tight, since there are potential branching programs that can compute a function. Hence, in order to have stronger lower bounds we need to consider larger size parameters. The current framework requires us to find explicit functions that require -size larger than the given size parameter. Hence, the best we can hope for without improving the state-of-the-art lower bounds is to show that is ETH-hard. Here the trivial upper bound is , which while still quasipolynomial, is significantly stronger.
7.3 Results for nondeterministic BP minimization
Currently our hardness result only holds for the problem of minimizing deterministic branching programs. But we believe that similar ideas could be used to show that the problem of minimizing nondeterministic branching programs for partial functions is also ETH-hard. Nondeterministic branching programs () differ from deterministic ones by an additional type of nodes, called nondeterministic nodes, that do not have labels. Such nodes correspond to nondeterministic guessing of a branching program. Note, that if we consider a complexity measure that counts all nodes in an other than sinks, then our main result immediately extends to the ETH-hardness of minimizing and - for all .
However, the standard complexity measure for s is the minimal number of nodes that query a variable (so all nodes other than sinks and nondeterministic nodes) in an computing the function. So one of the approaches to showing the ETH-hardness of the Partial Minimum Nondeterministic Branching Program Size Problem could be in showing that existence of the with nondeterministic nodes for is equivalent to the existence of the deterministic for . Namely by adding the nondeterministic node we do not make it easier for an to compute . Could this result be obtained with an additional careful analysis?
7.4 Could we adapt the -hardness proof of Hirahara?
Hirahara in [7] showed a randomized reduction from -hard problem that is a variation of Minimum Monotone Satisfying Assignment called Collective Minimum Weight Monotone Satisfying Assignment (CMMSA) problem to . One of the possible directions of strengthening our result from the ETH-hardness to -hardness is by adapting this framework for branching programs. Indeed, the reduction from [7] seems to be robust enough to correctly work on all no-instances of CMMSA even for branching programs. But the same reduction does not immediately go through for the yes-instances. The main obstacle that we see is in showing an analogue of the Uhlig’s lemma [25, 26] for branching programs. This lemma states that in order to compute the direct product of the same function (of arbitrary circuit complexity) on subexponentially many inputs, it is sufficient to use one circuit of size . Such efficient compression is an important step for constructing a small circuit corresponding to the yes-instances of CMMSA. It is not known whether an analogue of Uhlig’s lemma exists for branching programs. Rao and Sinha [19] showed a variation of the direct-sum theorem for read-once BPs, hence disproving the analogue of Uhlig’s lemma in setting. We believe that showing an analogue of Uhlig’s lemma for general branching programs in the specific setting required by the Hirahara’s proof is a promising direction for showing -hardness of .
References
- [1] Beate Bollig. On the minimization of (complete) ordered binary decision diagrams. Theory of Computing Systems, 59(3):532–559, 2016. doi:10.1007/s00224-015-9657-x.
- [2] Beate Bollig and Ingo Wegener. Improving the variable ordering of obdds is np-complete. IEEE Trans. Computers, 45(9):993–1002, 1996. doi:10.1109/12.537122.
- [3] Magnus Gausdal Find, Alexander Golovnev, Edward A. Hirsch, and Alexander S. Kulikov. A better-than-3n lower bound for the circuit complexity of an explicit function. In FOCS, pages 89–98. IEEE Computer Society, 2016. doi:10.1109/FOCS.2016.19.
- [4] S.J. Friedman and K.J. Supowit. Finding the optimal variable ordering for binary decision diagrams. IEEE Transactions on Computers, 39(5):710–713, 1990. doi:10.1109/12.53586.
- [5] Oliver Giel. Branching program size is almost linear in formula size. Journal of Computer and System Sciences, 63:222–235, 2001. doi:10.1006/JCSS.2001.1762.
- [6] Ludmila Glinskih and Artur Riazanov. MCSP is hard for read-once nondeterministic branching programs. In Armando Castañeda and Francisco Rodríguez-Henríquez, editors, LATIN 2022: Theoretical Informatics - 15th Latin American Symposium, Guanajuato, Mexico, November 7-11, 2022, Proceedings, volume 13568 of Lecture Notes in Computer Science, pages 626–640. Springer, 2022. doi:10.1007/978-3-031-20624-5_38.
- [7] Shuichi Hirahara. Np-hardness of learning programs and partial MCSP. In 63rd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2022, Denver, CO, USA, October 31 - November 3, 2022, pages 968–979. IEEE, 2022. doi:10.1109/FOCS54457.2022.00095.
- [8] Shuichi Hirahara, Igor C. Oliveira, and Rahul Santhanam. Np-hardness of minimum circuit size problem for OR-AND-MOD circuits. In Rocco A. Servedio, editor, 33rd Computational Complexity Conference, CCC 2018, June 22-24, 2018, San Diego, CA, USA, volume 102 of LIPIcs, pages 5:1–5:31. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2018. doi:10.4230/LIPICS.CCC.2018.5.
- [9] Rahul Ilango. Constant depth formula and partial function versions of MCSP are hard. In FOCS, pages 424–433. IEEE, 2020. doi:10.1109/FOCS46700.2020.00047.
- [10] Rahul Ilango. The minimum formula size problem is (ETH) hard. In 62nd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2021, Denver, CO, USA, February 7-10, 2022, pages 427–432. IEEE, 2021. doi:10.1109/FOCS52979.2021.00050.
- [11] Valentine Kabanets and Jin-yi Cai. Circuit minimization problem. In STOC, pages 73–79. ACM, 2000. doi:10.1145/335305.335314.
- [12] Jiatu Li and Tianqi Yang. 3.1n - o(n) circuit lower bounds for explicit functions. In Proceedings of the 54th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2022, pages 1180–1193, New York, NY, USA, 2022. Association for Computing Machinery. doi:10.1145/3519935.3519976.
- [13] Daniel Lokshtanov, Dániel Marx, and Saket Saurabh. Slightly superexponential parameterized problems. SIAM J. Comput., 47(3):675–702, 2018. doi:10.1137/16M1104834.
- [14] William J. Masek. Some np-complete set covering problems. Unpublished Manuscript., 1979.
- [15] Cody D. Murray and R. Ryan Williams. On the (non) np-hardness of computing circuit complexity. Theory of Computing, 13(4):1–22, 2017. doi:10.4086/toc.2017.v013a004.
- [16] Eduard I. Nechiporuk. On a boolean function. Dokl. of USSR Academy of Sciences, 169(4):765–766, 1966. URL: http://mi.mathnet.ru/dan32449.
- [17] Arlindo L. Oliveira, Luca P. Carloni, Tiziano Villa, and Alberto L. Sangiovanni-Vincentelli. Exact minimization of binary decision diagrams using implicit techniques. IEEE Trans. Computers, 47(11):1282–1296, 1998. doi:10.1109/12.736442.
- [18] Vaughan R. Pratt. The effect of basis on size of boolean expressions. In FOCS, pages 119–121. IEEE Computer Society, 1975. doi:10.1109/SFCS.1975.29.
- [19] Anup Rao and Makrand Sinha. A direct-sum theorem for read-once branching programs. In Klaus Jansen, Claire Mathieu, José D. P. Rolim, and Chris Umans, editors, Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, APPROX/RANDOM 2016, September 7-9, 2016, Paris, France, volume 60 of LIPIcs, pages 44:1–44:15. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2016. doi:10.4230/LIPICS.APPROX-RANDOM.2016.44.
- [20] Alexander A. Razborov and Steven Rudich. Natural proofs. J. Comput. Syst. Sci., 55(1):24–35, 1997. doi:10.1006/JCSS.1997.1494.
- [21] Michael E. Saks and Rahul Santhanam. Circuit lower bounds from np-hardness of MCSP under turing reductions. In Shubhangi Saraf, editor, 35th Computational Complexity Conference, CCC 2020, July 28-31, 2020, Saarbrücken, Germany (Virtual Conference), volume 169 of LIPIcs, pages 26:1–26:13. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2020. doi:10.4230/LIPICS.CCC.2020.26.
- [22] Detlef Sieling. The complexity of minimizing and learning obdds and fbdds. Discrete Applied Mathematics, 122(1):263–282, 2002. doi:10.1016/S0166-218X(01)00324-9.
- [23] Yasuhiko Takenaga and Shuzo Yajima. Hardness of identifying the minimum ordered binary decision diagram. Discret. Appl. Math., 107(1-3):191–201, 2000. doi:10.1016/S0166-218X(99)00226-7.
- [24] Craig A. Tovey. A simplified np-complete satisfiability problem. Discret. Appl. Math., 8(1):85–89, 1984. doi:10.1016/0166-218X(84)90081-7.
- [25] Dietmar Uhlig. On the synthesis of self-correcting schemes from functional elements with a small number of reliable elements. Mathematical Notes of the Academy of Sciences of the USSR, 15.6:558–562, 1974.
- [26] Dietmar Uhlig. Networks computing boolean functions for multiple input values. Proceedings of the London Mathematical Society Symposium on Boolean Function Complexity, pages 165–173, 1992.
- [27] Ingo Wegener. Branching Programs and Binary Decision Diagrams. Society for Industrial and Applied Mathematics, 2000. doi:10.1137/1.9780898719789.
