Lifting to Randomized Parity Decision Trees
Abstract
We prove a lifting theorem from randomized decision tree depth to randomized parity decision tree (PDT) size. We use the same property of the gadget, stifling, which was introduced by Chattopadhyay, Mande, Sanyal and Sherif [ITCS 23] to prove a lifting theorem for deterministic PDTs. Moreover, even the milder condition that the gadget has minimum parity certificate complexity at least suffices for lifting to randomized PDT size.
To improve the dependence on the gadget in the lower bounds for composed functions, we consider a related problem whose inputs are certificates of . It is implicit in the work of Chattopadhyay et al. that for any function , lower bounds for the -depth of give lower bounds for the PDT size of . We make this connection explicit in the deterministic case and show that it also holds for randomized PDTs. We then combine this with composition theorems for -depth, which follow by adapting known composition theorems for decision trees. As a corollary, we get tight lifting theorems when the gadget is Indexing, Inner Product or Disjointness.
Keywords and phrases:
Parity decision trees, compositionCategory:
RANDOMFunding:
Farzan Byramji: Supported by NSF Award AF: Medium 2212136.Copyright and License:
2012 ACM Subject Classification:
Theory of computation Oracles and decision treesAcknowledgements:
FB thanks Sreejata Kishor Bhattacharya, Eric Blais, Zachary Chase, Arkadev Chattopadhyay, Jyun-Jie Liao, Shachar Lovett, Jackson Morris and Anthony Ostuni for discussions and feedback at various stages of this work. The authors would like to thank the anonymous reviewers for helpful comments.Editors:
Alina Ene and Eshan ChattopadhyaySeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
Lifting theorems provide a way to convert lower bounds for a function in a weak model of computation to lower bounds in a stronger model of computation by composing with a function , typically called a gadget. Given functions and , their composition is defined by
Typically such lifting theorems show that for certain choices of the gadget , the two-party communication complexity of in some model is lower bounded by the complexity of in a related query model [36, 27, 25, 28, 14, 31]. These have several applications and have led to the resolution of some long-standing problems [36, 26, 22, 15, 33]. An important challenge in the area is to decrease the gadget size to a constant. Current proofs require the gadget size to be logarithmic in the input length of the outer function.
As a stepping stone towards query-to-communication lifting theorems with improved gadget size, we may consider the problem of lifting to models which lie between communication protocols and decision trees. One such natural model is that of parity decision trees (PDTs). A parity decision tree is a decision tree where each node queries a parity for some and the sum is over . While being interesting on its own, another motivation for proving PDT lower bounds comes from proof complexity. The minimum size of a refutation of an unsatisfiable CNF formula in the proof system tree-like Resolution over parities () is (essentially) equal to the minimum size of a deterministic parity decision tree solving the related false clause search problem for [29]. Lifting theorems for deterministic parity decision trees using constant size gadgets were recently proved by Chattopadhyay, Mande, Sanyal and Sherif [16] and independently by Beame and Koroth [6] which gave a direct way to transform tree-like Resolution lower bounds to tree-like lower bounds.
As a next step, we may ask for lifting theorems for randomized parity decision trees. While lower bounds for randomized PDTs do not seem to directly imply lower bounds for stronger proof systems, lower bound techniques against randomized PDTs (along with several other ideas) have been recently used to prove lower bounds against certain subsystems of (dag-like) [20, 11, 3]. (More precisely, they use distributional lower bounds against deterministic PDTs.)
In this work, we prove a lifting theorem from randomized decision tree (DT) depth to randomized parity decision tree (PDT) size with constant size gadgets. For a function , we use to denote the deterministic DT depth of and to denote the -error randomized DT depth of . Similarly, we use and to denote the corresponding size measures. We use in the superscript to denote the analogous PDT measures. For example, denotes the minimum size of any randomized PDT computing to error .
To prove the lifting theorem for randomized PDTs, we use the same property of the gadget, stifling, which was introduced by [16] to prove their lifting theorem for deterministic PDTs. A function is said to be -stifled if for all of size and , there is some way to fix all bits other than those in to force the function to output . A function which is -stifled is also simply called stifled. Some examples of stifled functions include Inner Product, Indexing and Majority.
Theorem 1.
For any function , any stifled function ,
where the implicit constant depends on the gadget size .
Our results also hold for relations (like most other lifting theorems) and partial , but we mostly focus on total functions in this section for simplicity.
Let us mention two applications of the above lifting theorem and its underlying ideas to regular (see the full version for details). In their proof of an exponential separation between regular and general Resolution, Bhattacharya, Chattopadhyay and Dvořák [11] required a randomized PDT lifting theorem and so they used a lifting theorem for randomized communication complexity [14]. This lifting theorem requires a logarithmic size gadget with a fairly large multiplicative constant which implies that the number of clauses in the resulting formula (and thereby the upper bound) is a large polynomial. By instead using the above lifting theorem for randomized PDTs with a constant size gadget, the lifted formula has the same number of clauses as the base formula (up to a constant factor) thereby improving the separation. We also show that ideas similar to those in the simulation can be used to improve the known lower bound for the bit pigeonhole principle in regular from [20] to .
As warm-up to proving the lifting theorem for randomized PDTs, we also consider the simpler question of which gadgets allow lifting from randomized DT depth to randomized DT size, that also does not seem to have been considered before. In the deterministic case, Alekseev, Filmus and Smal [2] (also [19]) showed that a resistant gadget allows lifting DT depth to DT size, where a gadget is resistant if fixing a single bit of the input cannot fix the value of the function. We observe that their ideas can also be used to prove the analogous statement for randomized decision trees.
Theorem 2.
For any function , resistant function ,
Theorems 1 and 2 are in fact slightly stronger than stated since they lift to rank, a measure which lower bounds size. This is also true of the deterministic PDT lifting theorem [16, 6] and the simulation-based proof of the deterministic DT size lifting theorem [2], though they do not explicitly mention it. PDT rank also lower bounds depth in subspace decision trees. A subspace decision tree is a decision tree where each internal node can query the indicator of an affine subspace.
Let us recall the definition of rank, introduced by Ehrenfeucht and Haussler [21]. It will be convenient to work with the following alternative way of looking at rank. Consider decision trees where at each node, one of the two outgoing edges is marked, and define the cost of such a marked decision tree to be the maximum number of marked edges along any root-to-leaf path. The rank of a function , , is the minimum cost of such a marked decision tree computing . Compared to size, rank is closer in spirit to depth, which better motivates some of the ideas discussed later.
The general question of whether randomized DT size satisfies a composition theorem ‘Does hold for all and up to polylog factors?’ was recently asked by Dahiya [18]. The corresponding question in the deterministic case has a positive answer, as shown by Dahiya and Mahajan [19].111Here we require , where is the input length for . Considering and shows that some such condition is necessary [2]. But this still leaves open the possibility that there are other functions with which allow lifting. We discuss this in more detail in the full version. They actually show that deterministic DT rank satisfies a composition theorem which implies the composition theorem for size.
The composition question is interesting even in the most basic setting of decision tree depth. While a composition theorem for deterministic depth has been known for long [38, 40, 32], the case of randomized depth is more subtle. It is still unknown whether we have for all total functions and . In fact, it is known that the statement is false in its most general form for the composition of a relation and a partial function [23] and even for the composition of partial functions [7]. There is a long line of work [1, 9, 4, 24, 23, 5, 7, 8, 13, 37] studying this question and proving lower bounds on of the form or where is some complexity measure.
Theorem 2 can be used to show that a composition theorem for rank implies one for depth, or, taking the contrapositive, counterexamples for depth also provide counterexamples for rank. Still, some of these composition theorems [9, 23, 8] can be adapted to give analogous composition theorems for randomized DT rank (see Appendix B in the full version).
Motivated by the work on the composition question for ordinary decision trees, we try to better understand the dependence on the inner function in the lower bounds on PDT rank for composed functions. [16] proved that for any -stifled , and the dependence on stifling in this lower bound cannot be improved since for some functions such as Indexing, is -stifled and .
We observe that the ideas in [16] also work with an adaptive version of stifling. To state this notion precisely, we consider the following task. Let be a Boolean function. Given query access to a certificate of , recognize whether is a -certificate or a -certificate. Let denote this problem. Now instead of counting all queries in the cost, only include the queries which evaluate to . The minimum number of ’s required to solve is denoted by . Using this notion, we show the following.
Theorem 3.
For all functions ,
This is inspired by discussion at the end of the talk [39], where it is shown that for the Inner Product gadget we can get linear dependence on the gadget size even though Inner Product is not -stifled. Observe that if is -stifled, then since the first queries made by an algorithm could all be and it has still not determined whether is a -certificate or a -certificate. Thus, the above lower bound is always at least as good as the one obtained from stifling. When is Inner Product or Disjointness on bits, is not -stifled but , which shows that this bound can sometimes be better.
We also prove a randomized analogue of Theorem 3.
Theorem 4.
For all functions ,
Here is the natural analogue of the linearized complexity measure introduced in [8] where an inner composition theorem was proved. For a function ,
where varies over randomized decision trees on , varies over inputs in the domain of , denotes the expected number of ’s seen when running on and .
Using this, we can show that when the inner function is Inner Product, Disjointness or Indexing, the upper bound for is the best one can do. For these inner functions, the deterministic PDT rank is equal to the randomized PDT rank (up to constants) so the upper bound is simply .
Corollary 5.
For , for all functions ,
While Theorems 3 and 4 do imply the lifting theorems mentioned earlier222Strictly speaking, Theorem 4 does not seem to directly imply Theorem 1 because of the additive constant loss in . However, one can use other measures for which an inner composition theorem holds, like sabotage complexity, in place of to recover Theorem 1., they still work in the standard basis. It is natural to consider analogues of the above measures which work with parity certificates instead of ordinary certificates, and indeed we can prove analogues of the above results using such measures. These lifting theorems can be found in the full version.
Theorems 3 and 4 can be viewed as improving the quantitative dependence on the gadget in the lifting theorems obtained via stifling. We can also consider the qualitative question of whether a more general property than stifling allows lifting to parity decision trees. In this direction, Alekseev, Filmus and Smal [2] completely classified gadgets based on what gadgets allow polynomial lifting, (for some ). However, this does not answer the question of which gadgets allow linear lifting, . We observe by considering a mild generalization of stifling that for any gadget whose minimum parity certificate complexity is at least , . This can be seen as the natural parity analogue of the statement that if has minimum certificate complexity at least , then for all , [2, 19].
The similarities go further. The class of gadgets which are not already captured by the above condition (possibly after first moving to a subfunction) are the ones which satisfy . If is a total function which cannot be computed by a single parity query and satisfies , we show that there is some gadget such that understanding whether allows lifting to PDT rank is equivalent to understanding whether allows lifting to DT rank. More precisely, we have the following.
Proposition 6.
Let be a total function which is not a parity. Then one of the following holds:
-
and for all functions , we have and
. -
and there exists some such that for all functions , we have and .
1.1 Techniques
The simulation for proving the lifting theorem for randomized PDTs with stifled gadgets builds on the ideas of [16]. Let us start by recalling their main idea for simulating a parity query efficiently. On an input for , we will simulate a parity decision tree for . For concreteness, suppose the parity at the root of is . While this parity depends on multiple blocks, we would like to simulate it while making just one query to . To do this, we localize the parity query to a single bit, say , which is informally thought of as being responsible for the value of the whole parity. Specifically, we view for some as fixing where we think of as still being free. At this point, we no longer have control of but since is stifled, we can fix the remaining bits in block to force . So we only need to make one actual query to simulate a parity query.
Now suppose we wish to simulate a randomized PDT. To ensure correctness, following the usual simulation framework for randomized communication protocols and decision trees, we now require suitable hard distributions and on the -inputs and -inputs respectively of the gadget . On an input for , we simulate a parity decision tree for on the distribution . Continuing with the above example, suppose we wish to simulate the parity on the distribution where is unknown. In general, simulating this parity by only querying one bit of seems hard but consider the following very special case. On querying , suppose we find that the corresponding distribution of block is such that is a uniform random bit which is independent of all the other bits, i.e. for some distribution on . Then irrespective of the distributions of and , we know that this parity is equally likely to be or and so we can just move to a uniform random child. The remaining bits in are set according to . So at least in this special case, we could simulate the parity with just one query.
To actually use the above observation in our simulation, we rely on the following idea of Bhattacharya, Chattopadhyay and Dvořák [11]. If is a stifling gadget of constant size , the uniform distributions on and have the following useful property. For any , with constant probability, the bits other than form a certificate of in which case the bit is uniformly distributed. Using this property, we can now simulate a parity query across multiple blocks by a constant number of queries in expectation since each time we make an actual query, with good probability, we will be able to simulate the current parity without making any more queries.
Next we briefly describe the ideas behind the PDT simulation theorems with improved dependence on the gadget. We focus here on the deterministic case; the randomized case follows a similar proof outline. For the deterministic PDT lifting theorem, we proceed in two steps. First, we observe that the proof in [16] implicitly uses a general reduction showing that for all functions, . In particular, does not need to be of composed form. This lower bound even works for relations , once we suitably define . In Appendix A of the full version, we use this lemma and other ideas in this work to give simple proofs of some known lower bounds for tree-like .
With the lower bound in hand, the next step is to simply note that satisfies a composition theorem analogous to usual decision tree complexity. Combining these gives Theorem 3. The simulation in [16] can be understood as instead using the relation whenever is -stifled, which can be seen as some analogue of .
1.2 Related work
In independent work, Podolskii and Shekhovtsov [34] have also proved a lifting theorem for randomized PDTs. In fact, they lift to the stronger model of semistructured communication protocols where one party is restricted to sending parities and the other can send arbitrary messages. The gadgets they allow are certain generalizations of Indexing where each index points to a distinct parity and their lifting theorem has the right dependence on the gadget size for such gadgets. This class naturally captures gadgets like Indexing and Inner Product, and by considering suitable reductions, their lifting theorem also applies to other gadgets. Some of the underlying ideas for simulating parities in their work are similar to ours, though it seems that our techniques do not directly imply their result for PDTs with the correct dependence on the gadget size and vice-versa.
Since we focus on the simpler model of PDTs, our proof of the lifting theorem using stifling gadgets is quite short, and we also provide a refined classification of gadgets for when lifting to PDTs is possible. It would be interesting to give a simulation unifying the lower bounds from [34] and our lower bounds for composed problems via -depth. We suspect that by combining techniques useful for composition in ordinary decision trees, with techniques used in query-to-communication lifting, it may be possible to find broader classes of gadgets which allow lifting to semi-structured protocols.
Besselman et al. [10] have recently obtained direct sum theorems for randomized PDT depth. They show that a direct sum theorem holds when the lower bound is proved via discrepancy or against product distributions. The direct sum question can be seen as the special case of composition where the outer function is the identity function. Our results using -depth (and its parity generalization) also give direct sum results for PDTs, where direct sum theorems hold for -depth by adapting the proofs for ordinary decision trees [30].
These direct sum results are incomparable to those of [10]. In one direction, for the Majority function, we have while any PDT solving MAJ on the uniform distribution requires depth . On the other hand, there are functions for which randomized -depth is polynomially larger than the PDT depth on product distributions. Such functions can be obtained by lifting such a separation for ordinary decision trees using a stifled gadget. The NAND tree provides such a separation for ordinary decision trees [37]. Similarly, in the deterministic case, the NAND tree also shows that deterministic -depth can sometimes be quadratically larger than (parity) certificate complexity.
1.3 Organization
2 Preliminaries
For a positive integer , we use to denote the set . All logs are to the base . We use to denote the Hamming weight of a string . Let be a relation. Let be a partial function on some domain (typically ). Then the composed relation is defined as follows. If for , there is some such that , then for all (in this case, we think of as lying outside the domain). Otherwise if and only if . For a relation and input , we sometimes use to denote the legal outputs on . A partial function is also sometimes interpreted as the relation where is related to if is in the domain of and otherwise is related to all possible outputs .
We use standard notation in most places to only represent universal constants and when required, we will explicitly note which parameters need to be large for the inequalities to hold. Additionally, if the constant depends on some parameter or function, this will be indicated by a subscript. We now define the query complexity measures used in this work. Refer to [12] for a survey about some of these measures.
Decision trees.
A deterministic decision tree on is a binary rooted tree with leaves being labeled from some set and internal nodes labeled by each with two outgoing edges labeled and . On an input , starting at the root, we repeatedly follow the edge according to the value of where is the label of the current node, until we reach a leaf. The label of the leaf is the output of on , which we denote by .
The cost of on , (or sometimes ), is the number of queries made by on . The depth of is defined by . The size of , denoted , is the number of leaves of .
The rank of is defined in the following way. Consider markings of the edges of such that one of the outgoing edges from each internal node is marked. For such a marked tree, the cost associated with this marking is the maximum number of marked edges on any root-to-leaf path. The rank of a tree is the minimum cost of any marking of . This definition of rank is equivalent to the more common bottom-up definition and the equivalence is proved in [17] but the idea also appears implicitly in prior work. For such a marked tree , we will use to denote the number of marked edges on the root-to-leaf path taken by .
For a relation , a decision tree is said to compute if for all , . Define to be the minimum depth of a deterministic decision tree computing . Define and to be the minimum size and rank respectively of a decision tree computing .
A randomized decision tree is a probability distribution over deterministic decision trees. We will use the same notation as in the deterministic case to denote the worst-case depth, size or rank of a randomized decision tree. We also use the corresponding expected measures described next. On input , we define . Define . Similarly and .
For a relation , a randomized decision tree is said to compute to error if for all , . We use to denote the worst case analogues of for randomized decision trees that compute to error . The corresponding expected measures are denoted by . We omit the subscript when dealing with error .
For any decision tree , . This directly implies for a randomized decision tree , by applying the above relation to each tree in the support of . To get the corresponding relation for the expected complexity measures, we use Jensen’s inequality to get , which in our notation is . This inequality does not depend on what queries are allowed and also holds for other kinds of decision trees (like parity decision trees).
A parity decision tree (PDT) is like a decision tree but the internal nodes can now query parities. We will denote a parity in different ways, for some or as where is the indicator vector for . The notation for parity decision trees is similar to that for (ordinary) decision trees. We will use in the superscript to denote the parity analogue of an ordinary query complexity measure. For example, denotes the minimum expected rank of a parity decision tree computing to error . When dealing with a parity on inputs with an underlying block structure , for , we use to denote the projection of onto the block.
We now define -depth and -depth. The -depth of an ordinary decision tree is the maximum number of edges labeled on any root-to-leaf path in . We use to denote the minimum -depth of a deterministic decision tree for , and similar notation with in the superscript for other -query complexity measures. The measures related to -depth are defined similarly.
We next mention two standard techniques for proving lower bounds on deterministic decision tree depth and rank. To prove lower bounds on for a relation , it suffices to give an Adversary strategy in the Querier-Adversary game for the relation . In this game, Adversary has a hidden string and Querier’s goal is to find some related to according to while making as few queries as possible. This technique is complete in the sense that if , then there is an Adversary strategy scoring points. This game also works for other deterministic query complexity measures by changing the kinds of queries Querier is allowed to make.
A similar game can be used to characterize the rank of a relation. In the Prover-Delayer game [35] for relation , similar to the Querier-Adversary game, Prover makes queries to a hidden string and Delayer responds by revealing the corresponding bits of , except for the following change. Delayer may instead choose to respond with , which is interpreted as Prover getting to decide how to fix the queried bit. Delayer gets to know what bit Prover picks in this case. The game continues in this manner until Prover can correctly output an which is related to . Delayer’s score is the number of ’s announced during the game. The maximum score guaranteed by a Delayer strategy for is equal to the rank of (see [19] for a proof).
The Prover-Delayer game can be equivalently described in a way closer to the Querier-Adversary game in the following way. Now instead of just picking some to query, Querier also picks a bit and Adversary gets a point only if the announced value is equal to . The best score achievable by an Adversary strategy in this game is equal to the best score of a Delayer strategy. This equivalent view can be seen as the natural game corresponding to the description of rank using marked decision trees.
By changing the allowed queries, the Prover-Delayer game can capture rank in other query models. For instance, rank in PDTs is captured by the parity Prover-Delayer game [29].
Certificate complexity.
A partial assignment is a string in . Say that is consistent with if for all , or . We will sometimes interpret a partial assignment as the subcube it defines, . So we write to express that is consistent with .
is said to be a certificate for a relation if there exists some , such that for all , . For a partial function and , a partial assignment is said to be a -certificate if for all , (interpreting as a relation). In other words, we require that for all , . The size of a certificate is the number of non-’s in it. The certificate complexity of a relation at , denoted , is defined as the minimum size of a certificate which is consistent with . The certificate complexity of , is the maximum certificate complexity of any input. The minimum certificate complexity of , , is the minimum size of a certificate for .
When working with a partial function , we will sometimes only be interested in certificates whose corresponding subcubes are completely contained in the domain of . We will call these domain certificates. We will drop the word domain when it is clear from context or when working with total functions.
A parity certificate for is given by a collection of linear equations on , such that there exists for which for all satisfying the equations in , we have . The size of a parity certificate is the number of equations in it. We may always assume that the linear forms involved in a parity certificate are linearly independent, since we can remove redundant equations without changing the defined affine subspace. The minimum parity certificate complexity of , , is the minimum size of a parity certificate for . A domain parity certificate of a partial function is a parity certificate for whose corresponding affine subspace is completely contained in the domain of .
3 Lifting theorems for randomized rank
In this section, we describe simple simulation theorems which lift randomized decision tree depth to rank in randomized decision trees and randomized parity decision trees.
3.1 Lifting to randomized ordinary decision tree rank
Alekseev, Filmus and Smal [2] showed that resistant gadgets suffice for lifting decision tree depth to size by generalizing Urquhart’s argument for the XOR gadget [41]. A function is said to be -resistant if .
This also follows from results of Dahiya and Mahajan [19] who show more generally that and .
We prove the following for randomized decision trees.
Theorem 8.
Suppose is -resistant. For any relation ,
By standard arguments, we get lifting in the worst case if we incur an additive loss in the error. This additive loss can be removed by standard amplification when the outer relation is a function and the error is a constant to get Theorem 2.
For the simulation, it will be convenient to work with an equivalent distributional description of resistant functions.
Definition 9 (balanced function).
A function is -balanced for if for every , there is a distribution supported on such that for each , for each , . A function is balanced if it is -balanced for some .
Note that a balanced function is necessarily resistant. It is also easy to see that being resistant is a sufficient condition for being balanced, as shown below.
Observation 10.
If is -resistant, then it is -balanced.
Proof.
For , the distribution witnessing that is -balanced is defined in the following way. Select a subset of of size uniformly at random. For each , pick uniformly at random (independently of other bits). Finally set all remaining bits so that the resulting string is a -input for . This last step can be performed by the assumption that is -resistant. Each is included in with probability and conditioned on being included in the first step, it is fixed to with probability .
We now show that any balanced gadget can be used for lifting to randomized decision tree rank. Using the distributions coming from the above observation, the simulation below is equivalent to applying a suitable random projection as in the second proof of the depth to size lifting theorem in [2]. However, it is analyzed differently for which presenting it as a simulation is more convenient.
Proposition 11.
Let be -balanced for some . Then for all relations ,
Proof.
We will show how to simulate a randomized decision tree computing with error probability by a randomized decision tree to compute with the same error probability. The expected depth of on any input will be at most .
Let and be the distributions on and respectively showing that is -balanced. For each , let be the distribution on defined by independently sampling for each , the block from . The simulation essentially samples , where is the input on which we wish to compute , and executes on . Since is unknown, the individual blocks of are sampled as they are queried by the decision tree . Since computes correctly for each with probability , the probability that outputs incorrectly on input is at most where the inner probability is over the randomness of . So computes to error .
We now describe in more detail. Since a randomized decision tree is a distribution over deterministic decision trees , it is enough to show how to simulate a deterministic decision tree while making at most queries in expectation. Suppose queries at the root. Then we query and sample . Now that is known, we move to the appropriate child. In the future, if makes any queries to bits of , we move according to the already sampled . We repeat this process until we reach a leaf at which point we output the label of the leaf reached. Note that this procedure also generates where since .
To estimate the number of queries made to , we will show next that starting at any node in during the simulation, the number of queries made until we cross a marked edge is at most in expectation. At any node, one of the outgoing edges is marked. By the assumption on , whenever a query is made, we follow the marked edge with probability at least . Thus, it takes at most queries in expectation to cross a marked edge. (During the simulation, we may reach a node where we directly move to the unmarked child with probability because the bit there had already been sampled earlier, but in this case no new query is made at that node.)
To get the total number of queries made in expectation when simulating , define for each , the random variable counting the number of queries made between crossing the marked edge and crossing the marked edge during the simulation. If we have reached a leaf of without crossing edges, then . Then we have for all by what was argued above.
Now the total number of queries made is since we must reach a leaf in after crossing at most many marked edges. By linearity of expectation The total number of queries made when simulating is at most .
3.2 Lifting to randomized parity decision tree rank
We now prove that stifled gadgets allow lifting to randomized parity decision tree rank.
Definition 12 (stifled functions).
A function is -stifled if for every subset of of size and each , there is a domain -certificate of which leaves free, i.e. for all .
Similar to the case of ordinary decision trees in the previous subsection, it will be convenient to work with an equivalent property arising from certain distributions on the -inputs and -inputs of the gadget. The following definition is a slight generalization of balanced functions considered in [11].
Definition 13 (affine balanced functions).
A function is -affine balanced if for each , there is a distribution supported on such that for each , there exist distributions on and on such that can be written as the mixture . Here is a uniform random bit independent of and we think of as the bit and as assigning bits in .
The above definition says we can sample from in the following way:
-
With probability , sample .
-
With probability , set uniformly at random, and independently .
Observation 14.
If is -stifled, then it is -affine balanced.
Proof.
The distribution witnessing that is -affine balanced is defined in the following way. Select a subset of of size uniformly at random. Fix the bits outside to a domain -certificate for (which can be done by the assumption that is -stifled). Finally for each , pick independently and uniformly at random. Each is included in with probability and conditioned on being included in the first step, it is fixed to each with probability independent of the other bits.
We now prove the lifting theorem for randomized PDTs. In the proof below, we do not give a truly online simulation but after each query, we simplify the PDT being simulated. This is primarily done to make it easy to verify correctness and analyze the number of queries made. We could alternatively have given a simulation closer to [16, 6] by keeping a list of parity queries made during the simulation.
Proposition 15.
Let be a -affine balanced function. For any relation ,
Proof.
Let be a randomized parity decision tree computing . For the induction below, it will be convenient to allow each parity query to also involve a constant term. Let and be distributions on and respectively showing that is -affine balanced. For each , define to be the distribution on defined by independently sampling for each , the block from . We will define a randomized decision tree computing by simulating on the distribution . The correctness of follows from the correctness of .
is defined in the following way. First sample a deterministic parity decision tree from . We simulate it by a randomized decision tree in the following way. Suppose the query at the root of involves a variable . Query the variable and suppose it was . We will now set the variables in the block according to the distribution in the following way. Recall that can be written as a mixture where is a distribution on strings in and is a distribution on domain -certificates that leave the bit free. We set as follows:
-
1.
With probability , set according to
-
2.
With the remaining probability , set bits according to the distribution and independently “set” for a uniform random bit .
In the second case, even though the parity may depend on variables from blocks other than , we may informally think of it equivalently as fixing . Note that irrespective of the distribution of blocks other than , it is indeed true that the above parity is equally likely to be or if we are the second case since is a uniform random bit (even after conditioning on all the other blocks) which appears in the parity. Additionally, note that after conditioning on and the other , the distribution on all blocks other than is still since block is independent of the rest.
Once we have set as above (where possibly is a linear form depending on other blocks) we substitute them in the tree and simplify appropriately. Specifically if any query node becomes a constant we remove it and directly attach the appropriate child to its parent. In particular, if we are in the second case, then the query at the root is set to a random and we move to that child. Note that this simplification preserves the action of the tree on the distribution when conditioned on the revealed .
Since the distribution on the other blocks stays , we can now repeat this process with the query at the new root (which may be the same as the previous one if we were in case one, with all variables from block removed). This is done until has become just a leaf and we give the same output in .
We now analyze the expected number of queries made by in simulating and show that it is at most on any input . We will show by induction that a PDT of rank at most which only depends on variables from at most blocks is simulated using at most queries in expectation.
Since a PDT with rank or which does not depend on any blocks is just a leaf, the statement holds whenever or . Suppose the statement holds for all pairs with . Let be a PDT of rank at most and depending on at most blocks. Suppose is the first variable queried by the above simulation because of some appearing in the query at the root.
Consider what happens after we simplify the tree based on the sampled . In all cases, the rank of the resulting tree, say is at most since the rank cannot increase by removing parts of the tree, and the number of blocks on which the tree depends has decreased by . Since case 2 while sampling happens with probability , with probability at least we go down the marked edge and the rank of is at most . Thus, the expected number of queries made in simulating is at most
| (by induction) |
Since the simulation of a randomized PDT corresponds to a distribution over trees simulating the deterministic PDTs , we get that on any input , the expected number of queries made is at most which is if we take an optimal RPDT for .
Remark 16.
As pointed out by a reviewer, we can alternatively get a bound on the number of queries made during the simulation in Proposition 15 in the following way. Each time an actual query is made, with probability at least , we go down the marked edge in the PDT being simulated, thereby contributing to the number of marked edges crossed. This implies that the expected number of marked edges crossed in the PDT when simulating it on the distribution is at least times the expected number of queries made which gives what we wanted. This argument also shows that instead of considering expected rank of the PDT we could have considered the expected cost on the worst-case input.
Combining Observation 14 and Proposition 15, we get the following lifting theorem for randomized PDTs.
Theorem 17.
Suppose is -stifled. For any relation ,
Remark 18.
The factor loss in the lower bound in Theorem 17 is necessary at least when we allow to be a partial function as the following example shows. We will take the outer function to be parity and the inner function to be approximate majority which is defined as follows: is if , if and otherwise. Note that here denotes the ends rather than the gap.
When and , the lower bound from Theorem 17 is just a constant. For this regime of parameters, there is a PDT computing which makes parity query. For each block , pick uniformly. For each , except with probability at most , we have . The PDT simply outputs the parity of . The error probability is at most by the union bound.
4 Parity decision tree lower bounds via -depth
4.1 Reduction to deterministic -depth and the Blocker-Certifier game
The proof of the lifting theorem for deterministic PDT size [16] implicitly contains a claim which reduces the task of proving lower bounds on PDT rank to the simpler task of proving lower bounds in a certain query model where one can only query one coordinate at a time but the input is a partial assignment instead of a binary string. This reduction works for all relations and, in particular, does not need the problem to be of composed form.
To describe the reduction, we need some definitions. Let be a relation. Define the relation as follows. For every ,
In words, is a correct output on if there is some consistent with for which is a correct output according to relation .
For a Boolean function , has the following simple description. The input to the partial function is promised to be a domain certificate of and the goal is to output whether it is a -certificate or a -certificate. Recall that for a partial function , for to be a domain certificate, we require that the subcube corresponding to is completely contained in the domain of .
Similar to -depth and -depth for ordinary decision trees, we may define -depth for decision trees on . For a relation , let be the smallest number of ’s any deterministic decision tree (which is only allowed to query an index at a time) computing must see in the worst case. Similarly let and be the analogous randomized query complexity measures when computing to error .
Since we mainly care about the -depth of relations of the form , we now introduce a game capturing , called the Blocker-Certifier game. This is essentially obtained by specializing the usual Querier-Adversary game corresponding to decision tree depth to our setting. However, since the score only depends on the number of ’s, we may allow the Adversary to fix positions to or before they are queried (similar to some Delayer strategies in the Prover-Delayer game) and then Querier only picks a coordinate to be fixed to . In the game below, Blocker’s role is similar to that of Querier (or Prover) and Certifier corresponds to an Adversary (or Delayer).
Let denote , the complement of . The Blocker-Certifier game for is played on a string which is initially . The game is played in rounds. In a round,
-
1.
Certifier picks a subset (possibly empty) and for each , sets for some .
-
2.
Blocker picks an such that and sets .
The game ends when we have the following situation. There is some , such that for every way of fixing the remaining ’s in to bits , there is a way to fix ’s in to bits such that the resulting string satisfies . In other words, the game has not ended if for every , Certifier can fix the remaining ’s to bits to get an -certificate for . Certifier’s score is the number of ’s in at the end of the game. The Blocker-Certifier value is the maximum score guaranteed by a Certifier strategy for the Blocker-Certifier game on . The equivalence between the Blocker-Certifier game and the usual Querier-Adversary game in this setting (or equivalently the definition of -depth) is proved in the full version.
We can now relate PDT rank for a relation and the -depth of . In the proof below, we use a Certifier strategy in the Blocker-Certifier game to give a parity Delayer strategy, but the argument can also be expressed as a simulation.
Lemma 19 (implicit in [16]).
For any relation ,
Proof.
Suppose we have a Certifier strategy for the Blocker-Certifier game on scoring points. We will use this to give a Delayer strategy for the parity Prover-Delayer game on achieving the same score.
Delayer essentially imitates the Certifier strategy by localizing the parity queries of Prover to the input so that they may be treated as positions that have been touched by Blocker in the Blocker-Certifier game. Delayer will have fixed some bits in to or while some positions would have been marked (denoted ) based on the queries made by Prover. Each such marked position corresponds to a linear equation coming from a parity query, where is marked and none of the positions in were marked at the time of the query.
We now explain this in detail. We will use to denote the string in the Blocker-Certifier game. will denote a collection of linear equations as explained above, which starts off empty. In the beginning, Delayer fixes bits in exactly according to the move made by Certifier at the start of the Blocker-Certifier game. On a parity query , Delayer first simplifies this parity query according to previously fixed bits to get a parity where and all variables in are still free. If , then Delayer simply responds with . Otherwise arbitrarily mark some and respond with . Suppose Prover responds with . Then the equation is added to .
Next in the Blocker-Certifier game, Blocker sets to to which Certifier responds by (possibly) fixing some other bits of . As before, Delayer fixes the variables in in the same way as . Later queries of the Prover are handled in essentially the same way as before, but the simplification now also has to remove any marked variables appearing the query by substituting suitable parities using the appropriate equations in .
We claim the Prover-Delayer game cannot end unless the corresponding Blocker-Certifier game is over. Suppose less than variables have been marked so far. For every , we will create an input consistent with all the parity queries made so far such that . Since fewer than variables are set to in the Blocker-Certifier game, there is a way to fix the remaining free variables in such that for all inputs consistent with the obtained partial assignment , . Now consider the string obtained by fixing all the marked bits according to the equations in . Since the marked variables are the pivots of these equations, such an extension indeed exists. By construction, this satisfies all the parity queries made so far and is consistent with the partial assignment . Thus Delayer can always score at least points.
To get a lifting theorem for PDT rank, the above lemma can now be combined with a lower bound on the Blocker-Certifier value for composed problems. First, note that since in the problem , we are only required to be correct when each block lies in the domain of , i.e. is a domain certificate of . To get the lifting theorem for any -stifled gadget , [16] use that . This inequality is in the same spirit as or . Similar to the case of decision tree depth or rank for composed problems, we can get an essentially tight lower bound on the -depth of composed problems.
Lemma 20.
For any relation and any function ,
This is proved in the same way as the usual composition theorem for deterministic (ordinary) decision tree depth [38, 40, 32]. For completeness, we sketch a proof of a more general composition theorem in Appendix B of the full version which implies Lemma 20.
Theorem 21.
For any relation and any function ,
The unique disjointness function is an example of a function for which the Blocker-Certifier value is much larger than how stifled it is. Recall that is the partial function such that where we are promised that there is at most one such that . This is a subfunction of both inner product and disjointness. It is easy to see that UDISJ is not -stifled, since there is no -certificate which leaves both and free.
On the other hand, there is a simple Certifier strategy in the Blocker-Certifier game333This strategy is essentially from discussion after the talk [39], but we have been unable to recognize who suggested it. for which achieves score . Initially Certifier does not fix any bits. Suppose Blocker sets (the case is analogous). Then Certifier responds by setting to ensure that . Certifier follows this strategy until the end of the game ensuring that for each , either both are unset or at least one of them is fixed to . We claim that the game cannot end before rounds. Indeed if fewer than rounds have taken place, then there is some such that . Moreover Certifier’s strategy ensures that wherever this is not the case, we have or . Therefore, for any , by setting and all other unset bits to , we obtain a domain -certificate for UDISJ which is consistent with all the moves made so far.
4.2 Reduction to randomized -depth
The following lemma is the randomized analogue of Lemma 19.
Lemma 22.
For any relation ,
Proof.
Let be a randomized PDT computing to error . We will give a randomized decision tree for computing with expected -depth at most . To do this, on any input , we will simulate on the distribution which is the uniform distribution over all strings in the subcube defined by .
By the definition of , makes an error only when makes an error on the input in being simulated. Therefore,
We now describe how simulates . First sample a deterministic PDT . The tree will keep track of a list of linear equations , one for each that has already been queried. The linear form on the right hand side of any such linear equation does not depend on any of the variables that have previously been queried. From this description, it is clear that these equations are linearly independent. Moreover, for each such , if , the corresponding equation in is exactly . We will additionally maintain the invariant that the system is equivalent to the system defined by all the parities from the root to the current node when combined with equations for all which have already been queried and are not .
Starting at the root of , the tree performs the following steps until a leaf is reached in .
-
1.
Let (for some ) be the query at the current node of . Iteratively perform substitutions using the equations in until the query has been simplified to which does not contain any variables that have already been queried and . Set which will later store which have already been queried.
-
2.
Repeat the following
-
Pick any and query . If , go to step 3(a). Otherwise add to , and to .
until all have been queried. When this happens, go to step 3(b).
-
-
3.
-
(a)
() Pick uniformly at random. Move to the child of corresponding to . Add to the equation .
-
(b)
(none of , is ) Since all ’s in the parity have been determined, move to the appropriate child .
-
(a)
The output is the same as the label of the leaf reached in .
Lemma 23 (proved later) shows that each leaf of is reached with the correct probability according to the distribution . As argued earlier, this implies the correctness of the tree.
We now analyze the expected -depth of the above randomized decision tree simulating . Note that in each round, the tree sees one if we reach step 3(a) and otherwise no ’s. We will keep track of the number of marked edges seen as a measure of progress. In step 3(b), the number of ’s seen has not changed this round. On the other hand, in step 3(a), since we move to a random child, with probability at least we move down the marked edge in . Thus, in expectation, after ’s, we move down a marked edge in . By linearity of expectation, after at most -queries in expectation, a leaf is reached since the maximum number of marked edges on any root to leaf path is .
Lemma 23.
Let be a deterministic PDT on . Let . Let be the event that node of is visited by the randomized procedure described in the proof of Lemma 22 when run on input . Let be the event that for a random , running on reaches . Then for every , we have .
Proof.
Fix . The proof is by induction on the depth of . The statement holds when is the root since in this case, .
Now suppose has depth at least . Let be its parent. If , then by induction and, therefore, . Hence, we may assume that . We can write and . So it suffices to prove that .
Let be the list of equations at the begin of the round where the current node is during the execution of the decision tree simulation with as the input string. Let be the set of all that were queried before reaching and which are not . Let be the parity query at and let be such that is the child corresponding to . So is the probability that for a random , conditioned on all the equations from the root to being satisfied. Note that since whenever , we may additionally condition on any subset of these fixed ’s being the corresponding ’s. By the invariants, the system of equations is equivalent to the system containing equations describing the parities from the root to as well as the queries made to so far. Therefore, by abuse of notation, we may express as where we view as the event that all equations in hold. Moreover under , the parity is equal to for some as in step 1 of the round. So we have .
Now we only need to verify that when simulating the query at node we go to with the correct probability . There are two cases to consider:
-
1.
For all , we have . In this case, all are queried and step 3(b) is executed in the round starting at . So we move to the correct child with probability .
-
2.
Step 3(a) is executed in the round starting at . In this case, there is some such that . Since is independent of by construction, .
This finishes the proof.
We now combine Lemma 22 with a composition theorem for randomized -depth. Recall that a tight composition theorem does not hold in general for ordinary decision trees when composing a relation with a partial function and so we cannot have such a statement for -depth (by lifting with, say, ). However, we can still adapt known randomized composition theorems to the setting of -depth. In Appendix B of the full version, we adapt the composition theorem of [8] to prove a composition theorem for a general class of decision trees. This composition theorem provides the best dependence on the inner function up to some loss by a constant multiplicative factor and an additive constant.
Lemma 24 (following [8]).
For any relation , any function
,
Theorem 25.
For any relation , any function ,
Using Theorem 25 or some other related composition theorem, we can show that, for instance, when the inner function is UDISJ or IND, then the obvious upper bound on PDT rank of is optimal. Note that both and are total functions extending . So the lower bound for UDISJ also implies the lower bounds for inner product and disjointness in Corollary 5.
Corollary 26.
For any relation , for any ,
Proving the lower bound for UDISJ will suffice to prove it also for IND since contains as a subfunction. By Theorem 25, it is sufficient to prove . Instead of showing this directly, we will show that the simpler quantity sabotage -complexity of is .
Sabotage complexity was first defined by Ben-David and Kothari [9] for ordinary decision trees and here we consider its natural -analogue. Sabotage complexity is the expected zero-error query complexity of the following task. Let be a partial function. Given a string , where we interpret as representing that the coordinate is free, find a in under the promise that is consistent with some -input and some -input . It can alternatively be characterized as [23, Theorem B.1], where varies over distributions on pairs in , varies over deterministic decision trees solving and denotes the number of marked edges (-queries) on the path from the root to the node where and separate. More precisely, is the unique node in such that both and reach but they disagree on the query made at node .
The composition theorem using sabotage complexity [9] is straightforward to adapt to -decision trees, , so we omit it.
Lemma 27.
.
Proof.
For brevity, let denote . The hard distribution is generated as follows. First sample in the following way. Pick uniformly. Set to or uniformly. For each , independently set to or uniformly. Finally, obtain by replacing the in by and by replacing the by . Observe that after conditioning on , the distribution on is exactly what we would get if we performed the above procedure for . This will let us use induction.
Let . We will show that by induction. The base case is clear. For the induction step, we start by making some simplifying assumptions about since we only care about the cost of on the distribution . Since our distribution is invariant under permuting blocks and permuting bits within a block, we may assume that the query at the root in is (we use to denote the queries in to avoid confusion with ). In the subtree where , for any query to , we remove it and directly attach its parent to the subtree where . We do the same with the roles of and interchanged. Note that this does not affect correctness of on since in any pair in the support of , if , then also , and similarly the other way around. Also the cost of does not increase by performing this simplification.
By the observation above, since the distribution conditioned on is identical to , the subtrees where and give trees solving the separation task on the distribution . Therefore, we have the recurrence,
which by induction gives .
Proof of Corollary 26.
The upper bounds follow from simulating a randomized decision tree for by using a deterministic tree for the inner function at each node of .
For the lower bounds, as stated earlier, a lower bound for also implies the same lower bound for . So we only need to show the lower bound for . By combining and Lemma 27, we get . Now using this with Lemma 22, we get .
Remark 28.
The simulation using stifling gadgets, Theorem 17, can be understood as using the fact that for a -stifled function , . Indeed, if we couple the distributions of certificates underlying and used in that proof according to the set of coordinates that are fixed, any decision tree correctly computing must see a on the first query with probability .
References
- [1] Scott Aaronson, Shalev Ben-David, and Robin Kothari. Separations in query complexity using cheat sheets. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, pages 863–876, 2016. doi:10.1145/2897518.2897644.
- [2] Yaroslav Alekseev, Yuval Filmus, and Alexander Smal. Lifting Dichotomies. In Rahul Santhanam, editor, 39th Computational Complexity Conference (CCC 2024), volume 300 of Leibniz International Proceedings in Informatics (LIPIcs), pages 9:1–9:18, Dagstuhl, Germany, 2024. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.CCC.2024.9.
- [3] Yaroslav Alekseev and Dmitry Itsykson. Lifting to bounded-depth and regular resolutions over parities via games. In Proceedings of the 57th Annual ACM Symposium on Theory of Computing, STOC ’25, pages 584–595, New York, NY, USA, 2025. Association for Computing Machinery. doi:10.1145/3717823.3718150.
- [4] Anurag Anshu, Dmitry Gavinsky, Rahul Jain, Srijita Kundu, Troy Lee, Priyanka Mukhopadhyay, Miklos Santha, and Swagato Sanyal. A composition theorem for randomized query complexity. In 37th IARCS Annual Conference on Foundations of Software Technology and Theoretical Computer Science (FSTTCS 2017). Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2018. doi:10.4230/LIPIcs.FSTTCS.2017.10.
- [5] Andrew Bassilakis, Andrew Drucker, Mika Göös, Lunjia Hu, Weiyun Ma, and Li-Yang Tan. The power of many samples in query complexity. In 47th International Colloquium on Automata, Languages, and Programming (ICALP 2020). Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2020. doi:10.4230/LIPIcs.ICALP.2020.9.
- [6] Paul Beame and Sajin Koroth. On Disperser/Lifting Properties of the Index and Inner-Product Functions. In Yael Tauman Kalai, editor, 14th Innovations in Theoretical Computer Science Conference (ITCS 2023), volume 251 of Leibniz International Proceedings in Informatics (LIPIcs), pages 14:1–14:17, Dagstuhl, Germany, 2023. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ITCS.2023.14.
- [7] Shalev Ben-David and Eric Blais. A new minimax theorem for randomized algorithms. In 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), pages 403–411. IEEE, 2020.
- [8] Shalev Ben-David, Eric Blais, Mika Göös, and Gilbert Maystre. Randomised composition and small-bias minimax. In 2022 IEEE 63rd Annual Symposium on Foundations of Computer Science (FOCS), pages 624–635. IEEE, 2022. doi:10.1109/FOCS54457.2022.00065.
- [9] Shalev Ben-David and Robin Kothari. Randomized query complexity of sabotaged and composed functions. Theory of Computing, 14(5):1–27, 2018. doi:10.4086/toc.2018.v014a005.
- [10] Tyler Besselman, Mika Göös, Siyao Guo, Gilbert Maystre, and Weiqiang Yuan. Direct sums for parity decision trees. In 40th Computational Complexity Conference (CCC 2025), Leibniz International Proceedings in Informatics (LIPIcs), Dagstuhl, Germany, 2025. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.CCC.2025.16.
- [11] Sreejata Kishor Bhattacharya, Arkadev Chattopadhyay, and Pavel Dvořák. Exponential Separation Between Powers of Regular and General Resolution over Parities. In Rahul Santhanam, editor, 39th Computational Complexity Conference (CCC 2024), volume 300 of Leibniz International Proceedings in Informatics (LIPIcs), pages 23:1–23:32, Dagstuhl, Germany, 2024. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.CCC.2024.23.
- [12] Harry Buhrman and Ronald de Wolf. Complexity measures and decision tree complexity: a survey. Theoretical Computer Science, 288(1):21–43, 2002. doi:10.1016/S0304-3975(01)00144-X.
- [13] Sourav Chakraborty, Chandrima Kayal, Rajat Mittal, Manaswi Paraashar, Swagato Sanyal, and Nitin Saurabh. On the composition of randomized query complexity and approximate degree. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2023). Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2023. doi:10.4230/LIPIcs.APPROX/RANDOM.2023.63.
- [14] Arkadev Chattopadhyay, Yuval Filmus, Sajin Koroth, Or Meir, and Toniann Pitassi. Query-to-communication lifting using low-discrepancy gadgets. SIAM Journal on Computing, 50(1):171–210, 2021. doi:10.1137/19M1310153.
- [15] Arkadev Chattopadhyay, Michal Kouckỳ, Bruno Loff, and Sagnik Mukhopadhyay. Simulation theorems via pseudo-random properties. computational complexity, 28:617–659, 2019. doi:10.1007/S00037-019-00190-7.
- [16] Arkadev Chattopadhyay, Nikhil S Mande, Swagato Sanyal, and Suhail Sherif. Lifting to parity decision trees via stifling. In 14th Innovations in Theoretical Computer Science Conference (ITCS 2023). Schloss Dagstuhl – Leibniz Zentrum für Informatik, 2023. doi:10.4230/LIPIcs.ITCS.2023.33.
- [17] Arjan Cornelissen, Nikhil S Mande, and Subhasree Patro. Improved quantum query upper bounds based on classical decision trees. In 42nd IARCS Annual Conference on Foundations of Software Technology and Theoretical Computer Science (FSTTCS 2022). Schloss Dagstuhl – Leibniz Zentrum für Informatik, 2022. doi:10.4230/LIPIcs.FSTTCS.2022.15.
- [18] Yogesh Dahiya. Exploring Size Complexity and Randomness in the Query Model. PhD thesis, HBNI, 2024. URL: https://www.imsc.res.in/xmlui/handle/123456789/881.
- [19] Yogesh Dahiya and Meena Mahajan. On (simple) decision tree rank. Theoretical Computer Science, 978:114177, 2023. doi:10.1016/J.TCS.2023.114177.
- [20] Klim Efremenko, Michal Garlík, and Dmitry Itsykson. Lower bounds for regular resolution over parities. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, pages 640–651, 2024. doi:10.1145/3618260.3649652.
- [21] Andrzej Ehrenfeucht and David Haussler. Learning decision trees from random examples. Information and Computation, 82(3):231–246, 1989. doi:10.1016/0890-5401(89)90001-1.
- [22] Ankit Garg, Mika Göös, Pritish Kamath, and Dmitry Sokolov. Monotone circuit lower bounds from resolution. In Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, pages 902–911, 2018. doi:10.1145/3188745.3188838.
- [23] Dmytro Gavinsky, Troy Lee, Miklos Santha, and Swagato Sanyal. Optimal composition theorem for randomized query complexity. Theory of Computing, 19(1):1–35, 2023. doi:10.4086/TOC.2023.V019A009.
- [24] Mika Göös, TS Jayram, Toniann Pitassi, and Thomas Watson. Randomized communication versus partition number. ACM Transactions on Computation Theory (TOCT), 10(1):1–20, 2018. doi:10.1145/3170711.
- [25] Mika Göös, Pritish Kamath, Toniann Pitassi, and Thomas Watson. Query-to-communication lifting for p np. computational complexity, 28:113–144, 2019. doi:10.1007/S00037-018-0175-5.
- [26] Mika Göös and Toniann Pitassi. Communication lower bounds via critical block sensitivity. SIAM Journal on Computing, 47(5):1778–1806, 2018. doi:10.1137/16M1082007.
- [27] Mika Goos, Toniann Pitassi, and Thomas Watson. Deterministic communication vs. partition number. SIAM Journal on Computing, 47(6):2435–2450, 2018. doi:10.1137/16M1059369.
- [28] Mika Göös, Toniann Pitassi, and Thomas Watson. Query-to-communication lifting for bpp. SIAM Journal on Computing, 49(4):FOCS17–441, 2020. doi:10.1137/17M115339X.
- [29] Dmitry Itsykson and Dmitry Sokolov. Resolution over linear equations modulo two. Annals of Pure and Applied Logic, 171(1):102722, 2020. doi:10.1016/J.APAL.2019.102722.
- [30] Rahul Jain, Hartmut Klauck, and Miklos Santha. Optimal direct sum results for deterministic and randomized decision tree complexity. Information Processing Letters, 110(20):893–897, 2010. doi:10.1016/J.IPL.2010.07.020.
- [31] Shachar Lovett, Raghu Meka, Ian Mertz, Toniann Pitassi, and Jiapeng Zhang. Lifting with Sunflowers. In Mark Braverman, editor, 13th Innovations in Theoretical Computer Science Conference (ITCS 2022), volume 215 of Leibniz International Proceedings in Informatics (LIPIcs), pages 104:1–104:24, Dagstuhl, Germany, 2022. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ITCS.2022.104.
- [32] Ashley Montanaro. A composition theorem for decision tree complexity. Chicago Journal of Theoretical Computer Science, 2014(6), 2014. doi:10.4086/cjtcs.2014.006.
- [33] Toniann Pitassi and Robert Robere. Lifting nullstellensatz to monotone span programs over any field. In Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, pages 1207–1219, 2018. doi:10.1145/3188745.3188914.
- [34] Vladimir Podolskii and Alexander Shekhovtsov. Randomized lifting to semi-structured communication complexity via linear diversity. In 16th Innovations in Theoretical Computer Science Conference (ITCS), LIPIcs. Schloss Dagstuhl, 2025. doi:10.4230/LIPIcs.ITCS.2025.78.
- [35] Pavel Pudlák and Russell Impagliazzo. A lower bound for dll algorithms for k-sat (preliminary version). In Proceedings of the eleventh annual ACM-SIAM symposium on Discrete algorithms, pages 128–136, 2000. URL: http://dl.acm.org/citation.cfm?id=338219.338244.
- [36] Ran Raz and Pierre McKenzie. Separation of the monotone nc hierarchy. In Proceedings 38th Annual Symposium on Foundations of Computer Science, pages 234–243. IEEE, 1997. doi:10.1109/SFCS.1997.646112.
- [37] Swagato Sanyal. Randomized query composition and product distributions. In 41st International Symposium on Theoretical Aspects of Computer Science (STACS 2024). Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024. doi:10.4230/LIPIcs.STACS.2024.56.
- [38] Petr Savický. On determinism versus unambiquous nondeterminism for decision trees. Technical Report TR02-009, Electronic Colloquium on Computational Complexity (ECCC), 2002. URL: https://eccc.weizmann.ac.il//report/2002/009/.
- [39] Suhail Sherif. Lifting to parity decision trees via stifling (with applications to proof complexity). URL: https://www.youtube.com/watch?v=PeZVs6WUf-4.
- [40] Avishay Tal. Properties and applications of boolean function composition. In Proceedings of the 4th Conference on Innovations in Theoretical Computer Science (ITCS), pages 441–454, 2013. doi:10.1145/2422436.2422485.
- [41] Alasdair Urquhart. The depth of resolution proofs. Studia Logica, 99:349–364, 2011. doi:10.1007/S11225-011-9356-9.
