Doubly Sub-Linear Interactive Proofs of Proximity
Abstract
We initiate a study of doubly-efficient interactive proofs of proximity, while focusing on properties that can be tested within query-complexity that is significantly sub-linear, and seeking interactive proofs of proximity in which
-
1.
The query-complexity of verification is significantly smaller than the query-complexity of testing.
-
2.
The query-complexity of the honest prover strategy is not much larger than the query-complexity of testing.
We call such proof systems doubly-sublinear IPPs (dsIPPs).
We present a few doubly-sublinear IPPs. A salient feature of these IPPs is that the honest prover does not employ an optimal strategy (i.e. a strategy that maximizes the verifier’s acceptance probability). In particular, the honest prover in our IPP for sets recognizable by constant-width read-once oblivious branching programs uses a distance-approximator for such sets.
Keywords and phrases:
Interactive Proof Systems, Interactive Proofs of Proximity, Query Complexity, Read Once Branching Programs, Sub-linearCopyright and License:
2012 ACM Subject Classification:
Theory of computation Interactive proof systemsEditors:
Raghu MekaSeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
This work combines the mind-frames of interactive proofs of proximity (i.e., IPPs) and doubly-efficient interactive proofs (de-IPs), while giving the notion of doubly-efficient a new meaning that focuses on the query-complexity of the honest prover strategy. Specifically, we focus on properties that can be tested by reading a small portion of the input (equiv., have query-complexity that is significantly sub-linear), and seek interactive proofs of proximity in which
-
1.
The query-complexity of verification is significantly smaller than the query-complexity of testing.
-
2.
The query-complexity of the honest prover strategy is not much larger than the query-complexity of testing.
In addition, we may seek analogous relations for the time-complexities; yet, we recall that, in the context of property testing, time-complexity is secondary to query-complexity (see [4, Sec. 1.3.1]).
A salient feature of (almost all) the IPPs that we present is that the honest prover does not employ an optimal strategy, because it cannot afford to read the entire input (per the second query-complexity condition). In particular, while an optimal prover strategy for the verifiers that we present achieves perfect completeness, our honest provers don’t.
1.1 Our focus: query-efficient generation of proofs of proximity
Focusing on properties that can be tested in sub-linear query-complexity, we initiate a study of interactive proofs of proximity in which verification requires less queries than testing whereas proving does not require much more queries than testing. We recall the wider context first.
Property testing.
The focus of property testing[6, 22] is on approximate decision procedures that read small portion of their input (see [4] for an introduction to the subject). For a predetermined property (i.e., a set of valid objects), approximate decision means distinguishing between objects that have the property and objects that are far from any object having this property.111Distances are measured according to the relative Hamming measure; that is, is considered -far from if and differ on more than locations. Such procedures, called testers, are probabilistic and obtain local views of the object by performing queries; that is, the object is seen as a function and the testers get oracle access to this function (and thus may be expected to read only part of the object).
The insistence on procedures that read a small portion of the input reflects envisioning applications in which the input is huge and fetching it entirely is infeasible. In some of these applications, one may afford computation time that is almost linear in the size of the input, although fetching the entire input is still deemed infeasible.
Proofs.
A generic question is whether proofs can offer verification that is more efficient than the corresponding decision. In the context of property testing, this refers to interactive proofs of proximity (IPPs)[3, 21]. In such proof systems, verification should require significantly less queries than testing, and suitable notions of completeness and soundness should hold.222In addition, one requires low communication complexity. In particular, this does not allow the prover to send the entire input to the verifier (a possibility that is not relevant in the context of double-efficiency that we consider here). Specifically, completeness means that inputs that have the property should be accepted with high probability (when the prover uses an adequate strategy), whereas soundness means that inputs that are far from the property should be rejected with high probability (no matter what strategy is employed by the prover). Indeed, in the completeness condition we consider a honest prover, whereas in the soundness condition we consider a cheating one.
Doubly-efficient proofs.
Seeking to utilize proof systems in reality prohibits the use of arbitrary honest provers. Such applications require that the honest prover strategy be relatively efficient, a requirement that is captured by the term double-efficiency, which refers to the complexities of both the verifier and the honest prover. Given our postulate that it is infeasible to fetch the entire input and our focus on query complexity, this means that the honest prover strategy should also have query complexity that is significantly smaller than linear. We call such proof systems doubly-sublinear IPPs (dsIPPs).
Note that dsIPPs may exist only for properties that can be tested using small query complexity. This is the case because the interaction between the verifier and the honest prover (along with the queries they make) can be emulated by the tester. Hence, we focus on properties that can be tested using small query complexity and ask for which of these properties we can obtain dsIPPs?
We stress that the notion of double-efficiency employed here is different from the notion employed in the context of IPs (see[10, 19, 5]) and also in studies of IPPs that made reference to doubly-efficient IPPs (e.g.[20]). In these prior cases, the reference was to polynomial running time of the honest prover, whereas we focus on its query complexity and on the case that it is small (and in particular is sub-linear).
1.2 Our Main Results
We construct doubly-sublinear IPPs (dsIPPs) for several problems. Specifically, for any property that can be decided by a constant-width read-once oblivious branching program (ROOBP), for approximating the input’s Hamming weight, for telling whether a given function is a permutation, and for a relaxation of the graph bipartiteness problem in the bounded-degree model.
As mentioned earlier, the honest prover strategies in (almost all) the dsIPPs that we present do not employ an optimal strategy, because they cannot afford to read the entire input. In particular, while optimal prover strategies for these IPPs achieves perfect completeness (i.e., the verifier always accepts inputs that have the property), our honest provers do not achieve perfect completeness.
1.2.1 Protocol for Read-Once Oblivious Branching Programs
We construct a dsIPP for any property that can be decided by constant-width read-once oblivious branching programs (ROOBPs). Recall that in such branching programs, in each layer, all vertices are labelled by the same input variable (which is the one being read), and each input variable labels the vertices of at most one layer. A branching program decides a property if it accepts all inputs in the property and rejects all inputs that are not in the property. See Section 4.1 for a fuller discussion on ROOBPs.
Newman [16] showed that any property that can be decided by a constant-width ROOBP has a tester that makes queries, where the exponent of the polynomial is linear in the width of the branching program. We construct a dsIPP for any such property, where the verifier’s query complexity is only , which is optimal.333Consider the set property , the input and a honest prover strategy that refers to the input but is also invoked as a cheating strategy on a random input that is at Hamming distance from . Then, an -query verifier cannot distringuish the two cases.
Theorem 1 (dsIPP for ROOBPs, informal).
Let be a property that can be decided by a constant-width ROOBP, and let and denote the input length and the proximity parameter. Then, for every , there exists an -round dsIPP for such that the verifier has query complexity , the honest prover has query complexity , and the communication complexity is .
See Corollary 20 for a full and formal statement. Evidently, our protocol allows for a trade-off between the number of rounds (on one hand) and the query complexity of the prover and the communication (on the other hand). In particular, a logarithmic number of rounds suffices for getting the prover’s query complexity and the communication to depend only poly-logarithmically on . The verifier’s runtime can be linear in the communication complexity assuming access to a suitable procedure specifying the structure of the ROOBP (see below).444We remark that for a general non-uniform ROOBP no verifier can have a sublinear running time: for example, one layer in the ROOBP might make it go into a state that rejects all inputs. Without reading the entire ROOBP, the verifier has no hope of detecting whether there is a “universal rejection” layer. The prover’s runtime depends on the computational complexity of tolerant testing for properties decided by ROOBPs, see below.
Our protocol builds on an IPP of Goldreich, Gur and Rothblum[7], in which the honest prover reads the entire input. On the other hand, the IPP of[7] works also for unbounded width (but its communication complexity grows logarithmically with the width). We note that our dsIPP is actually tolerant (see below), whereas this was not the case for the protocol of [7] (or the tester of [16], see below).
A reduction to tolerant testing.
An interesting feature of our IPP is that the honest prover strategy is reduced to tolerant testing (or distance approximation) for a property that can be decided by an ROOBP of (at most) the same width and length. A tolerant tester [17] for a property gets proximity parameters and is supposed to accept (w.h.p.) if the input is -close to the property, and to reject (w.h.p.) if the input is -far from the property. The complexity is usually measured as a function of the gap (and possibly also of the input length).
Theorem 1 follows by plugging any tolerant tester for constant-width ROOBPs into our general reduction (which is stated in Theorem 2). While, as noted above, Newman’s tester for ROOBPs[16] is not tolerant, we show that that it can be made tolerant. This contribution of the current work is presented in the full version of the article.
Our reduction of the prover’s strategy to tolerant testing ROOBPs actually gives a tolerant dsIPP of ROOBPs, where Tolerant IPPs are defined analogously to tolerant testers.555Specifically, if the input is -close to the property, then the honest prover should convince the verifier to accept (w.h.p.); but if the input is -far from the property, then the verifier should reject (w.h.p.) regardless of the prover’s strategy. We stress that, while in the context of testing, tolerant testing and distance approximation are equivalent (under suitable parameters), this is not the case for dsIPPs (see Section 2 and Remark 8). Fixing the ROOBP under consideration, the reduction assumes that the honest prover has access to a distance approximator that gets an ROOBP of (at most) the specified width and length, and approximates the distance of an input from the corresponding property (i.e., the property decided by the ROOBP). We need the deviation of the distance approximation to be , and its error probability to be sufficiently small. The approximator only needs to operate on “sub-ROOBPs” of the original ROOBP.
Before stating the reduction, we also specify the access the verifier and the prover need to the ROOBP in order to get sublinear runtimes. We assume that the prover and the verifier have (unit cost) access to two procedures that specify the structure of the ROOBP as follows. The first procedure, given two nodes and in the ROOBP, returns a bit indicating whether there is a path from to . The second procedure gets as input the index of a layer and returns the vertices in that layer. (We emphasize that these procedures do not depend on the input to the ROOBP: they ignore edge labels and refer to the ROOBP as a directed graph.)
Theorem 2 (Reducing tolerant dsIPP to distance approximation for ROOBPs, informal).
For any constant , let be a property that can be decided by a width ROOBP, and let denote the proximity parameters. Suppose that there exists a distance approximator for -width ROOBP (and its subgraphs) with query-complexity and time-complexity , when denotes the approximation parameter, denotes the error probability bound, and the ROOBP has length . Then, for every , there exists an -round tolerant dsIPP for such that
-
The verifier’s query-complexity is , and the honest prover’s query-complexity is , where and .
-
The communication complexity is .
-
Assuming that the verifier and the prover have access to procedures specifying the structure of the ROOBP (see above), the verifier’s runtime equals the communication complexity (up to constant factors), and the honest prover’s runtime is .
See Theorem 19 for a full and formal statement. As discussed above, Theorem 1 is obtained from the above reduction by plugging in a distance approximator for constant-width ROOBPs, which we construct based on Newman’s (non-tolerant) property tester. While the reduction of Theorem 2 only has polynomial dependence on the width of the ROOBP, Newman’s tester and our tolerant version of it have an exponential dependence on the width of the ROOBP (which is the reason that Theorem 1 holds only for constant width).666In contrast, Theorem 2 holds also for varying width, but in that case the complexities have a multiplicative factor of . See the details in Section 4.6.
1.2.2 Protocol for Hamming Weight
We construct a dsIPP for the Hamming weight (HW) problem, where the input is a string and a claim about its Hamming weight. Note that Hamming weight cannot be computed by constant-width ROOBP [18], so the results of Section 1.2.1 do not give a dsIPP for HW.
Approximating the (relative) Hamming weight up to distance requires queries in the standalone setting. We construct a dsIPP with verifier queries (which is optimal).
Theorem 3 (dsIPP for Hamming weight, informal).
Let and denote the input length and the proximity parameter for the Hamming weight problem (HW). Then, for every , there exists an -round dsIPP for HW such that the verifier has query complexity and the honest prover’s query complexity and its runtime are both . Furthermore, the verifier’s runtime and the communication complexity are both .
See Theorem 10 for a full and formal statement. This protocol also obtains a trade-off between the number of rounds and the honest prover’s runtime. Taking the number of rounds to be logarithmic in gives a honest prover with query complexity and runtime that are poly-logarithmic in . The honest prover’s query complexity and runtime grow cubically with (the reciprocal of) the proximity parameter. We wonder whether this dependence can be improved to quadratic (which would be optimal [1]).
Rothblum, Vadhan and Wigderson [21] also constructed an IPP for Hamming weight, but did not have a sub-linear honest prover. Our HW protocol borrows ideas from the IPP for ROOBP of[7]. Indeed, one can get an IPP for HW directly from their ROOBP protocol, by applying it to the -width ROOBP that computes HW. However, their protocol does not have a sub-linear prover.
1.2.3 Protocol for PERM
For , the set PERM consists of all permutations over . Here we consider the problem of testing (resp., verifying) whether a function is in PERM. Note that PERM has a tester of complexity , which we outline below, whereas a query lower bound of follows from[13].777See also a direct proof in[23, Apdx A]. Furthermore, PERM has two different (1-round) IPPs, which were presented in[13] and[7], respectively. We review them next:
-
1.
In the IPP of[13] the verifier selects uniformly at random a point , and sends to the prover, who is supposed to return its -preimage; the verifier accepts if and only if the prover’s answer is mapped by to .
-
2.
In the IPP of[7], the verifier selects uniformly at random a point , queries on it, and sends to the prover, who is supposed to return its -preimage; the verifier accepts if and only if the prover’s answer equals .
(This IPP, unlike the previous one, utilizes prover-oblivious queries.)
In both cases, is always accepted, whereas functions that are -far from PERM are rejected with probability . (In both cases, the protocol is repeated times to yield an IPP.)
More importantly, in both cases, the honest prover finds the required -preimage by querying on all points (or practically so). Our initial conjecture was that an IPP for PERM cannot have a verifier that uses queries and an honest prover that makes queries. Interestingly, this “working conjecture” is wrong.
Theorem 4 (dsIPP for PERM).
For every , there exists a 1-round IPP for PERM that has a verifier that uses queries and an honest prover that uses queries.
The communication and time complexity of both parties is .
Sketch of the proof of Theorem 4.
The straightforward tester for PERM consists of selecting random points in , querying the function on them, and rejecting if and only if collisions are found (among the -images).
Evidently, any is accepted with probability 1, whereas (as can be seen later) any that is -far from PERM is rejected with high constant probability. A similar analysis for the soundness case implies that for any that is -far from PERM, if we select random points, then we expect to see collisions and that these collisions involve disjoint pairs of points. This leads us to the following protocol.
-
1.
The verifier selects (distinct) random points in , denoted , and sends them to the prover,
-
2.
The prover queries the input on these sample points, and sends the answers to the verifier.
-
3.
The verifier rejects if it sees a collision (i.e., if for some ).
-
4.
Otherwise, it sub-samples of the original points, queries on each of these samples, and accept if and only if all answers match the prover’s answers (i.e., if for the ’s it sub-sampled).
Clearly, is always accepted. Turning to the soundness analysis, we fix an arbitrary function that is -far from PERM and let denote the set of points that form collisions under . Then, ; actually, , because modifying to a permutation requires changing its value on all but a single point in for every .
An analogous consideration applies to the sample of points, denoted , that is sent to the prover. Specifically, the answers provided by the prover must disagree with the values of on at least points, because for each the prover must provide different answers to all the other points in that are in (since otherwise the verifier rejects in Step 2). Towards upper-bounding , we observe that it is upper-bounded by , where is defined such that for every it holds that (i) whenever and (ii) for each that forms a collision under , the restriction of to is 2-to-1 except on at most one element (in case is odd). Now, we lower-bound by the number of collisions of under (i.e., )
First note that the number of collisions of under is at least , which is at least . Hence, with high probability over the choice of , we get collisions of under . In this case, there are disjoint pairs of points in such that the elements in each pair have the same -image (and hence also the same -image).
Wishing to avoid rejection in Step 2, the prover must cheat on the values of the foregoing points (in ), but (with high probability) at least one of these points will appear in the sub-sample that the verifier selects in Step 4 (since the sub-sample rate is ). Hence, in Step 4, when quering the function on this sub-sample, the verifier detects this cheating and rejects (w.h.p.). This completes the proof of Theorem 4.
1.2.4 Protocol for Bipartiteness
We construct a dsIPP for a relaxation of graph bipartiteness in the bounded degree model. In this model (see, e.g.,[4, Sec. 9.1]), the input is an undirected -vertex graph of constant maximum degree . The input graph is represented by its incidence function such that is the -th neighbor of the vertex (or if has less than than neighbors). The prover and the verifier have query access to this function. The distance between two -vertex graphs is the ratio (over ) of the number of edges on whose presence or absence they disagree.
The promise problem we consider is that the input graphs are rapidly-mixing graphs; that is, graphs in which a random lazy walk of logarithmic length starting at any vertex reaches any other vertex with probability . (An -step lazy walk is described by a -long sequence over such that indicates a step that moves from the current vertex to its neighbor (in case has at least neighbors) and staying in place if has less than neighbors.)
Theorem 5 (dsIPP for bipartiteness).
Let and denote the number of vertices and the proximity parameter. Then, there exists a 1-round dsIPP for bipartiteness on rapidly-mixing graphs in the bounded-degree graph model in which the verifier’s query complexity is , and the honest prover’s query and time complexity are . Furthermore, the verifier’s runtime is , and the communication complexity is .
Related work.
Goldreich and Ron [9] showed a lower bound on the query complexity of testing bipartiteness.888Their lower bound, which is stated for general 3-regular graphs, also holds under the additional promise that the graphs are rapidly-mixing (see the exposition in [4, Sec. 9.3.1]). In particular, the distribution of NO cases is concentrated on expander graphs (see [4, Clm. 9.18.1]). A similar analysis shows that the YES cases are bipartite expanders (w.h.p.) A later work of the same authors[8] shows a tester with query complexity. We remark that their tester works for general graphs, without needing the rapidly-mixing condition. Rothblum, Vadhan and Wigderson [21] constructed an IPP for an intermediate relaxation of bipartiteness, where the YES case includes all the bipartite graphs, but the NO case includes only the graphs that are both far from bipartite and rapidly-mixing. In their IPP, the verifier’s query complexity is , but the honest prover is not sub-linear: it has to read the entire graph. Our proof of Theorem 5 uses the [21] protocol, but shows a sublinear-time implementation for the honest prover, where the implementation works for graphs that are both rapidly-mixing and bipartite.
Proof of Theorem 5: The sublinear-time honest prover.
We first recall the[21] protocol. In the basic test, the verifier picks a random vertex , and performs a lazy random walk of length , which is long enough to guarantee the “rapidly-mixing” condition (see above), reaching a vertex . The verifier sends and to the prover, asks the prover to recover the parity of a (simple) path that leads from to , and accepts if and only if the answer equals the parity of the number of “real moves” in the lazy walk (i.e., moves in which the walk did not stay in the current vertex). If the graph is bipartite, then the prover, who can read the entire graph, can always answer correctly by checking whether and are on the same side of the bipartition. As shown in[21], if the graph is rapidly mixing and -far from bipartite, then a cheating prover will fail with probability . In order to reduce the soundness error (from to ), this basic test is repeated times in parallel (and the verifier accepts if and only if it accepted in all invocations of the basic test).
We construct a sub-linear time honest prover for the basic test. Upon receiving the vertices and , the prover performs independent random walks of length starting at , as well as independent random walks of length starting at . Let (resp., ) denote the set of vertices traversed in (the union of all) the walks that started at (resp., ). The prover checks if there is a vertex in the intersection of and (i.e., a vertex encounted in a walk that started at as well as in a walk that started at ). If so, then the prover has found a walk from to (going through ), and it replies with the parity of the number of real moves on this walk.
Soundness follows directly from the soundness of the [21] protocol (since we did not change the verifier). We now argue for completeness (alas not perfect completeness). If the graph is bipartite, then all paths from to have the same parity. Thus, whenever the prover finds a path from to (i.e., when the intersection is non-empty), then it returns the correct parity. It remains to show that in each of execution of the basic test (i.e., for each and sent by the verifier), the intersection of and is non-empty with probability at least . We show this in Claim 6, which is where we use the condition that the graph is rapidly-mixing, and it follows that the verifier accepts in all repetitions with probability at least .
Claim 6.
If the graph is rapidly-mixing, then, for every two vertices , the probability, over the honest prover’s random walks, that is at most .
Proof.
We focus on the sets and of terminal vertices reached by the random walks (i.e., the final vertex in each walk), and show that these subsets intersect with probability at least . Using the rapidly-mixing condition, observe first that taking independent walks from , with probability at least , the number of distinct vertices reached (i.e., the size of the set ) is . Assuming that this event occurs, and recalling that each of the random walks starting at is rapidly-mixing, we infer that, with probability at least , such a walk terminates in a vertex that resides in . Recalling that there are walks starting at , the probability that they all land outside of is thus at most . The claim follows.
1.3 Further Related Work
Goldwasser, Rothblum, Shafer and Yehudayoff [11] studied interactive proofs for approximate verification of the results of a machine-learning computation on an unknown underlying distribution. The verifier and the honest prover both have sampling access to the unknown distribution (in some of their results the honest prover also has membership queries, see also [12]). While their model is quite different from ours, there is a similarity in spirit: if we think of the unknown distribution as a “huge input” to the proof system, then they are also concerned with proof systems where verification is more efficient than the computation being verified (a learning problem), and proving is not much more expensive than performing the same computation. In particular, proving is sub-linear in the size of the “input” (the unknown distribution). The settings, however, are quite different: both in the access to the input (querying an unknown function vs. sampling from an unknown distribution), and in the types of tasks studied (approximate decision vs. machine learning). We stress that other recent works on verifying properties of unknown distributions [2, 14, 15] do not study honest provers with sublinear sample complexity.
1.4 Technical Overview
In this section provide overviews of our dsIPPs for Hamming Weight and for properties that are decidable by ROOBPs. Recall that the dsIPPs for PERM and for a relaxed version of Bipartiutness were sketched in prior sections.
We start by outlining our IPPs for Hamming Weight (HW), and then extend its underlying ideas to obtain IPPs for ROOBPs. Both IPPs follow an abstract idea that underlies the IPPs of[21, 7], but deviate from it in a significant manner. Loosely speaking, the IPPs of[21, 7] rely on partitioning the original property to sub-properties that refer to parts of the original input. However, while the IPPs of[21, 7] call for an honest prover strategy that finds optimal partitions of the property to sub-properties, we shall use partitions that are sub-optimal, because these sub-optimal partitions can be found more efficiently (by the honest prover).
IPPs for Hamming Weight (HW).
Our -round IPPs proceed by recursion, where for a parameter (to be set to ), the current input is partitioned to equal-length blocks, denoted . Clearly, the Hamming weight of , denoted , equals the sum of the Hamming weights of the ’s (i.e., ). Hence, wishing to prove that equals , an optimal prover determines the corresponding weights of the ’s, denoted , sends the ’s to the verifier, who selects a random , and the parties proceed to prove that the weight of equals . (Indeed, after iterations, the verifier can check the claim, which refers to a single bit, by making a single query.) Needless to say, determining these ’s requires reading the entire input , which we want to avoid. Instead, we consider a query-efficient honest prover that obtains approximations of the desired ’s (or obtains the exact value when , which happens after recursion steps). This requires relaxing the verification procedure so that it does not reject in case the sum does not equal (but is rather only close to it).
As in[21], the soundness analysis relies on the fact that if , then, for every sequence of ’s such that , it holds that ; equivalently, if is -far from having weight , then, for every sequence of ’s such that , on the average, is -far from having weight . Note, however, that if we allow to deviate from by say , then, on the average, is -far from having weight . Hence, at the bottom of the recursion, the expected distance of single bits from their claimed weight is , which means that the verifier rejects with such probability. Using and repeating the protocol times, we obtain the desired IPP.
IPPs for constant-width ROOBP.
As in the case of HW, we wish to proceed by recursion. Indeed, in the -round IPP of[7], on current input , an optimal prover determines the path in the current ROOBP that accepts , which (as before) is partitioned to equal-length strings . This path defines sub-ROOBPs that each accept the corresponding ’s. Alas, determining these sub-ROOBPs requires reading the entire input , which we want to avoid.
Using the fact that the ROOBP has bounded width, it follows that, for each , there is a bounded number of sub-ROOBPs (i.e., the square of the width bound) such that is accepted by (at least) one of them. Using a distance approximator for (constant-width) ROOBPs, the honest prover can find, for each , a sub-ROOBP such that is close to being accepted by this sub-ROOBP. This means that the foregoing description has to be modified: In subsequent iteration of the recursion it does not necessarily hold that the current is accepted by the current ROOBP; it is only the case that the current is close to being accepted by the current ROOBP. But in this case, for , it does not necessarily hold that each is close to being accepted the corresponding sub-ROOBPs, but it is rather the case that their average distance from these sub-ROOBPs is small; that is, if is -close to the current ROOBP, then for some such that , each is -close to a corresponding sub-ROOBP, but it is not necessarily the case that for each .
Recall that the honest prover does not find these ’s, but rather finds their approximate values (as well as the corresponding sub-ROOBPs). Specifically, for width bound , the honest prover considers an auxilary graph with layers, index , such that the layer of contains the vertices that are at the layer of the current ROOBP. Each pair of vertices in adjacent layers of represents a possible sub-ROOBP, where pairs in layers and of represent a sub-ROOBP that reads . For each such pair, the honest prover estimates the distance of from being accepted by the corresponding sub-ROOBP, where these estimates are obtained by invoking the distance-approximator (for ROOBPs). Using a shortest path algorithm, the honest prover finds sequence of intermediate vertices in such that the average distance of the ’s from the corresponding sub-ROOBPs, denoted , is minimal, and sends to the verifier along with the corresponding distances. (As in the case of HW, the verifier will select uniformly at random, and the parties will proceed to prove the corresponding claim (i.e., that is -close to being accepted by the corresponding sub-ROOBP).)
2 Preliminaries and Definitions
For strings , the relative Hamming distance between and is the fraction of coordinates in which they disagree (we often refer to this as the distance for short). If this distance is at most , then we say that is -close to , and otherwise we say that is -far from . We define the distance of from a (non-empty) set as its distance from the closest in , and we define the string being -close and -far from the set analogously. We extend these definitions from strings to functions by identifying a function with its truth table.
A property (or a language) is a set of strings of varying lengths (i.e. in ). The approximate decision problem for is deciding, for specified proximity parameters, whether an input string is close or far from the property. See [4] for an introduction to property testing: the field that studies algorithms of sublinear query complexity for such approximate decision problems.
Definition 7 (Interactive Proof of Proximity (IPP, tolerant)).
An IPP is a protocol between two probabilistic parties, a prover and a verifier who both get an input length and a joint input . The verifier has query access to (and explicit access to the input length ). The parties interact and at the end of the interaction the verifier accepts or rejects. The protocol is a (tolerant) IPP for a property and for proximity parameters if for every input length and every input :
-
Completeness: If is -close to the property (in relative Hamming distance), then the verifier, after interacting with the prover , accepts with probability at least (the probability is over the prover’s and the verifier’s coin tosses). If the verifier, interacting with the honest prover, accepts every input in the language with probability 1 then we say that the IPP has perfect completeness.
In this work we focus on IPPs where the honest prover only has query access to the input (similarly to the verifier, it gets the input length as an explicit input).
-
Soundness: If is -far from the property , then for every cheating strategy , the verifier , after interacting with the prover , rejects with probability at least (the probability is over the verifier’s coin tosses, w.l.o.g. the cheating prover can be taken to be deterministic, and can have explicit access to the entire input).
A non-tolerant IPP is one where . In this case, we refer to a single proximity parameter (the parameter is implicit). Testers are viewed as a special case of IPPs in which there is no prover (or no interaction with it).
The protocol’s complexity measures include the (honest) prover’s and the verifier’s query complexities and runtimes, the communication complexity and the round complexity (the number of back-and-forth communication rounds). These complexities are typically measured as a function of the gap between the proximity parameters and of the input length.
Our focus in this work is on protocols where the query complexities of the verifier and the honest prover are as small as possible. In particular, we focus on properties that have testers of sublinear query complexity and require that the verifier’s query complexity should be smaller than the tester’s. The prover’s query complexity should be as close as possible to the tester’s. We refer to these as doubly sub-linear IPPs (dsIPPs, see the discussion in the Introduction).
Remark 8.
We emphasize that, while in the context of testing, tolerant testing and distance approximation are equivalent [17] (under suitable parameters), this is not the case for IPPs. The reason for this divergence is that a tolerant IPP only provides a one-sided guarantee: it can affirmatively convince the verifier of upper bounds on the input’s distance from the property, but it does not convince the verifier of a lower bound on the distance.
Remark 9.
We remark that in the study of standard IPPs (where the honest prover can read the entire input), if we allow linear communication complexity, then the query complexity can always be tiny. This is because the prover can send the entire input to the verifier. The verifier receives this alleged input and verifies that it is close to the real input by checking consistency for a few random locations (these are the only queries made to the real input). If the alleged input is close to the real input, then the verifier can simply check if the alleged input is in the property (or, for tolerant protocols, that it is at the appropriate distance from the property). For dsIPPs it is not clear that linear communication always allows for a sublinear verifier. In particular, the aforementioned strategy cannot be utilized because the honest prover doesn’t know the entire input and cannot send it. Thus, dsIPPs with linear (or even polynomial) communication may be an interesting object for further study.
3 dsIPP for the Hamming Weight Problem
In this section, we present a doubly sub-linear IPP (dsIPP) for the Hamming Weight Problem. Specifically, given a claimed weight , a proximity parameter and an oracle access to an input , we present an -round IPP, with and with poly-logarithmic communication complexity, in which verifies that using query complexity of , the honest prover uses poly-logarithmic query complexity. We note that we use to represent the number of (pairs) of message exchanges (and not the total number of messages sent).
First, in Section 3.1, we present a standard IPP for the Hamming Weight Problem. This IPP (which is not doubly sub-linear) uses a divide-and-conquer approach, and in it the prover computes exact weights for sub-sequences of the input . Then, in Section 3.2, we present a warm-up towards a dsIPP: We amend the standard IPP by allowing to approximate the Hamming weight of the sub-sequences, and relaxing the accepting conditions of in a way that doesn’t compromise the completeness and soundness of the protocol. Unfortunately, in this IPP the prover still has query complexity of . Lastly, in Section 3.3, we present the actual dsIPP, which applies the IPP presented in Section 3.2 recursively, thus improving the query complexity of without compromising the query complexity of .
3.1 The Standard IPP
Given some , we want to verify that . First, we observe that when we partition to consecutive equally-length parts, , the following holds:
-
On the one hand, if , then .
-
On the other hand, if , then for any such that , it holds that .
Given the above, we consider the following IPP: sends to . Then, verifies that , and if so, selects at random , and verifies that (by reading the entire ). Note that if and interacts with , then for every , it holds that , and the verifier always accepts. However, if and , it holds that . Then, the average deviation of the claimed weight of from the actual weight of translates to a deviation on a corresponding fraction of random ’s, hence guarantees that the verifier rejects with probability at least . To increase the rejection probability, we can repeat this procedure for times.
The query complexity of the verifier is the length of a single times the number of repetitions, which gives . However, to compute , the prover must access to the entire input , which yields query complexity of .
3.2 A Warm-Up towards a dsIPP
In this section, we aim to achieve poly-logarithmic query complexity for the honest prover without compromising the query complexity of the verifier : We amend the interaction described in Section 3.1 by allowing to approximate the Hamming weight of the sub-parts of . However, at the end of this section, we explain why still has query complexity of . In Section 3.3, we show how to solve this issue and present the actual dsIPP.
First, we observe that for a deviation parameter and an error probability parameter , there exists a ()-approximator for the Hamming weight of with query complexity and error probability : Given the input and the deviation parameter , the approximator chooses, uniformly at random, indices of , and outputs the (normalized) number of indices for which . Using Chernoff bound, we can show that with probability at least , the output of the approximator deviates from by at most .
Based on this approximator, we amend the interaction such that the prover provides approximated weights to the verifier, by using the approximator with some predefined error probability and deviation parameter (we fix both parameters later). That is, sends to , such that for every , is a -approximation for . We also change the verifier’s accepting conditions in the interaction, so that allows this deviation (and doesn’t reject ’s approximations). That is, now verifies that , chooses uniformly at random, and verifies that (by reading the entire ). Similar to the standard IPP, we repeat the protocol for times.
3.2.1 Correctness
The completeness of the protocol still holds (but with a completeness error): If and we interact with , then for a single invocation of the protocol the following holds: For every , with probability at least , it holds that , and therefore also , and the verifier accepts. Since we invoke the protocol for times, then the verifier accepts in all invocations with probability at least where . Thus, we can set , and the verifier accepts with high constant probability in all invocations.
To show the soundness of the protocol also holds, we observe that if , then for any such that , it holds that . Then, the average deviation of the claimed weight of from the actual weight of , beyond the allowed deviation, translates to a deviation on a corresponding fraction of random ’s. Since the verifier rejects if , we can set and infer that the verifier still rejects with probability at least in a single invocation. Since we invoke the protocol for times, we get that the verifier rejects with high constant probability in at least one of the invocations.
3.2.2 Query Complexity
Now, let us analyze the query complexity of and :
-
The query complexity of is , because we read a sub-sequence of the input of length for repetitions.
-
The query complexity of is , since we invoke the approximator with error probability parameter for times in each of the repetitions.
Since we can -approximate the Hamming weight of without any interaction using query complexity , we can perform -test for the Hamming weight of using query complexity . Therefore, for the IPP to be meaningful, we want the query complexity of to be . To achieve this in our setting, we must require . However, this causes the query complexity of to be , whereas we want our IPP to be doubly sub-linear. In the next section, we extend the foregoing IPP by applying it recursively, and show that we can get both query complexity of for , and poly-logarithmic query complexity for .
3.3 The Actual dsIPP
In Section 3.2, we aim to achieve poly-logarithmic query complexity for the honest prover , without compromising the query complexity of the verifier . Unfortunately, for reasons explained at Section 3.2.2, still has query complexity of .
To solve this, we extend the IPP presented in Section 3.2 by applying it recursively: That is, after selects at random , instead of directly computing the Hamming weight of , both parties recursively run the protocol on with the claimed Hamming weight , and do it for rounds. Only in the base of the recursion, computes the Hamming weight of the input (at this level), and accepts accordingly. This gives us an improvement in the query complexity of the verifier, as the length of the input after rounds is (), while the number of times the prover needs to run the approximator grows only to (in every repetition). To make sure that the query complexity of is , we set .
Indeed, we now allow the prover to deviate from the claimed weight for times: At each of the recursion levels, as well as in the base level. Thus, to make sure that soundness still holds (i.e. the total allowed deviation is not too big), we set the verifier’s allowed deviation (i.e. ) to be smaller (by a factor of ).
To make sure that completeness still holds, we first change the error probability parameter of the approximator. now runs the approximator for a total of times, and so we set . In addition, we observe the following: Assume interacts with , with input and weight parameter , such that . Then, when we are not in the first recursion level, then the weight parameter provided to the protocol is the approximated weight provided by in the previous recursion level. Therefore, the weight parameter in the current recursion level might deviate from the actual weight of the input in this level. Since also performs weight approximations in the current recursion level, we need to account for both deviations. Details follow.
Protocol 1.
dsIPP for the Hamming Weight Problem
-
Query Access Input:
-
Deviation Parameter:
-
Other Parameters: Weight , initial number of rounds and remaining number of rounds
-
1.
Set , , and .
-
2.
: If then accept iff .
-
3.
:
-
(a)
For every , approximate up to a deviation factor of with probability at least : Choose uniformly at random indices of and set the approximation as the estimated fraction of indices for which .
-
(b)
Send to .
-
(a)
-
4.
:
-
(a)
Verify that , otherwise reject.
-
(b)
Choose uniformly at random and send to .
-
(a)
-
5.
Both parties recursively invoke the protocol with input , deviation parameter , weight , and remaining number of rounds .
-
1.
Repeat the protocol for times, and accept iff accepts in all repetitions.
We show that Protocol 1 is a dsIPP for the Hamming Weight Problem: In Section 3.3.1, we show the correctness of the protocol, and in Section 3.3.2, we present the query complexity of the protocol.
3.3.1 Correctness
Completeness.
Assume and interacts with . We observe that at any recursion level (but the base level), the following claim holds: If our input is and our weight parameter is such that , then with probability at least , the prover sends ) to such that the following holds:
-
1.
The verifier doesn’t reject in Step (i.e., ).
-
2.
The protocol is recursively invoked with some and for which .
The claims follows from the fact that runs the approximator on every with deviation parameter and error probability parameter . Then, since we start with , with probability at least , the verifier doesn’t reject in Step in all the recursion levels. By the second item of the claim, the base of the recursion is also invoked with an input and weight parameter is such that , and the verifier accepts. Since we repeat the protocol for times, we get that the verifier accepts in all invocations with probability at least where . Then, since we set , we get that accepts with high constant probability in all invocations.
Soundness.
Assume . We observe that in each recursion level, we lose at most an additive factor of . Since we set , we get that after rounds, we are still at distance at least from a valid assertion. Thus, the verifier rejects with probability at least in each invocation, and therefore rejects with high constant probability in at least one of the invocations.
3.3.2 Computational Complexity
We present the query, communication and time complexity of Protocol 1. Let us start by analyzing the query complexity of (denoted ) and (denoted ):
-
, because we read a sub-sequence of the input of length for times.
-
is the query complexity of the approximator times the number of calls makes to the approximator in all repetitions:
-
–
The query complexity of the approximator is , because invokes the approximator with error probability parameter and deviation parameter .
-
–
calls the approximator for times in each of the repetitions.
Since we set , we get that .
-
–
Hence, regardless of how we set , the query complexity of the verifier is . Indeed, the larger we set , the smaller gets, but this causes the round complexity to grow. Specifically, to minimize , we can set , and we get .
Now, the communication of Protocol 1 is : In each of the repetitions there are rounds, and in each round the prover sends weight approximations that can be represented by bits (because the approximations are up to a factor of ). Again, since we set , we can set and get poly-logarithmic communication complexity.
Lastly, the time complexity of the honest prover (denoted ) is the same as its query complexity (i.e., ), whereas the time complexity of the verifier (denoted ) is , because in each of the repetitions there are rounds, and in each round the verifier performs simple calculations on values that can be represented by bits.
We conclude with the following theorem, that summarizes what we achieved in this section.
Theorem 10.
For every function , there exists an -round dsIPP for the Hamming Weight Problem with , and communication complexity of . In addition, the time complxity of the dsIPP is and .
4 Tolerant dsIPP for ROOBPs of Constant Width
In this section, we present a tolerant dsIPP for ROOBPs of constant width. Before we review our plan in Section 4.2, let us recall some standard definitions related to ROOBPs.
4.1 Definitions
Definition 11 (ROOBP).
A branching program (BP) on variables is a directed acyclic graph that has a unique source vertex (denoted ) with in-degree and a unique sink vertex (denoted ) with out-degree 0. Each non-sink vertex is labeled by an index , and has outgoing edges, which are labeled by either or . An input defines a walk on starting at the source vertex, such that at every vertex labeled by , the step taken is on the edge labeled by . The output of the branching program on input , denoted , is defined as if the walk reached the sink vertex, and if it got “stuck” before reaching it (i.e., the walk reached vertex labeled by , and there was no outgoing edge labeled by ).
An oblivious BP is a BP in which the nodes are partitioned into levels, , and edges are going only from one level to nodes in the consecutive level. In addition, all the vertices of some level are associated with the same index. Therefore, we can say that the level itself is associated with some index. A read-once oblivious BP (ROOBP) is a BP in which no two levels are associated with the same index. An ROOBP of constant width is an ROOBP in which every level has at most vertices.
From now on, let be an ROOBP with variables and width at most .
Remark 12.
By the above, there’s a 1-1 correspondence between an path in and an accepting input for . In that case, we say that the path is associated with the accepting input.
Remark 13.
We say that two vertices and in are connected only if they are connected via a directed path. If , we say that is forwards-connected to , and is backwards-connected to .
Remark 14.
Without loss of generality, we assume:
-
1.
For every , is associated with the index . Otherwise, we can change the input to an input by reordering its indices accordingly.
-
2.
depends on the entire input. Otherwise, we can change the input to , and then will depend on every index of .
-
3.
There exists a path between the source and the sink of ; this is equivalent to assuming there exists an accepted by .
Definition 15 (Absolute and Relative Distance).
We denote the absolute distance between two strings by , and their relative distance by . Assume there exist an accepting input for , an ROOBP of length . We denote the absolute distance of a string from by . Similarly, the relative distance of from is denoted by .
4.2 Our Plan
For a fixed ROOBP of width , proximity parameters , and oracle access to an input , we present an -round IPP, with and with poly-logarithmic communication complexity, in which uses query complexity of , the honest prover uses poly-logarithmic query complexity, and the following holds:
-
If , then when communicating with , with probability at least the verifier accepts .
-
If , then when communicating with any prover , with probability at least the verifier rejects .
We note that we use to represent the number of (pairs) of message exchanges (and not the total number of messages sent).
Our plan is as follows: First, in Section 4.3, we present a standard tolerant IPP for the ROOBP Problem. This IPP (which is not doubly sub-linear) uses a divide-and-conquer approach, and in it the prover partitions the ROOBP to a sequence of sub-ROOBPs (see Definition 16) according to the accepting path associated with an accepting input that minimizes the distance to in (see Definition 17). Then, in Section 4.4, we present a warm-up towards the tolerant dsIPP: We amend the standard tolerant IPP by allowing to partition the ROOBP according to an approximated path, and relaxing the accepting conditions of in a way that doesn’t compromise the completeness and soundness of the protocol. Unfortunately, in this IPP the prover still has query complexity of . Lastly, in Section 4.5, we present the actual tolerant dsIPP, which applies the IPP presented in Section 4.4 recursively, thus improving the query complexity of without compromising the query complexity of .
4.3 The Standard Tolerant IPP
Given some ROOBP , and proximity parameters , we want to verify that . In the following tolerant IPP, that adapts the non-tolerant IPP from [7], we decompose to consecutive equally-length sub-ROOBPs, . Therefore, we start with defining the notion of a sub-ROOBP.
Definition 16 (sub-ROOBP).
For , let and . We define as a sub-ROOBP of , with the following properties:
-
The source (resp., sink) vertex of is (resp., ).
-
is of length .
-
If is the input for , then is the input for .
Indeed, there is more than one way we can decompose to consecutive equally-length sub-ROOBPs, because for each sub-ROOBP there could be up to possible source-sink pairs. Therefore, we continue with defining the notion of (what we consider) a valid decomposition of to sub-ROOBPs: A decomposition that doesn’t cause us to “lose” any distance (i.e. the average distance of from all the sub-ROOBPs in the decomposition is at least the distance of from ). We can achieve this by enforcing the decomposition to follow some path associated with some accepting input for , since for any accepting input we have .
Definition 17 (-Decomposition).
Let be an ROOBP, and let be an accepting path in . The ()-decomposition of is a sequence of sub-ROOBPs of , denoted ), such that for every , the source of is the ’th vertex in , and the sink of is the ’th vertex in .
Now, we observe that when we partition to consecutive equally-length parts, , the following holds:
-
On the one hand, if is -close to , then for , the -decomposition of constructed according to an accepting input that minimizes the distance to , it holds that .
-
On the other hand, if , then for any , a -decomposition of constructed according to some accepting path in , it holds that .
Given the above, we consider the following IPP: The prover finds the -decomposition of constructed according to an accepting input that minimizes the distance to . Then, checks the distance of from each of the sub-ROOBPs in the decomposition, and sends a succinct representation of the decomposition (for example, the sink of every sub-ROOBP), as well as the obtained distances. After receives some decomposition and claimed distances , verifies that is a valid ()-decomposition of , and that the average of the claimed distances is at most . Lastly, selects at random , and verifies that (by reading the entire ).
Note that if is -close to and interacts with , then , and for every , it holds that , and the verifier always accepts. However, if , then for any valid decomposition and claimed distances sent by some prover such that , it holds that . Then, the average deviation of the actual distance of from , from the claimed distance of from , translates to a deviation on a corresponding fraction of random ’s, which guarantees that the verifier rejects with probability at least . To increase the rejection probability, we can repeat this procedure for times.
We stress that can verify the validity of a ()-decomposition without any query access to the input . Thus, the query complexity of the verifier is the length of a single times the number of repetitions, which gives . However, to find an accepting input that minimizes the distance to , the prover must access to the entire input , which yields query complexity of .
4.4 A Warm-Up towards a tolerant dsIPP
In this section, we aim to achieve poly-logarithmic query complexity for the honest prover without compromising the query complexity of the verifier : We amend the interaction described in Section 4.3 by allowing to divide the ROOBP according to an accepting input that approximates the distance of to . However, at the end of this section, we explain why still has query complexity of . In Section 4.5, we show how to solve this issue and present the actual tolerant dsIPP.
In the full version of this paper, we show the existence of a -distance-approximation algorithm for ROOBPs of constant width with query complexity that is independent of the length of the input, . That is, we show there exists an algorithm that given an ROOBP of width , an input , deviation parameter and error probability parameter , outputs such that with probability at least it holds that , and this algorithms has query complexity of .
Based on this distance-approximation algorithm, we amend the interaction as follows: performs distance approximation (with parameters and we fix later) for every (-decomposition (constructed according to every accepting path ). Since performs distance approximation also for the decomposition of constructed according to an accepting input that minimizes the distance to , can just choose the the )-decomposition which yields the minimal (average) distance approximation. That is, chooses for which is minimal.
Even though there could be many possible )-decompositions, the total number of sub-ROOBPs in all the -decompositions is upper-bounded by : For every , there are at most possible source-sink pairs for the ’t sub-ROOBP. Thus, using the union bound, with probability at least , all approximations deviate from the actual distance by at most , and the following holds:
-
For every , it holds that .
-
If , then since chooses the )-decomposition which yields the minimal (average) distance approximation, we get that .
Lastly, we also change ’s accepting conditions in the interaction, so that allows the deviation. That is, after receives and , in addition to verifying that is a valid decomposition, verifies that , and after chooses uniformly at random, verifies that (by reading the entire ). Similar to the standard IPP, we repeat the protocol for times.
4.4.1 Correctness
We claim that the completeness of the protocol still holds (but with a completeness error). If interacts with , then for a single invocation of the protocol, using the union bound, with probability at least , all approximations deviate from the actual distances by at most . Then, if , it holds that , and for every it holds that , and the verifier accepts. Since we invoke the protocol for times, then for , the verifier accepts in all invocations with probability at least . Thus, we can set , and the verifier accepts with high constant probability in all invocations.
To show the soundness of the protocol also holds, assume , and let be a decomposition that sends, along with the respective claimed distance approximations . If the decomposition is valid, then . Thus, if , we get that . Then, the average deviation of the actual distance of from , from the claimed distance of from , beyond the allowed deviation, translates to a deviation on a corresponding fraction of random ’s. Since the verifier rejects if , we can set , and infer that the verifier still rejects with probability at least in a single invocation. Since we invoke the protocol for times, we get that the verifier rejects with high constant probability in at least one of the invocations.
4.4.2 Query Complexity
Now, let us analyze the query complexity of and :
-
Because can verify that the ()-decomposition that sends is valid without any query access to the input , the query complexity of is : reads a sub-sequence of the input of length for repetitions.
-
The query complexity of is , because the distance approximation algorithm with deviation parameter and error probability parameter has query complexity of , and invokes the algorithm for times in each of the repetitions.
Since we can -approximate without any interaction using query complexity , we can perform tolerant ()-test for ROOBPs using query complexity . Therefore, for the IPP to be meaningful, we want the query complexity of to be . To achieve this in our setting, we must require . However, this causes the query complexity of to be , whereas we want our IPP to be doubly sub-linear. In the next section, we extend the foregoing IPP by applying it recursively, and show that we can get both query complexity of for , and poly-logarithmic query complexity for .
4.5 The Actual dsIPP
In Section 4.4, we aim to achieve poly-logarithmic query complexity for the honest prover , without compromising the query complexity of the verifier . Unfortunately, for reasons explained at Section 4.4.2, the prover still has query complexity of .
To solve this, we extend the IPP presented in Section 4.4 by applying it recursively: That is, after selects at random , instead of directly checking the distance of from , both parties recursively run the protocol on with , and do it for rounds. Only in the base of the recursion, assuming the input is and the ROOBP is , checks the distance of from and accepts accordingly. This gives us an improvement in the query complexity of the verifier, as the length of the input after rounds is (), while the number of times the prover needs to run the distance approximation algorithm grows only to (in every repetition). In order to make the query complexity of independent of the input length, we set .
Now, we observe the following: Assume interacts with , with input and ROOBP , such that . Then, since we are only guaranteed that , when both parties recursively apply the protocol with input and ROOBP , we can’t guarantee that (or even guarantee that ).
Therefore, we amend the protocol as follows: In addition to the input and the ROOBP , we add a claimed distance parameter, , that represents ’s claim regarding the distance of from . In the first recursion level, we set , since claims that . Then, at each recursion level, assume our input , our ROOBP is , our (new) claimed distance parameter is , and sends the decomposition and the respective approximations . Then, verifies that is a valid decomposition and that (i.e., we replace in the previous condition with the new parameter ). Lastly, for some both parties recursively run the protocol on with and the claimed distance , and at the base of the recursion, verifies that the distance of the input from the ROOBP does not exceed the claimed distance by more than .
Now, on every recursion level, only claims that the average distance doesn’t grow too much with respect to the approximated distance in the previous recursion level. Since uses a distance approximation algorithm with some error probability parameter , and chooses the ()-decomposition of that yields the minimal average distance approximation from , the claim holds for all recursion levels with probability at least , and at the base level, the verifier accepts. Since invokes the protocol for times, then for , the verifier accepts in all invocations with probability at least . Thus, we can set , and the verifier accepts with high constant probability in all invocations. Lastly, we note that at each recursion level, we want to account for two possible deviations: In the approximation that provides for the previous recursion level (i.e. ), and in the approximations that performs in the current recursion level (i.e. ).
For soundness, we observe that allows the prover to exceed from the initial claimed distance (i.e., ) by at most for times: At each of the recursion levels, as well as in the base level. Thus, to make sure that soundness still holds (i.e. does not allow the prover to exceed too much in distance), we set to be smaller (by a factor of ). Details follow.
Protocol 2.
Tolerant dsIPP for the ROOBP Problem
-
Query Access Input:
-
Proximity Parameters:
-
Massive Parameter: ROOBP of width
-
Other Parameters: Claimed distance (initialized to in the first round), initial number of rounds and remaining number of rounds
-
1.
Set , , and .
-
2.
: If then accept iff .
-
3.
:
-
(a)
For every , and for every and , approximate the distance of from with deviation parameter and error probability parameter .
-
(b)
For every ()-decomposition of , , let be the respective distance approximations obtained in the previous step.
-
(c)
Find a decomposition for which is minimal, and send (a succinct representation of) , along with , to .
-
(a)
-
4.
:
-
(a)
Verify that is a valid ()-decomposition, and that . Otherwise, reject.
-
(b)
Choose uniformly at random and send to .
-
(a)
-
5.
Both parties recursively invoke the protocol with input , ROOBP , claimed distance , and remaining number of rounds .
-
1.
Initiate the protocol with , repeat the protocol for times, and accept iff accepts in all repetitions.
Remark 18.
can efficiently find a decomposition in Step as follows: Construct a graph with being all the vertices of in the levels , and are connected in with edge-weight if performs distance approximation for in Step and it results in . Then, every ()-decomposition considered in Step has a corresponding path in , and the weight of the path equals the sum of distance approximations for the sub-ROOBPs in the decomposition. Thus, can choose the decomposition with the shortest path in .
We show that Protocol 2 is a dsIPP for ROOBPs: In Section 4.5.1, we show the correctness of the protocol, and in Section 4.6.1, we present the query complexity of the protocol.
4.5.1 Correctness
Completeness.
Assume and interacts with . We observe that at any recursion level (but the base level), the following claim holds: If our input is , our ROOBP is and our claimed distance parameter is such that , then with probability at least , the prover sends ) and ) to such that the following holds:
-
1.
The verifier doesn’t reject in Step (i.e., ).
-
2.
The protocol is recursively invoked with some , and for which .
The claims follows from the fact that runs the distance approximation algorithm with deviation parameter and with error probability parameter . Since we start with , then with probability at least , the verifier doesn’t reject in Step in all the recursion levels. By the second item of the claim, the base of the recursion is also invoked with an input , ROOBP and claimed distance parameter such that , and the verifier accepts. Since we repeat the protocol for times, we get that for , the verifier accepts in all invocations with probability at least . Then, since we set , we get that accepts with high constant probability in all invocations.
Soundness.
Assume . We observe that in each round, we lose at most an additive factor of in distance (i.e. we allow the prover to “exceed” the claimed distance up to an additive factor). Since we start with claimed distance , and since we set , we get that after rounds, we are still at distance at least from a valid assertion. Hence, the verifier rejects with probability at least in each invocation, and rejects with high constant probability in at least one of the invocations.
4.6 Computational Complexity (of Protocol 2)
The query and communication complexity of our tolerant dsIPP are our main complexity measures. Indeed, the communication complexity is , because in each of the repetitions there are rounds, and in each round the prover sends vertices and distance approximations to that can be represented by bits. Since we set , we can set and get poly-logarithmic communication complexity.
We continue in Section 4.6.1, where we calculate the query complexity of both the verifier and the prover. We show that if we set , then uses query complexity of , and the honest prover uses poly-logarithmic query complexity.
We finish by calculating the time complexity of our tolerant dsIPP. We show that, if we set , and assume both and have black-box access to two procedures that refer to the input ROOBP (see Section 4.6.2 for more details), then the verifier has time complexity of . In addition, if the honest prover uses a distance approximation algorithm for ROOBPs with time complexity , then has time complexity of .
4.6.1 Query Complexity
Let us analyze the query complexity of (denoted ) and (denoted ):
-
Because can verify that the ()-decomposition sends in every round is valid without any query access to the input , we have : reads a sub-sequence of the input of length for times.
-
uses a distance approximation algorithm with deviation parameter and error probability parameter , with query complexity . Then, the query complexity of is times the number of calls makes to in all the repetitions. Since calls the approximator for times in each of the repetitions, and we set , we get that .
Hence, regardless of how we set , the query complexity of the verifier is . Indeed, the larger we set , the smaller gets, but this causes the round complexity to grow. To minimize , we can set , and we get .
In the full version of this paper, we show the existence of a distance approximation algorithm with query complexity of . If uses this distance approximation algorithm, we get . Again, we can set , and if we assume is constant, we get poly-logarithmic query complexity.
4.6.2 Time Complexity
Indeed, assuming a standard access to the input ROOBP (i.e. the “Massive Parameter”), the time complexity of both the verifier (denoted ) and the honest prover (denoted ) is :
-
Given a sub-sequence of vertices in , the verifier has to verify that every is in the ’th level of , and that . Since and , this requires accessing the entire ROOBP, thus yields time complexity of .
-
approximates the distance of sub-parts of from sub-ROOBPs of of length for times. Using the distance approximation algorithm presented in the full version of this paper, this also yields time complexity of .
However, we can take an alternative approach in the analysis: Assume both and have black-box access to the following two procedures, that refer to the structure of the input ROOBP, and are independent of the specific input :
-
1.
Given two vertices and , determine whether in .
-
2.
Given , list the (up to ) vertices that are in level of .
In addition, assume uses a distance approximation algorithm with deviation parameter and error probability parameter , with time complexity (where is the length of the input and is the width of the ROOBP). Now, for , we can analyze the time complexity of and in the ’th round of the protocol:
-
calls the first two procedures for times, obtains values that can be represented by bits, and performs simple calculations on the obtained values.
-
calls the first two procedures for times, and calls the distance approximation algorithm for at most times on inputs of length . By doing the foregoing, obtains values that can be represented by ) bits, and performs simple calculations on the obtained values. This yields time complexity of .
Since and perform the foregoing for every in each of the invocations of the protocol, and since we set , we get that:
To minimize the time complexity, we can set , and assuming is constant, we get and . We conclude with the following theorem and corollary.
Theorem 19.
For every function , there exists an -round tolerant dsIPP for ROOBPs of width with proximity parameters . In the tolerant dsIPP, the honest prover uses a distance approximation algorithm with deviation parameter and error probability parameter such that the query complexity of is and the time complexity of is . Then, the computational complexity of the tolerant dsIPP is as follows:
-
Query Complexity: and .
-
Communication Complexity: .
-
Time Complexity: If both and have black-box access to procedures that refer to the structure of the input ROOBP, then:
-
–
-
–
-
–
In the full version of this paper, we show the existence of a distance approximation algorithm with time complexity of . If uses this distance approximation algorithm, we get the following corollary.
Corollary 20.
For every function , there exists an -round tolerant dsIPP for ROOBPs of width with , , and communication complexity of . If both and have black-box access to procedures that refer to the structure of the input ROOBP, then and .
References
- [1] Ran Canetti, Guy Even, and Oded Goldreich. Lower bounds for sampling algorithms for estimating the average. Inf. Process. Lett., 53(1):17–25, 1995. doi:10.1016/0020-0190(94)00171-T.
- [2] Alessandro Chiesa and Tom Gur. Proofs of proximity for distribution testing. In Anna R. Karlin, editor, 9th Innovations in Theoretical Computer Science Conference, ITCS 2018, January 11-14, 2018, Cambridge, MA, USA, volume 94 of LIPIcs, pages 53:1–53:14. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2018. doi:10.4230/LIPICS.ITCS.2018.53.
- [3] Funda Ergün, Ravi Kumar, and Ronitt Rubinfeld. Fast approximate probabilistically checkable proofs. Inf. Comput., 189(2):135–159, 2004. doi:10.1016/J.IC.2003.09.005.
- [4] Oded Goldreich. Introduction to Property Testing. Cambridge University Press, 2017. doi:10.1017/9781108135252.
- [5] Oded Goldreich. On doubly-efficient interactive proof systems. Found. Trends Theor. Comput. Sci., 13(3):158–246, 2018. doi:10.1561/0400000084.
- [6] Oded Goldreich, Shafi Goldwasser, and Dana Ron. Property testing and its connection to learning and approximation. J. ACM, 45(4):653–750, 1998. doi:10.1145/285055.285060.
- [7] Oded Goldreich, Tom Gur, and Ron D. Rothblum. Proofs of proximity for context-free languages and read-once branching programs. Inf. Comput., 261:175–201, 2018. doi:10.1016/J.IC.2018.02.003.
- [8] Oded Goldreich and Dana Ron. A sublinear bipartiteness tester for bounded degree graphs. Comb., 19(3):335–373, 1999. doi:10.1007/S004930050060.
- [9] Oded Goldreich and Dana Ron. Property testing in bounded degree graphs. Algorithmica, 32(2):302–343, 2002. doi:10.1007/S00453-001-0078-7.
- [10] Shafi Goldwasser, Yael Tauman Kalai, and Guy N. Rothblum. Delegating computation: Interactive proofs for muggles. J. ACM, 62(4):27:1–27:64, 2015. doi:10.1145/2699436.
- [11] Shafi Goldwasser, Guy N. Rothblum, Jonathan Shafer, and Amir Yehudayoff. Interactive proofs for verifying machine learning. In James R. Lee, editor, 12th Innovations in Theoretical Computer Science Conference, ITCS 2021, January 6-8, 2021, Virtual Conference, volume 185 of LIPIcs, pages 41:1–41:19. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2021. doi:10.4230/LIPICS.ITCS.2021.41.
- [12] Tom Gur, Mohammad Mahdi Jahanara, Mohammad Mahdi Khodabandeh, Ninad Rajgopal, Bahar Salamatian, and Igor Shinkar. On the power of interactive proofs for learning. In Bojan Mohar, Igor Shinkar, and Ryan O’Donnell, editors, Proceedings of the 56th Annual ACM Symposium on Theory of Computing, STOC 2024, Vancouver, BC, Canada, June 24-28, 2024, pages 1063–1070. ACM, 2024. doi:10.1145/3618260.3649784.
- [13] Tom Gur, Yang P. Liu, and Ron D. Rothblum. An exponential separation between MA and AM proofs of proximity. Comput. Complex., 30(2):12, 2021. doi:10.1007/S00037-021-00212-3.
- [14] Tal Herman and Guy N. Rothblum. Verifying the unseen: interactive proofs for label-invariant distribution properties. In Stefano Leonardi and Anupam Gupta, editors, STOC ’22: 54th Annual ACM SIGACT Symposium on Theory of Computing, Rome, Italy, June 20 – 24, 2022, pages 1208–1219. ACM, 2022. doi:10.1145/3519935.3519987.
- [15] Tal Herman and Guy N. Rothblum. Doubley-efficient interactive proofs for distribution properties. In 64th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2023, Santa Cruz, CA, USA, November 6-9, 2023, pages 743–751. IEEE, 2023. doi:10.1109/FOCS57990.2023.00049.
- [16] Ilan Newman. Testing membership in languages that have small width branching programs. SIAM J. Comput., 31(5):1557–1570, 2002. doi:10.1137/S009753970038211X.
- [17] Michal Parnas, Dana Ron, and Ronitt Rubinfeld. Tolerant property testing and distance approximation. J. Comput. Syst. Sci., 72(6):1012–1042, 2006. doi:10.1016/J.JCSS.2006.03.002.
- [18] Pavel Pudlák. A lower bound on complexity of branching programs (extended abstract). In Michal Chytil and Václav Koubek, editors, Mathematical Foundations of Computer Science 1984, Praha, Czechoslovakia, September 3-7, 1984, Proceedings, volume 176 of Lecture Notes in Computer Science, pages 480–489. Springer, 1984. doi:10.1007/BFB0030331.
- [19] Omer Reingold, Guy N. Rothblum, and Ron D. Rothblum. Constant-round interactive proofs for delegating computation. SIAM J. Comput., 50(3), 2021. doi:10.1137/16M1096773.
- [20] Guy N. Rothblum and Ron D. Rothblum. Batch verification and proofs of proximity with polylog overhead. In Rafael Pass and Krzysztof Pietrzak, editors, Theory of Cryptography – 18th International Conference, TCC 2020, Durham, NC, USA, November 16-19, 2020, Proceedings, Part II, volume 12551 of Lecture Notes in Computer Science, pages 108–138. Springer, 2020. doi:10.1007/978-3-030-64378-2_5.
- [21] Guy N. Rothblum, Salil P. Vadhan, and Avi Wigderson. Interactive proofs of proximity: delegating computation in sublinear time. In Dan Boneh, Tim Roughgarden, and Joan Feigenbaum, editors, Symposium on Theory of Computing Conference, STOC’13, Palo Alto, CA, USA, June 1-4, 2013, pages 793–802. ACM, 2013. doi:10.1145/2488608.2488709.
- [22] Ronitt Rubinfeld and Madhu Sudan. Robust characterizations of polynomials with applications to program testing. SIAM J. Comput., 25(2):252–271, 1996. doi:10.1137/S0097539793255151.
- [23] Hadar Strauss. Emulating computationally sound public-coin ipps in the pre-coordinated model. Electronic Colloquium on Computational Complexity (ECCC), TR24-131, 2024. URL: https://eccc.weizmann.ac.il/report/2024/131/.
