The Planted Orthogonal Vectors Problem
Abstract
In the -Orthogonal Vectors (-OV) problem we are given sets, each containing binary vectors of dimension , and our goal is to pick one vector from each set so that at each coordinate at least one vector has a zero. It is a central problem in fine-grained complexity, conjectured to require time in the worst case.
We propose a way to plant a solution among vectors with i.i.d. -biased entries, for appropriately chosen , so that the planted solution is the unique one. Our conjecture is that the resulting -OV instances still require time to solve, on average.
Our planted distribution has the property that any subset of strictly less than vectors has the same marginal distribution as in the model distribution, consisting of i.i.d. -biased random vectors. We use this property to give average-case search-to-decision reductions for -OV.
Keywords and phrases:
Average-case complexity, fine-grained complexity, orthogonal vectorsCopyright and License:
2012 ACM Subject Classification:
Theory of computation Computational complexity and cryptographyAcknowledgements:
We are grateful to Andrej Bogdanov, Antoine Joux, Moni Naor, Nicolas Resch, Nikolaj Schwartzbach, and Prashant Vasudevan for insightful discussions.Funding:
Work supported by European Research Council (ERC) under the EU’s Horizon 2020 research and innovation programme (Grant agreement No. 101019547) and Cariplo CRYPTONOMEX grant. Part of this work was done when the first author was visiting Bocconi University.Editors:
Anne Benoit, Haim Kaplan, Sebastian Wild, and Grzegorz HermanSeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
The security of cryptographic systems crucially relies on heuristic assumptions about average-case hardness of certain computational problems. Sustained cryptanalysis alongside technological advances such as large-scale quantum computers, put these hardness assumptions under constant risk of being invalidated. It is therefore desirable to try to design cryptographic schemes based on new computational problems, preferably ones whose hardness is well-studied.
The field of computational complexity developed over the last fifty years a good understanding of the hardness of certain problems – e.g., SAT is widely believed to require at least superpolynomial, maybe even exponential time [28] – however these are worst-case problems, and hence unsuitable for direct use as a basis for cryptography.
Fine-grained complexity [27] is a younger branch of computational complexity that studies “hardness of easy problems”, i.e., problems known to be solvable in polynomial time but supposedly not faster than some specified polynomial, say not faster than in cubic time. It gives rise to fine-grained cryptography [6, 23, 25], the idea that it might be possible to build cryptography, notably public-key encryption, based on conjectured average-case hardness of polynomial-time problems studied in fine-grained complexity. These problems are easier than NP-hard ones, but for polynomials of sufficiently high degree may still be hard enough to give honest parties an adequate advantage over malicious attackers.
1.1 The k-Orthogonal Vectors Problem
The Orthogonal Vectors (OV) problem [29], together with its generalization -OV, is one of the three main hard problems studied in fine-grained complexity, alongside 3SUM and APSP [27]. Arguably, among the three, OV is the one whose (worst-case) hardness we understand the most – in particular because it is implied by the Strong Exponential Time Hypothesis (SETH) [16], which is about the very well studied SAT problem.
We say that vectors are orthogonal if, for all , , meaning that for every coordinate there is at least one zero entry among the vectors. For , let each be a collection of -dimensional binary vectors, which we view as matrices in . We denote by the -th vector of . The -Orthogonal Vectors problem (-OV) asks whether there exist such that are orthogonal.
Worst-case complexity.
The naive algorithm solves -OV in time . For any fixed constant , the algorithms by Abboud et al. [1] and Chan and Williams [8] solve OV in dimension in time with . However, Gao et al. [13] conjecture that no such strongly subquadratic algorithm exists for superlogarithmic dimension . This conjecture (known as Low-dimension Orthogonal Vector Conjecture) is also implied by SETH [29]. Both the upper bound for and the SETH-implied hardness for generalize to -OV, for any constant , where the running time barrier is [27].
Average-case complexity.
For cryptographic purposes we care about average-case hardness – because we want to be able to efficiently sample instances that are hard to solve (in contrast to only having an existential knowledge that there are some hard instances). Moreover, the sampler shall correctly tell (with good probability) whether its output is a yes- or no-instance.
One way to achieve this is to embed a solution in an instance sampled from a distribution that generates no-instances with good probability. This method of planting a solution has been applied to a number of problems, e.g., -Clique [18] and (in the fine-grained setting) -SUM [12, 2] (a generalization of 3SUM) and Zero--Clique [23] (a generalization of Zero-Triangle, which is a problem harder than both 3SUM and APSP), but not for (-)OV. The following question remains wide-open [6, 11, 10]:
How to plant orthogonal vectors (so that they are hard to find)?
1.2 Our results
We propose a way of planting a solution in -OV instances where each vector entry is i.i.d. according to a -biased111We say that a random bit is -biased if it equals with probability and equals with probability . coin flip, for an appropriately chosen value of so that the planted solution is the only one in the instance, with good probability. We conjecture that solving these instances requires time on average.
Let us remark that all our results are already nontrivial for , i.e., for the Orthogonal Vectors problem. However, from the point of view of cryptographic applications, larger values of are more interesting (as they potentially offer a bigger advantage for the honest parties), so we present all our results in full generality.
Superlogarithmic dimension.
The -OV problem might have appeared as a poor candidate for a fine-grained average-case hard problem, as Kane and Williams [19] showed that for any fixed , -OV instances of i.i.d. -biased entries can be solved in time for some by AC0 circuits. However, such instances are only nontrivial for ,222For larger (resp. smaller) , almost all instances will be no-instances (resp. yes-instances). a parameter setting which can be anyway solved in time , even in the worst case, using the algorithm of Chan and Williams [8]. To obtain a candidate hard distribution based on i.i.d. entries, we therefore choose to sample the entries as with subconstant probability , which leads to nontrivial instances in the superlogarithmic dimension regime . In the full version of the paper we present another simple argument why a logarithmic dimension is not sufficient, further justifying our choice.
The -wise independence.
Our planting procedure has the following notable property: any subset of (or less) out of the vectors that form the planted solution has the marginal distribution identical to that of independently sampled vectors with i.i.d. -biased random entries. In particular, each individual vector of the solution has the same marginal distribution as any other vector in the instance. This would not be true if we planted random vectors conditioned on orthogonality (i.e., a type of solution that may appear spontaneously with small probability), because such vectors tend to be sparser than the expectation. This sparsity is what makes the Kane–Williams algorithm [19] work, and lack thereof makes our instances immune to that algorithmic idea.333Note though that the Kane–Williams algorithm does not run in truly subquadratic time in superlogarithmic dimension anyway, so above all it is the high dimension, not the -wise independence, that makes our distribution immune to all known attacks.
We note that the -wise independence property holds “for free” in natural distributions for -SUM [2, 12] and Zero--Clique [23] because of the natural symmetry of cyclic groups . However, it is a priori unclear how to get it for -OV.
Finally, in Theorem 7, we argue that our distribution is the unique distribution over -OV instances that has this property, explaining the title of this paper.
Search-to-decision reductions.
To demonstrate the usefulness of the -wise independence property, we give a fine-grained average-case search-to-decision reduction for our conjectured hard -OV distribution. Actually, we give two such reductions. The first one, in Section 6, is very simple, but it introduces an overhead in the failure probability, so it is relevant only if the decision algorithm succeeds with probability higher than . The other reduction, in Section 7, looses only a constant factor in the failure probability. Even though we present both reductions specifically for -OV, they are likely to generalize to any planted problem with the -wise independence property.
Planting multiple solutions.
In the full version of the paper, we also argue that -wise independence allow planting more than one solution in a single instance, which we believe might be useful for building cryptographic primitives.
1.3 Technical overview
Planting.
How do we generate orthogonal vectors such that any of them look innocent? First of all, we can focus on generating a single coordinate, and then repeat the process independently for each of the coordinates. Consider the joint distribution of i.i.d. -biased random bits. We need to modify it to set the probability of ones to . If we just do it, and scale up the remaining probabilities accordingly, the probability of ones turns out wrong. After we fix that, the probability of ones is off, and so on, in a manner similar to the inclusion-exclusion principle. By doing this mental exercise we end up with a formula for the joint distribution of bits in a single coordinate of the vectors to be planted. How do we actually sample from this distribution? Since it has the -wise independence property, the following approach must work: First sample i.i.d. -biased bits, and then sample the -th bit with probability depending on the number of ones among the first bits. In Section 3 we show how to set this last probability exactly.
Search-to-decision reductions.
Both our reductions are based on the same basic idea: In order to find the planted solution, we replace some of the vectors in the input instance with newly sampled vectors with i.i.d. -biased entries and run the decision algorithm on such a modified instance. If at least one of the planted vectors got resampled, the resulting instance has the same distribution as if no planting occurred (thanks to the -wise independence), and so the decision algorithm returns no with good probability. Otherwise the planted solution is still there and the decision algorithm likely says yes.
Our first reduction (see Section 6) applies this idea to perform a binary search. It introduces a factor of overhead in the running time and also in the failure probability, because we need to take a union bound over all invocations of the decision algorithm returning correct answers.
Our second reduction (see Section 7) is an adaptation of a search-to-decision reduction for -SUM due to Agrawal et al. [2]. In short, the reduction repeatedly resamples a random subset of vectors, runs the decision algorithm, and keeps track for each of the original vectors, how many times the decision algorithm returned yes when this vector was not resampled. Statistically, this count should be larger for vectors in the planted solution. A careful probabilistic analysis shows that this is indeed the case.
1.4 Open problems
Hardness self-amplification.
Could a black-box method increase the success probability of algorithms solving (the search or decision variants of) the planted -OV problem, at a small cost in the running time? If so, the lack of algorithms with high success probability for planted -OV would then suggest that no algorithm can solve the problem even with just a small success probability – a property desirable, e.g., from the point of view of potential cryptographic applications.
Such hardness self-amplification was recently shown for (both the search and decision variants of) the planted clique problem by Hirahara and Shimizu [15]. In the world of fine-grained complexity, Agrawal et al. [2] showed hardness self-amplification for the planted -SUM search problem and closely related problems. Hardness self-amplification for planted -OV remains an open problem.
Because of the overhead in the failure probability induced by both of our search-to-decision reductions, hardness self-amplification for the decision variant of the planted -OV problem in particular would mean that the hardness of the search problem could be based on a weak conjecture about the hardness of the decision problem, such as Conjecture 10.
Fine-grained asymmetric cryptography.
A key goal of fine-grained cryptography is to devise an advanced asymmetric cryptography scheme – such as public key encryption – whose security is based on hardness of a well understood problem from fine-grained complexity. So far the closest to this goal seems to be the key exchange protocol due to LaVigne, Lincoln, and Vassilevska Williams [23], which is based on hardness of the planted Zero--Clique problem. Despite being based on a parameterized problem (that allows for arbitrary polynomial -hardness by simply choosing a large enough ), the protocol offers only quadratic security, i.e., breaking the encryption takes only quadratically more than it takes to encrypt and decrypt a message. This limitation seems inherent to the protocol because it is based on a similar idea as Merkle puzzles [24].
It is an open problem if fine-grained cryptography with superquadratic security is possible. We believe that -OV could be a good hard problem for that purpose, because of a different structure, which addition-based problems, like -SUM and Zero--Clique, are lacking.
We remark that the key exchange protocol of LaVigne, Lincoln, and Vassilevska Williams [23] can be adapted to work with our planted -OV instead of the planted Zero--Clique problem, but naturally the protocol’s security remains quadratic. One needs new techniques to break the quadratic barrier.
In recent work, Alman, Huang, and Yeo [5] show that if one-way functions do not exist then average-case hardness fine-grained assumptions on planted -SUM and Zero--Clique are false for sufficiently large constant . It might be possible to generalize their results to -OV. However, a construction of public-key encryption from planted -OV would be interesting even in a world where one-way functions do exist, as they are not known to imply superquadratic-gap public-key encryption [17].
Faster algorithms for average-case OV.
Algorithms for random OV instances seem to be underexplored. Up until recently [4] it was not known if the average-case OV admits even a subpolynomial improvement compared to the worst case. With this paper we hope to inspire more research in this direction. We would even be happy to see our conjecture refuted.
A natural starting point for such an attack is the recent Alman–Andoni–Zhang algorithm [4] for random OV. It works in time for dimension , so it is not truly subquadratic for . However, in their setting, the hardest case is when the probability of a one entry is chosen to make the expected number of orthogonal pairs a constant, while in our setting we use a higher value of to lower the expectation to inverse polynomial – which means that in our setting the orthogonal pair “stands out more”. It might seem plausible to adjust internal parameters of the algorithm (in particular, the so-called group size) to better exploit our setting. However, under closer inspection, it turns out that the key quantity in the analysis of the algorithm is the expected inner product of two random vectors, equal to , which happens to be in both settings. Refuting our conjecture likely requires a new technical development beyond such an adjustment.
Finally, let us point out a related problem: the planted approximate maximum inner product problem, often referred to as the light bulb problem. Unlike planted OV, it is known to admit truly subquadratic -time algorithms [26, 20, 21, 3]. This is in contrast with the worst-case complexities of the two problems, which are known to be equivalent under fine-grained reductions [9, 22].
A worst-case to average-case reduction.
In the opposite direction than the previous open problem, one could try to show that our conjectured hardness of the planted -OV problem is implied by one of well-studied worst-case hypotheses in fine-grained complexity, e.g., SETH. This would require a worst-case to average-case reduction. So far, in fine-grained complexity, such reductions are only known for algebraic or counting problems [6, 14, 7, 11, 10], but not for decision nor search problems like ours.
2 The model distribution
Fix , and let for . We define the family of model distributions that generate matrices where all entries are i.i.d. -biased bits with probability
As will become apparent later, for the planting algorithm to work it is crucial that , but thanks to this holds for large enough .
We show that the model distribution indeed generates no-solutions with good probability.
Lemma 1.
A -OV instance sampled from the model distribution is a no-instance with probability at least .
Proof.
For a -OV instance , a fixed combination of vectors (where ) is orthogonal iff, for each coordinate , not all of the vectors feature a one in that coordinate. Since i.i.d. -biased bits are all ones with probability , the probability that are orthogonal (determined by the all-ones event not occurring in any of the coordinates) is:
By linearity of expectation, the expected value for the number of solutions among all possible combinations of vectors, denoted by , is
By Markov’s inequality, this is also a bound on the probability of any solution occurring, i.e., . Therefore, an instance sampled from is a no-instance with probability at least .
We remark that one can make the probability of sampling a no-instance arbitrarily high. Indeed, in order to get the probability it suffices to replace with in the formula for the probability parameter . However, having in mind the cryptographic motivation, seems to be a reasonable default choice for the failure probability of the sampler, because with the same probability the attacker can just guess the solution.
3 The planted distribution
To plant a solution at locations in an instance sampled from , we apply the following randomized algorithm.
.
-
1.
For each coordinate :
-
(a)
Let be the number of ones among .
-
(b)
If is even, flip with probability . (Here we need .)
-
(a)
-
2.
Return .
We justify this way of planting in Section 4. For now, observe that if all vectors feature a one at coordinate , we have and flips the final bit to a zero with probability
On the other hand, if the last coordinate is the single zero alongside ones, then is odd and will never break orthogonality by flipping the last bit to a one. Thus, outputs a yes-instance of -OV with a solution at . We call the vectors at these positions the planted vectors.
We sample yes-instances of -OV by planting a solution in an instance at locations chosen uniformly at random.
Distribution .
-
1.
Sample from .
-
2.
Sample uniformly at random from .
-
3.
Return .
The above observation about immediately yields the following.
Lemma 2.
A -OV instance sampled from the planted distribution is a yes-instance with probability .
4 The -wise independence of planted vectors
Our method of planting orthogonal vectors arises from the idea that for any planted problem, any proper “piece” of the planted solution should be indistinguishable from any comparable piece of the instance as a whole, conditioned on the latter still being consistent with being a part of a solution itself.
For example, in the case of planting a -clique in a graph this requirement is trivial. Indeed, the projection of the clique onto a smaller subset of vertices yields a -clique, which are exactly those subgraphs of of size which could feasibly belong to a solution.
In contrast to the previous example, in the case of -SUM, any set of elements in an instance could feasibly be part of a solution, as one can always construct a th number such that . Thus, by the principle we described, to plant a solution in an instance with i.i.d. uniformly random elements, the marginal distribution of the distribution of planted solutions given by any projection to elements should itself be uniformly random. This holds true in the case of the planted -SUM [2], where the planted solution is distributed uniformly over the set of all -tuples that form valid -SUM solutions. The case of planted Zero--Clique [23] is analogous. For both of these problems, planting by inserting elements drawn from the model distribution conditioned on them forming a solution yields a distribution that follows the described principle.
This is different from the -OV problem with a model distribution of i.i.d. vector entries. Here, as with -SUM and Zero--Clique, any set of elements (in this case vectors) could form a solution to -OV. All that is needed is for the last vector to feature a zero in all those coordinates where the other all were one. However, sparse vectors are far more likely to be part of a solutions than dense ones. Therefore, conditioning i.i.d. -biased vectors on being orthogonal yields a distribution which does not follow our principle: projecting onto any subset of vectors results in vectors that are on average sparser than (and thus different from) i.i.d. -biased vectors. As we will show now, our method of planting does satisfy this principle: Any subset of planted vectors are independent and identically distributed -biased vectors.
Let and . Recall that both sampling from the model distribution and the planting by are independent and identical for each coordinate . Hence, all -bit sequences , for all , are independent and identically distributed, according to a distribution whose probability density function we denote by .
Lemma 3.
Let and let be the number of ones in . Then
Proof.
Fix a coordinate . Let be the random variable denoting the entries of the -th coordinate among the vectors at locations before planting. We proceed by case distinction.
Case 1.
If is even, the probability of occurring in the given coordinate of the planted solution is given by
Case 2a.
If is odd and for some , then may occur either directly in the model instance, or by (for which is even) occurring in the model instance and flipping the final bit:
Case 2b.
Similarly, if is odd and for some , then can occur either directly in the model instance or by flipping the final bit of the sequence :
Remark 4.
Despite acting only on the last collection , Lemma 3 implies that the resulting distribution is invariant under permutation of the sequence of the collections .
Having as the distribution of planted vectors, rather than, e.g., the -vector joint model distribution conditioned on orthogonality, ensures -wise independence among the planted vectors. I.e., the projection of planted vectors onto any subset of size is identically distributed to vectors from the model distribution.
Lemma 5 (-wise independence).
Marginalizing any one of the bits of yields independent -biased bits.
Proof.
By Remark 4 we may assume w.l.o.g. that the last bit is the one marginalized out. The lemma then follows from the definition of , as the first entries of any coordinate in the planted vectors are unchanged from the model instance, and are therefore independent -biased bits.
This property is useful in bounding the probability of a planted instance containing a solution besides the planted one.
Lemma 6.
A -OV instance sampled from the planted distribution has more than one solution with probability less than .
Proof.
While the vectors at positions are guaranteed to form a solution, by -wise independence, all combinations of of these vectors and non-planted vectors form a set of independent -biased vectors which is therefore a solution to the -OV problem with probability . By linearity of expectation,
and the claim follows from Markov’s inequality.
4.1 Uniqueness
Our way of planting is unique in the following sense.
Theorem 7.
Let be a probability distribution such that and that marginalizing any one of the bits yields independent -biased bits. Then .
Proof.
We show that for all . Let denote the number of ones in . We proceed by induction over , i.e., the number of zeros in .
Base case: .
Then and . Thus .
Inductive case: .
We assume w.l.o.g. that the zeros are the last bits of , i.e., . Marginalizing the final bits of yields independent -biased bits, whereby the probability of all remaining bits being ones is
Thereby,
| (by the induction hypothesis) | ||||
| where the sum term is merely the probability of in the marginal distribution of , which by Lemma 5 in turn consists of independent -biased bits. Hence, | ||||
5 Conjectured hard problems
In this section we formally define the problems that we conjecture to require time.
Definition 8 (Solving planted decision -OV).
Let be an algorithm that given a -OV instance outputs either 0 or 1. For , we say solves the decision problem with success probability , if for both and large enough ,
where randomness is taken over both the instance and the random coins used by .
Similarly, we define a notion of recovering a solution from a planted instance.
Definition 9 (Solving planted search -OV).
Let be an algorithm that given a -OV instance outputs a tuple . For a given , we say solves the planted search problem with success probability if for large enough ,
where randomness is taken over both the instance and the random coins used by .
Now we are ready to formally state our main conjecture.
Conjecture 10.
For any and , there exists no algorithm that solves the planted decision -OV problem with any constant success probability in time .
6 Search-to-decision reduction via binary search
We reduce the search problem of finding the planted solution to the decision problem of determining whether an instance contains a planted solution. This means that given a decision algorithm that can correctly distinguish whether an instance was sampled from the model or planted distribution with sufficient probability, one can recover the planted secret through this reduction. The reduction introduces a factor increase in both the running time and error probability of the algorithm.
The idea is to find each planted vector using something akin to binary search on each collection . We can split into two partitions of roughly equal size and run the decision algorithm twice, on instances where one of the two partitions is first replaced by newly sampled -biased vectors. The vector planted in is guaranteed to be replaced in one of these cases, and by -wise independence the resulting instance follows the model distribution. The search space is thus cut in half and we can recurse on this smaller search space to eventually find the planted vector.
Theorem 11 (Search-to-decision reduction).
Let and let be an algorithm that solves the planted decision -OV problem with success probability in time . Then there exists an algorithm that solves the planted search -OV problem with success probability at least in expected time .
Proof.
Consider an instance . First let us focus only on recovering the location of the planted vector in the first collection . The reduction begins with the “full” search space , and narrows it down by half in each iteration, so that the desired is recovered after iterations.
At each iteration, the current search space is arbitrarily partitioned in two sets of equal size (up to one vector when is odd). The decision algorithm is then executed on two new instances, where the respective sets of vectors in are replaced with newly sampled -biased vectors.
By the -wise independence, if the vector belonging to the solution is replaced, all vectors are independently and identically distributed -biased vectors, i.e., the instance is distributed according to . On the other hand, if the solution survives resampling, the instance remains distributed according to . Therefore, the output of is used to decide which of the two partition blocks should be assumed as the new search space.
The reduction is correct if decides correctly at every iteration. Of course, might fail with probability . By a union bound over all invocations of this happens with probability at most . Thereby recovers the location of the first planted vector with success probability at least .
As for the runtime, with runtime is invoked times, and across all iterations vectors are resampled in total. Since a single -biased bit can be sampled in expected time , sampling a -dimensional vector takes time in expectation. Therefore, recovering the location of the first planted vector takes time .
The same process is repeated another times to recover the locations of the planted vectors among . As is constant, this does not increase the running time asymptotically but the success probability drops to .
7 Search-to-decision reduction via counters
We present a second search-to-decision reduction, adapted from that of Agrawal et al. [2] for planted -SUM. As in the method in Section 6, we use the fact that an algorithm for the decision -OV problem, when given a planted -OV instance with some of the vectors resampled, correctly detects whether any of the planted vectors were among the resampled vectors. However, instead of iteratively narrowing a pool of candidate vectors, we iterate this process on the entire instance, and, for each vector , we keep count of the number of iterations in which survived and ’s output was 1. After iterations we output the vectors with the highest counts among each of the collections, which, as we show, coincides with the planted solution (with good probability).
Theorem 12 (Search-to-decision reduction).
For any , if there exists an algorithm that solves the planted decision -OV problem with success probability at least in time , then there exists an algorithm that solves the planted search -OV problem with success probability at least in expected time .
In more detail, let be the following randomized algorithm, which takes a -OV instance and resamples some of the vectors:
Algorithm .
-
1.
For each and :
-
(a)
With probability , replace by a newly-sampled -biased vector
-
(a)
-
2.
Output
For , let indicate the indices of the vectors of which are replaced by . For a vector in and a given execution of , we say survives if does not replace . We say survives if survives for each , i.e., .
Now, let be the binary random variable indicating whether survives. Our chosen probability for to resample a vector yields the following.
Lemma 13.
For a -OV instance and any , survives with probability one half, i.e., .
Proof.
Each vector is independently picked to be replaced with probability . The chance of all vectors surviving is
As laid out before, our search algorithm repeatedly executes and then the given decision algorithm . The hope is that the output of the latter correlates with the survival of the planted solution. We therefore also keep track of which vectors survive whenever believes there is still a planted solution in the instance output by .
Algorithm .
-
1.
Initialize a counter for each and .
-
2.
Repeat times:
-
(a)
-
(b)
-
(c)
If :
-
i.
Set for every that was not replaced by
-
i.
-
(a)
-
3.
Set for each
-
4.
Output
This works well for instances where is good at detecting whether a particular solution survives. To capture this notion, we say an instance is good, if it has only one solution, at some location , and the output of indicates whether survives except for a small constant probability, i.e.,
where the probability is taken over the internal randomness used by .
In the following, let be a randomized algorithm that outputs a planted instance sampled from as well as the location of the planted solution.
Lemma 14.
If solves the planted decision -OV problem with success probability at least , an instance is good except with probability at most .
Proof.
An instance contains only a single solution except with probability less than . We will now show that for such with only the planted solution, the decision algorithm correctly detects if this solution survives except with probability , from which the claim follows by a union bound.
First, consider the following distribution:
Distribution .
-
1.
-
2.
-
3.
Output
Observe that, if we condition on , the planted solution survives , which merely resamples some of the i.i.d. -biased vectors in the instance. Thus, is distributed according to . On the other hand, if , at least one of the planted vectors is replaced by a newly-sampled -biased vector. By -wise independence, the subset of planted vectors that survive are i.i.d. -biased vectors, as are all other vectors in . Hence, conditioning on yields the model distribution .
Therefore, for the algorithm , which solves the planted decision -OV problem with success probability at least , we have
Now, let be the random variable that, for a given instance with a solution planted at , denotes the probability of , where the randomness is taken over the internal coins used by . Then
By Markov’s inequality,
Next, observe that is the only solution in the instance output by with good probability,
Thus, by a union bound over this and our result in the first step, we find that an instance is good except with probability at most :
We now show that the search algorithm performs well on this large fraction of good instances.
Lemma 15.
Let be an algorithm that solves the planted decision -OV problem with success probability . Then fails to recover the solution of good instances with probability less than .
Proof.
Let be a good instance with its only solution at . After iterations, we expect the counters for vectors in the solution to be the highest. Using the fact that , we find that for non-planted vectors, i.e., where ,
| () | ||||
| On the other hand, for planted vectors, i.e., where , | ||||
| () | ||||
Picking the highest counters is guaranteed to yield the solution if the ranges of the counters of the planted and non-planted vectors do not overlap, that is to say if no counter deviates from its expected value by half the difference of (the bounds on) the two expected values ( ‣ 7) and ( ‣ 7), which we denote by :
Each counter is the sum of i.i.d. binary random variables. By a Chernoff bound, a counter for a vector which is not in the planted solution, i.e., , exceeds its expected value by with probability at most
Similarly, a counter for a vector that is part of a planted solution () falls short of its expected value by with at most the same probability
For a choice of iterations, both of these probabilities are at most . If none of the counters deviate by , the ranges of counters for vectors at and those for vectors not at are disjoint and selecting for the highest counters is guaranteed to yield . Thus, by a union bound over all counters, we fail to recover with probability at most .
We can now complete the proof of Theorem 12.
Proof of Theorem 12.
By Lemma 14, an instance is good except with probability at most . By Lemma 15, is able to recover the solution from a good instance except with probability at most . By a union bound against the instance not being good or failing on a good instance, solves the planted search -OV problem with success probability at least .
In each of the iterations, we execute both and once and update the counters. Each execution of samples, in expectation, new vectors, each of which can be sampled in expected time as explained in the proof of Theorem 11. runs in time and updating the counters takes linear time, for the total expected runtime of .
References
- [1] Amir Abboud, Richard Ryan Williams, and Huacheng Yu. More applications of the polynomial method to algorithm design. In Proceedings of the Twenty-Sixth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2015, pages 218–230. SIAM, 2015. doi:10.1137/1.9781611973730.17.
- [2] Shweta Agrawal, Sagnik Saha, Nikolaj I. Schwartzbach, Akhil Vanukuri, and Prashant Nalini Vasudevan. k-SUM in the sparse regime: Complexity and applications. In Advances in Cryptology – CRYPTO 2024 – 44th Annual International Cryptology Conference, volume 14921 of Lecture Notes in Computer Science, pages 315–351. Springer, 2024. doi:10.1007/978-3-031-68379-4_10.
- [3] Josh Alman. An illuminating algorithm for the light bulb problem. In 2nd Symposium on Simplicity in Algorithms, SOSA 2019, volume 69 of OASIcs, pages 2:1–2:11. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/OASICS.SOSA.2019.2.
- [4] Josh Alman, Alexandr Andoni, and Hengjie Zhang. Faster algorithms for average-case orthogonal vectors and closest pair problems. In 2025 Symposium on Simplicity in Algorithms (SOSA), pages 473–484. SIAM, 2025. doi:10.1137/1.9781611978315.35.
- [5] Josh Alman, Yizhi Huang, and Kevin Yeo. Fine-grained complexity in a world without cryptography. In Advances in Cryptology – EUROCRYPT 2025 – 44th Annual International Conference on the Theory and Applications of Cryptographic Techniques, volume 15607 of Lecture Notes in Computer Science, pages 375–405. Springer, 2025. doi:10.1007/978-3-031-91098-2_14.
- [6] Marshall Ball, Alon Rosen, Manuel Sabin, and Prashant Nalini Vasudevan. Average-case fine-grained hardness. In Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2017, pages 483–496. ACM, 2017. doi:10.1145/3055399.3055466.
- [7] Enric Boix-Adserà, Matthew S. Brennan, and Guy Bresler. The average-case complexity of counting cliques in Erdős-Rényi hypergraphs. In 60th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2019, pages 1256–1280. IEEE Computer Society, 2019. doi:10.1109/FOCS.2019.00078.
- [8] Timothy M. Chan and R. Ryan Williams. Deterministic APSP, orthogonal vectors, and more: Quickly derandomizing Razborov-Smolensky. ACM Trans. Algorithms, 17(1):2:1–2:14, 2021. Announced at SODA 2016. doi:10.1145/3402926.
- [9] Lijie Chen and Ryan Williams. An equivalence class for orthogonal vectors. In Proceedings of the Thirtieth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2019, pages 21–40. SIAM, 2019. doi:10.1137/1.9781611975482.2.
- [10] Mina Dalirrooyfard, Andrea Lincoln, Barna Saha, and Virginia Vassilevska Williams. Average-case hardness of parity problems: Orthogonal vectors, k-SUM and more. In Proceedings of the 2025 Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2025, pages 4613–4643. SIAM, 2025. doi:10.1137/1.9781611978322.158.
- [11] Mina Dalirrooyfard, Andrea Lincoln, and Virginia Vassilevska Williams. New techniques for proving fine-grained average-case hardness. In 61st IEEE Annual Symposium on Foundations of Computer Science, FOCS 2020, pages 774–785. IEEE, 2020. doi:10.1109/FOCS46700.2020.00077.
- [12] Itai Dinur, Nathan Keller, and Ohad Klein. Fine-grained cryptanalysis: Tight conditional bounds for dense k-SUM and k-XOR. J. ACM, 71(3):23, 2024. Announced at FOCS 2021. doi:10.1145/3653014.
- [13] Jiawei Gao, Russell Impagliazzo, Antonina Kolokolova, and Ryan Williams. Completeness for first-order properties on sparse structures with algorithmic applications. ACM Trans. Algorithms, 15(2):23:1–23:35, 2019. Announced at SODA 2017. doi:10.1145/3196275.
- [14] Oded Goldreich and Guy N. Rothblum. Counting t-cliques: Worst-case to average-case reductions and direct interactive proof systems. In 59th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2018, pages 77–88. IEEE Computer Society, 2018. doi:10.1109/FOCS.2018.00017.
- [15] Shuichi Hirahara and Nobutaka Shimizu. Hardness self-amplification: Simplified, optimized, and unified. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, pages 70–83. ACM, 2023. doi:10.1145/3564246.3585189.
- [16] Russell Impagliazzo, Ramamohan Paturi, and Francis Zane. Which problems have strongly exponential complexity? J. Comput. Syst. Sci., 63(4):512–530, 2001. Announced at FOCS 1998. doi:10.1006/JCSS.2001.1774.
- [17] Russell Impagliazzo and Steven Rudich. Limits on the provable consequences of one-way permutations. In Proceedings of the 21st Annual ACM Symposium on Theory of Computing, pages 44–61. ACM, 1989. doi:10.1145/73007.73012.
- [18] Ari Juels and Marcus Peinado. Hiding cliques for cryptographic security. Designs, Codes and Cryptography, 20(3):269–280, 2000. Announced at SODA 1998.
- [19] Daniel M. Kane and Richard Ryan Williams. The orthogonal vectors conjecture for branching programs and formulas. In 10th Innovations in Theoretical Computer Science Conference, ITCS 2019, volume 124 of LIPIcs, pages 48:1–48:15. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPICS.ITCS.2019.48.
- [20] Matti Karppa, Petteri Kaski, and Jukka Kohonen. A faster subquadratic algorithm for finding outlier correlations. ACM Trans. Algorithms, 14(3):31:1–31:26, 2018. Announced at SODA 2016. doi:10.1145/3174804.
- [21] Matti Karppa, Petteri Kaski, Jukka Kohonen, and Padraig Ó Catháin. Explicit correlation amplifiers for finding outlier correlations in deterministic subquadratic time. Algorithmica, 82(11):3306–3337, 2020. Announced at ESA 2016. doi:10.1007/S00453-020-00727-1.
- [22] Karthik C. S. and Pasin Manurangsi. On closest pair in euclidean metric: Monochromatic is as hard as bichromatic. In 10th Innovations in Theoretical Computer Science Conference, ITCS 2019, volume 124 of LIPIcs, pages 17:1–17:16. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2019. doi:10.4230/LIPICS.ITCS.2019.17.
- [23] Rio LaVigne, Andrea Lincoln, and Virginia Vassilevska Williams. Public-key cryptography in the fine-grained setting. In Advances in Cryptology – CRYPTO 2019 – 39th Annual International Cryptology Conference, volume 11694 of Lecture Notes in Computer Science, pages 605–635. Springer, 2019. doi:10.1007/978-3-030-26954-8_20.
- [24] Ralph C. Merkle. Secure communications over insecure channels. Commun. ACM, 21(4):294–299, 1978. doi:10.1145/359460.359473.
- [25] Alon Rosen. Fine-grained cryptography: A new frontier? IACR Cryptol. ePrint Arch., page 442, 2020. URL: https://eprint.iacr.org/2020/442.
- [26] Gregory Valiant. Finding correlations in subquadratic time, with applications to learning parities and the closest pair problem. J. ACM, 62(2):13:1–13:45, 2015. Announced at FOCS 2012. doi:10.1145/2728167.
- [27] Virginia Vassilevska Williams. On some fine-grained questions in algorithms and complexity, pages 3447–3487. World Scientific, 2018. doi:10.1142/9789813272880_0188.
- [28] R. Ryan Williams. Some estimated likelihoods for computational complexity. In Computing and Software Science – State of the Art and Perspectives, volume 10000 of Lecture Notes in Computer Science, pages 9–26. Springer, 2019. doi:10.1007/978-3-319-91908-9_2.
- [29] Ryan Williams. A new algorithm for optimal 2-constraint satisfaction and its implications. Theor. Comput. Sci., 348(2-3):357–365, 2005. doi:10.1016/J.TCS.2005.09.023.
