Abstract 1 Introduction 2 Preliminaries 3 Evaluating Integer-Valued Functions 4 Approximating Real-Valued Functions References

Hardness Amplification for Real-Valued Functions

Yunqi Li ORCID Department of Computer Science, National University of Singapore, Singapore Prashant Nalini Vasudevan ORCID Department of Computer Science, National University of Singapore, Singapore
Abstract

Given an integer-valued function f:{0,1}n{0,1,,m1} that is mildly hard to compute on instances drawn from some distribution D over {0,1}n, we show that the function g(x1,,xt)=f(x1)++f(xt) is strongly hard to compute on instances (x1,,xt) drawn from the product distribution Dt. We also show the same for the task of approximately computing real-valued functions f:{0,1}n[0,m). Our theorems immediately imply hardness self-amplification for several natural problems including Max-Clique and Max-SAT, Approximate #SAT, Entropy Estimation, etc..

Keywords and phrases:
Average-case complexity, hardness amplification
Copyright and License:
[Uncaptioned image] © Yunqi Li and Prashant Nalini Vasudevan; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Computational complexity and cryptography
Funding:
Both authors are supported by the National Research Foundation, Singapore, under its NRF Fellowship programme, award No. NRF-NRFF14-2022-0010.
Related Version:
Full Version: https://eccc.weizmann.ac.il/report/2025/087/
Editors:
Srikanth Srinivasan

1 Introduction

Hardness amplification is the process of taking a computational problem Π and a distribution D of instances over which Π is mildly hard, and constructing a problem Π and distribution D over which Π is much harder. There has been extensive work in the past studying hardness amplification for various computational tasks such as computing Boolean functions [26, 17, 12, 19, 18, 5], inverting efficient functions [26, 11], distinguishing between distributions [9], deciding languages contained in complexity classes like 𝖭𝖯 [21, 15, 23, 14, 6], 𝖤𝖷𝖯 [24], #𝖯 [20, 8], or 𝖯 [4, 13, 7], solving optimization problems [10], for specific interesting or structured problems [3, 16, 1], etc..

In this paper, we study hardness amplification for the task of evaluating integer- or real-valued functions. We first describe our setting and results, and then demonstrate various corollaries of our theorems that motivate studying such problems.

Evaluating Functions

In our first result, we consider functions f:{0,1}nm, where m denotes the set of integers {0,1,,m1}, and the computational task is to compute f(x) given an input x{0,1}n. Here, we only use m to denote set without involving any modulo operations. Suppose we are given such an f and a distribution D on {0,1}n over which computing f is (1δ)-hard – meaning that no small circuit can correctly compute f(x) with probability greater than (1δ) when x is drawn from D. Our objective is to construct a function g and distribution D such that computing g is η-hard over D for some small η. Further, we would like g to have the same type as f – to take bit-strings as input and produce integers from a bounded range as output.

For Boolean functions, Yao’s XOR Lemma [26] shows that such amplification can be achieved by having g be the XOR of multiple instances of f; and similar results are also known if g is the Recursive Majority-of-3 of such instances [21]. We show that in this case of integer-valued functions, having g be the sum (over integers) of multiple instances of f achieves the same. For t, denote by (𝖲𝖴𝖬tf) the function that takes t inputs x1,,xt{0,1}n, and outputs the sum if(xi).

Theorem 1.1 (Simplification of Theorem 3.1).

Suppose a function f:{0,1}nm is (1δ)-hard to compute over a distribution D for circuits of size s. Then, for t, the function (𝖲𝖴𝖬tf) is (2mtδ)-hard to compute over the product distribution Dt for circuits of size (cm2t2δ2log(t))s, as long as t>cm2δ, where c and c are some universal constants.

In a typical application of this theorem (see Section 1.1 for examples), one might take δ to be a small constant, m to be some polynomially large value in n, assume (1δ)-hardness for s being any arbitrary polynomial in n, and set t to be ω(m2) but still some polynomial in n. The theorem would then imply that (𝖲𝖴𝖬tf) is 1/poly(n)-hard for all polynomial-sized circuits.

To place the O(m/tδ)-hardness we obtain in context, observe that it is not possible to show that summation generically amplifies such a function to hardness less than Θ(1/tδ). Consider, for example, a hypothetical function that takes values in {0,1}, is easy to compute on (12δ) fraction of inputs, and on the remaining 2δ is optimally hard (so it is not possible to do better than random guessing). This function is (1δ)-hard. Given t random inputs from the hard distribution, the hardness comes only from about δt of the inputs. And simply guessing randomly on each of these will yield the correct value for the sum of their outputs with probability Ω(1/tδ).

Approximating Functions

In our second result, we extend the above hardness amplification to the task of approximately evaluating bounded real-valued functions. Here we consider functions f:{0,1}n[0,m), and the task is, for some approximation parameter ϵ+, to compute some value in the range (f(x)ϵ,f(x)+ϵ) given input x. We show that summation again amplifies hardness, though with slightly different dependence on the various parameters, and also now depending on the ϵ.

Theorem 1.2 (Simplification of Theorem 4.1).

Suppose a function f:{0,1}n[0,m) is (1δ)-hard to ϵ-approximate over a distribution D for circuits of size s. Then, the function (𝖲𝖴𝖬tf) is 10(mϵtδ)1/2-hard to ϵ-approximate over the product distribution Dt for circuits of size (cmϵ(tδ)3/2log(t))s, as long as t>cm2ϵ2δ, where c and c are some universal constants.

Here too, in the applications we show, parameters are set as described for Theorem 1.1 earlier, with ϵ=Θ(1).

Paper Outline

In the rest of this section, we describe various corollaries of the above theorems (Section 1.1) and provide an overview of the proofs of these theorems (Section 1.2). In Section 2, we set up the definitions, conventions, and hardcore lemmas needed in the rest of the paper. In Sections 3 and 4, we state and prove more comprehensive versions of Theorems 1.1 and 1.2, respectively.

1.1 Corollaries

Functions mapping bit-strings to integers or real numbers, even within limited ranges, are quite general and capture a variety of natural problems whose complexity is of significant interest – essentially any problem Π whose solution Π(x) for an instance x is a bounded integer or real number. For such problems, our results roughly say that computing the sum of solutions to t instances is a much harder problem. Such a statement is not particularly meaningful in general, but things become much more interesting if the problem Π also happens to admit an additively homomorphic self-reduction.

That is, suppose there is an efficient algorithm R such that for any inputs x1,,xt, we have Π(R(x1,,xt))=iΠ(xi). In this case, the problem of computing the sum of solutions of t instances can be reduced back to the solving the problem Π itself on a single instance. Then, our results can be used to show that mild hardness of Π implies strong hardness of Π itself, possibly on a different distribution over instances (in what is sometimes referred to as hardness self-amplification). And this requirement is weak enough that many natural and important problems have such self-reductions. Below, we show three examples, each of which is qualitatively distinct from the others.

Optimization Problems

Various natural optimization problems can be cast in terms of computing a polynomially bounded integer-valued function of the input, and further be shown to possess simple additively homomorphic self-reductions. For example, consider the Max-Clique problem where, given (the adjacency matrix of) a graph G, the task is to compute the size of its largest clique. Given graphs G1,,Gt, we can create a new graph consisting of one copy of each Gi, with edges between every pair of vertices that are not from the same graph. The size of the maximum clique in this composite graph is simply the sum of the maximums in all the Gi’s.

Another example is the 𝖬𝖺𝗑𝖲𝖠𝖳 problem, where given a CNF formula ϕ, the task is to find the maximum number of clauses satisfied by any assignment to its variables. Given t formulas ϕ1,,ϕt on disjoint sets of variables, the maximum number of clauses of the formula (ϕ1ϕt) that can be satisfied is simply the sum of the maximums of all the ϕi’s. Note that both of these problems also happen to be 𝖭𝖯-hard. We get the following from Theorem 1.1, following the arguments above.

Corollary 1.3.

The following holds for any problem Π{𝖬𝖺𝗑𝖲𝖠𝖳,𝖬𝖺𝗑𝖢𝗅𝗂𝗊𝗎𝖾}. If there is a family of distributions on which Π is 0.9-hard, then there is a family of distributions on which Π is O(n0.49)-hard (where n is the instance size). Further, if the former family is efficiently sampleable, then so is the latter.

In these corollaries, we consider the asymptotic hardness of these problems rather than for a fixed input size, which is why we need to consider a family of distributions – one distribution for each value of the instance size parameter n – rather than a single distribution. And by efficiently sampleable, we mean sampleable by a polynomial-sized family of circuits. Similarly, when we just say η-hard, we mean η-hard for families of circuits of size any polynomial in the instance size parameter n. More detailed and careful statements of all the corollaries are presented in the full version, along with sketches of their proofs. These are all asymptotic statements, and so involve applying our theorems (which are stated for arbitrary input lengths and circuit sizes) for all members of families of functions and circuits.

Hardness amplification for optimization problems, including the above examples, was studied in [10]. However, they considered the task of actually finding the maximum clique, the maximally satisfying assignment, etc., and their results are incomparable to the corollary above.

Entropy Estimation

A natural problem (that has incidentally been of some significance in cryptography [25]) is that of estimating the Shannon entropy of a distribution given its sampling algorithm (say, as a circuit). Given a distribution over {0,1}m, its entropy is some real number in the range [0,m], and Shannon entropy is also conveniently additive: for any random variables X and Y, we have H(X,Y)=H(X)+H(Y). We cannot use Theorem 1.1 here because the entropy is not integer-valued, but we can use Theorem 1.2 to show hardness amplification for the task of approximately computing the entropy of a given distribution.

In this case, we can in fact go further. A related decision problem, called the Entropy Difference problem, is known to be complete for the complexity class 𝖲𝖹𝖪, which consists of problems that possess statistical zero-knowledge proofs [22]. In this problem, given sampling algorithms for two distributions D0 and D1, and promised that their entropies are separated by a gap of at least 1, the task is to tell which distribution has larger entropy. If this problem is even mildly average-case hard over some distribution of instances (D0,D1), then the task of computing the entropy of distributions to within ±1/2 is mildly average-case hard for the distribution given by sampling (D0,D1) as above and randomly outputting one of the two distributions. These observations, together with Theorem 1.2, give us the following.

Corollary 1.4.

If there is a problem in 𝖲𝖹𝖪 that is 0.9-hard over some family of distributions, then there exists a family of distributions on which O(1)-approximating Shannon entropy is O(n0.24)-hard. Further, if the former family of distributions is efficiently sampleable, then so is the latter.

Approximate Counting

Given as input the description of a Non-deterministic Turing Machine M and an input x for it, define f(M,x) to be the number of accepting paths in the execution of M given input x. This function captures the defining problem of the complexity class #𝖯. The same can be done with the #𝖯-complete problem #𝖲𝖠𝖳, which is the problem of counting the number of satisfying assignments to a given Boolean formula.

In both these cases, however, the function can take exponentially many values in the instance size, and so trying to apply our theorems would not give meaningful bounds. However, if we instead consider the function g(M,x)=log2(f(M,x)), this function is real-valued and lies in a polynomially bounded range (as long as the formula is guaranteed to be satisfiable). Further, additive ±ε approximations to g are equivalent to multiplicative 2±ϵ approximations to f. Theorem 1.2 now implies the following.

Corollary 1.5.

Suppose there is a family of distributions over satisfiable Boolean formulas on which multiplicatively approximating the number of satisfying assignments to within a factor of 2 is 0.9-hard. Then there is a family of distributions on which the same task is O(n0.24)-hard (where n is the instance size). Further, if the former is efficiently sampleable, then so is the latter.

1.2 Technical Overview

In this section, we give an overview of our analysis of the hardness amplification of summation. We focus on hardness amplification of evaluating integer-valued functions (Theorem 1.1).

Hardness of {𝟎,𝟏}-valued functions

We will start by showing something even simpler – hardness amplification for functions that only take two different values. Our techniques here are inspired by ideas in the proof of existing hardness amplification theorems for problems in 𝖭𝖯 [21].

Suppose that there is a function f:{0,1}n{0,1}, and a corresponding distribution H over {0,1}n, on which the function f is strongly average-case hard; that is, there is some small γ=1/poly(n) such that for any circuit C of size at most s,

PrxH[C(x)=f(x)]<12(1+γ),

For simplicity, we assume that f is balanced over H. That is,

PrxH[f(x)=0]=PrxH[f(x)=1].

For some t, consider the function g:{0,1}tnt+1, where g(x1,,xt)=f(x1)++f(xt). In the following, we will show that the average-case hardness of g improves polynomially in t. Particularly, for any circuits of size at most approximately s, we have:

PrxHt[C(x)=g(x)]<(tt2)2t(1+t2γ). (1)

Before we show this, let us understand the best hardness we can hope to show. A simple algorithm for computing g is to just always output the value that g is most likely to take. Since f is assumed to be balanced over H, this value would be t2.

PrxHt[g(x)=t2]=(tt2)2tΘ(1t).

If the function f had been optimally (1/2)-hard, then this would also be the best possible algorithm for g. What we have is that f is (1+γ)/2-hard for some small γ, indicating that f is still almost optimally hard. We essentially show that in this case the above algorithm is still nearly the best possible algorithm g.

We start by observing that the hardness of f implies the indistinguishability of the distributions H conditioned on different outputs – H|f(x)=0 and H|f(x)=1. That is, for any circuit C of size at most s,

|PrxH[C(x)=1f(x)=1]PrxH[C(x)=1f(x)=0]|<γ. (2)

For function g, consider the performance of any circuit C^ for computing g. We claim that, for any kt and any v in the output range of C^, the following holds:

|PrxHt[C^(x)=vg(x)=k+1]PrxHt[C^(x)=vg(x)=k]|<γ. (3)

To prove it, assume

|PrxHt[C^(x)=vg(x)=k+1]PrxHt[C^(x)=vg(x)=k]|γ. (4)

The distribution xHt|g(x)=k is equivalent to the distribution sampled as follows:

  1. 1.

    Independently sample x1,,xkH|f(x)=1, and xk+1,,xtH|f(x)=0 independently

  2. 2.

    Sample a uniformly random permutation π from all possible permutations over t coordinates

  3. 3.

    Output π(x1,,xt)

Using this observation and the linearity of expectation, (4) implies that there must exist a fixed (x1,,xk,xk+2,,xt) and a permutation π, such that

|PrxH,x^H^(x)[C^(x^)=vf(x)=1]PrxH,x^H^(x)[C^(x^)=vf(x)=0]|γ.

where H^(x) denotes π(x1,,xk,x,xk+2,xt). Then, a circuit C:{0,1}n{0,1} can be constructed by taking x1,,xk,xk+2,,xt and permutation π as non-uniform advice and working as follows: on the input x{0,1}n, outputs 1 iff C^(π(x1,,x,,xt)) outputs v. The size of C is approximately the size of C^, and we have

|PrxH[C(x)=1f(x)=1]PrxH[C(x)=1f(x)=0]|γ,

which contradicts the hardness of f as captured by (2).

Based on (3) and a simple telescoping argument, we further obtain, for any i,jt+1 and any v in the output range of C^,

|PrxHt[C^(x)=vg(x)=i]PrxHt[C^(x)=vg(x)=j]|<|ij|γ. (5)

We can now bound the probability that a circuit C^ can correctly compute the function g as:

PrxHt[C^(x)=g(x)] =k=0tPrxHt[g(x)=k]PrxHt[C^(x)=k|g(x)=k]
=k=0t(tk)2tPrxHt[C^(x)=k|g(x)=k]
<k=0t(tk)2t(PrxHt[C^(x)=k|g(x)=t2]+|kt2|γ)
(tt2)2tk=0tPrxHt[C^(x)=k|g(x)=t2]+k=0t(tk)2t|kt2|γ
(tt2)2t(1+t2γ).

where the second line follows from the fact that f is balanced over H, the third line follows from (5), the fourth line from the maximality of the central binomial co-efficient, the fifth line from computing the sum of the series there, and the last line from the fact that the events in the probability expressions are disjoint.

The above approach to bounding the probability of computing the sum of t independent instances of a function whose value between two possible outputs is strongly hard to decide is at the core of the proofs of our results.

Reducing to two outputs

Since our amplification approach is based on the strong indistinguishability of distributions over the pre-image sets of two outputs, for any evaluation problem, we will identify such a pair of indistinguishable pre-image sets based on assumption that the evaluation problem is hard. For simplicity, we will take the distribution over which the problem is hard to be the uniform distribution over {0,1}n.

Consider a function f:{0,1}nm that is (1δ)-hard for circuits of size s. For every a,bm, ab, we define a computation problem in which we only consider the correctness on inputs whose output belongs to {a,b}. We formalize the above problem by defining the relations Ra,b over {0,1}n×m:

  • If f(x){a,b}, then (x,y)Ra,b if and only if y=f(x);

  • If f(x){a,b}, then (x,y)Ra,b for every ym.

For any x{0,1}n, denote by Ra,b(x) the set of ym such that (x,y)Ra,b. We show that the hardness of f implies that there must exist some abm and some distribution over their pre-image sets for which it is hard to distinguish whether f(x) is a or b. More precisely, we show that there exists a pair (a,b) such that Ra,b is (1δ/(m2))-hard for circuits of size sm2.

Amplifying using hardcore sets

Pick a relation Ra,b that has such hardness. As its hardness essentially comes from the hardness of deciding between two possible outputs a and b, we can extend existing proofs of the hardcore lemma for Boolean functions (e.g. that of [17]) to obtain a hardcore set111Actually, what we obtain are hardcore distributions, but we assume these are sets in this overview for simplicity. for Ra,b. This is a set H{0,1}n of density at least δ=δ/(m2) such that the relation Ra,b is (1+γ)/2-hard over random inputs from H for circuits of size roughly γ2s/m2. Further, we can ensure that all inputs xH are such that f(x){a,b}, and with only a small loss, we can also ensure that this set is balanced between the outputs a and b.

The rest of the argument is quite standard. Given t inputs x1,,xt sampled uniformly at random from {0,1}n, with high probability roughly at least a δ fraction of these will fall in H. Looking at just these δt inputs, we are essentially back in the case discussed at the beginning of this overview – that of a function that has two possible outputs, with inputs being sampled from a hard distribution balanced between these outputs. Applying the amplification arguments there to this subset of inputs, we get from (1) that computing the sum of the f(xi)’s for the xi’s that fall in H is roughly (1/δt+δtγ)-hard for circuits of size roughly γ2s/m2. In order to compute the sum of all the f(xi)’s, the sum corresponding to the above δt inputs needs to be computed. So this hardness carries over to computing 𝖲𝖴𝖬tf as well. Setting γ=1/δt now gives us Theorem 1.1.

2 Preliminaries

2.1 Notations

For m, denote the set {0,,m1} by m (note that this is just a set, not the ring of integers modulo m). For v,ϵ, we use [v±ϵ) to denote the interval [vϵ,v+ϵ) and use round brackets for open intervals and square brackets for closed intervals.

For distributions H,G over the same domain {0,1}n, for any integers 0kt, the symbol Πt(Hk,Gtk) stands for the following distribution: sample x1,,xk from H independently and sample xk+1,,xt from G independently, sample a permutation π over t entries uniformly randomly, and output π(x1,,xt).

For n,t and some alphabet Σ,Σ, given functions f:{0,1}nΣ and g:ΣtΣ, denote the function that outputs g(f(x1),,f(xt)) over input (x1,,xt) by gf. For a relation R𝒳×𝒴, for any x𝒳, we define R(x)={y:y𝒴(x,y)R}.

We represent real numbers using strings. In each case, the range of relevant real numbers is some [0,m) that will be clear from the context, and the string is to be interpreted as a fixed-point representation of numbers in that range. That is, for any y[0,m), y is evaluated as y12logm1+y22logm2++ylogm20+ylogm+121+, where yi{0,1} and y is represented by (y1,y2,).

2.2 Average-Case Hardness

The average-case hardness of a problem is defined with respect to a distribution over its input domain. We start by formally defining the hardness of functions and relations.

Definition 2.1 (Hardness of Evaluating Functions).

For any δ(0,1), n,l,s, consider a function f:{0,1}nΣ, where Σ is an output domain that can be encoded by {0,1}l and l=log|Σ|. For a distribution D over {0,1}n, f is called δ-hard on D for circuits of size s if, for any circuit C:{0,1}n{0,1}l of size at most s, we have

PrxD[C(x)=f(x)]<δ.
Definition 2.2 (Hardness of Satisfying Relations).

For any δ(0,1), n,l,s, consider a relation R{0,1}n×Σ, where Σ is an output domain that can be encoded by {0,1}l and l=log|Σ|. For a distribution D over {0,1}n, R is called δ-hard on D for circuits of size s, if for any circuit C:{0,1}n{0,1}l of size at most s,

PrxD[C(x)R(x)]<δ.
Definition 2.3.

For a function f:{0,1}n[0,m), the approximation problem with distance d is denoted by a relation Rfd{0,1}n×, where

Rfd={(x,y):|f(x)y|<d}.

Similarly, the closed approximation is defined by

R^fd={(x,y):|f(x)y|d}.
Definition 2.4 (Hardness of Approximating Functions).

For any δ(0,1), α,s and m,ϵ, consider a function f:{0,1}n[0,m). For a distribution D over {0,1}n, f is called δ-hard to approximate on D with accuracy α and distance ϵ for circuits of size s, if the relation R{0,1}n×{0,1}α is δ-hard on D for circuits of size s, where

R={(x,y):|f(x)y|<ϵ},

and real value y is encoded by a binary of length α.

2.3 Hardcore Lemmas

Impagliazzo’s hardcore lemma [17] implies the existence of a strongly hard subset within an instance space where the Boolean function is only mildly hard on average. In this section, we first extend the hardcore lemma to a more general setting, for relations with a closure property under majority. Then, we will demonstrate how to transform a hard distribution into a balanced one, to facilitate our subsequent proofs of hardness amplification.

The following definition of density is utilized to measure the flatness of the hardcore distribution obtained, ensuring that it can be reintegrated into the original distribution [2, Chapter 19].

Definition 2.5 (Relative Density).

For δ(0,1] and distributions X,Y on {0,1}n, X is called δ-dense with respect to Y, if for any x{0,1}n, we have

Pr[X=x]1δPr[Y=x].

The closure property under majority is highly useful for identifying the hardcore distribution of functions or relations. Several prior studies have relied on this closure property to prove the existence of hardcore distribution [17, 19, 18, 5]. We begin by formally defining majority combiner.

Definition 2.6 (Majority Combiner).

For a relation R{0,1}n×Σ and t, a circuit M:ΣtΣ is called a majority combiner for relation R over t coordinates, if for any x{0,1}n and any y1,,ytΣ, such that |{i:i{1,,t}yiR(x)}|>t/2, we have M(y1,,yt)R(x).

The key idea behind Impagliazzo’s hardcore lemma is that, if for any distribution H with certain density, there always exists a circuit of slightly smaller size that solves the problem with probability more than 12+γ, for some γ(0,1), then we can construct a larger circuit by taking the majority vote of multiple carefully chosen circuits, which can result in a high probability of agreement with function f, leading to a contradiction.

Lemma 2.7 (Extended Hardcore Lemma).

For n, consider a relation R{0,1}n×{0,1}m; define G={x:y{0,1}m,yR(x)}. Consider any distribution D over {0,1}n. Consider any δ(0,1), γ(0,1/2), and large enough s. Let t=8log(2/γδ)γ2 be an integer. If there exists a majority combiner for relation R over t coordinates of size at most s/2, and R is (1δ)-hard on D for circuits of size s, then there is a distribution H over {0,1}nG which is δ-dense with respect to D, such that R is (12+γ)-hard on H for circuits of size γ2s16log(2/γδ).

 Remark 2.8.

The proof of Lemma 2.7 follows [17, 2] via the Min-Max theorem with some subtle adjustments, which can be found in the full version of this work. The best-known parameter for the small circuit size in the lemma is O(γ2slog(1/δ)), whereas the bound we present here is O(γ2slog(2/γδ)). In fact, the proof in [5], which uses a multiplicative weight update method, can be directly adapted to our extended setting. For simplicity, we provide a proof for a slightly weaker bound, which suffices for our purpose.

2.4 Balancing Hardcore Distributions

Definition 2.9 (Balanced Distribution).

For a domain Σ, given a distribution X on {0,1}n and y0,y1Σ, a relation R{0,1}n×Σ is called balanced on X around {y0,y1}, if

PrxX[y0R(x)]=PrxX[y1R(x)]=12 and PrxX[y0R(x)y1R(x)]=0.

Adapting the approach in [21], we derive the following lemma, which will be useful in our later arguments. We refer the reader to the full version of this paper for the proof.

Lemma 2.10 (Balanced Hardcore).

For γ(0,12), δ(0,1) and s, for a relation R{0,1}n×Σ and distributions H,D over {0,1}n, suppose that there exist a,bΣ, ab, such that

  1. (1)

    For any x𝖲𝗎𝗉𝗉(H), there is exactly one of the following holds: aR(x) or bR(x).

  2. (2)

    Letting Da denote the distribution xD|aR(x)bR(x) and Db denote the distribution xD|bR(x)aR(x); Da,Db has δ-density with respect to D.

If H is a δ-dense distribution with respect to D and R is 12(1+γ)-hard on H for circuits of size s, then there is a distribution H with density δ with respect to D, on which R is balanced around {a,b} and (12+γ)-hard for circuits of size s.

3 Evaluating Integer-Valued Functions

We now present our hardness amplification result for evaluating integer-valued functions. Given a function f, which is somewhat hard to evaluate on average, it is feasible to show that the function 𝖲𝖴𝖬tf possesses a strong average-case hardness, where 𝖲𝖴𝖬t represents the summation function over t coordinates. We state our main theorem below.

Theorem 3.1.

For δ(0,1), m,s,t and a distribution D over {0,1}n, for any large enough s, consider a function f:{0,1}nm that is (1δ)-hard on D for circuits of size s, define a function g:({0,1}n)ttm as follows:

g(x1,,xt)=i=1tf(xi).

Then, for γ(0,1), for large enough s, g is η-hard on Dt for circuits of size s, where

η =eμ/4+(μμ2)2μ(1+μ2γ),μ=tδm(m1),
s =γ2s512m2log(4m2/γδ).

For sufficiently small γ, the dominant term of η is (μμ/2)/2μ=Θ(1/μ). Taking t large enough enables us to establish hardness amplification. To prove it, we will first construct the hardcore distribution for the function f, show that summation effectively amplify the hardness on this hard distribution and then generalize the result to the original distribution.

Intuitively, the hardcore distribution of any Boolean-valued function is straightforward, as the output is restricted to either yes or no. However, for integer-valued functions, the structure of a hard set is inherently more complicated. We characterize the hardcore of integer-valued functions by defining a set of new problems, simply considering two values in the output domain and an input value is only considered relevant if its corresponding output matches one of those. We show that there is a pair of values such that the resulting problem is hard. Then, we extract a hardcore from this hard problem, which possesses a good structure corresponding to the original function.

Lemma 3.2.

For δ(0,1), m,s and a distribution D over {0,1}n, consider a function f:{0,1}nm. For any a,bm and ab, define the relation Ra,b{0,1}n×{0,1}logm as follows:

  • If f(x){a,b}, then (x,y)Ra,b if and only if y=f(x);

  • If f(x){a,b}, then (x,y)Ra,b for any y{0,1}logm.

For sm2logm, if f is (1δ)-hard on D for circuits of size s, there exists a pair of a,bm,ab, such that Ra,b is (12δm(m1))-hard on D for circuits of size sm2.

This lemma suggests that if the function f is hard to evaluate on average, then there must exist output values a,b, such that distinguishing their pre-images is also hard on average. We defer the proof to Section 3.1, and our hardcore construction is shown below.

Lemma 3.3 (Hardcore for Integer-Valued Functions).

For δ,γ(0,1), m,s and a distribution D, consider a function f:{0,1}nm, which is (1δ)-hard on D for circuits of size s. If s is sufficient large, there exist a,bm,ab and a 2δm(m1)-dense distribution H (with respect to D), on which f is balanced around {a,b} and 12(1+γ)-hard for circuits of size γ2s256m2log(4m2/γδ).

Proof.

Suppose f:{0,1}nm is (1δ)-hard on D for circuits of size s, by Lemma 3.2, if sm2logm, there exists a pair of a,b,ab, such that Ra,b is (12δm(m1))-hard on D for circuits of size sm2. This hardness implies that PrxD[f(x)=a]>2δm(m1), then the distribution D|f(x)=a is 2δm(m1)-dense with respect to D (and the same holds for b).

For relation Ra,b, the majority gate is a natural choice for the combiner. For some γ(0,1), let t=128log(4m2/γδ)γ2 be an integer, if the size of majority-of-t circuits is less than s2m2, by Lemma 2.7, there is a 2δm(m1)-dense hardcore distribution H (with respect to D) over {x:f(x){a,b}}{0,1}n, on which Ra,b is (12+γ4)-hard for circuits of size γ2s256m2log(4m2/γδ). By Lemma 2.10, we can construct a distribution H with density 2δm(m1) (with respect to D), on which f is balanced around {a,b} and 12(1+γ)-hard for circuits of size γ2s256m2log(4m2/γδ).

We believe that this hardcore distribution exhibits a desirable structure, given that the hardness of evaluating function f on it can be captured by indistinguishability. In the following, we consider the inputs sampled from the hardcore distribution. It is a natural way to amplify the average-case hardness of function f, by constructing the function 𝖲𝖴𝖬tf. If f is hard on distribution H on which there are two possible outputs, then the best algorithm with restricted running time would not perform significantly better than random guessing. Consequently, the almost optimal way for guessing the value of 𝖲𝖴𝖬tf on Ht is output the one with the highest probability of occurring.

Lemma 3.4.

For γ(0,1), m,s,t and a,bm, consider a hardcore distribution H{0,1}n, on which function f:{0,1}nm is balanced around {a,b} and 12(1+γ)-hard for circuits of size s. Define the function g:({0,1}n)ttm as follows:

g(x1,,xt)=i=1tf(xi).

For k,0<kt and any other distribution G over ({0,1}n)tk, for sufficient large s, specifically slog(tm), the function g is η-hard on Πt(Hk,G) for circuits of size s2, where

η=(kk2)2k(1+k2γ).

This lemma demonstrates that summation can effectively amplify the hardness over the hardcore distribution, for which the proof will be presented in Section 3.2. However, the hardcore lemma only ensures the existence of a hard distribution without guaranteeing efficient sampling. Therefore, to derive a more meaningful hardness result, we will eventually focus on the hardness on the original distribution.

One direct approach is to embed this distribution into the original one with some probability mass parameterized by its relative density. When sampling instances from the original distribution a sufficient number of times, the number of instances sampled from the hard one will concentrate around the expected value. Based on this observation, we proceed to prove our main theorem.

Theorem 3.1. [Restated, see original statement.]

For δ(0,1), m,s,t and a distribution D over {0,1}n, for any large enough s, consider a function f:{0,1}nm that is (1δ)-hard on D for circuits of size s, define a function g:({0,1}n)ttm as follows:

g(x1,,xt)=i=1tf(xi).

Then, for γ(0,1), for large enough s, g is η-hard on Dt for circuits of size s, where

η =eμ/4+(μμ2)2μ(1+μ2γ),μ=tδm(m1),
s =γ2s512m2log(4m2/γδ).
Proof.

Theorem 3.1 For a function f:{0,1}nm, which is (1δ)-hard on D for circuits of size s, by Lemma 3.3, there exists a,bm,ab and a 2δm(m1)-dense distribution H with respect to D, such that f is balanced around {a,b} and 12(1+γ)-hard on H for circuits of size s^=γ2s256m2log(4m2/γδ).

H has density δ^=2δm(m1), then there exists a distribution G over {0,1}n, such that D=δ^H+(1δ^)G. For t, we have

Dt=k=0t(tk)δ^k(1δ^)tkΠt(Hk,Gtk).

Therefore, for any large enough s, and any circuit C of size s=s^2, by Lemma 3.4,

PrxDt[C(x)=g(x)]
=k=0t(tk)δ^k(1δ^)tkPrxΠt(Hk,Gtk)[C(x)=g(x)]
k=0μ1(tk)δ^k(1δ^)tk+k=μt(tk)δ^k(1δ^)tk(μμ2)2μ(1+μ2γ)
<k=0μ1(tk)δ^k(1δ^)tk+(μμ2)2μ(1+μ2γ)
<eμ/4+(μμ2)2μ(1+μ2γ),

where μ=tδ^2. The last inequality is obtained by Chernoff bound, where the first term is equivalent to the probability that a binomial distribution with parameter (n,δ^) samples a value less than μ.

3.1 Proof of Lemma 3.2

Lemma 3.2. [Restated, see original statement.]

For δ(0,1), m,s and a distribution D over {0,1}n, consider a function f:{0,1}nm. For any a,bm and ab, define the relation Ra,b{0,1}n×{0,1}logm as follows:

  • If f(x){a,b}, then (x,y)Ra,b if and only if y=f(x);

  • If f(x){a,b}, then (x,y)Ra,b for any y{0,1}logm.

For sm2logm, if f is (1δ)-hard on D for circuits of size s, there exists a pair of a,bm,ab, such that Ra,b is (12δm(m1))-hard on D for circuits of size sm2.

Proof.

Assume that, for any a,bm, a<b, there is a circuit Ca,b:{0,1}n{0,1}logm of size sm2, such that

PrxD[Ca,b(x)Ra,b(x)]12δm(m1).

Let Cb,a=Ca,b. Then, we can construct a circuit C as follows:

  1. 1.

    Input: x{0,1}n

  2. 2.

    For i{0,1,,m1}:

    • If for every ji, Ci,j(x)=i holds, return i.

  3. 3.

    Return .

During the process, C compares each output of Ci,j(x) with i, which means there are Θ(m2) comparison of Θ(logm) bits, so the circuit complexity of C is at most (m2)sm2+cm2logm, where c is a constant. If sm2logm, the size of C is at most s.

By taking the union bound, we have

PrxD[(a,b),a<b:Ca,b(x)Ra,b(x)]a<bPrxD[Ca,b(x)Ra,b(x)]δ.

For any input x, there exists at most one i, such that for every ji, Ci,j(x) outputs i. If not, we have Ci,j(x)=i=j, which leads to a contradiction.

If every Ca,b(x) outputs a correct value in Ra,b(x), there exists a unique i=f(x), such that Ci,j(x)Ri,j(x) for every ji, then C computes f(x) correctly.

PrxD[C(x)=f(x)]PrxD[(a,b),a<b:Ca,b(x)Ra,b(x)]1δ,

which contradicts the fact that f is (1δ)-hard for circuits of size s.

3.2 Proof of Lemma 3.4

Lemma 3.4. [Restated, see original statement.]

For γ(0,1), m,s,t and a,bm, consider a hardcore distribution H{0,1}n, on which function f:{0,1}nm is balanced around {a,b} and 12(1+γ)-hard for circuits of size s. Define the function g:({0,1}n)ttm as follows:

g(x1,,xt)=i=1tf(xi).

For k,0<kt and any other distribution G over ({0,1}n)tk, for sufficient large s, specifically slog(tm), the function g is η-hard on Πt(Hk,G) for circuits of size s2, where

η=(kk2)2k(1+k2γ).
Proof.

Consider a hardcore distribution H and a function f:{0,1}nm, such that f is balanced around {a,b} and 12(1+γ)-hard on H for circuits of size s. Let Ha denote the distribution xH|f(x)=a and Hb denote xH|f(x)=b, we have H=12Ha+12Hb. For any i,k, denote Π(Hai,Hbki) by Hik Then, consider any x^=(x^1,,x^tk)({0,1}n)tk and a permutation π of t entries, we claim the following fact.

Claim 3.5.

If slog(tm), for any circuit C:({0,1}n)t{0,1}logtm of size at most s2 and any ik, vtm, we have

|Prx¯Hk[C(π(x¯,x^))=vx¯Hi+1k]Prx¯Hk[C(π(x¯,x^))=vx¯Hik]|<γ.

By Claim 3.5 (which is proven below), combined with triangle inequality, for any i,jk,

Prx¯Hk[C(π(x¯,x^))=v|x¯Hik]<Prx¯Hk[C(π(x¯,x^))=vx¯Hjk]+|ij|γ.

Let Δ=i=1tkf(x^i). For any circuit C:({0,1}n)t{0,1}logtm of size at most s2,

Prx¯Hk[C(π(x¯,x^))=g(π(x¯,x^))]
=i=0kPrx¯Hk[x¯Hik]Prx¯Hk[C(π(x¯,x^))=g(x¯,x^)|x¯Hik]
=i=0kPrx¯Hk[x¯Hik]Prx¯Hk[C(π(x¯,x^))=ai+b(ki)+Δ|x¯Hik]
<12ki=0k(ki)(Prx¯Hk[C(π(x¯,x^))=ai+b(ki)+Δ|x¯Hk2k]+|ik2|γ)
12k((kk2)i=0kPrx¯Hk[C(π(x¯,x^))=ai+b(ki)+Δ|x¯Hk2k]+k2γ(kk2))
(kk2)2k(1+k2γ).

Hence, for any distribution G over ({0,1}n)tk,

Prx¯Π(Hk,G)[C(x¯)=g(x¯)](kk2)2k(1+k2γ).

3.2.1 Proof of Claim 3.5

Claim 3.5. [Restated, see original statement.]

If slog(tm), for any circuit C:({0,1}n)t{0,1}logtm of size at most s2 and any ik, vtm, we have

|Prx¯Hk[C(π(x¯,x^))=vx¯Hi+1k]Prx¯Hk[C(π(x¯,x^))=vx¯Hik]|<γ.
Proof.

Without loss of generality, suppose there exists a circuit C:({0,1}n)t{0,1}logtm of size at most s2 and ik, ktm, such that

|Prx¯Hk[C(π(x¯,x^))=vx¯Hi+1k]Prx¯Hk[C(π(x¯,x^))=vx¯Hik]|γ.

Then, there must exists a tuple (x1,x2,,xk1) and a permutation πk of k entries, such that

|PrxH[C(π(H¯(x),x^))=v|xHa]PrxH[C(π(H¯(x))=v|xHb]|γ,

where H¯(x) denotes πk(x,x1,,xk1).

Construct a circuit C, implementing C by taking an input of length n, along with fixed π,x^ and πk,(x1,,xk1),v as the non-uniform advice, outputs a if and only if C outputs v and outputs b otherwise. Then,

|PrxH[C(x)=axHa]PrxH[C(x)=bxHb]|γ.

Therefore, the probability that C correctly compute f is

PrxH[C(x)=f(x)]=12Pr[C(x)=a|xHa]+12Pr[C(x)=b|xHb]12(1+γ),

which contradict to f is 12(1+γ)-hard on H.

4 Approximating Real-Valued Functions

In this section, we extend our results to real-valued functions. To avoid any precision loss introduced by encoding, we assume that for any open interval of length ϵ2, there is a value that can be encoded by {0,1}α in the interval. Therefore, for any approximation considered below, we always assume α>log(m/ϵ)+4.

Theorem 4.1.

For δ(0,1), m,ϵ, α,s,t, α>log(m/ϵ)+4 and a distribution D over {0,1}n, consider a function f:{0,1}n[0,m) that is (1δ)-hard to approximate on D with accuracy α and distance ϵ, for circuits of size s. For any large enough t, define a function g:({0,1}n)t[0,tm) as follows:

g(x1,,xt)=i=1tf(xi).

Then, for γ(0,1), for any large enough s, g is η-hard to approximate on Dt with accuracy α and distance ϵ, α>log(tm/ϵ)+4, for circuits of size s, where

η =eμ/4+(μμ2)2μ(6+μ2γ),μ=ϵm2tδ,
s =γ2ϵ256m2tδlog(8tm2/ϵ2γ2δ)s.

When γ is small enough, η will be dominated by (μμ/2)/2μ=Θ(1/μ). For large enough t, we can effectively obtain a function with strong average-case hardness to approximate by taking the summation of multiple copies of f.

Lemma 4.2.

For δ(0,1), m,ϵ, α,s, α>log(m/ϵ)+4 and a distribution D over {0,1}n, consider a real-valued function f:{0,1}n[0,m). For any l, let d=ϵl, which denotes the radius of the partitioned intervals. For any a,b{d,3d,,(2m2d1)d},ab, define the relation Ra,b{0,1}n×{0,1}α as follows:

  • If f(x)[a±d) or f(x)[b±d), then (x,y)Ra,b if and only if y(f(x)±ϵ);

  • If f(x)[a±d) and f(x)[b±d), then (x,y)Ra,b for any y{0,1}α.

For any integer l>1, for large enough s, if f is (1δ)-hard to approximate on D with accuracy α and distance ϵ for circuits of size s, there exist a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, such that Ra,b is (14d2δm2)-hard on D, for circuits of size d2sm2.

The proof is deferred to Section 4.1. Following the approach used in the integer case, we proceed with the construction of the hardcore distribution. This hard distribution maintains a good structure, on which f will be balanced and mapped to [a±d) or [b±d), for some fixed a,b, which implies the closed approximation of f with distance d is balanced on H around {a,b}.

Lemma 4.3 (Hardcore for Real-Valued Functions).

For δ,γ(0,1), m,ϵ, α,s, α>log(m/ϵ)+4 and a distribution D over {0,1}n, consider a function f:{0,1}n[0,m), which is (1δ)-hard to approximate D with accuracy α and distance ϵ for circuits of size s. For any integer l3, let d=ϵl, there exist a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, and a 4d2δm2-dense distribution H (with respect to D), on which the closed approximation of f with distance d is balanced around {a,b} and the approximation of f with accuracy α and distance ϵ is 12(1+γ)-hard for circuits of size γ2d2s256m2log(2m2/d2γδ).

Proof.

Suppose f:{0,1}n[0,m) is (1δ)-hard on D for circuits of size s, by Lemma 4.2, for integer l>1, d=ϵl, if s is sufficiently large, there exist a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, such that Ra,b (as defined in Lemma 4.2) is (14d2δm2)-hard on D for circuits of size d2sm2.

For Ra,b, the majority combiner can be constructed by taking the middle point. By Lemma 2.7, when s is large enough, we have a 4d2δm2-dense distribution H (with respect to D) over {x:f(x)[a±d)[b±d)}{0,1}n, on which Ra,b is (12+γ4)-hard for circuits of size s^=γ2d2s256m2log(2m2/d2γδ), as well as the approximation of f with accuracy α and distance ϵ is hard on H. If l3, then ba>2d, the intervals [a±d] and [b±d] are disjoint. Then, by Lemma 2.10, we can construct a distribution H with density 4d2δm2, on which the closed approximation of f with distance d is balanced around {a,b} and the approximation of f with accuracy α and distance ϵ is 12(1+γ)-hard for circuits of size s^.

Analogously, we prove that summation suffices to achieve amplification.

Lemma 4.4.

For γ(0,1), m,ϵ and α,s,t, α>log(m/ϵ)+4, for large enough l, let d=ϵl, consider a hardcore distribution H{0,1}n and a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, on which the closed approximation of function f:{0,1}n[0,m) with distance d is balanced around {a,b} and f is 12(1+γ)-hard to approximate with accuracy α and distance ϵ for circuits of size s. For any integer t, define a function g:({0,1}n)t[0,tm) as follows:

g(x1,,xt)=i=1tf(xi).

For k,0<kt and any other distribution G over ({0,1}n)tk, for any large enough s, function g is η-hard to approximate with accuracy α and distance ϵ on Πt(Hk,G) for circuits of size s2, where α>log(tm/ϵ)+4 and

η=(kk2)2k(3kl+3+k2γ).

We postpone the proof to Section 4.2 and now present the proof for Theorem 4.1.

Theorem 4.1. [Restated, see original statement.]

For δ(0,1), m,ϵ, α,s,t, α>log(m/ϵ)+4 and a distribution D over {0,1}n, consider a function f:{0,1}n[0,m) that is (1δ)-hard to approximate on D with accuracy α and distance ϵ, for circuits of size s. For any large enough t, define a function g:({0,1}n)t[0,tm) as follows:

g(x1,,xt)=i=1tf(xi).

Then, for γ(0,1), for any large enough s, g is η-hard to approximate on Dt with accuracy α and distance ϵ, α>log(tm/ϵ)+4, for circuits of size s, where

η =eμ/4+(μμ2)2μ(6+μ2γ),μ=ϵm2tδ,
s =γ2ϵ256m2tδlog(8tm2/ϵ2γ2δ)s.
Proof.

For α>log(m/ϵ)+4 and a function f:{0,1}[0,m), which is (1δ)-hard for circuits of size s, by Lemma 4.3, for l, let d=ϵl, there exist a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, and a 4d2δm2-dense distribution H with respect to D, such that the closed approximation of f with distance d is balanced around {a,b} and that f is 12(1+γ)-hard to approximate with accuracy α and distance ϵ on H for circuits of size s^=γ2d2s256m2log(2m2/d2γδ).

Since H has relative density δ^=4d2δm2, there exists a distribution G over {0,1}n, such that D=δ^H+(1δ^)G. For t, we have

Dt=k=0t(tk)δ^k(1δ^)tkΠt(Hk,Gtk).

Therefore, for any large enough s,t, for any circuit of size s=s^2, by Lemma 4.4, we have

PrxDt[C(x)=g(x)] =k=0t(tk)δ^k(1δ^)tkPrxΠt(Hk,Gtk)[C(x)=g(x)]
<k=0μ(tk)δ^k(1δ^)tk+(μμ2)2μ(3μl+3+μ2γ)
eμ/4+(μμ2)2μ(3μl+3+μ2γ),

where μ=tδ^2=2td2δm2. Let l=μ, we have

μ=2td2δm2=2tϵ2δl2m2=ϵlm2tδ=l.

Then, l=ϵm2tδ,

μ=ϵm2tδ,s=γ2ϵ256m2tδlog(8tm2/ϵ2γ2δ)s.

4.1 Proof of Lemma 4.2

Lemma 4.2. [Restated, see original statement.]

For δ(0,1), m,ϵ, α,s, α>log(m/ϵ)+4 and a distribution D over {0,1}n, consider a real-valued function f:{0,1}n[0,m). For any l, let d=ϵl, which denotes the radius of the partitioned intervals. For any a,b{d,3d,,(2m2d1)d},ab, define the relation Ra,b{0,1}n×{0,1}α as follows:

  • If f(x)[a±d) or f(x)[b±d), then (x,y)Ra,b if and only if y(f(x)±ϵ);

  • If f(x)[a±d) and f(x)[b±d), then (x,y)Ra,b for any y{0,1}α.

For any integer l>1, for large enough s, if f is (1δ)-hard to approximate on D with accuracy α and distance ϵ for circuits of size s, there exist a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, such that Ra,b is (14d2δm2)-hard on D, for circuits of size d2sm2.

Proof.

Suppose that for any a,b{d,3d,,(2m2d1)d}, a<b, there is a circuit Ca,b:{0,1}n{0,1}α of size d2sm2, such that

PrxD[Ca,b(x)Ra,b(x)]14d2δm2.

For simplicity, let Cb,a=Ca,b. Taking a combination of circuits Ca,b, construct a new circuit C:{0,1}n{0,1}α as follows:

  1. 1.

    On input x, for any i{d,3d,,(2m2d1)d} in ascending order:

    • Compute Ci,j(x);

    • if Ci,j(x) outputs a value in (i±(ϵ+d)) for every ji, return maxji(Ci,j(x)), which is the maximum value among all the outputs given by Ci,j(x).

  2. 2.

    Return if no such i exists.

If s is large enough, specifically s(md)2α, the size of C is approximately (m2d2)d2sm2s. The performance of the circuit C is stated as follows.

Claim 4.5.

For any x{0,1}n, if Ca,b(x)Ra,b(x) for every distinct a,b, C(x)(f(x)±ϵ).

By Claim 4.5, we have

PrxD[C(x)(f(x)±ϵ)]PrxD[(a,b),a<b:Ca,b(x)Ra,b(x)]1(m2d2)4d2δm21δ.

The second inequality is obtained by taking the union bound. It contradicts the fact that f is (1δ)-hard to approximate with accuracy α and distance ϵ.

In the following, we will prove that the relation Ra,b (a<b) is potentially hard only if ba>(32l2)d. Assume the distance of two centers ba(32l2)d, construct a circuit Ca,b which outputs a value in (b(ϵd),a+(ϵd)), regardless of inputs. Since the interval length is

ab+2(ϵd)(32l2)d+2(ϵd)ϵ2,

there exists a value in this interval that can be encoded in {0,1}α.

Consider any input x, if f(x)[a±d), f(x)ϵ<a+dϵ<b(ϵd)<a+(ϵd)<f(x)+ϵ; if f(x)[b±d), f(x)ϵ<b(ϵd)<a+(ϵd)<bd+ϵf(x)+ϵ; otherwise, any output is in Ra,b(x). Then, the output is guaranteed to have Ca,b(x)Ra,b(x).

Therefore, there exist a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, such that Ra,b is (14d2δm2)-hard for circuits of size d2sm2.

4.1.1 Proof of Claim 4.5

Claim 4.5. [Restated, see original statement.]

For any x{0,1}n, if Ca,b(x)Ra,b(x) for every distinct a,b, C(x)(f(x)±ϵ).

Proof.

We partition the output space into multiple intervals with length 2d and define a set relation Ra,b. For each relation, we focus on the inputs whose outputs by f lie in the corresponding intervals [a±d) or [b±d). Recall the process of algorithm C: for i{d,3d,,(2m2d1)d} in ascending order, if for every ji, Ci,j(x)(i±(ϵ+d)) holds, then C stops and outputs maxji(Ci,j(x)).

Suppose that the algorithm returns when i=i^, then f(x)i^d. If f(x)<i^d, there exists i~<i^, such that f(x)[i~±d), since we assume the correctness of every Ca,b, then C should stop when i=i~, which results in a contradiction.

If f(x)[i^±d), Ci,j(x)(f(x)±ϵ) for every ji, then C(x)(f(x)±ϵ). If f(x)i^+d, suppose i~>i^ and f(x)[i~±d), then Ci^,i~(x)(f(x)±ϵ).

f(x)ϵ<Ci^,i~(x)C(x)<i^+d+ϵf(x)+ϵ.

4.2 Proof of Lemma 4.4

Lemma 4.4. [Restated, see original statement.]

For γ(0,1), m,ϵ and α,s,t, α>log(m/ϵ)+4, for large enough l, let d=ϵl, consider a hardcore distribution H{0,1}n and a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, on which the closed approximation of function f:{0,1}n[0,m) with distance d is balanced around {a,b} and f is 12(1+γ)-hard to approximate with accuracy α and distance ϵ for circuits of size s. For any integer t, define a function g:({0,1}n)t[0,tm) as follows:

g(x1,,xt)=i=1tf(xi).

For k,0<kt and any other distribution G over ({0,1}n)tk, for any large enough s, function g is η-hard to approximate with accuracy α and distance ϵ on Πt(Hk,G) for circuits of size s2, where α>log(tm/ϵ)+4 and

η=(kk2)2k(3kl+3+k2γ).
Proof.

For an integer l, let d=ϵl, consider a hardcore distribution H{0,1}n and a,b{d,3d,,(2m2d1)d}, ba>(32l2)d, such that, the closed approximation of function f:{0,1}n[0,m) with distance d is balanced around {a,b} on H. Denote distribution xH|f(x)(a±d) by Ha and distribution xH|f(x)(b±d) by Hb. Since f is balanced around {a,b} on distribution H, H=12Ha+12Hb. The hardness of approximating function f implies the indistinguishability of this two distribution Ha and Hb.

For t, let Hit=Πt(Hai,Hbti). Then, consider any fixed x^=(x^1,,x^tk)({0,1}n)tk and any permutation π of t coordinates, we have the following fact.

Claim 4.6.

For s,α, such that sα and (tm,α,ϵ) is valid, for any circuit C:({0,1}n)t{0,1}α of size at most s2, for any i,jk+1, such that |ij|=1 and any v,ϵ[0,tm),ϵϵ, we have

Prx¯Hkx¯π(x¯,x^)[C(x¯)(gt(x¯)+v±ϵ)|x¯Hjt]
Prx¯Hkx¯π(x¯,x^)[C(x¯)(gt(x¯)+v+v±(ϵ+2d))|x¯Hit]<γ

where v=(ji)(ab).

Let Δ=i=1tkf(x^i). For any large even t, for any circuit C:({0,1}n)t{0,1}α of size at most s2, the following holds,

Prx¯Hkx¯π(x¯,x^)[C(x¯)(gt(x¯)±ϵ)]
=i=0kPrx¯Hkx¯π(x¯,x^)[x¯Hik]Prx¯Hkx¯π(x¯,x^)[C(x¯)(gt(x¯)±ϵ)|x¯Hik]
=12ki=0k(ki)Prx¯Hkx¯π(x¯,x^)[C(x¯)(gk(x¯)+Δ±ϵ)|x¯Hik]
=12ki=k2k2(kk2+i)Prx¯Hkx¯π(x¯,x^)[C(x¯)(gk(x¯)+Δ±ϵ)|x¯Hk2+ik]
<12ki=k2k2(kk2+i)Prx¯Hkx¯π(x¯,x^)[C(x¯)(gk(x¯)+i(ab)+Δ±(ϵ+2d|i|))|x¯Hk2k] (6)
+12ki=k2k2(kk2+i)|i|γ
(kk2)2k(3kl+3+k2γ).

The inequality (4.2) is obtained by using Claim 4.6, we use the probability with a same condition x¯Hk2k to give an upper bound for the original term. The last inequality follows Claim 4.7. Therefore, for any distribution G over ({0,1}n)tk, we have

Prx¯Πt(Hk,G)[C(x¯)(gt(x¯)±ϵ)](kk2)2k(3kl+3+k2γ).

Claim 4.7.

For large enough l, for any circuit C, the following value is upper-bounded by (kk2)(3kl+3):

i=k2k2(kk2+i)Prx¯Hkx¯π(x¯,x^)[C(x¯)(gk(x¯)+i(ab)+Δ±(ϵ+2d|i|))|x¯Hk2k],

where ba>(32l2)d and ϵ=ld.

Proof.

Recall that ϵ denotes the tolerance of the approximation error, and for some integer l, we let d=ϵl. Since a,b are the multiples of d, let ba=l1d, for some positive integer l1. By our assumption, ba>(32l2)d, thus l1>32l2.

For simplicity, for any integer i1,i2, let

P(i1,i2)=Prx¯Hkx¯π(x¯,x^)[C(x¯)[gk(x¯)+Δ+i1d,gk(x¯)+Δ+i2d)|x¯Hk2k].

It is clear that P(i1,i2)+P(i2,i3)=P(i1,i3) for i1<i2<i3. Note that (kk2+i)=(kk2i), the value can be upper bounded by

i=k2k2(kk2+i)P(il1l2|i|,il1+l+2|i|)
= i=k2k2(kk2+i)j=il1l2|i|il1+l+2|i|1P(j,j+1)

where summation is justified by the basic property of probability. To calculate the above, we collect the sum of coefficients corresponding to each P(j,j+1). For each j, the term should be

i=if(j)ic(j)(kk2+i)P(j,j+1), (7)

for some functions if,ic. Let S(j)=i=if(j)ic(j)(kk2+i) denote the coefficient of P(j,j+1). Since the summation of all possible P(j,j+1) is at most 1, by the definition of probability, it is feasible to give an upper bound for the entire summation by computing the upper bound for S(j). For any i[if(j),ic(j)], i should satisfy:

il1l2|i|jil1+l+2|i|1 and k2ik2.

It is clear that S(j)=S(j1), since for any i[if(j),ic(j)],

il1l2|i|jil1+l+2|i|1
il1l2|i|j1il1+l+2|i|1.

Then, i[if(j1),ic(j1)], we derive S(j)=S(j1).

Recall that l1>32l2, assume l>6, l1l>1. When j0, we necessarily have il1+l+2|i|1. If there exists a negative i satisfies the inequality, then i(l12)+l(l12)+l<1, which leads to a contradiction. Therefore,

il1l2ijil1+l+2i1,

which is equivalent to

if(j)=jl+1l1+2ij+ll12=ic(j).

Therefore, we can give an upper bound for S(j),

S(j) =i=if(j)ic(j)(kk2+i)(j+ll12jl+1l1+2+1)(kk2+if(j))
(4j(l12)(l1+2)+3)(kk2+if(j)).

On the other hand, j can be upper bounded in terms of if(j), that is

jl+1l1+2<if(j)+1j<(if(j)+1)(l1+2)+l1<(if(j)+2)(l1+2).

Then, plug in if(j),

S(j)(4j(l12)(l1+2)+3)(kk2+if(j))<(4(if(j)+2)l12+3)(kk2+if(j)).

In the following, for 0ik2, we denote

T(i)=i+2l12(kk2+i). (8)

Since the maximum value of S(j) is at most the maximum value of 4T(i)+3(kk2), then

maxjS(j)=maxj0S(j)4maxiT(i)+3(kk2).

The first equality holds because S(j)=S(j1). To find the maximum value of T(i), we compare each pairs of adjacent terms in the sequence: T(i+1)>T(i) if and only if i<k+532. We have

maxiT(i)<k+532+2l12(kk2).

Therefore, for k12,

maxjS(j)<4k+532+2l12(kk2)+3(kk2)<3(kl+1)(kk2).

4.2.1 Proof of Claim 4.6

Claim 4.6. [Restated, see original statement.]

For s,α, such that sα and (tm,α,ϵ) is valid, for any circuit C:({0,1}n)t{0,1}α of size at most s2, for any i,jk+1, such that |ij|=1 and any v,ϵ[0,tm),ϵϵ, we have

Prx¯Hkx¯π(x¯,x^)[C(x¯)(gt(x¯)+v±ϵ)|x¯Hjt]
Prx¯Hkx¯π(x¯,x^)[C(x¯)(gt(x¯)+v+v±(ϵ+2d))|x¯Hit]<γ

where v=(ji)(ab).

Proof.

Without loss of generality, suppose it and j=i+1. Suppose that there is a circuit C of size at most s2 and v,ϵ[0,tm), ϵϵ, the inequality above does not hold. Then, there must exist a tuple (x1,,xk1) and a permutation πk of k entries, such that,

PrxHx¯πk(x1,,xk1,x)x¯π(x¯,x^)[C(x¯)(gt(x¯)+v±ϵ)|xHa]
PrxHx¯πk(x1,,xk1,x)x¯π(x¯,x^)[C(x¯)(gt(x¯)+v+v±(ϵ+2d))|xHb]γ.

In fact, gt(x¯)=f(x)+gk1(x1,,xt1)+gtk(x^), let Δ=g(x¯)f(x), which is a fixed value independent of x. Recall that ϵ is the tolerance of approximation error and d is the radius of the intervals around a and b, while letting d=ϵl for some large enough integer l. Define a circuit C:{0,1}n{0,1}α as follows:

  1. 1.

    Input: x{0,1}n.

  2. 2.

    Compute yC(π(x,x^)).

  3. 3.

    If y[a+Δ+v±(ϵ+d)), output a value in (a±ϵ4).

  4. 4.

    Otherwise, output a value in (b±ϵ4).

If s is large enough, the size of C is less than s. In the following, we will show that the circuit C can approximate function f with a good probability, then result in a contradiction.

For any xHa, which means f(x)[a±d), if C(x¯) output a value y in (gt(x¯)+v±ϵ) which is equivalent to (f(x)+Δ+v±ϵ), we necessarily have y[a+Δ+v±(ϵ+d)). For any value z(a±ϵ4), |f(x)z|ϵ4+dϵ, then

PrxH[C(x)(f(x)±ϵ)|xHa]PrxHx¯πk(x1,,xk1,x)x¯π(x¯,x^)[C(x¯)(gt(x¯)+v±ϵ)|xHa].

On the other hand, for any xHb, f(x)[b±d), if C(x¯) output a value y, such that y(gt(x¯)+v+v±(ϵ+2d)), which is equivalent to (f(x)+Δ+v+(ab)±(ϵ+2d)). The interval above covers the interval [a+Δ+v±(ϵ+d)), since

a+Δ+v+ϵ+df(x)+Δ+v+(ab)+ϵ+2d
a+Δ+vϵd>f(x)+Δ+v+(ab)ϵ2d.

The circuit C will output a value in (b±ϵ4). Therefore, the probability that C can successfully approximate function f on H is

PrxH[C(x)(f(x)±ϵ)]
=12PrxH[C(x)(f(x)±ϵ)|xHa]+12PrxH[C(x)(f(x)±ϵ)|xHb]
12PrxHx¯πk(x1,,xk1,x)x¯π(x¯,x^)[C(x¯)(gt(x¯)+v±ϵ)|xHa]
+12PrxHx¯πk(x1,,xk1,x)x¯π(x¯,x^)[C(x¯)(gt(x¯)+v+v±(ϵ+2d))|xHb].
12(1+γ),

which contradict to f is 12(1+γ)-hard on H.

References

  • [1] Shweta Agrawal, Sagnik Saha, Nikolaj I. Schwartzbach, Akhil Vanukuri, and Prashant Nalini Vasudevan. k-sum in the sparse regime: Complexity and applications. In Leonid Reyzin and Douglas Stebila, editors, Advances in Cryptology - CRYPTO 2024 - 44th Annual International Cryptology Conference, Santa Barbara, CA, USA, August 18-22, 2024, Proceedings, Part II, volume 14921 of Lecture Notes in Computer Science, pages 315–351. Springer, 2024. doi:10.1007/978-3-031-68379-4_10.
  • [2] Sanjeev Arora and Boaz Barak. Computational Complexity - A Modern Approach. Cambridge University Press, 2009. URL: http://www.cambridge.org/catalogue/catalogue.asp?isbn=9780521424264.
  • [3] Vahid R. Asadi, Alexander Golovnev, Tom Gur, and Igor Shinkar. Worst-case to average-case reductions via additive combinatorics. In Stefano Leonardi and Anupam Gupta, editors, STOC ’22: 54th Annual ACM SIGACT Symposium on Theory of Computing, Rome, Italy, June 20 - 24, 2022, pages 1566–1574. ACM, 2022. doi:10.1145/3519935.3520041.
  • [4] Marshall Ball, Alon Rosen, Manuel Sabin, and Prashant Nalini Vasudevan. Average-case fine-grained hardness. In Hamed Hatami, Pierre McKenzie, and Valerie King, editors, Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2017, Montreal, QC, Canada, June 19-23, 2017, pages 483–496. ACM, 2017. doi:10.1145/3055399.3055466.
  • [5] Boaz Barak, Moritz Hardt, and Satyen Kale. The uniform hardcore lemma via approximate bregman projections. In Claire Mathieu, editor, Proceedings of the Twentieth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2009, New York, NY, USA, January 4-6, 2009, pages 1193–1200. SIAM, 2009. doi:10.1137/1.9781611973068.129.
  • [6] Andrej Bogdanov and Luca Trevisan. On worst-case to average-case reductions for NP problems. SIAM J. Comput., 36(4):1119–1159, 2006. doi:10.1137/S0097539705446974.
  • [7] Enric Boix-Adserà, Matthew S. Brennan, and Guy Bresler. The average-case complexity of counting cliques in erdős-rényi hypergraphs. In David Zuckerman, editor, 60th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2019, Baltimore, Maryland, USA, November 9-12, 2019, pages 1256–1280. IEEE Computer Society, 2019. doi:10.1109/FOCS.2019.00078.
  • [8] Jin-yi Cai, Aduri Pavan, and D. Sivakumar. On the hardness of permanent. In Christoph Meinel and Sophie Tison, editors, STACS 99, 16th Annual Symposium on Theoretical Aspects of Computer Science, Trier, Germany, March 4-6, 1999, Proceedings, volume 1563 of Lecture Notes in Computer Science, pages 90–99. Springer, 1999. doi:10.1007/3-540-49116-3_8.
  • [9] Nathan Geier. A tight computational indistinguishability bound for product distributions. In Eike Kiltz and Vinod Vaikuntanathan, editors, Theory of Cryptography - 20th International Conference, TCC 2022, Chicago, IL, USA, November 7-10, 2022, Proceedings, Part II, volume 13748 of Lecture Notes in Computer Science, pages 333–347. Springer, 2022. doi:10.1007/978-3-031-22365-5_12.
  • [10] Elazar Goldenberg and Karthik C. S. Hardness amplification of optimization problems. In Thomas Vidick, editor, 11th Innovations in Theoretical Computer Science Conference, ITCS 2020, January 12-14, 2020, Seattle, Washington, USA, volume 151 of LIPIcs, pages 1:1–1:13. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2020. doi:10.4230/LIPICS.ITCS.2020.1.
  • [11] Oded Goldreich, Russell Impagliazzo, Leonid A. Levin, Ramarathnam Venkatesan, and David Zuckerman. Security preserving amplification of hardness. In 31st Annual Symposium on Foundations of Computer Science, St. Louis, Missouri, USA, October 22-24, 1990, Volume I, pages 318–326. IEEE Computer Society, 1990. doi:10.1109/FSCS.1990.89550.
  • [12] Oded Goldreich, Noam Nisan, and Avi Wigderson. On yao’s xor-lemma. In Oded Goldreich, editor, Studies in Complexity and Cryptography. Miscellanea on the Interplay between Randomness and Computation - In Collaboration with Lidor Avigad, Mihir Bellare, Zvika Brakerski, Shafi Goldwasser, Shai Halevi, Tali Kaufman, Leonid Levin, Noam Nisan, Dana Ron, Madhu Sudan, Luca Trevisan, Salil Vadhan, Avi Wigderson, David Zuckerman, volume 6650 of Lecture Notes in Computer Science, pages 273–301. Springer, 2011. doi:10.1007/978-3-642-22670-0_23.
  • [13] Oded Goldreich and Guy N. Rothblum. Counting t-cliques: Worst-case to average-case reductions and direct interactive proof systems. In Mikkel Thorup, editor, 59th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2018, Paris, France, October 7-9, 2018, pages 77–88. IEEE Computer Society, 2018. doi:10.1109/FOCS.2018.00017.
  • [14] Parikshit Gopalan and Venkatesan Guruswami. Hardness amplification within NP against deterministic algorithms. J. Comput. Syst. Sci., 77(1):107–121, 2011. doi:10.1016/J.JCSS.2010.06.008.
  • [15] Alexander Healy, Salil P. Vadhan, and Emanuele Viola. Using nondeterminism to amplify hardness. SIAM J. Comput., 35(4):903–931, 2006. doi:10.1137/S0097539705447281.
  • [16] Shuichi Hirahara and Nobutaka Shimizu. Hardness self-amplification: Simplified, optimized, and unified. In Barna Saha and Rocco A. Servedio, editors, Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, Orlando, FL, USA, June 20-23, 2023, pages 70–83. ACM, 2023. doi:10.1145/3564246.3585189.
  • [17] Russell Impagliazzo. Hard-core distributions for somewhat hard problems. In 36th Annual Symposium on Foundations of Computer Science, Milwaukee, Wisconsin, USA, 23-25 October 1995, pages 538–545. IEEE Computer Society, 1995. doi:10.1109/SFCS.1995.492584.
  • [18] Satyen Kale. Boosting and hard-core set constructions: a simplified approach. Electron. Colloquium Comput. Complex., TR07-131, 2007. URL: https://eccc.weizmann.ac.il/eccc-reports/2007/TR07-131/index.html.
  • [19] Adam R. Klivans and Rocco A. Servedio. Boosting and hard-core set construction. Mach. Learn., 51(3):217–238, 2003. doi:10.1023/A:1022949332276.
  • [20] Richard J. Lipton. New directions in testing. In Joan Feigenbaum and Michael Merritt, editors, Distributed Computing And Cryptography, Proceedings of a DIMACS Workshop, Princeton, New Jersey, USA, October 4-6, 1989, volume 2 of DIMACS Series in Discrete Mathematics and Theoretical Computer Science, pages 191–202. DIMACS/AMS, 1989. doi:10.1090/DIMACS/002/13.
  • [21] Ryan O’Donnell. Hardness amplification within np. J. Comput. Syst. Sci., 69(1):68–94, 2004. doi:10.1016/J.JCSS.2004.01.001.
  • [22] Amit Sahai and Salil P. Vadhan. A complete problem for statistical zero knowledge. J. ACM, 50(2):196–249, 2003. doi:10.1145/636865.636868.
  • [23] Luca Trevisan. On uniform amplification of hardness in NP. In Harold N. Gabow and Ronald Fagin, editors, Proceedings of the 37th Annual ACM Symposium on Theory of Computing, Baltimore, MD, USA, May 22-24, 2005, pages 31–38. ACM, 2005. doi:10.1145/1060590.1060595.
  • [24] Luca Trevisan and Salil P. Vadhan. Pseudorandomness and average-case complexity via uniform reductions. Comput. Complex., 16(4):331–364, 2007. doi:10.1007/S00037-007-0233-X.
  • [25] Salil Pravin Vadhan. A Study of Statistical Zero-Knowledge Proofs. PhD thesis, Harvard University, USA, 1999. AAI0801528.
  • [26] Andrew Chi-Chih Yao. Theory and applications of trapdoor functions (extended abstract). In 23rd Annual Symposium on Foundations of Computer Science, Chicago, Illinois, USA, 3-5 November 1982, pages 80–91. IEEE Computer Society, 1982. doi:10.1109/SFCS.1982.45.