Abstract 1 Introduction 2 Preliminary 3 Fooling the Functions of PTFs via Bounded Independence 4 Discretization References Appendix A Facts about Bump Function

A Pseudorandom Generator for Functions of Low-Degree Polynomial Threshold Functions

Penghui Yao ORCID State Key Laboratory for Novel Software Technology, New Cornerstone Science Laboratory, Nanjing University, Nanjing 210023, China
Hefei National Laboratory, Hefei 230088, China
Mingnan Zhao ORCID State Key Laboratory for Novel Software Technology, New Cornerstone Science Laboratory, Nanjing University, Nanjing 210023, China
Abstract

Developing explicit pseudorandom generators (PRGs) for prominent categories of Boolean functions is a key focus in computational complexity theory. In this paper, we investigate the PRGs against the functions of degree-d polynomial threshold functions (PTFs) over Gaussian space. Our main result is an explicit construction of PRG with seed length poly(k,d,1/ϵ)logn that can fool any function of k degree-d PTFs with probability at least 1ε. More specifically, we show that the summation of L independent R-moment-matching Gaussian vectors ϵ-fools functions of k degree-d PTFs, where L=poly(k,d,1ϵ) and R=O(logkdϵ). The PRG is then obtained by applying an appropriate discretization to Gaussian vectors with bounded independence.

Keywords and phrases:
Pseudorandom generators, polynomial threshold functions
Category:
Track A: Algorithms, Complexity and Games
Copyright and License:
[Uncaptioned image] © Penghui Yao and Mingnan Zhao; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Pseudorandomness and derandomization
Related Version:
Previous Version: https://arxiv.org/abs/2504.10904
Funding:
PY and MZ were supported by National Natural Science Foundation of China (Grant Nos. 62332009 and 12347104), Innovation Program for Quantum Science and Technology (Grant No. 2021ZD0302901), NSFC/RGC Joint Research Scheme (Grant No. 12461160276), Natural Science Foundation of Jiangsu Province (Grant No. BK20243060), and New Cornerstone Science Foundation.
Editors:
Keren Censor-Hillel, Fabrizio Grandoni, Joël Ouaknine, and Gabriele Puppis

1 Introduction

In computational complexity theory, derandomization is a powerful technique that aims to reduce randomness in algorithms without sacrificing efficiency or accuracy. A versatile approach for derandomization is to design explicit pseudorandom generators (PRGs) for notable families of Boolean functions. A PRG for a family of Boolean functions is able to consume few random bits and produce a distribution over high-dimensional vectors, which is indistinguishable from a target distribution, such as the uniform distribution over Boolean cube, by any function in the family. In this paper, we concern ourselves with the Gaussian distribution over n. Formally,

Definition 1.

Let {f:n{0,1}} be a family of Boolean functions. A function G:{0,1}rn is a pseudorandom generator for with error ϵ over Gaussian distribution 𝒩(0,1)n if for any f,

|𝔼su{0,1}r[f(G(s))]𝔼x𝒩(0,1)n[f(x)]|ϵ.

We call r the seed length of G. We also say G ϵ-fools over the Gaussian distribution.

There has been a considerable amount of research developing PRGs for various Boolean function families, including halfspaces, polynomial threshold functions and intersections of halfspaces. Let sign:{0,1} be the function such that sign(x)=1 iff x0. A halfspace is a Boolean function of the form f(x)=sign(a1x1++anxnb) for some a1,,an,b. Halfspaces are a fundamental class of Boolean functions which have found significant applications in machine learning, complexity theory, theory of approximation and more. A very successful series of work produced PRGs that ϵ-fools halfspaces with seed length poly-logarithmic in n and ϵ1 over both Boolean space [28, 5, 21, 7] and Gaussian space [19]. Polynomial threshold functions (PTFs) are functions of the form f(x)=sign(p(x)) where p is a polynomial. We call f is a degree-d PTF if p is a degree-d polynomial. PTFs are natural generalization for halfspaces since a halfspace is a degree-1 PTF. An explicit PRG that ϵ-fools PTFs over Boolean space has been achieved with seed length (d/ϵ)O(d)logn [21]. As for Gaussian space, a sequence of work [6, 12, 13, 14, 21, 15, 16, 24, 17] succeeds in giving a PRG with seed length polynomial in d, ϵ1 and logn [24, 17]. Another extension for halfspaces is intersections of k halfspaces which are polytopes with k facets. A line of work [8, 9, 27, 4, 25] results in PRGs with seed length polynomial in logk, logn and 1/ϵ over Boolean space [25] and over Gaussian space [4].

Considering the prosperity of PRGs for these functions families, we commence designing PRGs for functions of degree-d polynomial threshold functions.

Definition 2.

We say a function F:n{0,1} is a function of k degree-d PTFs if there exist k polynomials p1,,pk:n of degree d and a Boolean function f:{0,1}k{0,1} such that F(x)=f(sign(p1(x)),,sign(pk(x))).

This family consumes all three function families we discussed above. For example, it includes intersections of halfspaces by setting d=1 and f(x)=x1xk. The research on PRGs for functions of PTFs is driven by several motivations beyond its fundamental role in derandomization tasks. For instance, the collection of satisfying assignments of an intersection of k degree-2 PTFs corresponds to the feasible solutions set of an {0,1}-integer quadratic programing [22] with k constraints. The investigation into the structure of these sets has been a central focus of extensive research in areas including learning theory, counting, optimization, and combinatorics.

In this work, we consider building explicit PRGs for functions of degree-d PTFs over Gaussian space. Before presenting our main result, we briefly revisit relevant prior work on fooling functions of halfspaces.

Table 1: Related Work on PRGs for Intersections of PTFs.
Reference Function Family Seed Length
[8] Monotone functions of k halfspaces O((klog(k/ϵ)+logn)log(k/ϵ))
[9] Intersections of k δ-regular halfspaces
O(lognlogk/ϵ)
for δϵ5/(log8.1klog(1/ϵ))
[27] Intersections of k weight-t halfspaces poly(logn,logk,t,1/ϵ)
[25] Intersections of k halfspaces
polylogmϵ(2+δ)logn
for any absolute constant δ(0,1)
[4]
Intersections of k halfspaces
Arbitrary functions of k halfspaces
O(logn+poly(logk,1/ϵ))
O(logn+poly(k,1/ϵ))
[6] Intersections of k degree-2 PTFs O(lognpoly(k,1/ϵ))

1.1 Prior Work

The related work is summarized in Table 1. Gopalan, O’Donnell, Wu and Zuckerman [8] constructed PRGs for monotone functions of halfspaces. They modified the PRG for halfspaces in [21] and showed the modified PRG ϵ-fools any monotone function of k halfspaces over a broad class of product distributions with seed length O((klog(k/ϵ)+logn)log(k/ϵ)). When k/ϵlogcn any c>0, the seed length can be further improved to O(klog(k/ϵ)+logn).

Harsha, Klivans and Meka [9] considered designing PRGs for intersections of regular halfspaces (i.e., halspaces with low influence). A halfspace f(x)=sign(a1x1++anxnb) is δ-regular if iai4δ2iai2. They gave an explicit PRG construction for intersections of k δ-regular halfspaces over proper and hypercontractive distributions with seed length O(lognlogk/ϵ) when δ is no more than a threshold. Their proof is based on developing an invariance principle for intersections of regular halfspaces via a generalization of the well-known Lindeberg method [20] and an anti-concentration result of polytopes in Gaussian space from [18].

By extending the approach of [9] and combing the results on bounded independence fooling CNF formulas [1, 26], Servedio and Tan [27] designed an explicit PRG that ϵ-fools intersections of k weight-t halfspaces over Boolean space with poly(logn,logk,t,1/ϵ) seed length. A halfspace f(x)=sign(a1x1++anxnb) is said to be weight-t if each ai is an integer in [t,t].

As for intersections of k general halfspaces, O’Donnell, Servedio and Tan [25] gave a PRG construction over Boolean space with a polylogarithmic seed length dependence on k and n. Their proof involves a novel invariance principle for intersections of arbitrary halfspaces and a Littlewood–Offord style anticoncentration inequality for polytopes over Boolean space.

Concurrently, Chattopadhyay, De and Servedio [4] proposed a simple PRG that ϵ-fools intersections of k general halfspaces over Gaussian space, building upon the concept of Johnson-Lindenstrauss transform [10, 11]. The seed length is O(logn+poly(logk,1/ϵ)). Additionally, they show that the same PRG with seed length O(logn+poly(k,1/ϵ)) is able to fool arbitrary functions of k halfspaces.

Speaking of fooling functions of PTFs, the study by Diakonikolas, Kane and Nelson [6] stands out as the sole work that constructs a PRG for intersections of k degree-2 PTFs. Their PRG is specific to degree d2 with a O(lognpoly(k,1/ϵ)) seed length.

1.2 Main Result

In this work, we investigate the PRGs fooling any function of low-degree PTFs. The main result is the following.

Theorem 3 (Informal version of Theorem 20).

There exists an explicit PRG ϵ-fools any function of k degree-d PTFs over Gaussian space with seed length poly(k,d,1/ϵ)logn.

The proof is inspired by the PRG proposed in [13] and the work [17]. This theorem follows from two components.

(1) Bounded independence fools functions of 𝒌 degree-𝒅 PTFs

Consider the continuous random vector Y=1Li=1LYi where Yi is a R-wise independent standard Gaussian vector of length n. Every Yi,j is a standard Gaussian variable and for any degree-R polynomial f, 𝔼[f(Yi)]=𝔼y𝒩(0,1)n[f(y)]. We will prove that

Theorem 4.

(Informal version of Theorem 16) With R=O(logkdϵ) and L=poly(k,d,1ϵ), the distribution of Y ϵ-fools any function of k degree-d PTFs over Gaussian space.

The prior work [17] shows that bounded independence fools a single low-degree polynomial threshold function. This generalizes their work to the case of functions of k low-degree PTFs.

(2) Discretization of bounded independence Gaussians

An explicit PRG construction requires a discrete approximation to Gaussian vectors with bounded independence. The idea is to use a finite entropy random variable X to approximate Y. Previous work [13] uses the idea that a single Gaussian variable can be produced by two uniform random variables in [0,1] through the Box–Muller transform [2]. Therefore bounded independence Gaussian variables Yi can be generated by using bounded independence uniform random variables. Then by truncating these uniform [0,1] random variables to a sufficient precision, we obtain vectors Xi that serve as a discrete approximation of Yi. We prove that X also fools functions of k degree-d PTFs as long as X is a good approximation to Y.

Lemma 5 (Informal version of Lemma 19).

If Xi,j and Yi,j are sufficiently close with high probability, then X also fools functions of k degree-d PTFs.

2 Preliminary

Basic Notation

For n, [n] denotes the set {1,2,,n}. For αn and i[n], αi denotes the i-th coordinate of α, |α|=i=1n|αi| and α=max1in|αi|. For α,βn, αβ denotes the vector v such that vi=αiβi for all i[n], and αβ=i=1nαiβi. For αn, α!=i=1nαi!. When it is clear from the context, we will use both subscript and superscript as indices.

Derivatives and Multidimensional Taylor Expansion

For a function f:n and αn, we use αf to denote the partial derivative taken αi times in the i-th coordinate and define tf(x)=αn,|α|=t(αf(x))2. For f(a,b):n×n and α,βn, we use aαbβf to denote the partial derivative taken αi times in ai and βi times in bi. Using these notations, one has:

Theorem 6 (Multidimensional Taylor’s Theorem).

Let d and f:nn be a 𝒞d+1 function. Then for all x,yn,

f(y)=αn,|α|dαf(x)α!(yx)α+αn,|α|=d+1αf(z)α!(yx)α

where z=cx+(1c)y for some c(0,1).

Bump Function

Consider the bump function Ψ: defined by Ψ(x)={e1x21, if |x|<1,0, if |x|1. It is well known that this function is infinitely differentiable and the derivatives are bounded.

Fact 7.

For all t, |Ψ(t)(x)|t(3+o(1))t.

Let ρ be the smooth univariate function defined by ρ(x)={1, for x1,ee1(t1)21 for 0<x<1,0, for x0. It is easy to see ρ is obtained from Ψ via translation, stretch, and concatenation. We have

Fact 8.

For all t, |ρ(t)(x)|t(3+o(1))t.

Fact 9.

Let r(u,v)ρ(logulogv+c) for some constant c. Then we have that for all n,m, |nmr(u,v)unvm|(n+m)6(n+m)|u|n|v|m.

We include the proof for the above three facts in Appendix A for self-containment.

Gaussian Space and the Gaussian Noise Operator

We denote by y𝒩(0,1)n that y=(y1,,yn)n is a random vector whose components are independent standard Gaussian variables (i.e., with mean 0 and variance 1). We say a random vector Yn is a k-wise independent standard Gaussian vector if every component of Y is a standard Gaussian variable and 𝔼[p(Y)]=𝔼y𝒩(0,1)n[p(y)] for all polynomials p:n with degree at most k. For a function f:n on Gaussian space and 1p, the p-norm is denoted by fp=(𝔼y𝒩(0,1)n[|f(y)|p])1/p. For ρ[0,1], the Gaussian noise operator Uρ is the operator on the space of functions f:n defined by Uρf(x)=𝔼y𝒩(0,1)n[f(ρx+1ρ2y)].

The probabilists’ Hermite polynomials [23, Section 11] {Hj}j are defined by Hj(y)=(1)jφ(y)djφ(y)dyj where φ(y)=12πey22. The univariate Hermite polynomials {hj}j are defined by normalization: hj=1j!Hj. For a multi-index αn, the (multivariate) Hermite polynomial hα:n is hα(y)=j=1nhαj(yj). The degree of hα is |α|. The Hermite polynomials {hα}αn form an orthonormal basis for the functions over Gaussian space: 𝔼y𝒩(0,1)n[hα(y)hβ(y)]=1 iff α=β, and every degree-d polynomial f:n can be uniquely expanded as f(y)=αn,|α|df^(α)hα(y). We can also expand the function f(x+λy) in the Hermite basis in a manner similar to Taylor expansion.

Lemma 10 (Lemma 16 in [17]).

Suppose f(y)=αnf^(α)hα(y), we have f(x+λy)=αnαϕ(x)α!λ|α|/2hα(y), where ϕ(x)=U1λf(x1λ).

The function Uρf has the following expansion: Uρf(y)=αn,|α|dρ|α|f^(α)hα(y). The definition of Uρ can be extended to ρ>1 by its action on the Hermite polynomials: Uρhα(y)=ρ|α|hα(y). We will use the following hypercontractive inequality:

Theorem 11.

Let f:n and 2p, fpUp1f2.

For more details on analysis over Gaussian space, readers may refer to [23].

Low-Degree Polynomials

Low-degree polynomials are extensively studied in the literature. We list some results used in this paper. It is well-know that low-degree polynomials have the following anti-concentration property:

Lemma 12 (Theorem 8 in [3]).

Let p:n be a polynomial of degree d with p2=1. Then we have Prx𝒩(0,1)n[|p(x)|ϵ]=O(dϵ1/d).

Suppose p is a low-degree polynomial, the following gives an estimation on the deviation of p(x) caused by a small perturbation.

Lemma 13 (Lemma 22 in [13]).

Let p:n be a polynomial of degree d with p2=1. Suppose xn be a vector with xB(B>1). Let x be another vector such that xxδ<1. Then we have |p(x)p(x)|δnd/2O(B)d.

The magnitudes of the derivatives of a low-degree polynomial are likely to grow at a moderate rate with high probability. Formally,

Lemma 14 (Lemma 6 in [17]).

Let p:n be an arbitrary polynomial of degree d and y𝒩(0,1)n, the following holds with probability at least 1ϵd3:

tp(y)O(ϵ1)t1p(y) for all 1td.

The following lemma gives quantitative bounds on how much the derivatives tp(x+λy) are concentrated around those of ϕ(x)=𝔼y𝒩(0,1)n[p(x+λy)] when y𝒩(0,1)n.

Lemma 15 (Lemma 23 in [17]).

Let 0λ<1 and p:n be an arbitrary polynomial of degree d and ϕ(x)=U1λp(x1λ)=𝔼y𝒩(0,1)n[p(x+λy)]. For 0td and y𝒩(0,1)n,

(𝔼y𝒩(0,1)n[tp(x+λy)tϕ(x)R])1Rj=t+1d(λdR)jtjϕ(x)2.

3 Fooling the Functions of PTFs via Bounded Independence

In this section, we show that a random Gaussian vector matching certain moments fools any function of low-degree polynomial threshold functions. Formally, we prove

Theorem 16.

Fix a small constant 0<ϵ<1 and let R be an integer. Let p1,,pk:n be arbitrary polynomials of degree d and f:{0,1}k{0,1} be an arbitrary Boolean function. Define function

F(x)f(sign(p1(x)),,sign(pk(x))).

Let Y=1Li=1LYi where Yi is a 2dR-wise independent standard Gaussian vector of length n and L=Ω(k2d3R15ϵ2). Then, we have

|𝔼Y[F(Y)]𝔼y𝒩(0,1)n[F(y)]|=O(ϵkd3)+kdL2Ω(R).

The key idea in the proof of Theorem 16 is to analyze the derivatives of the disturbed function ϕi(x)=𝔼y𝒩(0,1)n[pi(x+λy)]. We will see that once the derivatives of ϕi are well-controlled by its preceding order derivative at x, tpi(x+λy) is concentrated around tϕi(x) for a random y, and pi(x+λy) and ϕi(x) share the same sign with high probability. Starting from this point, we use the mollifier introduced in [17]

G(x)i=1kt=0d1ρ(log(tpi(x)216ϵ2t+1pi(x)2)) (1)

to judge whether the derivatives are all well-controlled for all k polynomials. G(x)=0 as long as a certain order of derivative that is not controlled by its preceding order derivative. Our proof consists of following steps:

  • Approximation using the mollifier G: We first establish that

    |𝔼Y[F(Y)]𝔼y[F(y)]||𝔼Y[F(Y)G(Y)]𝔼y[F(y)G(y)]|.

    This approximation enables us to focus primarily on the analysis of F(y)G(y) in the subsequent steps.

  • Hybrid argument: Let λ=L1, y=λi=1Lyi where yi𝒩(0,1)n and Zi=λ(y1++yi1+Yi+1+YL). We will show

    𝔼[F(Zi+λYi)G(Zi+λYi)]𝔼[F(Zi+λyi)G(Zi+λyi)]. (2)

    Therefore by the triangle inequality, we have

    𝔼Y[F(Y)G(Y)] =𝔼[F(Z1+λY1)G(Z1+λY1)]
    𝔼[F(Z1+λy1)G(Z1+λy1)]=𝔼[F(Z2+λY2)G(Z2+λY2)]
    𝔼[F(ZL+λyL)G(ZL+λyL)]=𝔼[F(y)G(y)].

    To prove (2), we show for any fixed x,

    𝔼[F(x+λYi)G(x+λYi)]𝔼[F(x+λyi)G(x+λyi)].

    This is done by a case analysis:

    • The derivatives of all k polynomials ϕj(x) are well-controlled at point x. In this case, all pj(x+λYi) and pj(x+λyi) share the same sign with high probability. Thus, it is highly likely that F(x+λYi) and F(x+λyi) are nearly the same constant. It suffices to show Yi fools the mollifier function G(x+λyi).

    • At least one derivative is not controlled. In this case, we will show that G(x+λYi) and G(x+λyi) are 0 with high probability. This implies that F(x+λYi)G(x+λYi)=F(x+λyi)G(x+λyi)=0 with overwhelming probability.

In the subsequent sections, Section 3.1 first demonstrates that Yi is able to fool the mollifier function G when x is a well-behaved point. Section 3.2 shows the closeness of a single step in the hybrid argument. Lastly, we prove Theorem 16 using approximation and the hybrid argument in Section 3.3.

3.1 Fooling the Mollifier 𝑮

We begin with proving that a 2dR-wise independent standard Gaussian vector Y fools the mollifier function G(x+λy). To achieve this, we utilize the Taylor expansion to expand the mollifier function G(x+λy) up to a specified order. As a result, G(x+λy) is decomposed into two parts: a degree-d(R1) polynomial l(y) and a remainder term Δ(y). We mainly show that 𝔼[Δ] is negligible under both pseudorandom distribution and true Gaussian distribution. This leads us to the conclusion that 𝔼[G(x+λy)]𝔼[l(y)] and 𝔼[G(x+λY)]𝔼[l(Y)]. Furthermore, since l(y) has degree at most dR, it follows that 𝔼[l(y)]=𝔼[l(Y)]. Thus, we conclude that 𝔼[G(x+λy)]𝔼[G(x+λY)].

Lemma 17.

Fix a small constant 0<ϵ<1 and let R be an integer. Let p1,,pk:n be arbitrary polynomials of degree d. Define ϕi(x)U1λpi(x1λ)=𝔼y𝒩(0,1)n[pi(x+λy)] for all pi. Suppose that a fix point xn satisfies t+1ϕi(x)1ϵtϕi(x) for any 1ik and 0td1. Let Y be a 2dR-wise independent standard Gaussian vector of length n. For λ=O(k2d3R15ϵ2), we have

|𝔼Y[G(x+λY)]𝔼y𝒩(0,1)n[G(x+λy)]|=kd2Ω(R),

where G is defined in (1).

Proof.

Let σ(z)ρ(zlog16ϵ2) and by the definition of function G() defined in (1),

G(z)=i=1kt=0d1σ(logtpi(z)2logt+1pi(z)2).

Define variables {sit}1ik,0td1 and {rit}1ik,0td1 by letting sit=tpi(x+λy)2 and rit=t+1pi(x+λy)2 as functions of y. Apparently, we have

G(x+λy)=g(s,r)i=1kt=0d1σ(logsitlogrit).

Therefore, it is equivalent to prove

|𝔼y𝒩(0,1)n[g(s,r)]𝔼Y[g(s,r)]|=kd2Ω(R).

To this end, we expand g(s,r) into R-th order using the Taylor expansion at some point (a,b): g(s,r)=l(s,r)+Δ, where l(s,r) is a polynomial of y of degree at most dR and Δ is the remainder. Since Y is 2dR-wise independent, we know 𝔼y𝒩(0,1)n[l(s,r)]=𝔼Y[l(s,r)]. Therefore, it suffices to show 𝔼y𝒩(0,1)n[|Δ|] and 𝔼Y[|Δ|] are bounded by kd2Ω(R).

More specifically, we choose to expand g(s,r) at points ait=tϕi(x)2 and bit=t+1ϕi(x)2 via Theorem 6: g(s,r)=l(s,r)+Δ, where

l(s,r)=(αit)i[k],0td1kd(βit)i[k],0td1kd|α|+|β|<Rsαrβg(a,b)α!β!(sa)α(rb)β

and

Δ=(αit)i[k],0td1kd(βit)i[k],0td1kd|α|+|β|=Rsαrβg(s,r)α!β!(sa)α(rb)β.

for some (s,r) on the line segment joining (s,r) and (a,b). It is not hard to see that l(s,r) is a polynomial of y of degree at most d(R1). For the remainder term Δ, we will prove the following bound: 𝔼y𝒩(0,1)n[|Δ|]kd2Ω(R).

We now give bounds on 𝔼y𝒩(0,1)n[|Δ|]. The same argument applies to Y as well. Fix a small constant 0<δ<1kdR7. Let it be the event that

|tpi(x+λy)tϕi(x)|δtϕi(x).

Now let =1ik,0tdit. Note that

Δ=Δ𝟙+Δ𝟙¯=Δ𝟙+g(s,r)𝟙¯l(s,r)𝟙¯.

Here, 𝟙𝒜=1 when event 𝒜 occurs and 𝟙𝒜=0 otherwise. Therefore, by the triangle inequality, we have

𝔼y[|Δ|] 𝔼y[Δ𝟙]+𝔼y[g(s,r)𝟙¯]+𝔼y[|l(s,r)|𝟙¯]
𝔼y[Δ𝟙]+𝔼y[𝟙¯]+𝔼y[|l(s,r)|𝟙¯]
𝔼y[Δ𝟙](Term 1)+𝔼y[𝟙¯](Term 2)+𝔼y[l2(s,r)](Term 3)𝔼y[𝟙¯]. (Cauchy–Schwarz)

We are next to bound Term 13.

Bounding Term 1.

If event occurs, we have

|Δ| (αit)kd,(βit)kd|α|+|β|=R|sαrβg(s,r)|α!β!i=1kt=0d1|sitait|αit|ritbit|βit
(αit)kd,(βit)kd|α|+|β|=RR6Ri=1kt=0d1|sitait|αit|sit|αit|ritbit|βit|rit|βit
(αit)kd,(βit)kd|α|+|β|=RR6R(2δ+δ2(1δ)2)R=(R+2kd1R)R6R(4δ)R2R,

where the second inequality is from Fact 9, and the third one is true by the following facts:

  • (1δ)2aitsit(1+δ)2ait, (1δ)2bitrit(1+δ)2bit,

  • sitmin{sit,ait}(1δ)2ait and ritmin{rit,bit}(1δ)2bit, since (sit,rit) lies between (sit,rit) and (ait,bit).

This gives us 𝔼y[Δ𝟙]2R.

Bounding Term 2.

Since

tpi(x+λy)tϕi(x)tpi(x+λy)tϕi(x)

and

tϕi(x)tpi(x+λy)tpi(x+λy)tϕi(x),

we have

|tpi(x+λy)tϕi(x)|tpi(x+λy)tϕi(x) (3)

This gives us

Pry[it] Pry[tpi(x+λy)tϕi(x)δtϕi(x)]
1𝔼y[tpi(x+λy)tϕi(x)R]δRtϕi(x)R
1(j=t+1d(λdR)jtjϕ(x)2)R2δRtϕi(x)R
1(1δ)R(j=1dt(λdRϵ2)j)R21(2λdRδ2ϵ2)R

where the second inequality is from Markov’s inequality, the third one is from Lemma 15 and the fourth one holds since xn satisfies t+1ϕi(x)1ϵtϕi(x) for any 1ik and 0td1. For λ=O(k2d3R15ϵ2), we have Pry[it]12R. Consequently, Pry[]1kd2R. Therefore, we have 𝔼y[𝟙¯]kd2R.

Bounding Term 3.

We next to upper bound 𝔼y[l2(s,r)]. Note that

𝔼y[l2(s,r)]
|α|+|β|<R|α|+|β|<R|sαrβg(a,b)|α!β!|sαrβg(a,b)|α!β!𝔼y[|(sa)α(rb)β(sa)α(rb)β|]
q1<Rq2<R|α|+|β|=q1|α|+|β|=q2R6(q1+q2)𝔼y[i=1kt=0d1|sitait|αit|ait|αit|ritbit|βit|bit|βit|sitait|αit|ait|αit|ritbit|βit|bit|βit]()

By generalized Hölder’s inequality, for q1+q20

() i,t(𝔼y[|sitait|q1+q2])αitq1+q2|ait|αit(𝔼y[|ritbit|q1+q2])βitq1+q2|bit|βit
(𝔼y[|sitait|q1+q2])αitq1+q2|ait|αit(𝔼y[|ritbit|q1+q2])βitq1+q2|bit|βit.

Note that for 0<q2R,

𝔼y[(|tpi(x+λy)2tϕi(x)2|tϕi(x)2)q]
𝔼y[(2|tpi(x+λy)tϕi(x)|tϕi(x)+|tpi(x+λy)tϕi(x)|2tϕi(x)2)q]
𝔼y[(2tpi(x+λy)tϕi(x)tϕi(x)+tpi(x+λy)tϕi(x)2tϕi(x)2)q]
= j=0q2j𝔼y[tpi(x+λy)tϕi(x)2qjtϕi(x)2qj]
j=0q2j(4λdRϵ2)2qj=4qj=0q(λdRϵ2)2qj
24q(λdRϵ2)q(17λdRϵ2)q

where the first inequality is from |a2b2|=|ab||a+b||ab|(|a|+|b|)|ab|(2|a|+|ab|), the second one is from Eq.(3) and the third inequality is from Lemma 15. Therefore,

()i,t(17λdRϵ2)αit+βit+αit+βit=(17λdRϵ2)q1+q2.

Consequently,

𝔼y[l2(s,r)] q1<Rq2<R|α|+|β|=q1|α|+|β|=q2R6(q1+q2)(17λdRϵ2)q1+q2
q1<Rq2<R(R+2kd1)q1+q2R6(q1+q2)(17λdRϵ2)q1+q2
=q=02R2(q+1)((R+2kd1)R617λdRϵ2)q.

For λ=O(k2d3R15ϵ2) sufficiently small, we have 𝔼y[l2(s,r)]q=02R2(q+1)4q<2R.

Thus, putting everything together, we have

𝔼y[|Δ|] 𝔼y[Δ𝟙]+𝔼y[𝟙¯]+𝔼y[l2(s,r)]𝔼y[𝟙¯]
2R+kd2R+2Rkd2R
kd2Ω(R).

We now regard l(s,r) as a function of y, and from the above inequality we have

|𝔼y[G(x+λy)]𝔼y[l(s,r)]|𝔼y[|Δ|]kd2Ω(R).

Similarly, the same argument applying on 2dR-wise independent Gaussian vector Y gives us

|𝔼Y[G(x+λY)]𝔼Y[l(s,r)]|kd2Ω(R).

The lemma then follows from the fact that 𝔼y[l(s,r)]=𝔼Y[l(s,r)], since l(s,r) is a polynomial of y of degree at most d(R1).

3.2 A Single Step in the Hybrids

In this section, we analyze one single step in the entire hybrid argument. We will show that for any x, we have that 𝔼Y[F(x+λY)G(x+λY)]𝔼y[F(x+λy)G(x+λy)] for 2dR-wise independent Gaussian Y and true Gaussian y.

Let ϕi(x)=U1λpi(x1λ)=𝔼y[pi(x+λy)]. The proof proceeds through a case analysis based on the behavior of ϕi at the fixed point x. Specifically, we define x as well-behaved if t+1ϕi(x)1ϵtϕi(x) for all t[d] and i[k]. In other words, for each function ϕi, its t-th order derivatives are controlled by its (t1)-th order derivatives.

  • In the scenario where x is not well-behaved, we can identify an i0 and a t0 such that with at least probability 12R+1,

    t0+1pi0(x+λy)>14ϵt0pi0(x+λy).

    Thus, it is highly probable that the mollifier function G(x+λy)=0. So, the expectation of F(x+λy)G(x+λy) is no more that 2R+1. The same argument works for Y as well.

  • For the case that x is well-behaved, we will show that for all pi, pi(x+λy) and pi(x+λY) are nearly the same constant. This implies F(x+λy) and F(x+λY) are equal in most situations. Then it suffices to show Y fools the mollifier, as discussed in the previous section.

Lemma 18.

Fix a small constant 0<ϵ<1 and let R be an integer. Let p1,,pk:n be arbitrary polynomials of degree d and f:{0,1}k{0,1} be an arbitrary Boolean function. Define function

F(x)f(sign(p1(x)),,sign(pk(x))).

Let Y be a 2dR-wise independent standard Gaussian vector of length n. For any xn and λ=O(k2d3R15ϵ2)

|𝔼Y[F(x+λY)G(x+λY)]𝔼y𝒩(0,1)n[F(x+λy)G(x+λy)]|=kd2Ω(R),

where G is defined in (1).

Proof.

Let ϕi(x)=U1λpi(x1λ). Define x is good if for any 1ik and 0td1, t+1ϕi(x)1ϵtϕi(x). We prove this lemma by considering x is good or not.

We first consider that x is not good. In this case, we will show G(x+λy)=0 holds with high probability. Consequently, F(x+λy)G(x+λy) is zero with high probability. To this end, it suffices to find an i0 and a t0 such that

t0+1pi0(x+λy)>14ϵt0pi0(x+λy).

We choose an arbitrary i0 satisfying that there exists 0td1 such that t+1ϕi0(x) >1ϵtϕi0(x). Since x is not good, we know such i0 exists. And let t0 be the largest t such that t+1ϕi0(x)>1ϵtϕi0(x) holds. It is not hard to check

  • t0ϕi0(x)<ϵt0+1ϕi0(x),

  • t+1ϕi0(x)1ϵtϕi0(x) for tt0+1.

We are next to prove the following inequalities hold with high probability,

  1. (a)

    t0pi0(x+λy)<2ϵt0+1ϕi0(x)

  2. (b)

    t0+1pi0(x+λy)>12t0+1ϕi0(x)

It is easy to see (a) and (b) give us t0+1pi0(x+λy)>14ϵt0pi0(x+λy).

Showing (a).

By Markov’s inequality, we have

Pry[t0pi0(x+λy)t0ϕi0(x)ϵt0+1ϕi0(x)]
𝔼y[t0pi0(x+λy)t0ϕi0(x)R]ϵRt0+1ϕi0(x)R(t=t0+1d(λdR)tt0tϕ(x)2)R2ϵRt0+1ϕi0(x)R
(t=t0+1d(λdR)tt0(1ϵ2)tt01t0+1ϕ(x)2)R2ϵRt0+1ϕi0(x)R(t=1dt0(λdRϵ2)t)R2.

Here the second inequality is from Lemma 15. The third inequality uses the condition that t+1ϕi0(x)1ϵtϕi0(x) for tt0+1. Since λϵ2100dR, this probability is bounded by 2R. Therefore, with probability at least 12R,

t0pi0(x+λy)t0ϕi0(x)<ϵt0+1ϕi0(x).

Moreover, we know t0ϕi0(x)<ϵt0+1ϕi0(x). So, we have with probability at least 12R,

t0pi0(x+λy)t0ϕi0(x)+t0pi0(x+λy)t0ϕi0(x)<2ϵt0+1ϕi0(x).
Showing (b).

Similarly, we have

Pry[t0+1pi0(x+λy)t0+1ϕi0(x)12t0+1ϕi0(x)]
𝔼y[t0+1pi0(x+λy)t0+1ϕi0(x)R]2Rt0+1ϕi0(x)R(t=t0+2d(λdR)tt01tϕ(x)2)R22Rt0+1ϕi0(x)R
(t=t0+2d(λdRϵ2)tt01t0+1ϕ(x)2)R22Rt0+1ϕi0(x)R(4t=1dt01(λdRϵ2)t)R2.

Here the second inequality is from Lemma 15. The third inequality uses the condition that t+1ϕi0(x)1ϵtϕi0(x) for tt0+1. Since λϵ2100dR, this probability is bounded by 2R. Therefore, with probability at least 12R,

t0+1pi0(x+λy)t0+1ϕi0(x)t0+1pi0(x+λy)t0+1ϕi0(x)>12t0+1ϕi0(x).

Thus, combining (a) and (b), we have that with probability at least 122R,

t0+1pi0(x+λy)>12t0+1ϕi0(x)>14ϵt0pi0(x+λy),

and consequently G(x+λy)=0. This gives us the bound on the following expectation

𝔼y𝒩(0,1)n[F(x+λy)G(x+λy)]22R.

Since Y is dR-wise independent, the above argument still holds for Y. So,

|𝔼Y[F(x+λY)G(x+λY)]𝔼y𝒩(0,1)n[F(x+λy)G(x+λy)]|42R.

Now suppose that x is good. That is, for any 1ik and 0td1, t+1ϕi(x)1ϵtϕi(x). In this case, we will prove that the sign of pi(x+λy) is the same as the sign of ϕi(x) with high probability over the random variable y. Therefore, F(x+λy) is almost like a constant, since the value of F(x+λy) only depends on the signs of all pi(x+λy). To show the signs of pi(x+λy) and ϕi(x) are the same, it suffices to show pi(x+λy)ϕi(x)>0. Let

qi(y)pi(x+λy)ϕi(x)1=1ϕi(x)0<|α|dαϕi(x)α!λ|α|/2hα(y).

Here, we expand pi(x+λy)=ϕi(x)+0<|α|dαϕi(x)α!λ|α|/2hα(y) according to Lemma 10. We have by applying hypercontractive inequality in Theorem 11,

qi(y)RURqi(y)2 =1ϕi(x)0<|α|dαϕi(x)α!R|α|/2λ|α|/2hα(y)2
0<tdtϕi(x)2ϕi(x)2(λR)t0<td(λRϵ2)t.

In the last inequality, we use our assumption that tϕi(x)1ϵt1ϕi(x)|ϕi(x)|ϵt. Since λRϵ2 is sufficiently small, we have qi(y)R14. Therefore, by Markov’s inequality, we have Pry[|qi(y)|12]2Rqi(y)RR2R. This means that with probability at least 12R, we have |pi(x+λy)ϕi(x)1|12, and therefore pi(x+λy)ϕi(x)12. Thus, with probability at least 12R, the sign of pi(x+λy) is the same as the sign of ϕi(x). Then by a union bound,

Pry[1ik,sign(pi(x+λy))=sign(ϕi(x))]1k2R.

Let c=f(sign(ϕ1(x)),,sign(ϕk(x))) be a constant. We have

|𝔼y[F(x+λy)G(x+λy)]𝔼y[cG(x+λy)]|k2R.

The same argument applying on 2dR-wise independent Gaussian vector Y gives us

|𝔼Y[F(x+λY)G(x+λY)]𝔼Y[cG(x+λY)]|k2R.

By Lemma 17, we know |𝔼Y[G(x+λY)]𝔼y𝒩(0,1)n[G(x+λy)]|=kd2Ω(R). Therefore, we have

|𝔼Y[F(x+λY)G(x+λY)]𝔼y𝒩(0,1)n[F(x+λy)G(x+λy)]|=kd2Ω(R).

3.3 Proof of Theorem 16

Proof of Theorem 16.

By Lemma 14, the following holds with probability at least 1ϵkd3:

tpi(y)O(1ϵ)t1pi(y) for all 1td and 1ik.

Recall the function G(x)=i=1kt=0d1ρ(log(tpi(x)216ϵ2t+1pi(x)2)) as defined in (1). Note that G(x)=0 if there exists some i and t such that tpi(y)>O(1ϵ)t1pi(y). Thus, Pry[G(y)=1]1O(ϵkd3). We have

𝔼y[F(y)] =𝔼y[F(y)(1G(y))]+𝔼y[F(y)G(y)]
𝔼y[1G(y)]+𝔼Y[F(Y)G(Y)]+|𝔼y[F(y)G(y)]𝔼Y[F(Y)G(Y)]|
O(ϵkd3)+𝔼Y[F(Y)]+|𝔼y[F(y)G(y)]𝔼Y[F(Y)G(Y)]|.

Let y=1Li=1Lyi where yi𝒩(0,1)n and denote Zi=1L(y1++yi1+Yi+1+YL). We have

|𝔼y[F(y)G(y)]𝔼Y[F(Y)G(Y)]|
i=1L|𝔼Zi,Yi[F(Zi+1LYi)G(Zi+1LYi)]𝔼Zi,yi[F(Zi+1Lyi)G(Zi+1Lyi)]|
kdL2Ω(R)

where the last inequality is from Lemma 18. Thus,

𝔼y[F(y)]𝔼Y[F(Y)]+O(ϵkd3)+kdL2Ω(R).

And the other side follows from considering 1F(x).

4 Discretization

To give an explicit construction of a PRG, we need a discretization of R-wise independent Gaussian distributions. In this section, we show an algorithm which outputs L vectors {Xi}1iL approximating Yi, that is, |Xi,jYi,j| is sufficiently small. Before that, we first prove that if X and Y are close enough, then X also fools any function of low-degree polynomial threshold functions.

Lemma 19.

Let 0<ϵ,δ<1, and R be an integer. Let Y=1Li=1LYi where Yi is an R-wise independent Gaussian vector of length n for 1iL. Let p1,,pk:n be arbitrary polynomials of degree d and f:{0,1}k{0,1} be an arbitrary Boolean function. Define functions

F(x)f(sign(p1(x)),,sign(pk(x)))

Suppose that for any such function F,

|𝔼Y[F(Y)]𝔼y𝒩(0,1)n[F(y)]|ϵ.

Suppose that {Xi}1iL are random vectors of length n and there is a joint distribution over X and Y such that for each 1iL,1jn, Pr[|Xi,jYi,j|δ]1δ.

Let X=1Li=1LXi and we have that for any such function F

|𝔼X[F(X)]𝔼y𝒩(0,1)n[F(y)]|ϵ+k22kdδ1/dnLlog1δ+O(22knLδ).
Proof.

Let qi(x)=pi(x)+δ(nL)d/2(log1δ)d and F~(x)=f(sign(q1(x)),,sign(qk(x))). We will prove that

  1. (a)

    𝔼[F(X)]𝔼[F~(y)]+ϵ+O(22knLδ),

  2. (b)

    𝔼[F~(y)]𝔼[F(y)]+k22kdδ1/dnLlog1δ.

Combing (a) and (b), we have

𝔼[F(X)]𝔼[F(y)]+k22kdδ1/dnLlog1δ+ϵ+O(22knLδ).

The other side can be obtained in a similar way by considering 1F(x).

Proving (a).

Since F~ is a function of degree-d PTFs and Y fools such a function, we have 𝔼[F~(Y)]𝔼[F~(y)]+ϵ. Therefore, it suffices to prove that 𝔼[F(X)]𝔼[F~(Y)]+O(22knLδ).

Fix a set S[k]. Let PS(x)=iSsign(pi(x)) and QS(x)=iSsign(qi(x)). We first show that

𝔼[PS(X)]𝔼[QS(Y)]+O(nLδ).

Let event denote that for all i[L],j[n], |Yi,j|log1δ and |Xi,jYi,j|δ. By the tail bound of the standard Gaussian distribution, we have Pr[]1O(nLδ). We have

𝔼[PS(X)]=Pr[iSpi(X)0] Pr[iSpi(X)0]+Pr[¯]
Pr[iSpi(Y)δ(nL)d/2(log1δ)d]+O(nLδ)
=𝔼[QS(Y)]+O(nLδ),

where the second inequality is by Lemma 13 and viewing pi(1Li=1LXi) as a degree d function of nL variables.

Note that f(x)=S[k]f(𝟙S)iSxiiS(1xi) where 𝟙S denotes the length-k string with 1’s only at coordinates in S. This can be further simplified in the form

f(x)=S[k]cSiSxi,

with S|cS|22k. So, we have

𝔼[F(X)]=S[k]cS𝔼[PS(X)] =S[k]cS𝔼[QS(Y)]+S[k]cS(𝔼[PS(X)]𝔼[QS(Y)])
S[k]cS𝔼[QS(Y)]+O(22knLδ)=𝔼[F~(Y)]+O(22knLδ).
Proving (b).

Next we show 𝔼[F~(y)] and 𝔼[F(y)] are close. Note that

𝔼[QS(y)]=Pr[iSpi(y)δ(nL)d/2(log1δ)d]Pr[iSpi(y)0]+kdδ1/dnLlog1δ,

where the inequality is from Lemma 12. Thus, we have

𝔼[QS(y)]𝔼[PS(y)]+kdδ1/dnLlog1δ.

Furthermore, we know that

𝔼[F~(y)]=S[k]cS𝔼[QS(y)] =S[k]cS𝔼[PS(y)]+S[k]cS(𝔼[QS(y)PS(y)])
𝔼[F(y)]+k22kdδ1/dnLlog1δ.

We now prove the main theorem for constructing an explicit pseudorandom generator. The idea is that a standard Gaussian variable can be generated using two uniform [0,1] random variables through the Box–Muller transform [2]. Let Yi,j=2logui,jcos(2πvi,j) where ui,j and vi,j are uniform in [0,1]. Then Yi,j is a Gaussian variable. Thus, if we truncate ui,j and vi,j to a certain precision and produce Xi,j in a similar manner, X approximates Y with high probability.

Theorem 20.

There exists an explicit PRG which ϵ-fools any functions of any k degree-d polynomial threshold functions over 𝒩(0,1)n with seed length O(k5d11ϵ2logkdnϵ).

Proof.

In Theorem 16, set parameter ϵ as ϵkd3 and set R as Clogkdϵ for some large constant C. Then for L=Ck4d9ϵ2polylogkdϵ where C is a large constant, we have

|𝔼Y[F(Y)]𝔼y𝒩(0,1)n[F(y)]|O(ϵ).

Similar to the proof of Corollary 2 in [13], we can let Yi,j generated by

Yi,j=2logui,jcos(2πvi,j)

where ui=(ui,1,,ui,n) and vi=(vi,1,,vi,n) are 2dR-wise independent uniform [0,1] random vectors. Then let ui,j and vi,j be M=C′′kdlogkdnϵ-bit approximations to ui,j and vi,j (i.e., round ui,j and vi,j to multipels of 2M), where C′′ is a large constant, and let

Xi,j=2logui,jcos(2πvi,j).

Letting δ=Ω(2M/2), for the same reason as in the proof of Corollary 2 in [13], we have |Xi,jYi,j|<δ with probability at least 1δ. Then, by Lemma 19,

|𝔼X[F(X)]𝔼y𝒩(0,1)n[F(y)]|O(ϵ)+k22kdδ1/dnLlog1δ+O(22knLδ)=O(ϵ).

Note that Xi can be generated by 2dR-wise independent random variables ui,j and vi,j taken uniformly from {2M,22M,32M,,1} using O(dRM) randomness. Thus generating X uses O(LdRM)=O(k5d11ϵ2logkdnϵ) randomness.

References

  • [1] Louay M. J. Bazzi. Polylogarithmic independence can fool DNF formulas. SIAM Journal on Computing, 38(6):2220–2272, 2009. doi:10.1137/070691954.
  • [2] G. E. P. Box and Mervin E. Muller. A note on the generation of random normal deviates. The Annals of Mathematical Statistics, 29(2):610–611, 1958. doi:10.1214/aoms/1177706645.
  • [3] Anthony Carbery and James Wright. Distributional and Lq norm inequalities for polynomials over convex bodies in n. Mathematical Research Letters, 8:233–248, 2001. doi:10.4310/MRL.2001.v8.n3.a1.
  • [4] Eshan Chattopadhyay, Anindya De, and Rocco A. Servedio. Simple and efficient pseudorandom generators from Gaussian processes. In Proceedings of the 34th Computational Complexity Conference, Dagstuhl, DEU, 2020. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.CCC.2019.4.
  • [5] Ilias Diakonikolas, Parikshit Gopalan, Ragesh Jaiswal, Rocco A. Servedio, and Emanuele Viola. Bounded independence fools halfspaces. SIAM Journal on Computing, 39(8):3441–3462, 2010. doi:10.1137/100783030.
  • [6] Ilias Diakonikolas, Daniel M. Kane, and Jelani Nelson. Bounded independence fools degree-2 threshold functions. In 2010 IEEE 51st Annual Symposium on Foundations of Computer Science, pages 11–20, 2010. doi:10.1109/FOCS.2010.8.
  • [7] Parikshit Gopalan, Daniel M. Kane, and Raghu Meka. Pseudorandomness via the discrete fourier transform. SIAM Journal on Computing, 47(6):2451–2487, 2018. doi:10.1137/16M1062132.
  • [8] Parikshit Gopalan, Ryan O’Donnell, Yi Wu, and David Zuckerman. Fooling functions of halfspaces under product distributions. In 2010 IEEE 25th Annual Conference on Computational Complexity, pages 223–234, 2010. doi:10.1109/CCC.2010.29.
  • [9] Prahladh Harsha, Adam Klivans, and Raghu Meka. An invariance principle for polytopes. J. ACM, 59(6), 2013. doi:10.1145/2395116.2395118.
  • [10] William B Johnson, Joram Lindenstrauss, and Gideon Schechtman. Extensions of Lipschitz maps into Banach spaces. Israel Journal of Mathematics, 54(2):129–138, 1986. doi:10.1007/BF02764938.
  • [11] Daniel Kane, Raghu Meka, and Jelani Nelson. Almost optimal explicit johnson-lindenstrauss families. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, pages 628–639, Berlin, Heidelberg, 2011. Springer Berlin Heidelberg. doi:10.1007/978-3-642-22935-0_53.
  • [12] Daniel M. Kane. k-independent Gaussians fool polynomial threshold functions. In 2011 IEEE 26th Annual Conference on Computational Complexity, pages 252–261, 2011. doi:10.1109/CCC.2011.13.
  • [13] Daniel M. Kane. A small PRG for polynomial threshold functions of Gaussians. In Proceedings of the 2011 IEEE 52nd Annual Symposium on Foundations of Computer Science, pages 257–266, USA, 2011. IEEE Computer Society. doi:10.1109/FOCS.2011.16.
  • [14] Daniel M. Kane. A structure theorem for poorly anticoncentrated Gaussian chaoses and applications to the study of polynomial threshold functions. In 2012 IEEE 53rd Annual Symposium on Foundations of Computer Science, pages 91–100, 2012. doi:10.1109/FOCS.2012.52.
  • [15] Daniel M. Kane. A pseudorandom generator for polynomial threshold functions of Gaussian with subpolynomial seed length. In 2014 IEEE 29th Conference on Computational Complexity, pages 217–228, 2014. doi:10.1109/CCC.2014.30.
  • [16] Daniel M. Kane. A Polylogarithmic PRG for Degree 2 Threshold Functions in the Gaussian Setting. In David Zuckerman, editor, 30th Conference on Computational Complexity (CCC 2015), volume 33 of Leibniz International Proceedings in Informatics (LIPIcs), pages 567–581, Dagstuhl, Germany, 2015. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.CCC.2015.567.
  • [17] Zander Kelley and Raghu Meka. Random restrictions and PRGs for PTFs in Gaussian space. In Shachar Lovett, editor, 37th Computational Complexity Conference (CCC 2022), volume 234 of Leibniz International Proceedings in Informatics (LIPIcs), pages 21:1–21:24, Dagstuhl, Germany, 2022. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.CCC.2022.21.
  • [18] Adam R. Klivans, Ryan O’Donnell, and Rocco A. Servedio. Learning geometric concepts via Gaussian surface area. In 2008 49th Annual IEEE Symposium on Foundations of Computer Science, pages 541–550, 2008. doi:10.1109/FOCS.2008.64.
  • [19] Pravesh K. Kothari and Raghu Meka. Almost optimal pseudorandom generators for spherical caps: extended abstract. In Proceedings of the Forty-Seventh Annual ACM Symposium on Theory of Computing, pages 247–256, New York, NY, USA, 2015. Association for Computing Machinery. doi:10.1145/2746539.2746611.
  • [20] J. W. Lindeberg. Eine neue herleitung des exponentialgesetzes in der wahrscheinlichkeitsrechnung. Mathematische Zeitschrift, 15:211–225, 1922. doi:10.1007/BF01494395.
  • [21] Raghu Meka and David Zuckerman. Pseudorandom generators for polynomial threshold functions. SIAM Journal on Computing, 42(3):1275–1301, 2013. doi:10.1137/100811623.
  • [22] Jorge Nocedal and Stephen J. Wright, editors. Quadratic Programming, pages 438–486. Springer New York, New York, NY, 1999. doi:10.1007/0-387-22742-3_16.
  • [23] Ryan O’Donnell. Analysis of boolean functions. Cambridge University Press, 2014. doi:10.1017/CBO9781139814782.
  • [24] Ryan O’Donnell, Rocco A. Servedio, and Li-Yang Tan. Fooling Gaussian PTFs via local hyperconcentration. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 1170–1183, New York, NY, USA, 2020. Association for Computing Machinery. doi:10.1145/3357713.3384281.
  • [25] Ryan O’Donnell, Rocco A. Servedio, and Li-Yang Tan. Fooling polytopes. J. ACM, 69(2), January 2022. doi:10.1145/3460532.
  • [26] Alexander Razborov. A simple proof of Bazzi’s theorem. ACM Trans. Comput. Theory, 1(1), February 2009. doi:10.1145/1490270.1490273.
  • [27] R. A. Servedio and L. Tan. Fooling intersections of low-weight halfspaces. In 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS), pages 824–835, Los Alamitos, CA, USA, October 2017. IEEE Computer Society. doi:10.1109/FOCS.2017.81.
  • [28] Rocco A. Servedio. Every linear threshold function has a low-weight approximator. In 21st Annual IEEE Conference on Computational Complexity (CCC’06), pages 18–32, 2006. doi:10.1109/CCC.2006.18.

Appendix A Facts about Bump Function

  • Fact 7

    For all t, |Ψ(t)(x)|t(3+o(1))t.

Proof.

It is easy to check there exists a series of polynomials {Pt}t such that for x(1,1)

Ψ(t)(x)=Pt(x)(1x2)2tΨ(x).

Besides, {Pt}t has the following recursion:

P0(x)=1,P1(x)=2x,Pt=(1x2)2Pt1(x)+4(t1)x(1x2)Pt1(x)2xPt1(x).

The degree of Pt is at most 3t. Therefore we have

|Ψ(t)(x)|(maxx(1,1)|Pt(x)|)(maxx(1,1)Ψ(x)(1x2)2t).

Let f(x)=xex2t for x(1,+). We have f(x)f(2t)=2te by a simple calculation. Thus, we have that Ψ(x)(1x2)2t(2te)2t. We are left to bound maxx(1,1)|Pt(x)|.

Define P1 be the sum of absolute values of all coefficients of the polynomial P. Since x(1,1), we know maxx(1,1)|Pt(x)|Pt1. By the recursion, we have

Pt14Pt11+8tPt1120tPt11

where the last inequality is from Pt113tPt11 since the degree of Pt1 is at most 3t3. So we know maxx(1,1)20tt!. Therefore,

|Ψ(t)(x)|(2te)2t20tt!=t(3+o(1))t.

  • Fact 8

    For all t, |ρ(t)(x)|t(3+o(1))t.

Proof.

Directly from Fact 7 by observing that ρ(x)=eΨ(1x) for 0<x<1 and is constant elsewhere.

  • Fact 9

    Let r(u,v)ρ(logulogv+c) for some constant c. Then we have that for all n,m, |nmr(u,v)unvm|(n+m)6(n+m)|u|n|v|m.

Proof.

Let g(u,v)=logulogv+c. Then by the generalized chain rule for the derivative of the composition of two functions (also known as Faà di Bruno’s formula), we have

nmr(u,v)unvm= (a1,,an)na1+2a2++nan=n(b1,,bm)mb1+2b2++mbm=mn!m!i=1n(i!)aiai!i=1m(i!)bibi!
ρ(a1++an+b1++bm)(g(u,v))i=1n((1)ii!ui)aii=1m((1)i+1i!vi)bi.

Therefore

|nmr(u,v)unvm|nnmmn!m!(m+n)4(m+n)1|u|n|v|m(n+m)6(n+m)|u|n|v|m.