Abstract 1 Introduction 2 Organization 3 Notations and preliminaries 4 Proof Overview 5 Tools 6 Main results 7 Conclusions and further directions References

Random Restrictions of Bounded Low Degree Polynomials Are Juntas

Sreejata Kishor Bhattacharya ORCID School of Technology and Computer Science, Tata Institute of Fundamental Research, Mumbai, India
Abstract

We study the effects of random restrictions on low degree functions that are bounded on every point of the Boolean cube. Our main result shows that, with high probability, the restricted function can be approximated by a junta of arity that is just polynomial in the original degree. More precisely, let f:{±1}n[0,1] be a degree d polynomial (d2) and let ρ denote a random restriction with survival probability O(log(d)/d). Then, with probability at least 1dΩ(1), there exists a function g:{±1}n[0,1] depending on at most dO(1) coordinates such that fρg22d1Ω(1).

Our result has the following consequence for the well known, outstanding conjecture of Aaronson and Ambainis. The Aaronson-Ambainis conjecture was formulated to show that the acceptance probability of a quantum query algorithm can be well approximated almost everywhere by a classical query algorithm with a polynomial blow-up: it speculates that a polynomal f:{±1}n[0,1] with degree d has a coordinate with influence poly(1/d,𝖵𝖺𝗋[f]). Our result shows that this is true for a non-negligible fraction of random restrictions of f assuming 𝖵𝖺𝗋[f] is not too low.

Our work combines the ideas of Dinur, Friedgut, Kindler and O’Donnell [7] with an approximation theoretic result, first reported in the recent work of Filmus, Hatami, Keller and Lifshitz [8].

Keywords and phrases:
Analysis of Boolean Functions, Quantum Query Algorithms
Funding:
Sreejata Kishor Bhattacharya: Google Fellowship 2024.
Copyright and License:
[Uncaptioned image] © Sreejata Kishor Bhattacharya; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Quantum query complexity
Related Version:
Full Version: https://arxiv.org/abs/2402.13952
Acknowledgements:
I want to thank Francisco Escudero Gutierrez for pointing out typos and a small mistake in a previous version of this paper.
Editors:
Raghu Meka

1 Introduction

The analysis of Boolean functions is a vibrant field with connections and applications to diverse areas. While several outstanding results have been established for Boolean functions, the closely related class of bounded functions, i.e. those which take value in the real interval [0,1] at every point of the Boolean cube, remains far less understood. Indeed, such bounded functions are of great significance. As mentioned in Dinur, Friedgut, Kindler and O’Donnell [7], “such bounded functions often arise naturally, particularly as weighted averages of boolean functions; e.g., as Fourier transforms of boolean functions, as noise-convolutions of boolean functions, in the context of random walks on the discrete cube, and in hardness-of approximation theory in computational complexity”, and it is often important to understand when such functions depend essentially only on a few number of coordinates.

In this work, we study the effects of random restrictions on such bounded functions of low degree. The study of random restrictions of Boolean functions has been invaluable for establishing circuit lower bounds ( see for example, [9, 10, 21]), proof complexity (see for example, [5, 11]), de-randomization (see for example, [12]), and several other areas. Our main result shows a remarkable similarity in the behavior of bounded functions with their Boolean cousins, when hit with random restrictions.

By now, it is worth recalling that a classical result of Dinur, Friedgut, Kindler and O’Donnell [7] showed that bounded functions whose Fourier tail above degree k is exp(O(k2)) are approximable by juntas with arity exp(O(k)). They also showed that this result is tight in general. We are interested in the case when the function has small degree d. What additional structure follows from this stronger condition? To take cue on what to expect, let’s recall the case of Boolean functions. If f:{±1}n{0,1} is a degree d Boolean function, a seminal result of Beals, Buhrman, Cleve, Mosca, and de Wolf [4] shows that f can be represented by a DNF of width dO(1). By the celebrated Hastad’s switching lemma, this implies that if f is hit with a random restriction with survival probability 1/dO(1), with high probability f becomes constant. We show something similar is true for bounded functions as well: if f:{±1}n[0,1] is a degree d, bounded function and f is hit with a random restriction with survival probability O(log(d)/d), with high probability it can be well-approximated by a junta depending only on poly(d) coordinates. This is despite the fact that the expected number of variables that remain alive is O(nlogdd).

Theorem 1 (Main Theorem).

For any constants C1,C2>0, there exist constants C3,C4,C5>0 such that the following holds:
Let f:{±1}n[0,1] be a degree d polynomial and let ρ be a random restriction with survival probability log(d)C3d. With probability at least 1dC2, fρ is a (dC1𝖵𝖺𝗋[f],𝖵𝖺𝗋[f]2dC4) junta. Moreover, if J denotes the set of coordinates on which the junta depends, for each jJ we have 𝖨𝗇𝖿j[f]𝖵𝖺𝗋[f]2dC5.

This result has interesting consequences for the Aaronson-Ambainis conjecture, which we explain next.

One of the central open problems in the field of quantum query complexity is determining if there exists a partial function which is defined on a large fraction of the Boolean hypercube (say, constant) but whose quantum query complexity and classical query complexity are super-polynomially separated. The result of Beals et. al. [4] shows that no such separation is possible when the function is defined on the entire hypercube. On the other hand, functions for which we know such a separation (e.g. - Forrelation [1] , Bernstein-Vazirani [6]) are defined on an exponentially small fraction of the hypercube. A possible explanation as to why all known functions exhibiting large gaps between quantum and classical query complexity have very small support size would be the following folklore conjecture:

Conjecture 2.

Let Q be a quantum query algorithm with Boolean output on n qubits making q queries. Let P:{±1}n[0,1] be given by P(x)=Pr[Q outputs 1 on x]. For any ϵ>0, there exists a classical query algorithm A such that 𝖤[(A(x)Q(x))2]ϵ and A makes at most 𝗉𝗈𝗅𝗒(q,1ϵ) queries.

It is known that if Q makes at most q queries, then P is given by a polynomial of degree at most 2q. Although P has more structure than any arbitrary low degree bounded polynomial, it is further conjectured that such structure is not necessary. In other words, we forget the fact that P arises from a quantum query algorithm and instead try to construct a classical query algorithm for any bounded low-degree polynomial. This led to the following conjecture (also folklore).

Conjecture 3.

Let P:{±1}n[0,1] be a degree d polynomial. For any ϵ>0, there exists a classical decision tree T of depth at most 𝗉𝗈𝗅𝗒(d,1/ϵ) such that 𝖤[(P(x)T(x))2]ϵ.

Aaronson and Ambainis [2] proposed the following query algorithm to estimate P: suppose the variance of the function is sufficiently small. Then we terminate the query algorithm and output the average over the unqueried coordinates. If not, we query the coordinate with the highest influence and restrict the function according to the response received. We keep doing this until we have made too many queries or the variance has become sufficiently low. In order to show that this algorithm gives an accurate estimate, [2] observed that it is sufficient to prove the following conjecture.

Conjecture 4 (Aaronson-Ambainis conjecture).

Let f:{±1}n[0,1] be a degree d polynomial. Then, there exists a coordinate j such that 𝖨𝗇𝖿j[f]𝗉𝗈𝗅𝗒(1/d,𝖵𝖺𝗋[f]).

As a side remark, we mention that O’Donnell et al. [18] had shown previously that functions which can be approximated by decision trees have a coordinate with high influence. So conjectures 3 and 4 are equivalent.

The Aaronson-Ambainis Conjecture has received significant attention in the past few years. In 2006, a result of Dinur, Friedgut, Kindler and O’Donnell [7] showed that the conjecture is true if 𝗉𝗈𝗅𝗒(d) is replaced by exp(d). In 2012, Montanaro [16] proved the conjecture in the special case of block-multilinear forms where all coefficients have the same magnitude. In 2016, O’Donnell and Zhao [19] showed that it suffices to prove the conjecture for a special class of polynomials known as one-block decoupled polynomials. In 2020, Keller and Klein [13] claimed to have found a proof for the conjecture but their paper had a subtle flaw and turned out to be wrong. More recently, Lovett and Zhang [15] initiated a new line of attack using the notions of fractional block sensitivity and fractional certificate complexity. In 2022, Bansal, Sinha and de Wolf [3] proved that this conjecture is true for completely bounded block multilinear forms – a class of polynomials that captures a special kind of quantum query algorithms.

In this work, we show that Aaronson-Ambainis conjecture is true for a large fraction of random restrictions of f assuming 𝖵𝖺𝗋[f] is not too low. More precisely, we have:

Theorem 5.

There exist constants C1,C2>0 such that the following holds: let f:{±1}n[0,1] be a degree d polynomial (d2) with 𝖵𝖺𝗋[f]1/d. Let ρ denote a random restriction with survival probability log(d)C1d. Then,

𝖯𝗋[fρ has a coordinate with influence 𝖵𝖺𝗋[f]2dC2]𝖵𝖺𝗋[f]log(d)50C1d.

We view this result as evidence that the Aaronson-Ambainis Conjecture is true, and hope it gives new insights into the conjecture.

We also observe that our result implies one of the two main unconditional results proven in a recent paper of Lovett and Zhang [15] about the existence of small sensitive blocks albeit with a slightly different set of parameters.

Let f:{±1}n[0,1]. An input x{±1}n is said to be (r,ϵ) sensitive if there exists a y such that d(x,y)r and |f(x)f(y)|ϵ. [15] proves the following:

Theorem 6 (Lovett and Zhang [15]).

If f:{±1}n[0,1] has degree d, then at least Ω(𝖵𝖺𝗋[f]) fraction of the inputs are (r,ϵ) sensitive where ϵ=poly(𝖵𝖺𝗋[f]/d),r=poly(d,1/ϵ,log(n))

An immediate consequence of our result is that at least Ω((𝖵𝖺𝗋[f]/d)O(1)) fraction of inputs are (1,ϵ) sensitive where ϵ=poly(𝖵𝖺𝗋[f]/d). Thus, while we lose a bit in the fraction of sensitive inputs, we gain by letting our block size be exactly 1 instead of poly(d,1/ϵ,log(n)). It is interesting to remark that our techniques and those of Lovett and Zhang seem quite different. We further remark that [15] shows that if one can show every input has a small sensitive block, Aaronson-Ambainis conjecture will follow.

We believe that Theorem 5 opens up some possible lines of attack to the Aaronson-Ambainis conjecture:

  • Assuming a supposed counterexample f:{±1}n[0,1], modify it appropriately (e.g., by composing it with some appropriate gadget or applying a low noise operator) to get a function f~:{±1}n~[0,1] such that most of its random restrictions remain a counterexample. Combined with our result, this will prove Aaronson-Ambainis conjecture. This approach is discussed in a bit more detail in the conclusion.

  • If we can construct a decision tree that queries y (the input to the fixed part) in a few coordinates and outputs an influential coordinate of fy (the restricted function) with high probability, Aaronson-Ambainis conjecture will follow from the work of Keller and Klein. Again, this approach is discussed in a bit more detail in the conclusion.

2 Organization

We introduce notations and necessary preliminaries in section 3. We give a high level overview of our proof in section 4. In section 5 we compile some lemmas that will be needed in our main proof. Our main results are proven in section 6. Our main result is proven in section 6.2. It says that most random restrictions of a bounded low-degree function can be approximated by a small junta. In section 6.3 we prove that Aaronson-Ambainis conjecture is true for a non-negligible fraction of random restrictions.

3 Notations and preliminaries

Query algorithms

  1. 1.

    A classical query algorithm A (or equivalently, a decision tree) for computing a function f:{±1}n can access the input x{±1}n by adaptively issuing queries to its bits. We assume internal computations have no cost. The depth of the query algorithm/decision tree is the maximum number of bit queries issued on an input. We say A ϵ-approximates f if fA22=𝖤x{±1}n[(f(x)A(x))2]ϵ.

    For a partial function f:S({±1}n){0,1}, its classical query complexity D(f) is the smallest d for which there exists a decision tree T of depth d such that T(x)=f(x) for all xS.

  2. 2.

    A quantum query algorithm can access the input x{±1}n via an oracle Ox. The register has n qubits and some ancilla qubits. The oracle Ox applies the following unitary operation on the first n qubits:

    Ox|s1s2sn=(1)x,s|s1s2sn.

    The quantum query algorithm applies a sequence of unitary operators, where each operator is either Ox or an input-independent unitary operator U. In the end, it measures the first qubit and outputs the measurement result. The number of queries issued is the number of times Ox is applied.

    Notice that a quantum query algorithm Q naturally defines a function P:{±1}n[0,1]:

    P(x)=Pr[Q outputs 1 on input x].

    It is well-known that if Q makes q queries, then P is a degree 2q polynomial.

    For a function f:S({±1}n){0,1}, we define its quantum query complexity Q(f) to be the smallest q for which there exists a quantum query algorithm Q making q queries such that for all xS,

    Pr[Q outputs 1 on input x]{2/3 if f(x)=11/3 if f(x)=0.

Analysis of boolean functions

In this section we recall some results from analysis of boolean functions. A good reference is O’ Donnell’s textbook [17].

  1. 1.

    Any function f:{±1}n has a unique representation as f(x)=S[n]f^(S)χS(x) where χS(x)=iSxi. The coefficients f^(S) are the Fourier coefficients of f. The degree of f is max{|S||f^(S)0}.

  2. 2.

    The variance of f is

    𝖵𝖺𝗋x{±1}n[f(x)]=Sϕf^(S)2.
  3. 3.

    For a cooordinate i, the influence of the i’th coordinate is defined as

    𝖨𝗇𝖿i[f] =𝖤x{±1}n[(f(x)f(x(i))2)2]
    =iSf^(S)2.

    The total influence of f is

    𝖨𝗇𝖿[f]=i[n]𝖨𝗇𝖿i[f]=S|S|f^(S)2.

    From the Fourier expansion it is clear that if deg(f)d,𝖨𝗇𝖿[f]d𝖵𝖺𝗋[f].

  4. 4.

    Given two functions f,g{±1}n, we say g ϵapproximates f if fg22=𝖤x[(f(x)g(x))2]ϵ.

  5. 5.

    For a point x{±1}n and a subset S[n] and 1ρ1, we define a distribution Nρ,S(x) on {±1}n as follows:

    • The bits y1,y2,,yn are independent, and

      𝖯𝗋[yi=xi]={1 if iS(1+ρ)/2 if iS.

    When S=[n], we abbreviate Nρ,S(x) by Nρ(x).

  6. 6.

    For f:{±1}n and 1ρ1 define Tρf:{±1}n by

    Tρf(x)=𝖤zNρ(x)[f(z)].

    It is easy to see that the Fourier expansion of Tρf is given by

    Tρf(x)=S[n]ρ|S|f^(S)χS(x).
  7. 7.

    A function f:{±1}n is a junta of arity l or l-junta if there exists a subset S[n],|S|l such that f only depends on the coordinates in S. We say f is a (ϵ,l) junta if it can be ϵ-approximated by a l-junta, i.e., there exists a l-junta g such that fg22ϵ.

  8. 8.

    A restriction ρ=(S,y) of f:{±1}n is specified by a subset S[n] and an assignment y{±1}[n]S. Such a restriction naturally induces a function fρ:{±1}S. Sometimes we shall write fy instead of fρ (note that S is determined by y since y{±1}[n]S).

    By a random restriction with survival probability p, we mean sampling ρ=(S,y{±1}[n]S) where each coordinate i[n] is included in S with probability p independently, and each bit of y is independently set to a uniformly random bit.

 Remark 7.

Throughout the paper, all growing parameters (e.g., the degree d) will be assumed to be larger than some sufficiently big constant. This is to make the expressions look neat, as we will be replacing terms like (C1)kpoly(k) by (C2)k where C2>C1.

4 Proof Overview

The main result in this paper is a structural result for bounded low-degree functions similar in spirit to Hastad’s switching lemma [10]. Roughly speaking, we show that for come constant C>0, the following holds: let f:{±1}n[0,1] be a polynomial of degree d, and let ρ denote a random restriction with survival probability log(d)Cd. Then,

Pr[fρ is a (O(dC),O(dC)) junta]11dΩ(1).

Once this is established, we can prove Aaronson-Ambainis conjecture for random restrictions as follows: it is easy to see that

𝖯𝗋[𝖵𝖺𝗋[fρ]𝖵𝖺𝗋[f]log(d)2Cd]𝖵𝖺𝗋[f]log(d)2Cd.

This probability will be significantly more than the failure probability of the main result (1dΩ(1)) (this is the only place where we need the lower bound on 𝖵𝖺𝗋[f]). So for a Ω(𝖵𝖺𝗋[f]log(d)/d) fraction of random restrictions, the variance of fρ is high and fρ can be well approximated by a junta with arity 𝗉𝗈𝗅𝗒(d). This means one of the coordinates of the junta must have high influence. This concludes the proof.

Now we give a brief overview of how we prove the main result (Theorem 1). The starting point is the work by Dinur, Friedgut et al. [7] which states the following:

Theorem 8 (Dinur et al. [7]).

For any f:{±1}n[0,1], if

|S|>kf^(S)2exp(O(k2log(k)/ϵ)),

then f is a (ϵ,2O(k)/ϵ2) junta.

In other words, if the Fourier tail above a certain level k is bounded, then f can be well approximated by juntas of arity roughly 2O(k).

We start with the observation that random restrictions have bounded Fourier tails (Lemma 15): if the function has degree d and we make a random restriction with survival probability log(d)Cd where C>1 is a large constant, using Chernoff bound we can show that with very high probability the Fourier weight above level log(d) will be low; around the order of exp(Ω(Clog(d))). If we can manage to bring the Fourier weight above log(d) small enough so that Theorem 8 applies, then we will get that fρ can be well approximated by a 𝗉𝗈𝗅𝗒(d) junta. Unfortunately, if we try this, it turns out that we have to set the survival probability so low that on expectation the variance of fρ goes down significantly as well. In other words, while it is true that fρ can be well-approximated by juntas, it is for a trivial reason that its variance itself is very low. (And moreover, this is also true for functions f:{±1}n that are merely L2 bounded i.e, 𝖤[f2]1, so we would not be using any additional structure that arises from the fact that f is pointwise bounded.)

In order to make this approach work, we need to improve the tail bound exp(O(k2log(k))/ϵ) in Theorem 8. The problematic term is the quadratic exp(O(k2)) in the exponential. If the dominant term were to the order of exp(O(k)) instead, the calculations would go through. Can we hope to increase the tail bound to exp(O(k)) while paying a cost by increasing the junta arity? Unfortunately, again, this is not possible: [7] constructs a function which shows that the tail bound is essentially tight upto the log(k) factor - their function has f>k22exp(Θ(k2)) but approximating it to even 1/3 accuracy requires reading Ω(n) coordinates. Our key observation is that the function constructed by [7] has full degree whereas we are working with random restrictions of a low degree function, so in addition to the fact that the Fourier tail of fρ above level log(d) is very small, we also know that fρ has degree d. Can we hope to improve the tail bound in Theorem 8 if we have the additional restriction that the function is of degree d? Indeed, this turns out to be true. We prove the following result in section 6 (Theorem 28):

Theorem 9.

There exists a constant C such that the following is true:

If f:{±1}n[0,1] has degree d and |S|>kf^(S)2ϵCkdC, f is a (ϵ,ϵ2dCCk) junta.

Below we briefly discuss how we are able to improve the tail bound under the additional degree assumption. Dinur et al [7] prove their tail bound (Theorem 8) by showing the following result (we are omitting the exact quantitative parameters here for reading convenience).

Theorem 10.

Let h:{±1}n be a degree k function with 𝖤[h2]1 (but not necessarily pointwise bounded) which cannot be approximated by 2O(k) juntas to accuracy μ. Then, for any function g:{±1}n[0,1], E[(hg)2]ε.

Let’s sketch how [7] proves Theorem 8 using Theorem 10. They assume f is not approximable by 2O(k) juntas - their goal is then to show that the Fourier tail above level k, fk22, is large. To do so, they use Theorem 10: they take h to be the truncated function h=fk and g to be the original function g=f. This then lower bounds 𝖤[(ffk)2] which is precisely the Fourier tail above weight k. Thus, the distance lower bound ε in Theorem 10 governs the Fourier tail lower bound in Theorem 8. Since in Theorem 28 we have the additional information that f is of degree d, for our purposes it will suffice to lower bound the distance of h from bounded degree d functions, not necessarily all bounded functions. In section 6.2 we prove a result of the following form (again, we are omitting the exact parameters for reading convenience; for the exact statement see Theorem 26).

Theorem 11 (Informal version of Theorem 26).

Let h:{±1}n be a degree k function with 𝖤[h2]1 (but not necessarily pointwise bounded) which cannot be approximated by 2O(k) juntas to accuracy μ. Then, for any degree d function g:{±1}n[0,1], E[(hg)2]ε~.

The parameter ε~ in Theorem 11 is bigger than the corresponding ε parameter in Theorem 10 because we are only lower bounding the distance of h from bounded low-degree functions whereas Theorem 10 lower bounds the distance of h from arbitrary bounded functions. It turns out that the improvement in this parameter is sufficiently good for the calculations to go through.

In order to prove Theorem 11 we shall use the main idea of the proof of [7] along with a structural restriction for bounded low-degree functions proved first in Filmus et al. [8], and later used by Keller-Klein [13] and Lovett-Zhang [15]. Given any function f:{±1}n and x{±1}n define the block sensitivity of f at x to be

bs(f,x)=sup[j[k]|f(x)f(x(Bj))|]

where the supremum ranges over all partitions (B1,B2,,Bk) of the variables (x(Bj) denotes x with the coordinates in Bj flipped). Define the block sensitivity of f to be bs(f)=supxbs(f,x). We shall use the following fact about bounded low-degree functions:

Theorem 12.

If f:{±1}n[0,1] has degree d, bs(f)6d2.

We give a high-level overview of how we are able to improve upon the bound of ε using the fact that the block sensitivity of a bounded low degree function is small. At one point in their proof, [7] lower bounds the probability of a linear form of Rademacher random variables l(x1,x2,,xt)=a1x1++atxt exceeding a certain threshold times its standard deviation, i.e.,

𝖯𝗋[a1x1++atxtαa12++at2].

(In the proof, l(x1,,xt) is the linear part of a restriction of f: jUf^j(U)xj for some U[n].)

For each such point x where this linear form is high, [7] shows that many related points x must have f(x)>2. (To be precise, these related points are obtained by applying an appropriate noise operator to x.) Using this they conclude that f must deviate from the interval [0,1] too often and therefore cannot be approximated by any bounded function.

We follow the proof of [7] up until this point. Instead of directly lower bounding the probability that a1x1++atxt is high, we partition the set of variables [t] into L blocks B1,,BL (L is an appropriately chosen parameter) such that each block gets roughly same total weight: for all j[L],

iBjai2a12++at22L.

It will turn out that the ai’s are sufficiently small for such a partition to exist. For each block we lower bound the probability that the linear form restricted to this block is high:

𝖯𝗋[jBiajxjα~jBiaj2].

Now, on a random assignment z, the linear form restricted to many of these blocks will be high. Take such a block Bi: jBiajzjα~jBiaj2. For each such block we will be able to find a large number of related points zi such that |f(z)f(zi)| is large. (In our case, these related points are obtained by applying an appropriate noise operator restricted to Bi.) Crucially, these related points will differ from z only at Bi. Thus, we will find many points which differ from z at disjoint sets and whose f differ from z significantly. This will show that f cannot be too close to a bounded low degree function, because those functions have low block sensitivity.

Our advantage is that we need to set α so that we can conclude |f(z)f(z)| is only somewhat larger than Ω(d2/L) (as opposed to Ω(1) in [7]) - by setting L large enough this allows us to set a much smaller α and get rid of the quadratic exponential dependence.

5 Tools

In this section we compile some lemmas that we shall use in our proof.

A reverse Markov inequality

We will use the following simple inequality throughout the proof.

Lemma 13.

Let X be a random variable such that XM with probability 1. Let 𝖤[X]=μ>0. Then, 𝖯𝗋[Xμ/2]μ2M.

Proof.

Assume 𝖯𝗋[Xμ/2]<μ2M. Then,

𝖤[X]𝖯𝗋[Xμ/2]M+𝖯𝗋[Xμ/2]μ2<μ,

contradiction.

An anticoncentration inequality for linear forms of Rademacher random variables

Lemma 14.

There exists a universal constant K such that the following holds: let x1,,xn be independent Rademacher random variables and let l(x1,,xn)=a1x1++anxn. Let σ=a12++an2. Suppose |ai|σKt. Then,

Pr[l(x1,,xn)tσ]exp(Kt2).

Proof.

Equation 4.2 in [14].

Random restrictions have small tail

Lemma 15.

Let f:{±1}n have degree d, C>1 be a sufficiently large constant, and let ρ=(S,y{±1}[n]S) be a random restriction with survival probability log(d)Cd. Let k=log(d). Then,

𝖤[|T|>kfy^(T)2]exp(Clog(d)/8)𝖵𝖺𝗋[f].

Proof.

First suppose (S,y{±1}[n]S) is a fixed restriction. Note that for z{±1}S,

fy(z)=U[n]f^(U)χU(y,z)

so for T[S], fy^(T)=USf^(UT)χU(y). By Parseval’s theorem, for a fixed S,

𝖤y[fy^(T)2]=U[n]Sf^(UT)2.

Therefore, for a fixed S,

𝖤y[|T|>kfy^(T)2]=V[n],|VS|>kf^(V)2.

Randomizing over S again,

𝖤S,y[|T|>kfy^(T)2]=V[n]Pr[|VS|>k]f^(V)2.

Since f has degree d, we only need to worry about the terms where |V|d. Also, for |V|k the relevant probability is 0. Since each element is included in S with probability log(d)Cd, by Chernoff bound, for each V with |V|d,

Pr[|VS|>k]exp((C1)2k/4C)exp(Clog(d)/8).

Thus we get that

𝖤S,y[|T|>kfy^(T)2]exp(Clog(d)/8)Tϕf^(T)2=exp(Clog(d)/8)𝖵𝖺𝗋[f].

Random restrictions don’t have low variance

Lemma 16.

Let f:{±1}n be any function and let ρ=(S,y{±1}[n]S) be a random restriction with survival probability p. Then, 𝖤[𝖵𝖺𝗋[fρ]]p𝖵𝖺𝗋[f].

Proof.

Fix a restriction (S,y{±1}[n]S). For each TS, fy^(T)=U[n]Sf^(TU)χU(y). Thus, by Parseval’s theorem, for a fixed S,

𝖤y[𝖵𝖺𝗋[fy]]=T:TSϕf^(T)2.

Randomizing over S again,

𝖤S,y[𝖵𝖺𝗋[fy]] =TϕPr[TSϕ]f^(T)2
=Tϕ(1(1p)|T|)f^(T)2
Tϕpf^(T)2
=p𝖵𝖺𝗋[f].

Random restrictions with appropriate survival probability put large Fourier mass on the linear level

Lemma 17 (implicit in [7]).

Let f:{±1}n, J[n] and k be such that

2k|TJc|<2k+1f^(T)2μ.

Consider a random restriction ρ=(S,y{±1}[n]S) where each jJ is fixed and given a uniformly random assignment, and each iJc is kept alive with probability p=2k. Then,

𝖤[iSf^y({i})2]μ20.

Proof.

For a fixed (S,y{±1}[n]S) (note that JS=ϕ) and jS we have

fy^({j})=T[n],TS={j}f^(T)χT{j}(y).

By Parseval’s theorem, for a fixed S,

𝖤y[fy^({j})2]=T[n],TS={j}f^(T)2.

Randomizing over S,

𝖤[jSfy^({j})2] =T[n][j[n]Pr[TS={j}]]f^(T)2
2k|TJc|<2k+1|T|p(1p)|T|1f^(T)2.

By standard inequalities, for n[1/p,2/p), np(1p)n11/20. It follows that

𝖤y[fy^({j})2] μ20.

Some hypercontractive inequalities for low degree functions

These lemmas and their proofs can be found in [7].

Lemma 18 (Corollary 2.4 in [7]).

There exists a universal constant W>0 such that the following holds:

Let f:{±1}n be a degree k function. Let 𝖤[f2]=σ2. Then,

𝖤[f(z)2𝟣f(z)2Wkσ2]12𝖤[f(z)2].
Lemma 19 (Lemma 2.5 in [7]).

There exists a universal constant B>0 such that the following holds:

Let f:{±1}n be a degree k function. Let ρ[1/2,1/2] be a noise parameter, and x0{±1}n. Suppose 𝖤zNρ(x0)[f(z)f(x0)]=μ0. Then,

PrzNρ(x0)[f(z)f(x0)μ]1Bk.

The noise lemma

This is the main result of [7]. We use a slight variant. First we recall some known results from approximation theory.

Lemma 20.

For any k, there exist constants ρ1,ρ2,,ρk+1[1/2,1/2] with the following property: for any polynomial of degree k, p(x)=a0+a1x++akxk, there exists a j[k+1] such that p(ρj)a12(k+1).

Proof.

Page 112 in [20].

Now we state the lemma.

Lemma 21.

There exists a universal constant B>0 such that the following holds:

Consider a degree k polynomial f:{±1}n. Let S[n] and (x)=iSf^({i})xi. Consider an input x0{±1}n such that (x0)γ. Sample a z{±1}n by the following procedure:

  1. 1.

    Sample ρ{ρ1,,ρk+1} uniformly at random.

  2. 2.

    Sample zNρ,S(x0).

Then,

Pr[f(z)f(x0)γ2(k+1)]1(k+1)Bk.
 Remark 22.

Observe that z differs from x only in the coordinates of S. This will be crucial later on.

Proof.

Take B to be the same universal constant as in Lemma 19. By replacing f with an appropriate restriction if necessary, we can assume S=[n]. Consider the polynomial p(ρ)=Tρf(x0)f(x0). From the Fourier expansion of noise operator, we see that

p(ρ)=Sϕρ|S|f^(S).

This is a degree k polynomial in ρ with linear coefficient l(x0). By Lemma 20, there exists a h[k+1] such that p(ρh)γ/(2k+2). By Lemma 19,

PrzNρ(x0)[f(z)f(x0)γ2(k+1)|ρ=ρh]1Bk.

We choose ρ=ρh in step (1) with probability 1/(k+1), so

PrzNρ(x0)[f(z)f(x0)γ2(k+1)]1(k+1)Bk.

Partitioning a set of numbers in a balanced manner

We need an easy lemma about partitioning a set of weights none of which is too large into disjoint buckets where each bucket gets roughly the same total weight. We will later use this lemma on the set of small linear Fourier coefficients of a function.

Lemma 23.

Let a1,a2,,an be a set of non-negative real numbers and 1Ln. Suppose aia1+a2++an2L for all 1in. Then, there exists a partition (B1,B2,,BL) of [n] such that for all 1jL,

iBjaia1++an2L.

Proof.

Start with an arbitrary partition (B1,B2,,BL). Then, refine it iteratively according to the following algorithm.

Refinement algorithm: 1. Locate a j such that the condition is violated for j, i.e., iBjai<a1++an2L. If no such j exists, terminate. 2. Locate a k such that iBkaia1++anL. 3. Take an arbitrary lBk such that al0 and place it in Bj; BkBk{l} BjBj{l}
An appropriate k always exists in step (2) by an averaging argument. Since ala1++an2L, the size of Bk does not go below a1++an2L after step (3). It is easy to see this procedure must terminate. Formally, notice that the quantity

j[L]min(a1++an2LiBjai,0)

reduces by 12Lmin{ai|ai0} at each step, so at some point of time it must be 0 at which point the algorithm terminates and returns a valid partition.

Bounded low-degree functions have low block sensitivity

Lemma 24.

Let f:{±1}n[0,1] be a polynomial of degree d. Then, bs(f)6d2.

Proof.

Let B1,B2,,Bk be any partition of [n]. Our goal is to show that for all x{±1}n,

i=1k|f(x)f(x(Bi))|6d2.

Let g:{±1}k[0,1] be the function obtained from f obtained by identifying the variables of Bi with a single variable yi for 1ik. By Proposition 3.7 in [8], we have, for every y{±1}k,

i=1k|g(y)g(y(i))|6d2.

This implies

i=1k|f(x)f(x(Bi))|6d2,

as desired.

6 Main results

6.1 Improved tail bound for low degree functions

This section is the core technical part of our work. In Theorem 26 we show that if we have a function f:{±1}n (not necessarily bounded) with 𝖤[f2]1 which cannot be approximated by juntas, then f cannot be well-approximated by bounded low-degree functions. In Theorem 28 we use Theorem 26 on the truncated function fk to conclude that the tail bound appearing in [7] (Theorem 8) can be improved under the additional assumption that f has degree d.

 Remark 25.

For a subset J[n], consider the junta u:{±1}n which reads the coordinates of J and outputs the average over the unqueried coordinates. It is easy to see that u(x)=SJf^(S)χS(x), so uf22=SJf^(S)2. Thus, u approximates f if and only if SJf^(S)2 is small.

In fact, it is easy to see that there exists a function u depending only on coordinates of J such that fu22ϵ if and only if SJf^(S)2ϵ. This immediately follows from the Fourier expansion of fu.

Theorem 26.

There exists a constant C such that the following holds:
Let f:{±1}n be a degree k function (not necessarily bounded) with 𝖤[f2]1. Let J={j|𝖨𝗇𝖿j[f]θ} where θ=μ2CkdC. If SJf^(S)2μ, then for any degree d function g:{±1}n[0,1], 𝖤[(f(x)g(x))2]δ=μCkdC.

 Remark 27.

Notice here that although f is not pointwise bounded, g is.

Proof.

Let W,B,K be the universal constants from Lemma 18, Lemma 21 and Lemma 14 respectively. We take C to be a constant sufficiently larger than B,K,W.

There exists a t such that

2t|SJc|<2t+1f^(S)2μlog(k).

Let ρ=(U,y{±1}[n]U) (UJc) be a random restriction where each jJ is fixed to a uniformly random assignment, and survival probability for each jJ is 2t. By Lemma 17,

𝖤[jUfy^({j})2]μ20log(k).

Fix a U such that

𝖤y{±1}[n]U[jUfy^({j})2]μ20log(k).

By Parseval’s theorem we have for all jU,

𝖤[fy^({j})2]=SU={j}f^(S)2𝖨𝗇𝖿j[f].

For each y{±1}[n]U define 𝖲𝖬𝖠𝖫𝖫y={j|fy^({j})2Wk𝖨𝗇𝖿j[f]}. Observe that for all y,

j𝖲𝖬𝖠𝖫𝖫yfy^({j})2Wk𝖨𝗇𝖿[f]kWk(2W)k.

For each jU we have from Lemma 18

𝖤[f^y({j})2𝟣f^y({j})2Wk𝖨𝗇𝖿j[j]]12𝖤[f^y({j})2].

Thus,

𝖤y{±1}[n]U[j𝖲𝖬𝖠𝖫𝖫yf^y({j})2]μ40log(k),

so applying Lemma 13 2

𝖯𝗋y{±1}[n]U[j𝖲𝖬𝖠𝖫𝖫yf^y({j})2μ80log(k)]μ80log(k)(2W)kμ(3W)k.
22footnotetext: See remark 7.

Call y{±1}[n]U for which j𝖲𝖬𝖠𝖫𝖫yf^y({j})2μ80log(k) to be good. Let GOOD={y{±1}[n]U|y is good}. Let L=(2B)kd8𝖵𝖺𝗋[f]. For each good y, choose a partition 𝖣𝖨𝖵𝖨𝖣𝖤(y)=(B1,B2,,BL) of 𝖲𝖬𝖠𝖫𝖫y such that for all 1iL,

jBif^y({j})2μ160Llog(k).

(If there are multiple such partitions, choose any one of them and call it 𝖣𝖨𝖵𝖨𝖣𝖤(y).)

Our choice of parameters ensures that for all j𝖲𝖬𝖠𝖫𝖫y, f^y({j})2μ160Llog(k), so such a partition exists by Lemma 23. Let ρ1,ρ2,,ρk+1 be the constants from Lemma 20.

Suppose, for the sake of contradiction, there exists a degree d polynomial g:{±1}n[0,1] such that 𝖤[(f(x)g(x))2]δ. Throughout the rest of the proof, for a string s1{±1}[n]U and a string s2{±1}U, the pair (s1,s2) denotes the string s{±1}n which agrees with s1 on [n]U and with s2 on U. Consider the following randomized procedure which returns a real number.

Procedure 1: 1. Sample a y𝖦𝖮𝖮𝖣 uniformly at random. 2. Sample ρ{ρ1,ρ2,,ρk+1} uniformly at random. 3. Sample z{±1}U uniformly at random. 4. Let 𝖣𝖨𝖵𝖨𝖣𝖤(y)=(B1,B2,,BL). Sample z~(i)NBi,ρ(z) for 1iL. 5. Return i=1L|f(y,z)f(y,z~(i))|.
We estimate the probability that procedure 1 returns a number >15d2 in two different ways. First, we obtain a lower bound from the defintion of 𝖦𝖮𝖮𝖣. Then, we obtain an upper bound using the assumption that 𝖤[(f(x)g(x))2]δ and the fact that 𝖯𝗋[y is good] is somewhat large. These two bounds will contradict each other - and that will prove the theorem.

Lower bound.

Fix a y𝖦𝖮𝖮𝖣. Let 𝖣𝖨𝖵𝖨𝖣𝖤(y)=(B1,B2,,BL).

Let w=μ160Llog(k). For each i[L] we have

jBif^y({j})2w.

Choose α such that αw=100d4(2B)kL. Our choice of L ensures that α1. Moreover, our choice for influence threshold θ ensures that |f^y({j})|wKα for all jBi where K is the universal constant from the Lemma 14.

Therefore, we can apply Lemma 14 to obtain that

𝖯𝗋z{±1}Bi[jBizjf^y({j})100d4(2B)kL]exp(Kα2)1K1.

Here K1=exp(K) is an absolute constant.

By Lemma 21 applied on the restriction fy:{±1}U, as we sample z{±1}U u.a.r, ρ{ρ1,,ρk+1} u.a.r, z~(i)Nρ,Bi(z), we have that

𝖯𝗋[|f(y,z)f(y,z~(i))|30d3(2B)kL]1K1(k+1)Bk1(2B)k.

By linearity of expectation,

𝖤[|{i[L]||f(y,z)f(y,z~(i))|30d3(2B)kL}|]L(2B)k.

Using Lemma 13,

𝖯𝗋[|{i[L]||f(y,z)f(y,z~(i))|30d3(2B)kL}|L2×(2B)k]12×(2B)k.

Observe that

|{i[L]||f(y,z)f(y,z~(i))|30d3(2B)kL}|L2×(2B)ki[L]|f(y,z)f(y,z~(i))|15d3.

We conclude that for all y𝖦𝖮𝖮𝖣, as z,z~(1),,z~(L) are sampled as in Procedure 1,

𝖯𝗋[i[L]|f(y,z)f(y,z~(i))|15d3]12×(2B)k.

Thus, with probability at least 12×(2B)k, procedure 1 returns a number greater than 15d2( as 15d3>15d2).

Upper bound.

Since 𝖤[(f(x)g(x))2]δ and 𝖯𝗋y{±1}[n]U[y is good]μ/(3W)k, we have that

𝖤[(f(x)g(x))2|x[n]U is good]δμ(3W)k.

Now consider a uniformly sampled y𝖦𝖮𝖮𝖣. Observe that as we sample z{±1}U u.a.r, ρ{ρ1,,ρk+1} u.a.r and z~(i)Nρ,Bi(z), the marginal distribution of z~(i) is uniform on {±1}U. By Markov’s inequality, we have

𝖯𝗋[(f(y,z)g(y,z))21L2]L2δμ(3W)k

and for all i[L],

𝖯𝗋[(f(y,z~(i))g(y,z~(i)))21L2]L2δμ(3W)k.

By union bound, the probability that (f(y,z)g(y,z))21L2 or for some i, (f(y,z~(i))g(y,z~(i)))21L2 is at most (L+1)L2δμ(3W)k2L3δμ(3W)k. Our choice of δ ensures that this quantity is less than <12×(2B)k. Observe that if none of these bad events holds, since the block sensitivity of g is bounded above by 6d2 (Theorem 24), we have that

i[L]|g(y,z)g(y,z~(i))|6d2i[L]|f(y,z)f(y,z~(i))|6d2+1<15d2.

Thus, we conclude

𝖯𝗋[i[L]|f(y,z)f(y,z~(i))|>15d2]<2L3δμ(3W)k<12×(2B)k.

As promised, we get conflicting lower and upper bounds for the probability that procedure 1 returns a number >15d2. This is our desired contradiction.

Now we show that we can improve the tail bound of Theorem 8 under the additional assumption that f has low degree. This follows straightforwardly from Theorem 26.

Theorem 28.

There exists a universal constant C>0 such that the following is true:
Let f:{±1}n[0,1] be a degree d function. Let θ=𝖵𝖺𝗋[f]2dCCk and J={j|𝖨𝗇𝖿j[f]θ}. If SJf^(S)2μ, then |S|>kf^(S)2μdCCk.

Proof.

Assume |S|>kf^(S)2<μ/2 (otherwise we are done). Let C~ be the universal constant from Theorem 26.

The idea is to apply Theorem 26 to the truncated function

fk(x)=|S|kf^(S)χS(x).

Note that while fk is not pointwise bounded, it satisfies 𝖤[(fk)2]1 and 𝖨𝗇𝖿j[fk]𝖨𝗇𝖿j[f] for all j (this is clear from the Fourier expressions). Let H={j|𝖨𝗇𝖿j[fk]θ}. We have HJ, so

SHf^k(S)2SJf^(S)2μ2μ2.

Applying Theorem 26, we get that for any bounded degree d g:{±1}n[0,1], 𝖤[(f(x)g(x))2]μ2dC~C~k. Taking g to be our original function f, we get the desired tail lower bound:

𝖤[(ffk)2]μ2dC~C~k|S|>kf^(S)2>μ2dC~C~k.

Taking C to be a slightly larger constant than C~, we get that

|S|>kf^(S)2μdCCk.

6.2 Random restrictions can be approximated by juntas

In this section we use the fact that random restrictions have bounded tails to show that they can be approximated by juntas. This is our main result.

Theorem 29 (Restatement of Theorem 1).

For any constants C~1,C~2>0, there exist constants C~3,C~4,C~5>0 such that the following holds:
Let f:{±1}n[0,1] be a degree d polynomial and let ρ be a random restriction with survival probability log(d)C~3d. With probability at least 1dC~2, fρ is a (dC~1𝖵𝖺𝗋[f],𝖵𝖺𝗋[f]2dC~4) junta. Moreover, if J denotes the set of coordinates on which the junta depends, for each jJ we have 𝖨𝗇𝖿j[f]𝖵𝖺𝗋[f]2dC~5.

Proof.

We consider a random restriction with survival probability log(d)C~3d.

By Lemma 15, the expected Fourier tail of fρ above level log(d) is at most exp(C~3log(d)/8)𝖵𝖺𝗋[f]=𝖵𝖺𝗋[f]dC~3/8. By Markov’s inequality, with probability at least 1dC~3/16, the Fourier tail above log(d) is 𝖵𝖺𝗋[f]dC~3/16. Let C be the constant from Theorem 28. Let μ=𝖵𝖺𝗋[f]dC~3/16dCClog(d)=𝖵𝖺𝗋[f]dC~3/16d2C, θ=μ2dCClog(d)=μ2d2C and J={j|𝖨𝗇𝖿j[fρ]θ}. Let uρ:{±1}n[0,1] be the function which reads the coordinates in J and outputs the average of fρ over the coordinates in Jc. Choose C~3 large enough so that μdC~1𝖵𝖺𝗋[f]. Applying Theorem 28, we see that uρ approximates fρ to accuracy dC~1𝖵𝖺𝗋[f]. Using the fact that total influence is bounded by d, we see that uρ has arity 𝖵𝖺𝗋[f]2dCC~3 for a universal constant C. Taking (C~4,C~5)=(CC~3,C~3/322C), we are done.

6.3 Aaronson-Ambainis Conjecture is true for random restrictions

Theorem 30 (Restatement of Theorem 5).

There exist constants C1,C2>0 such that the following holds: let f:{±1}n[0,1] be a degree d polynomial (d2) with 𝖵𝖺𝗋[f]1/d. Let ρ denote a random restriction with alive probability log(d)C1d. Then,

𝖯𝗋[fρ has a coordinate with influence 𝖵𝖺𝗋[f]2dC2]𝖵𝖺𝗋[f]log(d)50C1d.

Proof.

Let M be a large constant. Apply Theorem 29 with (C~1,C~2)=(M,M) to get constants C~3,C~4,C~5. Let ρ be a random restriction with survival probability log(d)C~3d. By Lemma 16,

𝖤[𝖵𝖺𝗋[fρ]]𝖵𝖺𝗋[f]log(d)C~3d

so by Lemma 13,

𝖯𝗋[𝖵𝖺𝗋[fρ]𝖵𝖺𝗋[f]log(d)2C~3d]𝖵𝖺𝗋[f]log(d)2C~3d.

Since 𝖵𝖺𝗋[f]1/d, dM𝖵𝖺𝗋[f]log(d)10C~3d. By Theorem 29 and Remark 25, with probability at least 1dM, there exists a Jρ[n] such that every coordinate in Jρ has influence 𝖵𝖺𝗋[fρ]2dC5~ and

SJρfρ^(S)2 dM𝖵𝖺𝗋[f].

So with probability at least 𝖵𝖺𝗋[f]log(d)2C~3ddM𝖵𝖺𝗋[f]log(d)4C3~d, both these events (high variance of fρ and existence of Jρ) hold and we have that

SJρfρ^(S)2 𝖵𝖺𝗋[fρ]dM𝖵𝖺𝗋[f]𝖵𝖺𝗋[f]log(d)4C~3d.

In particular, we have that Jρϕ. Since for each jJρ we have 𝖨𝗇𝖿j[fρ]𝖵𝖺𝗋[fρ]2dC~5, we are done by taking (C1,C2)=(C~3,2+C~5).

7 Conclusions and further directions

In this paper, we showed that if f:{±1}n[0,1] is a degree d polynomial, a large fraction of its random restrictions have an influential coordinate.

It would be interesting to see if we can extend this to the full Aaronson-Ambainis conjecture. We describe a few potential approaches here.

  • Given a degree d polynomial f:{±1}n[0,1], we can lift it with a Boolean function g:{±1}m{±1}n each of whose coordinates gi is unbiased and given by a low degree function. Then, the lifted polynomial fg:{±1}m[0,1] will be a low degree polynomial. As long as the gi’s are pairwise independent, the variance of f will be preserved as well. Our result shows that a large fraction of random restrictions of fg have an influential coordinate. Can we construct g1,g2,,gn appropriately such that this allows us to conclude f must have an influential coordinate as well? The gi’s should introduce correlations between the different input bits of f so that most random restrictions of fgm look the same in some appropriate sense.

  • Our result shows that for a non-negligible fraction of y{±1}[n]U, there exists a j such that 𝖨𝗇𝖿j[fy]Var[f]2/dO(1). Can we construct a decision tree of height poly(d) which, on input y, outputs one of the influential coordinates of fy with high probability? If this is possible, this will imply Aaronson-Aambainis via the approach of Keller and Klein. We may additionally assume that for all j, Infj[f]=Ey[Infj[fy]]Var[f]2/d10000 (otherwise the result holds for f anyway). So basically the situation looks like this: for a non-negligible fraction of y, for some j, Infj[fy] is significantly higher than its expectation (Infj[f]) – and we want to probe y in a few bits to locate such a j.

References

  • [1] Scott Aaronson and Andris Ambainis. Forrelation: A problem that optimally separates quantum from classical computing. Electron. Colloquium Comput. Complex., TR14-155, 2014. URL: https://eccc.weizmann.ac.il/report/2014/155.
  • [2] Scott Aaronson and Andris Ambainis. The need for structure in quantum speedups. Theory Comput., 10:133–166, 2014. doi:10.4086/TOC.2014.V010A006.
  • [3] Nikhil Bansal, Makrand Sinha, and Ronald de Wolf. Influence in completely bounded block-multilinear forms and classical simulation of quantum algorithms. In Shachar Lovett, editor, 37th Computational Complexity Conference, CCC 2022, July 20-23, 2022, Philadelphia, PA, USA, volume 234 of LIPIcs, pages 28:1–28:21. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2022. doi:10.4230/LIPICS.CCC.2022.28.
  • [4] Robert Beals, Harry Buhrman, Richard Cleve, Michele Mosca, and Ronald de Wolf. Quantum lower bounds by polynomials. In 39th Annual Symposium on Foundations of Computer Science, FOCS ’98, November 8-11, 1998, Palo Alto, California, USA, pages 352–361. IEEE Computer Society, 1998. doi:10.1109/SFCS.1998.743485.
  • [5] Paul Beame, Russell Impagliazzo, Jan Krajícek, Toniann Pitassi, Pavel Pudlák, and Alan R. Woods. Exponential lower bounds for the pigeonhole principle. In S. Rao Kosaraju, Mike Fellows, Avi Wigderson, and John A. Ellis, editors, Proceedings of the 24th Annual ACM Symposium on Theory of Computing, May 4-6, 1992, Victoria, British Columbia, Canada, pages 200–220. ACM, 1992. doi:10.1145/129712.129733.
  • [6] Ethan Bernstein and Umesh V. Vazirani. Quantum complexity theory. SIAM J. Comput., 26(5):1411–1473, 1997. doi:10.1137/S0097539796300921.
  • [7] Irit Dinur, Ehud Friedgut, Guy Kindler, and Ryan O’Donnell. On the fourier tails of bounded functions over the discrete cube. In Proceedings of the Thirty-Eighth Annual ACM Symposium on Theory of Computing, STOC ’06, pages 437–446, New York, NY, USA, 2006. Association for Computing Machinery. doi:10.1145/1132516.1132580.
  • [8] Yuval Filmus and Hamed Hatami. Bounds on the sum of L1 influences. CoRR, abs/1404.3396, 2014. arXiv:1404.3396.
  • [9] Merrick L. Furst, James B. Saxe, and Michael Sipser. Parity, circuits, and the polynomial-time hierarchy. Math. Syst. Theory, 17(1):13–27, 1984. doi:10.1007/BF01744431.
  • [10] Johan Håstad. Almost optimal lower bounds for small depth circuits. In Juris Hartmanis, editor, Proceedings of the 18th Annual ACM Symposium on Theory of Computing, May 28-30, 1986, Berkeley, California, USA, pages 6–20. ACM, 1986. doi:10.1145/12130.12132.
  • [11] Johan Håstad. On small-depth frege proofs for PHP. In 64th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2023, Santa Cruz, CA, USA, November 6-9, 2023, pages 37–49. IEEE, 2023. doi:10.1109/FOCS57990.2023.00010.
  • [12] Russell Impagliazzo, Raghu Meka, and David Zuckerman. Pseudorandomness from shrinkage. In 53rd Annual IEEE Symposium on Foundations of Computer Science, FOCS 2012, New Brunswick, NJ, USA, October 20-23, 2012, pages 111–119. IEEE Computer Society, 2012. doi:10.1109/FOCS.2012.78.
  • [13] Nathan Keller and Ohad Klein. Quantum speedups need structure, 2019. arXiv:1911.03748.
  • [14] Michel Ledoux and Michel Talagrand. Probability in Banach Spaces. Springer Berlin Heidelberg, 1991. doi:10.1007/978-3-642-20212-4.
  • [15] Shachar Lovett and Jiapeng Zhang. Fractional Certificates for Bounded Functions. In Yael Tauman Kalai, editor, 14th Innovations in Theoretical Computer Science Conference (ITCS 2023), volume 251 of Leibniz International Proceedings in Informatics (LIPIcs), pages 84:1–84:13, Dagstuhl, Germany, 2023. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ITCS.2023.84.
  • [16] Ashley Montanaro. Some applications of hypercontractive inequalities in quantum information theory. Journal of Mathematical Physics, 53(12), December 2012. doi:10.1063/1.4769269.
  • [17] Ryan O’Donnell. Analysis of boolean functions, 2021. arXiv:2105.10386.
  • [18] Ryan O’Donnell, Michael E. Saks, Oded Schramm, and Rocco A. Servedio. Every decision tree has an influential variable. CoRR, abs/cs/0508071, 2005. arXiv:cs/0508071.
  • [19] Ryan O’Donnell and Yu Zhao. Polynomial bounds for decoupling, with applications. CoRR, abs/1512.01603, 2015. arXiv:1512.01603.
  • [20] Theodore J. Rivlin. Chebyshev polynomials – From approximation theory to algebra and number theory: 2nd edition. John Wiley & Sons Limited, 1990.
  • [21] Benjamin Rossman, Rocco A. Servedio, and Li-Yang Tan. An average-case depth hierarchy theorem for boolean circuits. In Venkatesan Guruswami, editor, IEEE 56th Annual Symposium on Foundations of Computer Science, FOCS 2015, Berkeley, CA, USA, 17-20 October, 2015, pages 1030–1048. IEEE Computer Society, 2015. doi:10.1109/FOCS.2015.67.