Characterizing the Distinguishability of Product Distributions Through Multicalibration
Abstract
Given a sequence of samples promised to be drawn from one of two distributions , a well-studied problem in statistics is to decide which distribution the samples are from. Information theoretically, the maximum advantage in distinguishing the two distributions given samples is captured by the total variation distance between and . However, when we restrict our attention to efficient distinguishers (i.e., small circuits) of these two distributions, exactly characterizing the ability to distinguish and is more involved and less understood.
In this work, we give a general way to reduce bounds on the computational indistinguishability of and to bounds on the information-theoretic indistinguishability of some specific, related variables and . As a consequence, we prove a new, tight characterization of the number of samples needed to efficiently distinguish and with constant advantage as
which is the inverse of the squared Hellinger distance between two distributions and that are computationally indistinguishable from and . Likewise, our framework can be used to re-derive a result of Halevi and Rabin (TCC 2008) and Geier (TCC 2022), proving nearly-tight bounds on how computational indistinguishability scales with the number of samples for arbitrary product distributions.
At the heart of our work is the use of the Multicalibration Theorem (Hébert-Johnson, Kim, Reingold, Rothblum 2018) in a way inspired by recent work of Casacuberta, Dwork, and Vadhan (STOC 2024). Multicalibration allows us to relate the computational indistinguishability of to the statistical indistinguishability of (for lower bounds on ) and construct explicit circuits to distinguish between and consequently (for upper bounds on ).
Keywords and phrases:
Multicalibration, computational distinguishabilityFunding:
Cassandra Marcussen: Supported in part by an NDSEG fellowship, and by NSF Award 2152413 and a Simons Investigator Award to Madhu Sudan.Copyright and License:
2012 ACM Subject Classification:
Theory of computation Circuit complexityAcknowledgements:
This project was inspired by a final project completed by the first two authors during Cynthia Dwork’s course “Topics in Theory for Society: The Theory of Algorithmic Fairness” at Harvard (Spring 2024). We thank Cynthia Dwork for initial discussions regarding multicalibration and its applications to complexity and hardness amplification. We thank the anonymous Eurocrypt and CCC reviewers for their suggestions and feedback. We thank Pranay Tankala for a correction regarding the statement of the Multicalibration Theorem.Editors:
Srikanth SrinivasanSeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
Given a sequence of samples promised to be drawn from one of two distributions , each defined on a domain , a well-studied problem in statistics is to decide which distribution the samples are from. Information theoretically, the maximum advantage in distinguishing given samples is , where denotes the result of independently sampling times from , and denotes the total variation distance. This advantage can then be related to the single-sample relation between in a few ways:
-
1.
An upper bound is given by the inequality:
(1) -
2.
To obtain a good 2-sided bound, we can use the Hellinger distance:
(2) where the Hellinger distance is defined as
In particular, Inequality (2) above shows that samples is both necessary and sufficient for distinguishing with constant advantage. In contrast, observe that Equation 1 in the first point above is actually not tight. To illustrate, we can consider the two distributions below:
-
1.
such that .
-
2.
such that .
Equation 1 implies only that , and thus does not rule out the possibility of distintiguishing with constant advantage after samples. However, as we see in the Hellinger-distance formulation of Equation 2, in this example samples is both necessary and sufficient to distinguish with constant advantage. Thus, a key advantage in using Hellinger distance to characterize statistical distinguishability is in getting instance-optimal characterizations of the statistical distinguishability for every pair of random variables.
In computer science, this task of distinguishing distributions arises in many domains, from cryptography to computational complexity. However, we typically work with the computational analogue of total variation distance known as computational indistinguishability. We say that are -indistinguishable by circuits of size if for every function computable by size circuits, . In this paper, we state most of our results using this concrete security formulation, where and are separate parameters. However, our results can be translated into the traditional asymptotic complexity formulation where , and for a security parameter and unspecified super-polynomial functions . In fact, we are interested in the setting of weak indistinguishability where is non-negligible () while remains super-polynomial (). We elaborate on the asymptotic formulations of our results in Section 1.2.
The classic hybrid argument of Goldwasser and Micali [6] shows that if are -indistinguishable for circuits of size , then are -indistinguishable for circuits of size , which corresponds to the weaker bound in Equation 1 above. However, a more refined understanding of (in)distinguishability under repeated samples has been elusive. Indeed, the work of Halevi and Rabin [9] was the first to show that are -indistinguishable for circuits of size (with subsequent improvements to the parameters being made by [5]). Note that when and , we can take and retain , so this bound agrees with the tighter bound in Equation 1 above up to an additive negligible term.
This bound on computational distinguishability can be viewed as an extension of Equation 1 to the computational setting. Similarly to Equation 1 from the information-theoretic setting, the characterization of [9, 5] is not an instance-optimal characterization of the computational distinguishability of random variables, and a computational analog of Equation 2 is still missing. In this work, we overcome this shortcoming by building on the key technique of multicalibration, as we explain below.
1.1 Main Results
In this work, we give a general way to reduce reasoning about computational indistinguishability to reasoning about the information-theoretic case. This also allows us to deduce as corollaries both the characterization achieved by [9, 5] and a computational analogue of Inequality (2) (with the total variation distance of replaced by the computational indistinguishability of and the Hellinger distance between replaced by a quantity we call the pseudo-Hellinger distance between ).
Theorem 1.1.
For every pair of random variables over , every integer and every , there exist random variables such that for every ,
-
1.
is -indistinguishable from by circuits of size , for each .
-
2.
and are -indistinguishable by circuits of size .
-
3.
and are -distinguishable by circuits of size .
-
4.
The statements above also hold with , but .
Remark 1.2.
More generally, we can define indistinguishability with respect to an arbitrary class of functions . See the full version of the paper for generalized versions of the different parts of the above statement.
Remark 1.3.
We can also generalize the above theorem to arbitrary product distributions. We prove this in the full version of the paper.
As elaborated upon in Section 1.3, the above theorem follows from assigning and to be a careful “mixture” of and . This mixture depends on carefully partitioning the domain space of the random variables using multicalibration. With these random variables, Item 2 follows directly from Item 1 via hybrid argument, so the value of Theorem 1.1 is that it provides not only an upper bound on (in)distinguishability but a matching lower bound, via Item 3. Thinking of , we see that up to a negligible additive term , and a polynomial change in circuit size ( vs. ), the computational indistinguishability of under multiple samples is tightly captured by the information-theoretic indistinguishability of under multiple samples. In particular, samples are both necessary and sufficient to distinguish with constant advantage. We can abstract this latter corollary as follows:
Definition 1.4.
For random variables over , , , the pseudo-Hellinger distance between and is the smallest such that there exist random variables such that:
-
1.
is -indistinguishable from for circuits of size .
-
2.
is -indistinguishable from for circuits of size .
-
3.
.
With this definition in hand, we derive the following characterization of the indistinguishability of , which is a computational analogue of Inequality (2):
Theorem 1.5.
If have pseudo-Hellinger distance , then for every :
-
1.
are -indistinguishable for circuits of size .
-
2.
are distinguishable by circuits of size .
Thus, the number of samples needed to distinguish with constant advantage is , where is the pseudo-Hellinger distance. As an immediate consequence, just like the traditional notion of Hellinger distance in the statistical distinguishability setting, our notion of pseudo-Hellinger distance gives an instance-optimal characterization of the computational distinguishability of any pairs of random variables. This is a key benefit of our work over prior works [9, 5].
To prove Part (1) of Theorem 1.5, we first observe that, by parts (1) and (2) of Definition 1.4 and a simple hybrid argument, the computational indistinguishability of and is bounded above by the computational indistinguishability of and , plus . Next, the computational indistinguishability of and is upper-bounded by the total variation distance between , which by Inequality (2) is bounded above by . This gives us the bound in Part (1) of the theorem.
To prove Part (2) of Theorem 1.5, we use Theorem 1.1 to obtain that are -indistinguishable from such that the computational distinguishability of by circuits of size is at least . In turn, Inequality (2) tells us that this is at least , which is then at least , where is the pseudo-Hellinger distance, over all that are -indistinguishable from .
Additionally, using Part 4 of Theorem 1.1, we can prove that there exists a such that samples are both necessary and sufficient to distinguish with constant advantage. When is uniform on , squared Hellinger distance can be related to the Rényi -entropy of , defined as follows:
Specifically, we have
This allows us to create another suitable abstraction for the case where is uniform, beginning with the following definition of pseudo-Rényi entropy.
Definition 1.6.
For random variable over , , , the pseudo-Rényi -entropy of is the smallest such that there exists a random variable such that:
-
1.
is -indistinguishable from for circuits of size .
-
2.
.
This definition yields the following characterization of the indistinguishability of from the uniform distribution:
Theorem 1.7.
If is a distribution over that has pseudo-Rényi -entropy and is the uniform distribution over , then for every :
-
1.
are -indistinguishable for circuits of size .
-
2.
are distinguishable by circuits of size .
The number of samples needed to distinguish from uniform is where is the gap between the pseudo-Rényi -entropy and its maximum possibly value (namely ).
As mentioned, we also can deduce Geier’s result [5] as stated above directly from Theorem 1.1. A formal statement and comparison can be found in the full version of the paper, but we note that Geier’s quantitative bound (the specific polynomial loss in circuit complexity) is better than ours. In addition, Geier proves a version of the result for uniform distinguishers, whereas we only work with nonuniform distinguishers (i.e., boolean circuits). Both of these limitations come from the currently known theorems about multicalibration, which is the main technical tool we use (as discussed below), and it is interesting problem to obtain multicalibration theorems that provide tight quantitative bounds and yield uniform-complexity results in applications such as ours. The benefit of using multicalibration is that it provides a direct translation between the computational and information-theoretic setting (e.g. as captured in Theorem 1.1), which not only yields existing results as corollaries but also offers new ones, such as Theorem 1.5.
1.2 Asymptotic Complexity Formulations
In the foundations of cryptography, it is common to state computational indistinguishability results in terms of asymptotic polynomial complexity. In this section, we demonstrate the extensibility of our results to the asymptotic setting by presenting Theorem 1.5 in such a way, along with concrete definitions of this asymptotic regime. We start by formalizing indistinguishability in the asymptotic setting (i.e. with respect to ensembles of random variables):
Definition 1.8.
Let and be ensembles of random variables, where each are supported on and . For , we say that if for all , there exists an such that for all , is -indistinguishable from by circuits of size .
Equivalently, we say if there exists , such that for all is -indistinguishable from with respect to size circuits.111The equivalence of these two formulations goes back to Bellare [1].
For simplicity, we will also use to denote that there exists some function , such that .
We also generalize the notion of pseudo-Hellinger distance (Definition 1.4) to the setting of ensembles of random variables:
Definition 1.9.
For ensembles of random variables and , we say that and have pseudo-Hellinger distance at most (denoted ) for a function , if there exist ensembles such that , , and , .
Finally, we require a notion of the sample complexity required to distinguish ensembles:
Definition 1.10.
We say that ensembles have computational sample complexity at least for if (where )222This choice of is arbitrary, and can be amplified through repetition.. We denote this as .
With these definitions, we can now state a corollary of our characterization of the indistinguishability of random variables in terms of their pseudo-Hellinger distance.
Corollary 1.11 (Asymptotic Formulation of Theorem 1.5).
Let be ensembles of random variables, let , let be a function such that , and let .
Then:
-
1.
If , then for some .
-
2.
If , then for some .
We delegate the proof of this to the full version of the paper (in its appendix).
1.3 Technical Overview
Our work builds on the recent work of Casacuberta, Dwork, and Vadhan [2]. Driving inspiration from [4], they showed how the recent notion of multicalibration from the algorithmic fairness literature [10] leads to simpler and more illuminating proofs of a number of fundamental known results in complexity theory and cryptography, such as the Impagliazzo Hardcore Lemma [12], the Dense Model Theorem [14], and characterizations of pseudoentropy [16]. We show how multicalibration can be used to derive new results about computational indistinguishability (Theorems 1.1 and 1.5) and in general reduce computational reasoning to information-theoretic reasoning. Specifically, applying the Multicalibration Theorem [10] to distinguishing problems in a way inspired by [2], we obtain the following. Let be the distribution where is sampled according to conditioned on , for some function . Let be the distribution over where for , the probability that is exactly .
Theorem 1.12.
For every pair of random variables , every positive integer , and every , there exist random variables and a function for such that:
-
(a)
is -indistinguishable from and is -indistinguishable from for circuits of size .
-
(b)
is identically distributed to and is identically distributed to .
-
(c)
For every , is identically distributed to .
-
(d)
is computable by circuits of size .
Because the above theorem shows that and are -indistinguishable for circuits of size , then by a simple hybrid argument, and are -indistinguishable for circuits of size . Thus, one direction of Theorem 1.1 clearly follows given Theorem 1.12: size circuits cannot distinguish better than .
However, the other direction is more subtle; namely, showing that one can distinguish with advantage approaching with small circuits. For this, we heavily rely on items (b), (c), and (d) from Theorem 1.12, as well as the property that . The intuition is the following: consider an element that is sampled from either , and let . By item (c), we know that once we condition on a value , are identically distributed. Thus, there is no information to be gained from the sample besides the label of the partition that is in (i.e., the value ). In fact, this gives us a simple recipe for the optimal distinguisher between : for each sample we see, let us compute the value , and see how much more likely it is to sample an element in under vs. . We can simply keep a running product over all of the samples of the form
If the above product is larger than , this means that a sequence of values is more likely under , and if it is less than , then this means a sequence of partitions of more likely under . In particular, calculating this simple expression and checking whether it is larger than is an optimal distinguisher for , as it is just the maximum likelihood decoder (that is to say, this decoder achieves distinguishing advantage between ). All that remains is to show that we can encode this distinguisher using small circuits. Here, we rely on point (d) of Theorem 1.12: computing the function can be done with circuits of small size. Thus, for any element , we compute the index of the partition that it is in , and feed this index into a look-up table (encoded non-uniformly) such that it returns the value . Across all elements we see, we then must only keep track of the product, and return whether the value it computes is . Here, we use the fact that has a small domain to prove that we can actually encode the look-up table mapping to the values without using too large of a circuit. Of course, there are also issues with overflow and underflow in performing the numerical calculations, but we show this can be computed formally with small circuits in the full version of the paper.
Finally, it remains to analyze the distinguisher we have constructed here for when we apply it to . Here, we rely on Part (b) of Theorem 1.12. Indeed, the distinguisher relies only on the probabilities (i.e., the probability mass placed on each value of ). But, the distributions place exactly the same probability mass on each part of the partition (formally, ). Thus, whatever distinguishing advantage our distinguisher achieves on is exactly the same as its distinguishing advantage over . This then yields the second part of Theorem 1.1.
1.3.1 Multicalibrated Partitions
We now review the concept of multicalibration and describe how we prove Theorem 1.12 from the Multicalibration Theorem.
Multicalibration is a notion that first arose in the algorithmic fairness literature. Roughly speaking, its goal is, given a predictor of an unknown function , and a domain (which is often thought of as a population of individuals), to ensure that every desired subpopulation receives calibrated predictions from . The subpopulations are specified by their characteristic functions , which came from a family . More formally, we say that for domain and distribution over , a function , and a class of functions , is a multicalibrated predictor for with respect to , if , and :
The sets define a partition of , which gives rise to the following more convenient formulation. Let denote the distribution conditioned on being in the subset of the domain.
Definition 1.13 (Multicalibration [10] as formulated in [2]).
Let be an arbitrary function, be a probability distribution over , and for an integer , let be the class of functions computable by size circuits. Let be constants. Then, we say that a partition of is -approximately multicalibrated for on if for every and every such that it holds that
where we define .
For the class of functions computable by size circuits, let denote the set of partitions such that there exists , and .
The result of Hébert-Johnson, Kim, Reingold, and Rothblum [10] can be stated as follows in the language of partitions:
Theorem 1.14 (Multicalibration Theorem [10]).
Let be a finite domain, for an integer , let be the class of functions computable by size circuits, let be an arbitrary function, be a probability distribution over , and . Then, there exists a -approximately multicalibrated partition of for on such that , where and .
In particular, one way to understand the above definition is that the partition breaks the domain into parts, such that within each part, for all functions , the function is -indistinguishable from its expected value.
Recently, this perspective on multicalibration has proven to be quite fruitful, with applications to machine learning [7], graph theory [4], and complexity theory and cryptography [4, 2]. Philosophically, the work of [2] showed how to reprove (and strengthen) theorems of average-case hardness and indistinguishability using multicalibration (specifically, the Impagliazzo Hardcore Lemma [12], the Dense Model Theorem [14], and characterizations of pseudoentropy [16]). One consequence of their work was a “loose” intuition that multicalibration translates questions about computational indistinguishability into questions about statistical indistinguishability. Our goal is to show this translation more clearly and generally, and underline directly how multicalibration reduces an understanding of computational indistinguishability to an understanding of statistical indistinguishability (as captured by Theorem 1.1).
1.3.2 Invoking Multicalibration
As mentioned above, multicalibration has already found a host of applications. However, for the specific purpose of distinguishing two distributions , it is not immediately clear how to apply this framework. Of course, we can naturally view each distribution as given by its probability mass function from to , but in order to create multicalibrated partitions, we need a function with respect to which the partitions should be calibrated. For this, our first observation is to use the following function :
Roughly speaking, the function is motivated by looking at the distribution . In this distribution, the probability of seeing an element is . However, we can also understand the sampling procedure as first choosing one of with probability , and then sampling from the chosen distribution. The probability that comes from is then , and similarly for . is thus measuring the relative fraction of the time that an element comes from .
Importantly, is now a well-defined function with respect to which we can construct a multicalibrated partition. Then we will define for by replacing with .
We define this procedure formally below:
Definition 1.15.
For a pair of random variables , positive integer , and parameter , we define the distribution , function , random variables as follows.
-
(a)
Let the distribution be as follows: To sample from , first pick uniformly at random. Output a sample . Note that
For a subset of the domain , let be the conditional distribution of over the set .
-
(b)
We then define the randomized function as:
-
(c)
Random variables : Consider the multicalibrated partition guaranteed by the Multicalibration Theorem when applied to the function , distribution over domain , class of functions computable by size circuits, and parameters and . Given the multicalibrated partition, construct the random variables as follows: to sample according to , first choose a piece of the partition , where is chosen with probability . Then sample .
With this definition, we can provide an informal overview of the proof of Theorem 1.12. We start with item (b), as its proof is essentially definitional. Recall that this item states that and are identically distributed (and analogously for ). This follows because in the construction of the random variable , we sample an element from part with probability , and thus necessarily, the probability mass placed on each part is identical between the two distributions. Next, we can also see that part (c) is definitional. In the construction of and , conditioned on being in the piece , each of the marginal distributions is exactly , and thus the marginal distributions match. Next, to conclude item (d) of Theorem 1.12, recall that the partition function is exactly returned by the Multicalibration Theorem. Here, we can take advantage of the fact that multicalibrated partitions are efficiently computable given oracle access to the underlying set of test functions (in this case, size circuits). By replacing these oracles with the size circuits, this yields an overall complexity of for the circuit size required to compute . Finally, it remains to prove item (a) of Theorem 1.12. Because the partition that is returned is multicalibrated with respect to our function , by applying Theorem 2.15 of [2] and following the approach of the DMT proof of [2], we can show that for every part and every circuit of size , the distributions and are close to indistinguishable (and likewise for ). By combining this “local” indistinguishability across all parts of the partition, this then yields the overall indistinguishability between .
It is worth comparing our use of the Multicalibration Theorem to the use of Impagliazzo’s Hardcore Theorem [12, 11] in some past works on computational indistinguishability and pseudorandomness [9, 13]. Applied to the same function above, the Hardcore Theorem says that if are -indistinguishable, for , then there is a subset such that such that and are -indistinguishable. Replacing and with their average as above, we obtain that are -indistinguishable from respectively, and such that Thus, the Hardcore Lemma tightly captures the single-sample computational indistinguishability of and by the single-sample information-theoretic indistinguishability of and . But it does not seem to capture enough information to obtain an instance-optimal characterization of the multiple-sample indistinguishability, as we do. Concretely, the Hardcore Lemma requires assuming some initial average-case hardness of (which comes from the weak indistinguishability of and ) and only gives us indistinguishability from an information-theoretic counterpart on a small subset , which is not guaranteed to be efficiently recognizable. In contrast, the partition given by the Multicalibration Theorem covers the entire domain with efficiently recognizable sets, and this is crucial for our instance-optimal characterization.
1.4 Organization of the paper
Sections 3 and additional sections from the full version prove the different components of Theorem 1.1. In Section 3, we apply the Multicalibration Theorem to prove Theorem 1.12. Theorem 1.12 Part (a) is equivalent to Theorem 1.1 Part (1). The full version presents a proof of Theorem 1.1 Parts (2) and (3), as well as a more general version of the statement we prove, which encompasses indistinguishability against families of functions beyond size circuits. Also in the full version are proofs of Theorem 1.1 Part (4), and a general version of the statement, again for broader families of functions, which all rely on Theorem 1.12.
The full version of the paper also shows how our results and analysis imply a result comparable to the main theorem proven in [9, 5], and how it can be used to prove Theorem 1.5 and Theorem 1.7, which characterize distinguishability in terms of pseudo-Hellinger distance and pseudo-Rényi -entropy (as defined in Definition 1.4 and Definition 1.6). Lastly, the full version also presents a generalization of our results to the distinguishability of general product distributions (instead of versus ).
2 Preliminaries
2.1 Distinguishability
Hellinger distance can be used to characterize the ability to distinguish two distributions when we receive multiple samples:
Combining the above gives us that
| (3) |
As a corollary, this implies that samples are necessary and sufficient to increase the total variation distance up to a constant.
Definition 2.2.
Let be a random variable over . The Rényi entropy of order of is defined as:
It can be shown that with equality if and only if is uniform on . More generally, the gap in this inequality exactly measures the Hellinger distance to the uniform distribution on .
Claim 2.3.
Let be a random variable over and let be the uniform distribution over . Then
Proof.
We can rewrite as follows:
Definition 2.4 (Statistical indistinguishability).
Random variables and are called -statistically indistinguishable if .
We define computational indistinguishability for a class of functions :
Definition 2.5 (Computational indistinguishability).
Random variables and over domain are -indistinguishable with respect to if for every :
Random variables and are -distinguishable with respect to if for every :
Definition 2.6 (Oracle-aided circuits and partitions).
For a class of functions, let be the class of functions that can be computed by an oracle-aided circuit of size with oracle gates, where each oracle gate is instantiated with a function from . Let denote the set of partitions such that there exists , and .
Note that when we restrict our attention to being the set of functions computable by size circuits, then is simply the set of functions computable by size circuits. The discussion in the introduction is using exactly this simplification.
2.2 Multicalibration
We now recall the definition of multicalibration [10, 7, 8, 3], but state it in its full generality (with respect to arbitrary classes of functions ):
Definition 2.7 (Multicalibration [10] as formulated in [2]).
Let be an arbitrary function, be a probability distribution over , and be a class of functions . Let be constants. A partition of is -approximately multicalibrated for on if for every and every satisfying it holds that
where we define .
As mentioned above, [10] shows how to construct multicalibrated partitions with low-complexity relative to the class of functions . Their main result can be stated as:
Theorem 2.8 (Multicalibration Theorem [10]).
Let be a finite domain, a class of functions, with , an arbitrary function, a probability distribution over , and . Then, there exists a -approximately multicalibrated partition of for on such that for , , and .
3 Applying the Multicalibration Theorem to Prove Theorem 1.12
In this section, we apply the Multicalibration Theorem to prove the following theorem, which is a generalization of Theorem 1.12. In later sections, we will use this theorem to prove the different components of Theorem 1.1.
Theorem 3.1 (General version of Theorem 1.12).
For every pair of random variables , every family of functions , and every , there exist random variables and a function for such that:
-
(a)
is -indistinguishable from and is -indistinguishable from for functions .
-
(b)
is identically distributed to and is identically distributed to .
-
(c)
For every , is identically distributed to .
-
(d)
is computable by functions in .
Definitions and notation used throughout the proof
The random variables and function in Theorem 3.1 are constructed as follows, by drawing a connection to the multicalibrated partition guaranteed by the Multicalibration Theorem.
Definition 3.2.
(General version of Definition 1.15) For a pair of random variables , class of functions, and parameter , we define the family of functions , distribution , function , random variables , and function as follows.
-
(a)
Let be a family of functions such that . For example, if corresponds to size circuits, then is the family of functions given by circuits of size at most , for some universal constant .
-
(b)
Let the distribution be as follows: To sample from , first pick uniformly at random. Output a sample . Note that
For a subset of the domain , let be the conditional distribution of over the set .
-
(c)
We then define the randomized function as:
-
(d)
Random variables : Consider the multicalibrated partition guaranteed by the Multicalibration Theorem when applied to the function , distribution over domain , class of functions , and parameters and . Given the multicalibrated partition, construct the random variables as follows: to sample according to , first choose a piece of the partition , where is chosen with probability . Then sample .
Let be the number of parts in this partition .
-
(e)
We let be the function that returns which part of the multicalibrated partition an element is in. That is, if for , .
Given this setup of , , and , we are now ready to prove the different parts of Theorem 3.1.
3.1 Proof of Part (a)
We first analyze the behavior of the random variables over parts of the partition given by the Multicalibration Theorem. We begin by studying the indistinguishability of versus , relying on the following lemma from [2].
Lemma 3.3 (Lemma 2.15 [2]).
Let be a class of functions that is closed under negation and contains the all-zero ( for all ) and all-one ( for all ) functions. Let be a distribution over and consider . Let be any family of functions such that for a universal constant .
Suppose that is identified with a randomized Boolean-valued function and is -indistinguishable from the constant function . Then the distribution is -indistinguishable from for .
Lemma 3.4.
For random variables and family of functions , consider the construction of the family of functions , the distribution , function , and partition as in Definition 3.2. For , let .
For all such that , is -indistinguishable from .
Proof.
Similarly to the approach in the proof of Theorem 5.3 (DMT++) in [2], we consider the distribution that picks at random and outputs a sample of , and the randomized function that outputs . This precisely corresponds to the choice of corresponding to and from Definition 3.2. As in the proof of Theorem 5.3 in [2], we will show that for all , and ; we can therefore use Lemma 3.3 (setting ) to transfer the indistinguishability of from over each part of the partition to the indistinguishability of and . More specifically, the Multicalibration Theorem guarantees that is a low-complexity partition with parts such that, on each with , is -indistinguishable from the corresponding constant function . Applying Lemma 3.3 with set to be therefore implies that for each part of the partition such that , is -indistinguishable from for any class of functions such that , for a universal constant .
We work out the details for showing below, and the proof of follows similarly. In what follows, let denote that the probability of an event is taken over the randomness of the randomized Boolean-valued function .
We see that
| (4) |
Now, the denominator equals:
Similarly, the numerator equals:
Therefore, Equation (4) gives:
which means that is equivalent to .
We next prove that we can represent as a convex combination of and .
Lemma 3.5.
For random variables and family of functions , consider the construction of the distribution and partition as in Definition 3.2. For every , define
Then .
Proof.
For ,
Relating this to the conditional distributions and of and , this expression equals:
Given the definition of and , we see that
| (5) |
Stated more concisely, . Note that , , and , so this is indeed a convex combination.
Next we prove that is computationally indistinguishable from and is computationally indistinguishable from on every part of the partition guaranteed by applying the Multicalibration Theorem to . We begin by noting that is -indistinguishable from , which is a simple consequence of the following observation.
Observation.
Note that, for all , , which can be seen as follows:
Lemma 3.6.
For random variables and family of functions , consider the construction of the family of functions , the distribution , function , and partition as in Definition 3.2. For , consider the definitions of and as in Lemma 3.5.
For all such that , is -indistinguishable from and is -indistinguishable from .
The idea behind the proof of this lemma is as follows. First, since for all , is equivalent to , we want to argue that must also be indistinguishable from on all such that . This is implied by the facts that is a convex combination of and and these two random variables are indistinguishable for these .
Proof.
By definition of , for :
From Lemma 3.5, this equals:
Define to be the random variable such that for , . Note that this equals . Recall that, from Lemma 3.4, for all such that is -indistinguishable from .
To show that is -indistinguishable from we need to bound, for every :
We see that
Similarly, we can also show that is -indistinguishable from .
We are finally ready to prove Theorem 3.1 part (a).
Proof.
For random variables and family of functions , consider the construction of the family of functions , the distribution , function , and partition as in Definition 3.2.
We use the analysis of indistinguishability of random variables over all such that to prove the indistinguishability of and globally over the domain. Because we have results about indistinguishability over parts of the partition that have enough weight with respect to , but not those whose weight is too small, we need to break up the analysis to handle both types of . Let be the set of such that .
We focus on the indistinguishability of from in the proof. The proof of indistinguishability of from follows by similar arguments.
| (6) |
Let us focus on the first of the two summations in Equation (6). We see
Applying Lemma 3.6, this is:
Let us now focus on the second of the two summations in Equation (6). We see
where the last inequality follows from observing that . Note that, by definition of and by the bounds on from the Multicalibration Theorem, is .
Combining the different components of the analysis above, we find that:
Therefore, is -indistinguishable from . Similarly, is -indistinguishable from .
Additionally, to conclude the theorem, note that by definition of , for the partition given by the Multicalibration Theorem that are defined according to, for all , .
3.2 Proof of Parts (b, c, d)
Here, we prove the remaining properties of Theorem 3.1:
Proof.
For random variables and family of functions , consider the construction of the family of functions , the distribution , function , partition , and function as in Definition 3.2. We show the following:
Claim 3.7 (Part (b) of Theorem 3.1).
is identically distributed to and is identically distributed to .
Proof.
We show this WLOG for . This follows by definition. Recall that to construct , we sample part with probability , and then replace the conditional distribution over with (as defined in Definition 3.2). Thus, for any , we have .
Claim 3.8 (Part (c) of Theorem 3.1).
For every in , is identically distributed to .
Proof.
Again, this is merely definitional. As in Definition 3.2, for any part , we define
and likewise
Thus, the two distributions have the same marginal when conditioned on being in any part . We conclude by recalling that the parts are exactly the pieces , for , hence yielding the statement.
Claim 3.9 (Part (d) of Theorem 3.1).
is computable by circuits in
Proof.
Recall that in Definition 3.2, is exactly the partition function that results from constructing a multicalibrated partition with parameters on the family (where ). As in Theorem 2.8, such a partition function can be computed by a function in
This yields the claim.
The proof of Theorem 3.1 then follows from the union of each of the individual claims.
References
- [1] Mihir Bellare. A note on negligible functions. J. Cryptol., 15(4):271–284, 2002. doi:10.1007/S00145-002-0116-X.
- [2] Sílvia Casacuberta, Cynthia Dwork, and Salil Vadhan. Complexity-theoretic implications of multicalibration. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, pages 1071–1082, 2024. doi:10.1145/3618260.3649748.
- [3] Sílvia Casacuberta, Cynthia Dwork, and Salil P. Vadhan. Complexity-theoretic implications of multicalibration. CoRR, abs/2312.17223, 2023. doi:10.48550/arXiv.2312.17223.
- [4] Cynthia Dwork, Daniel Lee, Huijia Lin, and Pranay Tankala. From pseudorandomness to multi-group fairness and back. In The Thirty Sixth Annual Conference on Learning Theory, pages 3566–3614. PMLR, 2023.
- [5] Nathan Geier. A tight computational indistinguishability bound for product distributions. In Theory of Cryptography Conference, pages 333–347. Springer, 2022. doi:10.1007/978-3-031-22365-5_12.
- [6] Shafi Goldwasser and Silvio Micali. Probabilistic encryption. J. Comput. Syst. Sci., 28(2):270–299, 1984. doi:10.1016/0022-0000(84)90070-9.
- [7] Parikshit Gopalan, Adam Tauman Kalai, Omer Reingold, Vatsal Sharan, and Udi Wieder. Omnipredictors. In Mark Braverman, editor, 13th Innovations in Theoretical Computer Science Conference, ITCS 2022, January 31 - February 3, 2022, Berkeley, CA, USA, volume 215 of LIPIcs, pages 79:1–79:21. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2022. doi:10.4230/LIPICS.ITCS.2022.79.
- [8] Parikshit Gopalan, Michael P. Kim, Mihir Singhal, and Shengjia Zhao. Low-degree multicalibration. In Po-Ling Loh and Maxim Raginsky, editors, Conference on Learning Theory, 2-5 July 2022, London, UK, volume 178 of Proceedings of Machine Learning Research, pages 3193–3234. PMLR, 2022. URL: https://proceedings.mlr.press/v178/gopalan22a.html.
- [9] Shai Halevi and Tal Rabin. Degradation and amplification of computational hardness. In Ran Canetti, editor, Theory of Cryptography, Fifth Theory of Cryptography Conference, TCC 2008, New York, USA, March 19-21, 2008, volume 4948 of Lecture Notes in Computer Science, pages 626–643. Springer, 2008. doi:10.1007/978-3-540-78524-8_34.
- [10] Úrsula Hébert-Johnson, Michael P. Kim, Omer Reingold, and Guy N. Rothblum. Multicalibration: Calibration for the (computationally-identifiable) masses. In Jennifer G. Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018, volume 80 of Proceedings of Machine Learning Research, pages 1944–1953. PMLR, 2018. URL: http://proceedings.mlr.press/v80/hebert-johnson18a.html.
- [11] Thomas Holenstein. Key agreement from weak bit agreement. In Harold N. Gabow and Ronald Fagin, editors, Proceedings of the 37th Annual ACM Symposium on Theory of Computing, Baltimore, MD, USA, May 22-24, 2005, pages 664–673. ACM, 2005. doi:10.1145/1060590.1060689.
- [12] Russell Impagliazzo. Hard-core distributions for somewhat hard problems. In Proceedings of IEEE 36th Annual Foundations of Computer Science, pages 538–545. IEEE, 1995. doi:10.1109/SFCS.1995.492584.
- [13] Ueli M. Maurer and Stefano Tessaro. A hardcore lemma for computational indistinguishability: Security amplification for arbitrarily weak prgs with optimal stretch. In Daniele Micciancio, editor, Theory of Cryptography, 7th Theory of Cryptography Conference, TCC 2010, Zurich, Switzerland, February 9-11, 2010. Proceedings, volume 5978 of Lecture Notes in Computer Science, pages 237–254. Springer, 2010. doi:10.1007/978-3-642-11799-2_15.
- [14] Omer Reingold, Luca Trevisan, Madhur Tulsiani, and Salil Vadhan. Dense subsets of pseudorandom sets. In 2008 49th Annual IEEE Symposium on Foundations of Computer Science, pages 76–85. IEEE, 2008. doi:10.1109/FOCS.2008.38.
- [15] Ton Steerneman. On the total variation and hellinger distance between signed measures; an application to product measures. Proceedings of the American Mathematical Society, 88(4):684–688, 1983.
- [16] Salil Vadhan and Colin Jia Zheng. Characterizing pseudoentropy and simplifying pseudorandom generator constructions. In Proceedings of the forty-fourth annual ACM symposium on Theory of computing, pages 817–836, 2012. doi:10.1145/2213977.2214051.
- [17] David Woodruff. Cs 15-859: Algorithms for big data: Lecture 9, 2019.
