Abstract 1 Introduction 2 Differential Privacy, Unbiasedness, and Deconvolution 3 Unbiased Estimation with Laplace Noise 4 Extension to Functions that are not Tempered Distributions 5 Numerical Results: Application to Mean Queries 6 Application to Slowly Scaling PRDP 7 Polynomial Functions under General Noise Distributions 8 Conclusions and Future Work References

Debiasing Functions of Private Statistics in Postprocessing

Flavio Calmon Harvard University, Cambridge, MA, USA Elbert Du ORCID Harvard University, Cambridge, MA, USA Cynthia Dwork Harvard University, Cambridge, MA, USA Brian Finley ORCID U.S. Census Bureau, Suitland, MD, USA Grigory Franguridi ORCID Center for Economic and Social Research, University of Southern California, Los Angeles, CA, USA
Abstract

Given a differentially private unbiased estimate q~=q(D)+ν of a statistic q(D), we wish to obtain unbiased estimates of functions of q(D), such as 1/q(D), solely through post-processing of q~, with no further access to the confidential dataset D. To this end, we adapt the deconvolution method used for unbiased estimation in the statistical literature, deriving unbiased estimators for a broad family of twice-differentiable functions – those that are tempered distributions – when the privacy-preserving noise ν is drawn from the Laplace distribution (Dwork et al., 2006). We further extend this technique to functions other than tempered distributions, deriving approximately optimal estimators that are unbiased for values in a user-specified interval (possibly extending to ±).

We use these results to derive an unbiased estimator for private means when the size n of the dataset is not publicly known. In a numerical application, we find that a mechanism that uses our estimator to return an unbiased sample size and mean outperforms a mechanism that instead uses the previously known unbiased privacy mechanism for such means (Kamath et al., 2023). We also apply our estimators to develop unbiased transformation mechanisms for per-record differential privacy, a privacy concept in which the privacy guarantee is a public function of a record’s value (Seeman et al., 2024). Our mechanisms provide stronger privacy guarantees than those in prior work (Finley et al., 2024) by using Laplace, rather than Gaussian, noise.

Finally, using a different approach, we go beyond Laplace noise by deriving unbiased estimators for polynomials under the weak condition that the noise distribution has sufficiently many moments.

Keywords and phrases:
Differential privacy, deconvolution, unbiasedness
Funding:
Flavio Calmon: This work was supported in part by Simons Foundation Grant 733782 and Cooperative Agreement CB20ADR0160001 with the United States Census Bureau. This material is also based upon work supported by the National Science Foundation under Grant No. CIF-2312667 and CIF-2231707.
Brian Finley: Works (articles, reports, speeches, software, etc.) created by U.S. Government employees are not subject to copyright in the United States, pursuant to 17 U.S.C. §105. International copyright, 2024, U.S. Department of Commerce, U.S. Government. Any opinions and conclusions expressed herein are those of the authors and do not reflect the views of the U.S. Census Bureau.
Copyright and License:
[Uncaptioned image] © Flavio Calmon, Elbert Du, Cynthia Dwork, Brian Finley, and Grigory Franguridi; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Security and privacy Data anonymization and sanitization
; Mathematics of computing Probability and statistics ; Theory of computation Theory of database privacy and security
Related Version:
Full Version: https://arxiv.org/pdf/2502.13314
Editors:
Mark Bun

1 Introduction

Differential privacy (DP) has become widely accepted as a “gold standard” of privacy protection in statistical analysis. In particular, it has been adopted by many companies such as Google, Meta, and Apple to protect customer data and by the U.S. Census Bureau to protect respondent data in the 2020 Census [1]. DP mechanisms work by introducing randomness into the computation of all statistics published from a protected database. This added noise guarantees that no attacker can confidently determine from the published statistics whether a particular record is included in the dataset, thereby preserving its privacy.

Among DP mechanisms, additive mechanisms are canonical and widely used. These simply add data-independent, zero-mean random noise to the statistics. For example, the Laplace mechanism adds Laplace-distributed noise and is one of the first and most fundamental DP mechanisms [6]. The scale of the added noise needs to be proportional to the statistics’ global sensitivity – the greatest amount by which the statistic could change upon the addition or deletion of a single record. Intuitively, this ensures that there is enough noise to mask the presence or absence of any particular record.

Often, however, the statistics to which noise is added differ from the final statistics of interest. In these cases, the statistics of interest must be estimated using the available noisy statistics. Suppose that the noisy statistic q~ is formed by applying an additive mechanism to the univariate statistic q, but that we want to learn f(q), not q. Even though q~ is unbiased for q, the plug-in estimator f(q~) is not generally unbiased for f(q). When unbiasedness is desired, other estimators must be used.

To address this problem, we first derive unbiased estimators for the Laplace mechanism, for a general class of twice-differentiable functions f – those which are also tempered distributions (Section 3). [11] develops recursive algorithms that are unbiased estimators for polynomials in Laplace variables. Our paper provides estimators for a large class of non-polynomial functions and gives a simple, closed-form estimator for polynomials. We also provide methods to adapt functions that are not tempered distributions in a way that permits unbiased estimation over a subset of q’s domain (Section 4). This extension lets us provide unbiased estimators for the case when f(q)=1/q, which, in turn, lets us provide unbiased estimators of ratio statistics. Such cases arise frequently in practice, as discussed below. Finally, we derive unbiased estimators for a very general class of additive mechanisms when f is a polynomial (see Section 7).

There are several reasons why noise may not be added directly to the statistic of interest and bias in the plug-in statistic must be considered. A leading case occurs when f(q) has a much higher global sensitivity than q. For example, when the domain of q includes 0 or values arbitrarily close to 0, the global sensitivity of f(q)=1/q is typically infinite and no amount of noise provides a finite DP guarantee. This same problem affects the many statistics which can be expressed as ratios of low-sensitivity statistics. For example, the mean is the ratio of a sum and a sample size. Likewise, in a simple linear regression of the regressand y on the regressor x, the ordinary least squares (OLS) estimator of the slope coefficient is the ratio of the empirical (co)variances Cov[x,y] and Var[x]. It is common, then, to add noise to the low-sensitivity statistics that form these ratios and use the plug-in estimator for the ratio statistic of interest. See, for example, [2] for such a treatment of the OLS estimator.

Noise may also be added to statistics that are not of direct interest because data curators, such as government agencies, may publish noisy microdata or a noisy predetermined set of aggregates for general-purpose use. For example, a researcher trying to learn the proportion of the population with doctoral degrees may only have access to published noisy totals of the general population and the population of degree holders. The plug-in estimator of the mean is, as above, the ratio of these noisy totals.

This situation may also arise because, under DP, there is a limited “privacy budget” which is drawn upon every time we use the raw data to release another (noisy) statistic. Splitting the budget among more statistics requires that more noise be added to each of them. This makes it beneficial to “re-use” statistics whenever possible. For example, in Section 5, we develop a DP mechanism that uses our results to provide private unbiased estimates of a mean and sample size. This mechanism obtains a noisy sample size via the Laplace mechanism and then re-uses it to estimate the denominator of the mean statistic. We find that this approach outperforms an alternative mechanism that uses an existing method from [12] to add noise directly to the mean query, without re-using the noisy sample size.

Again, these scenarios all have in common that the plug-in estimator f(q~) will typically be biased.111In fact, in the case with the Laplace mechanism and f(q)=1/q, the expectation and all higher moments of the plug-in estimator f(q~) fail to even exist, implying that the estimator has extremely fat tails and is very prone to returning extreme outliers. This affects the ratio statistics discussed above, as well. The unbiased estimator we develop for this case in Section 4.1 possesses finite moments of all orders. To see why unbiasedness is desirable, recall that the bias and variance of a sum of n uncorrelated estimates are respectively the sums of the estimates’ biases and variances. Accordingly, the sum’s bias increases at the rate O(n) while its standard deviation grows at the rate O(n). The sum’s overall RMSE therefore grows at the O(n) rate if the estimates are unbiased, but at the faster O(n) rate otherwise.

For example, consider the following simple example: suppose the true value of some quantity of interest is 1, but each time we try to learn the value of this quantity, we get a fresh draw from the distribution 𝒩(1,100). A mechanism that ignores the data and returns 0 has bias 1 and variance 0, resulting in an overall RMSE of 1. On the other hand, reporting the value of any single draw would have bias 0 and variance 100. On average, this will be off by around 10. Thus, on individual draws, the first mechanism is more accurate. However, if we take the mean of 10,000 such draws, the mechanism that always returned 0 still gives a mean of 0, which is still off by 1. On the other hand, the mean of the 10,000 draws now has variance 10010,000=1100, resulting in an RMSE of 110. That is, we would now expect this estimate to be off by around 0.1, which is a significant improvement over the biased estimator.

This makes unbiasedness very important in meta-analyses, which aggregate multiple estimates. It is also important when adding noise to a large number of quantities whose sums are of independent interest. This situation commonly arises in the local model of DP, where an extra layer of privacy is obtained by adding noise to every record even before entrusting it to the data curator. Sums or means using these noisy records could then be subject to severe error if the record-level estimates being summed are biased. For example, with network data, the count of k-stars (i.e., sets of k edges sharing a node) is a sum of polynomials in each node’s degree. In experiments with network data protected by local DP, [11] find that a mechanism that simply sums unbiased estimates of these polynomials outperforms the L2 error of prior work by factors as high as 5 orders of magnitude.

Likewise, unbiasedness is key when noise is added to disaggregate sums with the expectation that they can be aggregated further to obtain sums for larger groups. For example, [13] and [8] develop mechanisms for use with per-record DP – a variant of DP whose privacy guarantees differ between records, and which is being considered by the Census Bureau for use with its County Business Patterns (CBP) data product [3]. [8] develop transformation mechanisms for this purpose, which improve privacy guarantees by adding noise to concave functions of q rather than to q itself. Estimates of q must then be obtained from these noisy transformed values. The CBP data includes sums of employment and payroll, grouped by finely divided geographies and industry codes. If these transformation mechanisms were used for the CBP and the estimator of q for these sums were biased, further aggregates of these estimates to obtain, say, state-level sums would be subject to severe biases.

In Section 6, we apply our estimators to create variants of these transformation mechanisms that satisfy a stronger type of per-record DP guarantee than the ones originally proposed in [8].

In this paper, we make the following contributions:

  1. 1.

    We derive closed-form unbiased estimators for a large class of functions – twice-differentiable functions that are tempered distributions – when the Laplace mechanism is used. We also develop estimators that are unbiased for subsets of the statistic’s domain for functions that are not in this class.

  2. 2.

    We exposit the deconvolution method from the statistics literature (e.g., [15], page 185) for deriving unbiased estimators. This could be used to derive estimators for further functions and further mechanisms, and we believe its use in DP is novel and of independent interest.

  3. 3.

    We apply our unbiased estimators to create novel unbiased privacy mechanisms for per-record DP, a new variant of DP being considered for use by the Census Bureau [3].

  4. 4.

    We derive closed-form unbiased estimators for polynomial functions of statistics privatized using any of a large class of additive mechanisms.

2 Differential Privacy, Unbiasedness, and Deconvolution

The following is the definition of differential privacy, introduced in [6]:

Definition 1.

Datasets D and D are neighboring databases if they differ by the inclusion of at most 1 element.

Definition 2.

A mechanism is (ϵ,δ)-differentially private ((ϵ,δ)-DP) if, for any pair of neighboring datasets D,D and any measurable set of possible outcomes S, we have

Pr[(D)S]<eϵPr[(D)S]+δ.

Most of our work uses Fourier transforms [9]. The following definitions and theorems are adapted from the textbook treatment in [14].

Definition 3.

The Fourier transform of an absolutely integrable function f is

F[f(x)](y)=e2πiyxf(x)𝑑x.

We often denote the Fourier transform of f by f^(y). The Fourier transform also has an inverse:

F1[f^(y)](x)=e2πiyxf^(y)𝑑y.

There are many important functions we wish to compute unbiased estimates of which are not absolutely integrable. In particular, polynomials and the function f(q)=1/q are not absolutely integrable, so we must define the Fourier transform over a more general family of functions, tempered distributions.222Technically, 1/q is not a tempered distribution either, but only because it is poorly behaved at 0. This will be addressed in Section 4. Importantly, in this use, the term “distribution” does not refer to a probability distribution. Rather, it refers to a class of objects also known as “generalized functions.” For the purposes of this work, we can largely restrict ourselves to working with tempered distributions which are also functions, though there exist tempered distributions which are not functions, such as the Dirac delta “function”. Below, we specialize the relevant theory to the case of tempered distributions that are also functions, but the interested reader should see Appendix A of this paper’s full version (linked to on the title page) for the more general case.

For our purposes, then, tempered distributions can be thought of as functions that may not be absolutely integrable, but which grow no faster than a polynomial. Formally, this is expressed by the condition that the product of a tempered distribution and any function in the Schwartz space (defined below) is integrable.

Definition 4.

The Schwartz space S() is defined as follows:

S()={s:sC,supx|xms(n)(x)|<m,n},

where denotes the set of non-negative integers and s(n) denotes the nth derivative of s. That is, functions in S() are infinitely differentiable everywhere and they – along with all of their derivatives – go to 0 at a super-polynomial rate.

Note that all the functions sS() are absolutely integrable, so their Fourier transforms s^ exist. With the Schwartz space so defined, we introduce tempered distributions below.

Definition 5.

A function f is a tempered distribution if and only if, for all sS(),

f(x)s(x)𝑑x.

Definition 3 introduces the Fourier transform only for absolutely integrable functions. The following definition extends it to all tempered distributions.333To see that Definition 3 implies Definition 6 for absolutely integrable functions, note that the e2πixy term does not depend on the function f, so swapping the order of integration by Fubini’s theorem immediately gives us the equality in Definition 6.

Definition 6.

When it exists, f^ is the function such that for all sS(),

f^(x)s(x)𝑑x=f(x)s^(x)𝑑x.

Technically, the Fourier transform of a tempered distribution always exists and is a tempered distribution, but may not also be a function, even when the distribution being Fourier-transformed is a function. See Appendix A of this paper’s full version for details.

The deconvolution method we use to derive unbiased estimators in Section 3 is applicable because, as explained below, the requirement that an estimator be unbiased can be expressed in terms of a convolution.

Definition 7.

The convolution of functions f and g is

(fg)(x)=f(z)g(xz)𝑑z.

Critically, the Fourier transform of a convolution is the product of the convolved functions’ Fourier transforms.

Theorem 8 ([14] section 7.1 property c).
(fg)^(y)=f^(y)g^(y).

We will also need the following theorem to derive unbiased estimators for the case of Laplace noise.

Theorem 9 ([14] section 7.8 Example 5).

For any tempered distribution f, the Fourier transform of its kth derivative f(k) is f(k)^=(2πiy)kf^.

With the query q and its privacy-preserving noisy estimate q~, we say that an estimator g is unbiased for f(q) if

f(q)=𝔼[g(q~)|q]. (1)

By conditioning on the true query value, q, we treat the database as fixed. Our estimators, then, are unbiased with respect to the randomness in the 0-centered noise being added for privacy. Throughout the rest of this paper, all expectations are conditional on q unless otherwise noted and we suppress the extra conditioning notation so that 𝔼[.]𝔼[.|q].

Let the noise added for privacy be independent of the database and denote its PDF by r. The deconvolution method, as seen, for example, on page 185 of [15], starts by noting that if g is unbiased for f(q), then Equation 1 can be reexpressed in terms of a convolution:

f(q)=𝔼[g(q~)]=g(q~)r(q~q)𝑑q~=(gr)(q). (2)

With the unbiasedness equation in this form, Theorem 8 lets us Fourier-transform both sides to turn the convolution on the right-hand side into a simple multiplication. Finally, we simply solve for the Fourier transform of g in terms of the Fourier transforms of f and r and inverse-Fourier-transform the result. Formally,

f(q)=(gr)(q)f^(y)=g^(y)r^(y)g^(y)=f^(y)r^(y)g(x)=F1[f^(y)r^(y)](x), (3)

assuming the existence of all the involved Fourier and inverse Fourier transforms.

3 Unbiased Estimation with Laplace Noise

A standard mechanism for differential privacy perturbs the query with Laplace noise scaled to the global sensitivity of a query, which is the maximum difference between the query values on neighboring databases. That is, to achieve (ϵ,0)-DP when releasing the value of a query q with global sensitivity Δ, we can simply release q+Lap(0,Δϵ) [6].

Our primary contribution is deriving unbiased estimators for functions of q when we only have access to the value of q+Lap(0,b), for some noise scale parameter b. These estimators are unique (up to their values on a set of measure zero).

Theorem 10.

Let q~q+Lap(0,b).

  • For any twice-differentiable function f: that is a tempered distribution, f(q~)b2f′′(q~) is an unbiased estimator for f(q).

  • For any function f:, if two estimators g1(q~) and g2(q~) are unbiased for f(q), then g1 and g2 are equal almost everywhere.

Proof.

See Appendix B of this paper’s full version, linked to on the title page.

Some examples of unbiased estimators are given below.

Example 11.
  1. 1.

    Any power function f(q)=qk has unbiased estimator q~kb2k(k1)q~k2. In particular, for f(q)=cq for any constant c, the unbiased estimator is also cq~. Section 3.1 of [11] derives this estimator in the form of a recursive algorithm. We contribute the closed form here to simplify computation and facilitate intuitive understanding.

  2. 2.

    Within the set of twice differentiable functions f that are tempered distributions, Theorem 10 allows us to determine which functions are unbiased estimators of themselves. When f(q~) is unbiased for f(q), we have 𝔼[f(q~)]=𝔼[f(q~)b2f′′(q~)]𝔼[f′′(q~)]=0. By the second part of Theorem  10 and the unbiasedness of the zero function for zero, this implies f′′(x)=0 almost everywhere, so f(x) must be linear. The naive plug-in estimator, then, is biased for any nonlinear function in this class. This highlights the usefulness of Theorem 10.

    We can similarly characterize the f whose unbiased estimators are simply linear transformations of the plug-in estimator – that is, f for which 𝔼[αf(q~)+β]=f(q) for some α,β. By Theorem 10, these functions satisfy 𝔼[αf(q~)+β]=𝔼[f(q~)b2f′′(q~)], and so satisfy αf(x)+β=f(x)b2f′′(x) almost everywhere. When α1, solutions to this differential equation take the form444The case where α=1 and β=0 is dealt with above.

    f(x)=c1e1αbx+c2e1αbx+β1α.

    Using Euler’s formula, we can see that tempered distributions in this class include functions of the form f(x)=ccos(ux) and f(x)=csin(ux). Nonetheless, this is still a rather restricted class of functions.

 Remark 12.

When the function f is not twice differentiable but is a tempered distribution, an analog of Theorem 10 holds. This relies on the use of an alternative notion of the derivative that applies to all tempered distributions – the distributional derivative. For background on this derivative concept, see Appendix A of this paper’s full version.

In this case, we still have 𝔼[f(q~)]𝔼[b2f′′(q~)]=f(q), but the distributional derivative f′′ is a tempered distribution which is not a function. This does not give us an unbiased estimator, but instead we can rearrange to obtain the bias, as a function of q, of the naive plug-in estimator f(q~):

𝔼[f(q~)]f(q)=𝔼[b2f′′(q~)]=b2f′′(q+x)e|x|/b𝑑x. (4)

For example, let f(x)=|x|, f(x)=1 for x0 and 1 for x>0 (the discontinuity at 0 is irrelevant since {0} is a set of measure 0). Then f′′(x)=2δ(x), where δ is the Dirac delta function (defined in Example 29 in Appendix A of this paper’s full version). The bias is simply b22e|q|/b=be|q|/b.

Whether or not f is twice differentiable, Equation 4 suggests the intuition that the plug-in estimator will have greater bias when f has greater curvature near the true query value q.

4 Extension to Functions that are not Tempered Distributions

If the function qf(q) is not a tempered distribution, we can often bound the domain such that it is continuous and twice-differentiable in that domain. That is, suppose we know a priori that qL for some lower bound L. This is often the case in differential privacy, as DP-protected queries are commonly sums of nonnegative variables. Likewise, counts can often be lower bounded by 1. Then, suppose we replace the function f with some function

f~(q)={f(q)qLh(q)q<L, (5)

where h(L)=f(L) and h is twice differentiable with h(L)=f(L) and h′′(L)=f′′(L). The function f~ is thus twice differentiable. Assuming that h(q) and f(q) and their derivatives grow no faster than a polynomial as, respectively, q and q, f~ is a tempered distribution, as well. We can then apply Theorem 10 to get an unbiased estimator of f~, i.e.

𝔼[f~(q~)b2f~′′(q~)]=f~(q). (6)

With the assumption that qL, we have f~(q)=f(q), making this estimator unbiased for f(q), as well.

Example 13.

For f(q)=1q and L=1, we need to find some function h such that h(1)=1,h(1)=1,h′′(1)=2. An example of such a function is h(q)=1(q1)+(q1)2. We can generically use polynomials for h whenever f grows at most polynomially as q and is twice differentiable for qL.

We now focus on optimizing this method over polynomial extensions for a particular function of interest: f(q)=1/q.

4.1 Unbiased Estimation for 𝒇(𝒒)=𝟏/𝒒

We have shown that it is possible to construct a function that permits unbiased estimation as long as it is twice differentiable on some domain that the true query value is known to be in, and, if this domain is unbounded, as long as the function does not grow too quickly. In this section, we show how to optimally choose the function h in the above construction.

We restrict ourselves to polynomial functions h for two reasons. First, the solution among polynomials of fixed degree is efficiently computable. Second, when the optimal function h is a twice continuously differentiable tempered distribution, polynomials can approximate this function arbitrarily well, in the sense that the expected squared error of the polynomial-based estimator can be made arbitrarily close to optimal. This follows from Theorem 14.

Theorem 14 (Polynomial approximation).

Let L and let f:[L,) be twice differentiable and a tempered distribution. Let μ be a probability measure such that the integrals Lf(q)e(Lq)/b𝑑μ(q) and Le(Lq)/b𝑑μ(q) exist and are finite. With w:(,L], let f~[w] denote the function

f~[w](q)={f(q)qLw(q)q<L (7)

Let h:(,L] be an arbitrary twice continuously differentiable tempered distribution that satisfies h(L)=f(L), h(L)=f(L) and h′′(L)=f′′(L). Denote the estimator gf~[h]b2f~[h]′′ and denote its expected squared error by

αL(g(x)f(q))212be|xq|/b𝑑μ(q)𝑑x. (8)

There exists a sequence of polynomials (pK)K=1 over (,L] that satisfy pK(L)=f(L), pK(L)=f(L) and pK′′(L)=f′′(L) such that the sequence of associated estimators gKf~[pK]b2f~[pK]′′ satisfies

limKL(gK(x)f(q))212be|xq|/b𝑑μ(q)𝑑x=α. (9)

See Appendix C of this paper’s full version for a proof.

Now, letting h be a polynomial, suppose our estimator is g(x)=f~(x)b2f~′′(x) for f~ defined in Equation 5. For our error metric, we consider the estimator’s expected squared error, with the expectation taken over both the privacy noise and prior beliefs about q, reflected in the probability measure μ. We define our estimator as the solution to the following constrained optimization problem:

minf~L(f~(x)b2f~′′(x)f(q))212be|xq|/b𝑑μ(q)𝑑x (10)
=LL(f(x)b2f′′(x)f(q))212be|xq|/b𝑑μ(q)𝑑x
+mingLL(g(x)f(q))212be(xq)/b𝑑μ(q)𝑑x
subject to h(L)=f(L),h(L)=f(L), and h′′(L)=f′′(L).

Since the first double integral is constant with respect to h, optimizing this error metric is equivalent to optimizing

mingLL(g(x)f(q))212be(xq)/b𝑑μ(q)𝑑x, (11)

subject to the same constraints.

For simplicity, we shall now treat g as a function with domain (,L], as that is the only region on which we are optimizing, so g(x)=i=0kaixi. There is a one-to-one correspondence between polynomials g(x) and polynomials h(x)=i=0kbixi where ai=bib2(i+2)(i+1)bi+2. Thus, we are considering extensions of f~(q) where the part to the left of the lower bound L is a polynomial.

Theorem 15.

For any positive integer k, any real number L, and any function f which is twice differentiable on [L,), there is an algorithm that runs in time poly(k) which computes the polynomial g that minimizes

LL(g(x)f(q))212be(xq)/b𝑑μ(q)𝑑x

over polynomials of degree k, satisfying the constraints g(x)=h(x)b2h′′(x), h(L)=f(L),h(L)=f(L), and h′′(L)=f′′(L).

Proof.

See Appendix C.1 of this paper’s full version, linked to on the title page.

Corollary 16.

Provided that the optimal choice of h is a twice continuously differentiable tempered distribution, there exists an efficient algorithm to approximate the optimal unbiased estimator of f(q) given q+Z for ZLap(0,b) and the prior knowledge that qL.

This follows immediately from the fact that optimizing g also optimizes f~(q).

Note that this result can be easily extended to the cases where we have only an upper bound or both an upper and lower bound. If we only have an upper bound, everything works out exactly the same as if we only have a lower bound. If we have both, suppose we know that q[L,U] and define

f~(q)={h0(q)q>U,f(q)UqL,h1(q)q<L. (12)

Then the expected error (with the expectation over both the privacy noise and the prior on q) is

𝔼q,xq+Lap(0,b)[(f~(x)b2f~′′(x)f(q))2] (13)
=LU(f~(x)b2f~′′(x)f(q))212be|xq|/b𝑑μ(q)𝑑x.

Just like before, the error incurred by f~(x) on LxU is not affected by our choice of functions. Thus, we wish to compute

minh,h LLU(h1(x)b2h1′′(x)f(q))212be|xq|/b𝑑μ(q)𝑑x (14)
+ ULU(h0(x)b2h0′′(x)f(q))212be|xq|/b𝑑μ(q)𝑑x.

Since there is no interaction between h0 and h1, we can minimize these integrals independently in the same way as above.

5 Numerical Results: Application to Mean Queries

In this section, we illustrate the utility of our results by numerically comparing two mechanisms designed to return unbiased estimates of the sample size n and the mean m of an attribute c[0,1] in the database D. Sample sizes are published alongside any reported means in most research applications, making this a realistic use case. One mechanism, MU, returns an unbiased estimate of the mean using the results from Section 4.1. The other, MSS, uses the unbiased mean mechanism from [12] (see their Theorem D.6 and proof). To the best of our knowledge, this is the only published unbiased mechanism for means when the sample size is not treated as known. Both mechanisms use the Laplace mechanism with privacy budget ϵ1 to obtain the noisy sample size n~. Each mechanism then allocates a separate privacy budget ϵ2 to obtain a noisy mean. Both mechanisms have a total privacy budget of ϵ0=ϵ1+ϵ2.

Denote attribute c of record r by r.c and let g(q~;k,L) be the unbiased estimator of 1/q from Section 4.1 with the generic query q and polynomial extension of order k for qL. Algorithm 1 lays out MU. This algorithm applies the Laplace mechanism to the sum srDr.c and forms an unbiased estimate m~U of the mean m=s/n by multiplying the noisy sum s~ by g(n~;k,L). This is unbiased for m as long as nL.

For nL, the variance of Algorithm 1 is

𝕍[m~U] =(𝔼[s~]2+𝕍[s~])(𝔼[g(n~;k,L)]2+𝕍[g(n~;k,L)])𝔼[s~]2𝔼[g(n~;K,L)]2 (15)
=(s2+2/ϵ12)(𝔼[1/n]2+𝕍[g(n~;k,L)])s2/n2. (16)

For this section’s numerical results, we calculate 𝕍[g(n~;k,L)] numerically. Finally, we note that it is straightforward to show that all moments of m~U exist and are finite.

Algorithm 1 MU

Let T3 denote a random variable distributed according to a standard t distribution with 3 degrees of freedom. MSS is laid out in Algorithm 2. This algorithm first forms a version mSS of the mean query that simply equals 1 if n=0. It then scales the noise variable T3 in proportion to an upper bound on the query’s smooth sensitivity [12]. The final noisy mean m~SS is obtained by simply adding the scaled noise variable to mSS.

The scaling factor for the noise is τmax(eβ(n1),1/max(n,1)), where τ and β satisfy ϵ2=4β+2/(3τ). The standard t distribution with d degrees of freedom has variance d/(d2) giving m~SS a variance of

𝕍[mSS]=3τ2max(eβ(n1),1/max(n,1))2. (17)

Because the t distribution is symmetric, this mechanism is unbiased for s/n as long as n1. Unlike m~U, however, the third and higher moments of m~SS are infinite or do not exist. This is due to the t distribution’s very fat tails and implies that m~SS is more liable than m~U to produce extreme outliers.

Algorithm 2 MSS.

In our numerical evaluation of these mechanisms, we fix ϵ1=ϵ2=m=.5 for both mechanisms. For MSS, we follow [12] in setting β=ϵ2/12 and τ=3/ϵ2. For MU, we set L=1 so that both mechanisms are unbiased for n1 and set k=10.

With these settings, we compare the standard deviations (SDs) of the two mechanisms’ mean estimates for a range of sample sizes. Because the mechanisms are unbiased, this is equivalent to their root mean squared error. The sample size estimates of both mechanisms are the same, so we do not report their properties.

Figures 1 and 2 plot the mechanisms’ SDs as functions of n, with Figure 2 zooming in on larger values of n for clarity. It is immediately clear that the MU has a larger SD than MSS for n<13, and that this pattern reverses for larger n. For n19, however, both mechanisms have SDs greater than 1, making both unfit for most purposes at these sample sizes, given that the mean m has the domain [0,1].

Figure 3 shows the relative SD - that is, the ratio SD(MSS)/SD(MU) - as a function of n. For n13, the relative SD rises to a peak near 15 before settling down to an apparent constant of about 1.9 for n115.

Ultimately, MU appears to be the better mechanism for this setting; for any sample size where either mechanism returns useful results, MU has a substantially lower SD. The thinner tails of MU also recommend it as the better choice.

Refer to caption
Figure 1: Standard deviations of the mechanisms MSS and MU for a mean of n records in [0,1]. The mean is fixed at .5 and the mechanisms have a privacy budget of ϵ2=.5.
Refer to caption
Figure 2: Same as Figure 1, zoomed in to larger values of n (see the horizontal axis endpoints). Standard deviations of the mechanisms MSS and MU for a mean of n records in [0,1]. The mean is fixed at .5 and the mechanisms have a privacy budget of ϵ2=.5.
Refer to caption
Figure 3: Standard deviation of MSS divided by the standard deviation of MU for a mean of n records in [0,1]. The mean is fixed at .5 and the mechanisms have a privacy budget of ϵ2=.5.

6 Application to Slowly Scaling PRDP

In this section, we use our estimators to develop versions of unbiased privacy mechanisms from [8] that enjoy stronger privacy guarantees. Our estimators allow us to do so while maintaining the mechanisms’ unbiasedness.

(ϵ,0)-DP guarantees the upper bound ϵ on the privacy loss between any pair of neighboring databases. [8] develops mechanisms for a related privacy concept, per-record DP (PRDP) [13]. PRDP generalizes (ϵ,0)-DP by letting the privacy loss bound be a function of the record on which a given pair of neighbors differs. Semantically, this allows different records to have different levels of protection.

Denote a record in the database D by r and denote r’s attribute c by r.c[0,). PRDP was originally motivated by the need to protect data used to compute the sum query q(D)=r.cDr.c. Because the domain of c is unbounded, this sum can change by an arbitrarily large amount when a record is added or deleted. That is, the sum’s global sensitivity is infinite. This prevents commonly used privacy mechanisms, such as the Laplace mechanism, from providing a differential privacy guarantee with finite ϵ.

The traditional fix for this is to clip attribute c to lie in a bounded set before taking the sum. (ϵ,0)-DP can then be guaranteed by perturbing the sum with noise scaled in proportion to the width of the clipped data’s domain. Unfortunately, when the sum is dominated by a small number of large outliers, the outliers typically need to be clipped to drastically smaller values to preserve a reasonable balance of privacy loss and noise variance. This can induce catastrophic bias, rendering the clipped sums essentially useless. One might expect to see this type of behavior with income data, for example.

PRDP allows us to take a finer look at the privacy-utility tradeoff by recognizing that, even though outliers may suffer extreme privacy loss, the rest of the dataset may still enjoy strong privacy protections. Intuitively, a particular record’s privacy loss is proportional only to the amount by which the addition or deletion of that record can change the query. Queries may be highly sensitive to the presence of outliers while being relatively insensitive to typical records, leading different records to have different levels of privacy loss. The reassurance that the vast majority of the data may enjoy strong privacy guarantees whether or not the data is clipped may allow a data curator to reasonably decide against clipping if the resulting bias outweighs the enhanced privacy protection for a small number of records.

Below, we define PRDP.

Definition 17 (P-Per-Record Differential Privacy (P-PRDP) [13, 8]).

Let denote the symmetric set difference. The mechanism M satisfies per-record differential privacy with the policy function P (P-PRDP) if, for any record r; any pair of neighboring databases D,D such that DD={r}; and any measurable set of possible outcomes S, we have

Pr[(D)S]<eP(r)Pr[(D)S].

Ensuring strong privacy protection corresponds to ensuring that P is, in some sense, small. (ϵ,0)-DP is recovered by making the constant privacy guarantee P(r)=ϵ for all r, and strong privacy protection follows from a small ϵ. We cannot always make a guarantee this strong. Take the example where we want to publish a sum query on an unbounded attribute c that we are unwilling to clip. In this case, the privacy loss of the mechanisms that we will consider here is growing in r.c. Even though we cannot prevent P(r) from growing without bound in r.c, we can use mechanisms for which the growth rate is slow. [8] call such mechanisms “slowly scaling.” A slowly growing P narrows the gap in privacy losses between records with large and small values of c, letting a data curator more easily provide a desired level of protection for the bulk of the data without compromising too much on the privacy of outliers.

[8] introduces slowly scaling mechanisms, called transformation mechanisms, that work by adding Gaussian noise to a concave transformation f of the query (plus offset term) q(D)+a and then feeding the noisy value of f(q(D)+a) to an estimator g of q(D). By adding Gaussian noise, these mechanisms satisfy per-record zero-concentrated DP (PRzCDP), which is a weaker privacy guarantee than PRDP. PRzCDP relates to zero-concentrated DP [4, 7] in the same way that PRDP relates to ϵ-DP. The use of Gaussian noise also allowed [8] to draw on existing unbiased estimators from [16] to make their mechanism unbiased for a variety of transformation functions f.

Using the unbiased estimators from Theorem 10, we strengthen the transformation mechanisms to provide PRDP guarantees by adding Laplace, rather than Gaussian, noise, and we do so without losing the mechanism’s unbiasedness. Algorithm 3 lays out our transformation mechanism.

Algorithm 3 PRDP Transformation Mechanism.

To obtain the PRDP guarantee of Algorithm 3, we first need to define the per-record sensitivity [8], a record-specific analog of the global sensitivity.

Definition 18 (Per-Record Sensitivity [8]).

The per-record sensitivity of the univariate, real-valued query q for record r is

Δ(r)supD,D such that DD={r}|q(D)q(D)|.

Theorem 19 gives our most generic result on the PRDP guarantees of Algorithm 3. Theorems 19 and 20, their proofs, as well as Algorithm 3 are minimally modified from their analogs in [8], which use Gaussian, rather than Laplace noise. This is to facilitate the interested reader’s comparison of our results with theirs.

Theorem 19 (PRDP Guarantee for Transformation Mechanisms).

Assume the query value q(D)[0,); the offset parameter a; the noise scale parameter b(0,); the transformation function f:[a,) is concave and strictly increasing; and the estimator g:𝒢. Denote by Δf(r) the per-record sensitivity of the query f(q(D)+a), as defined in Definition 18. Algorithm 3(q(D),a,b,f,g) satisfies P-PRDP for P(r)=Δf(r)b.

See Appendix D of this paper’s full version (linked to on the title page) for the proof.

Probably the most common query encountered in applications of formal privacy, even as a component of other, larger queries, is the sum query. We now use the above result to derive the policy function for the transformation mechanism applied to a sum query.

Theorem 20 (Privacy of Transformation Mechanisms for Sum Query).

Let the assumptions of Theorem 19 hold, and further assume a0 and r.c0 for all records r. For the sum query q(D)=rDr.c, the per-record sensitivity of f(q(D)+a) is Δf(r)=f(r.c+a)f(a) and Algorithm 3(q(D),a,b,f,g) satisfies PRDP with the policy function P(r)=[f(r.c+a)f(a)]/b.

See Appendix D for the proof.

Critically, the policy function from Theorem 20 grows in r.c more slowly when the transformation function f grows more slowly. In the case where a=0 and f(x)=xk for some k1, the policy function is simply r.ck. Choosing larger values of k, then, forces the privacy loss to grow more slowly in r.c, reducing the gap in privacy losses between records with large and small values of c.

Applying our main results, we can obtain estimators such that the transformation mechanism gives us an unbiased estimate of q(D). In particular, polynomials555In the case of polynomials, [11] derived unbiased estimators which could also be used here. The estimator obtained from our Theorem 10 merely simplifies computation in this setting. are twice differentiable functions which are tempered distributions, so the following holds for f(x)=xk:

Corollary 21.

Given any function f such that f1 satisfies the conditions in Theorem 10, a, b(0,), estimator g:𝒢, and r.c0 for all records r, there exists an unbiased estimator for q(D) satisfying P-PRDP for P(r)=[f(r.c+a)f(a)]/b.

Proof.

The conditions for Theorem 20 hold, and Theorem 10 gives us a function g such that 𝔼[g(v~)]=q(D)+a is an unbiased estimator for f1. Therefore, g(v~)a is an unbiased estimator for q(D).

7 Polynomial Functions under General Noise Distributions

Additive mechanisms other than the Laplace mechanism, such as the discrete Gaussian or the staircase mechanisms, may be preferable in practice due to achieving higher accuracy while having similar privacy loss [5, 10]. In contrast to the Laplace case, these mechanisms may not admit tractable Fourier transforms, and hence unbiased estimators are generally not available in closed form. One exception is when the query of interest is a polynomial in one or many queries.

Using the following results to obtain an unbiased estimator of a polynomial that approximates a non-polynomial estimand may also allow users to obtain approximately unbiased estimators with great generality.

Theorem 22.

Suppose a mechanism takes as input q and outputs q~=q+Z for a random variable Z with at least p finite, publicly known moments. If f(q) is a polynomial in q of degree at most p, there exists an unbiased estimator g(q~) of f(q), which is itself a polynomial of degree at most p and is available in closed form.

Proof.

Suppose f(q)=n=0pbnqn. Let us find an unbiased estimator of the form g(q~)=n=0panq~p. We have

g(q+z)=n=0pank=0n(nk)qkznk. (18)

Denote μr=𝔼[zr] and take expectations to obtain

𝔼[g(q+z)] =n=0pank=0n(nk)qkμnk (19)
=n=0p(n0)anμn+(n=1p(n1)anμn1)q+(n=2p(n2)anμn2)q2+ (20)

For this polynomial to be equal to f, we need a=(a0,,ap) to solve Ma=b, where b=(b0,,bp) and

M=(μ0μ1μ2μ3μp0(11)μ0(21)μ1(31)μ2(p1)μp100(22)μ0(32)μ1(p2)μp2000(33)μ0(p3)μp30000(pp)μ0). (21)

Clearly, M is nondegenerate, and so the desired coefficients a exist and are unique.

We now extend this result to polynomials in multiple (univariate) queries, assuming that the noise variables added to each query are independent. The latter assumption, while seemingly restrictive, is typical for additive noise mechanisms in differential privacy.

Theorem 23.

Suppose a mechanism takes as input (q1,,qm) and outputs (q~1,,q~m)=(q1+Z1,q2+Z2,,qm+Zm) for independent random variables Z1,,Zm with finite, publicly known moments. If f(q1,,qm) is a polynomial in (q1,,qm), there exists an unbiased estimator g(q~1,,q~m) of f(q), which is itself a polynomial available in closed form.

Proof.

Clearly, it suffices to derive unbiased estimators for f(q1,,qm)=i=1mqipi. Let gi(q~i) be the unbiased estimator of qipi as in Theorem 22 and set g(q~1,,q~m)=i=1mgi(q~i). Since g1(q~1),,gm(q~m) are independent random variables, we have

𝔼[g(q~1,q~m)]=𝔼[i=1mgi(q~i)]=i=1m𝔼[gi(q~i)]=i=1mqipi=f(q1,,qm). (22)

8 Conclusions and Future Work

In this work, we have shown how to compute unbiased estimators of twice-differentiable tempered distributions evaluated on privatized statistics with added Laplace noise. In addition, we have proposed a method to extend this result to twice-differentiable functions which are not tempered distributions in a way that achieves approximately optimal expected squared error.

As the Laplace mechanism is simple and commonly used, these results are widely applicable to obtain unbiased statistics for free in postprocessing, which is particularly valuable due to the fact that aggregating unbiased statistics accumulates error more slowly than aggregating biased statistics. We have applied our results to derive a competitive unbiased algorithm for means and to derive unbiased transformation mechanisms for per-record DP mechanisms that enjoy stronger privacy protection than do analogs in previous work. Finally, we have derived an unbiased estimator for polynomials under arbitrary noise distributions with known moments, such as the discrete Gaussian mechanism or the staircase mechanism [5, 10].

We believe this paper opens several avenues for future research. These include the use of the deconvolution method to obtain unbiased estimators for other estimands and noise distributions. We believe a deconvolution method using multivariate Fourier transforms could also be used to obtain unbiased estimators of functions of multivariate queries. Although we did not attempt to optimize the numerical implementation in Section 5 of the integration in Section 4, we believe that an improved implementation could enable the practical use of higher-order polynomial extensions and further reduce error. In Section 7, we developed estimators that are exactly unbiased for polynomials that could approximate other functions of interest. Further work could elaborate on this process, developing concrete procedures for picking the approximating polynomial and deriving bounds on the resulting bias. Finally, future work could attempt to derive noise distributions that are optimal in the sense of minimizing the variances (or other utility metrics) of their unbiased estimators.

References

  • [1] John Abowd, Robert Ashmead, Ryan Cumings-Menon, Simson Garfinkel, Micah Heineck, Christine Heiss, Robert Johns, Daniel Kifer, Philip Leclerc, Ashwin Machanavajjhala, Brett Moran, William Sexton, Matthew Spence, and Pavel Zhuravlev. The 2020 Census Disclosure Avoidance System TopDown Algorithm. Harvard Data Science Review, (Special Issue 2), June 24 2022. URL: https://hdsr.mitpress.mit.edu/pub/7evz361i/release/2.
  • [2] Daniel Alabi, Audra McMillan, Jayshree Sarathy, Adam Smith, and Salil Vadhan. Differentially private simple linear regression, 2020. arXiv:2007.05157.
  • [3] Margaret Beckom, William Sexton, and Anthony Caruso. Researching formal privacy for the Census Bureau’s County Business Patterns program, 2023. URL: https://www.census.gov/data/academy/webinars/2023/differential-privacy-webinar.html.
  • [4] Mark Bun and Thomas Steinke. Concentrated differential privacy: Simplifications, extensions, and lower bounds. In Theory of Cryptography Conference, pages 635–658. Springer, 2016. doi:10.1007/978-3-662-53641-4_24.
  • [5] Clément L. Canonne, Gautam Kamath, and Thomas Steinke. The discrete Gaussian for differential privacy. In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS ’20, Red Hook, NY, USA, 2020. Curran Associates Inc.
  • [6] Cynthia Dwork, Frank McSherry, Kobbi Nissim, and Adam Smith. Calibrating noise to sensitivity in private data analysis. Journal of Privacy and Confidentiality, 7(3):17–51, May 2017. doi:10.29012/jpc.v7i3.405.
  • [7] Cynthia Dwork and Guy N Rothblum. Concentrated differential privacy. arXiv preprint arXiv:1603.01887, 2016.
  • [8] Brian Finley, Anthony M Caruso, Justin C Doty, Ashwin Machanavajjhala, Mikaela R Meyer, David Pujol, William Sexton, and Zachary Terner. Slowly scaling per-record differential privacy. arXiv preprint arXiv:2409.18118, 2024. doi:10.48550/arXiv.2409.18118.
  • [9] Jean Baptiste Joseph Fourier. The Analytical Theory of Heat. Cambridge Library Collection - Mathematics. Cambridge University Press, 2009.
  • [10] Quan Geng, Peter Kairouz, Sewoong Oh, and Pramod Viswanath. The staircase mechanism in differential privacy. IEEE Journal of Selected Topics in Signal Processing, 9(7):1176–1184, 2015. doi:10.1109/JSTSP.2015.2425831.
  • [11] Quentin Hillebrand, Vorapong Suppakitpaisarn, and Tetsuo Shibuya. Unbiased locally private estimator for polynomials of Laplacian variables. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 741–751, 2023. doi:10.1145/3580305.3599537.
  • [12] Gautam Kamath, Argyris Mouzakis, Matthew Regehr, Vikrant Singhal, Thomas Steinke, and Jonathan Ullman. A bias-variance-privacy trilemma for statistical estimation, 2023. doi:10.48550/arXiv.2301.13334.
  • [13] Jeremy Seeman, William Sexton, David Pujol, and Ashwin Machanavajjhala. Privately answering queries on skewed data via per-record differential privacy. Proc. VLDB Endow., 17(11):3138–3150, 2024. doi:10.14778/3681954.3681989.
  • [14] Gerrit van Dijk. Distribution Theory. De Gruyter, Berlin, Boston, 2013. doi:doi:10.1515/9783110298512.
  • [15] V.G. Voinov and M. Nikulin. Unbiased Estimators and Their Applications Volume 1: Univariate Case. Kluwer Academic Publishers, Dordrecht, 1993.
  • [16] Yasutoshi Washio, Haruki Morimoto, and Nobuyuki Ikeda. Unbiased estimation based on sufficient statistics. Bulletin of Mathematical Statistics, 6:69–93, 1956. URL: https://api.semanticscholar.org/CorpusID:55591271.