Approximating Klee’s Measure Problem and a Lower Bound for Union Volume Estimation
Abstract
Union volume estimation is a classical algorithmic problem. Given a family of objects , we want to approximate the volume of their union. In the special case where all objects are boxes (also called hyperrectangles) this is known as Klee’s measure problem. The state-of-the-art -approximation algorithm [Karp, Luby, Madras ’89] for union volume estimation as well as Klee’s measure problem in constant dimension uses a total of queries of three types: (i) determine the volume of ; (ii) sample a point uniformly at random from ; and (iii) ask whether a given point is contained in .
First, we show that if an algorithm learns about the objects only through these types of queries, then queries are necessary. In this sense, the complexity of [Karp, Luby, Madras ’89] is optimal. Our lower bound holds even if the objects are equiponderous axis-aligned polygons in , if the containment query allows arbitrary (not necessarily sampled) points, and if the algorithm can spend arbitrary time and space examining the query responses.
Second, we provide a more efficient approximation algorithm for Klee’s measure problem, which improves the running time from to . We circumvent our lower bound by exploiting the geometry of boxes in various ways: (1) We sort the boxes into classes of similar shapes after inspecting their corner coordinates. (2) With orthogonal range searching, we show how to sample points from the union of boxes in each class, and how to merge samples from different classes. (3) We bound the amount of wasted work by arguing that most pairs of classes have a small intersection.
Keywords and phrases:
approximation, volume of union, union of objects, query complexityFunding:
Kasper Green Larsen: Supported by a DFF Sapere Aude Research Leader Grant No. 9064-00068B.Copyright and License:
![[Uncaptioned image]](x1.png)
Yanheng Wang; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Computational geometryFunding:
Karl Bringmann and Yanheng Wang: This work is part of the project TIPEA that has received funding from the European Research Council (ERC) under the European Unions Horizon 2020 research and innovation programme (grant agreement No. 850979).Editors:
Oswin Aichholzer and Haitao WangSeries and Publisher:

1 Introduction
We revisit the classical problem of union volume estimation: given objects , we want to estimate the volume of .111Technically, the objects need to be measurable. In the most general form, can be any measurable sets in a measure space, and we want to estimate the measure of their union. However, this work only deals with boxes in (in our algorithm) and polygons in the plane (in our lower bound). This problem has several important applications such as DNF counting and network reliability; see the discussion in Section 1.2.
The state-of-the-art solution [19] works in a model where one has access to each input object by three types of queries: (i) determine the volume of the object, (ii) sample a point uniformly at random from the object, and (iii) ask whether a point is contained in the object. Apart from these types of queries, the model allows arbitrary computations. The complexity of algorithms is thus measured by the number of queries.
After Karp and Luby [18] introduced this model, Karp, Luby and Madras [19] gave an algorithm that -approximates the volume of objects in this model with constant success probability.222The success probability can be boosted to , adding a factor in time and query complexity. It uses queries plus additional time, which improves earlier algorithms [18, 22], and only asks containment queries of previously sampled points. The last 35 years have seen no algorithmic improvement. Is this classical upper bound best possible? We resolve the question in this work by providing a matching lower bound.
The union volume estimation problem was also studied recently in the streaming setting [25, 23], where a stream of objects arrive in order. When we are at position in the stream, we can only query object . This line of work yields algorithms with constant success probability that use queries and additional time per object (and the same bound also holds for space complexity). Summed over all objects, the bounds match the classical algorithm up to the factor. So, interestingly, even in the streaming setting the same upper bound can be achieved.333See also [30] for earlier work studying Klee’s measure problem in the streaming setting.
The perhaps most famous application of the algorithm by Karp, Luby, and Madras [19] is Klee’s measure problem [21]. This is a fundamental problem in computational geometry in which we are given axis-aligned boxes in and want to compute the volume of their union. An axis-aligned box is a set in the form , and the input consists of the corner coordinates of each box. A long line of research on this problem and various special cases (e.g., for fixed dimensions or for cubes) [31, 26, 11, 2, 1, 12, 32, 3] lead to an exact algorithm running in time for constant [13]. A conditional lower bound suggests that any faster algorithm would require fast matrix multiplication techniques [12], but it is unclear how to apply fast matrix multiplication to this problem. On the approximation side, note that the three queries can be implemented in time for any -dimensional axis-aligned box. Thus the union volume estimation algorithm can be applied, and it computes a -approximation to Klee’s measure problem in time , as has been observed in [4]. This direct application of union volume estimation was the state of the art for approximate solutions for Klee’s measure problem until our work. See Section 1.2 for interesting applications of Klee’s measure problem.
1.1 Our contribution
Lower bound for union volume estimation
Given the state of the art, a natural question is to ask whether the query complexity of the general union volume estimation algorithm of [19] can be further improved. Any such improvement would speed up several important applications, cf. Section 1.2. On the other hand, any lower bound showing that the algorithm of [19] is optimal also implies tightness of the known streaming algorithms (up to logarithmic factors), as the streaming algorithms match the static running time bound.
Previously, a folklore lower bound of was known. But between this and the upper bound there is still a large gap; consider for example the regime where the bounds are and , respectively. We close this gap by strengthening the lower bound to , thereby also showing optimality of [19]. Note that the lower bound is about query complexity in the model; it does not make any computational assumption. In particular, it holds even if the algorithm spends arbitrary time and space in computation.
Theorem 1.
Any algorithm that computes a -approximation to the volume of the union of objects via volume, sampling and containment queries with success probability at least must make queries.
We highlight that our lower bound holds even for equiponderous, axis-aligned polygons in the plane.
Upper bound for Klee’s measure problem
Our lower bound for union volume estimation implies that any improved algorithm for Klee’s measure problem must exploit the geometric structure of boxes. To this end, we exploit our knowledge of the corner coordinates and classify the boxes by shapes. In addition, we apply the geometric tool of orthogonal range searching. They allow us to approximate Klee’s measure problem faster than the previous algorithm. Here and throughout, we assume that the dimension is a constant; in particular, we suppress factors of in the time complexity.
Theorem 2.
There is an algorithm that runs in time and with probability at least computes a -approximation for Klee’s measure problem.
This is a strict improvement when is small. Consider for example the regime : Our algorithm runs in near-linear time, whereas the previous algorithm runs in quadratic time. We remark that the success probability can be boosted to any by standard techniques which incurs an extra factor in the running time. Finally, we highlight that the core of our algorithm is an efficient method to sample uniformly and independently with a given density from the union of the input boxes. This can be of independent interest outside the context of volume estimation.
1.2 Related work
A major application of union volume estimation is DNF counting, in which we are given a Boolean formula in disjunctive normal form and want to count the number of satisfying assignments. Computing the exact number of satisfying assignments is #P-complete, so it likely requires superpolynomial time. Approximating the number of satisfying assignments can be achieved by a direct application of union volume estimation, as described in [19]. Their algorithm remains the state of the art for approximate DNF counting, see e.g. [24]. This has been extended to more general model counting [27, 9, 24], probabilistic databases [20, 15, 28], and probabilistic queries on databases [7].
We also mention network reliability as another application for union volume estimation, which was already discussed in [19]. Additionally, Karger’s famous paper on the problem [17] uses the algorithm of [19] as a subroutine. However, the current state-of-the-art algorithms no longer use union volume estimation as a tool [8].
Finally, we want to draw a connection to the following well-known query sampling bound. Canetti, Even, and Goldreich [6] showed that approximating the mean of a random variable whose codomain is the unit interval requires queries, thus obtaining tight bounds for the sampling complexity of the mean estimation problem. Their bound generalize to on the number of queries needed to estimate the mean of a random variable in general. Before our work it was thus natural to expect that the dependence in the number of queries for union volume estimation is optimal. However, whether the factor is necessary, or the number of queries could be improved to, say, , was open to the best of our knowledge.
Klee’s measure problem is an important problem in computational geometry. One reason for its significance is that techniques developed for it can often be adapted to solve various related problems, such as the depth problem (given a set of boxes, what is the largest number of boxes that can be stabbed by a single point?) [13] or Hausdorff distance under translation in [14]. Moreover, various other problems can be reduced to Klee’s measure problem or to its related problems. For example, deciding whether a set of boxes covers its boundary box can be reduced to Klee’s measure problem [13]. The continuous -center problem on graphs (i.e., finding centers that can lie on the edges of a graph that cover the vertices of a graph) is reducible to Klee’s measure problem as well [29]. Finding the smallest hypercube containing at least points among given points can be reduced to the depth problem [16, 10, 13]. In light of this, it would be interesting to see if our approximation techniques generalize to any of these related problems.
2 Technical overview
We now sketch our upper and lower bounds at an intuitive level. The formal arguments will be presented in Section 3 and Section 4, respectively. In this paper we denote .
2.1 Upper bound for Klee’s measure problem
Due to our lower bound, we have to exploit the structure of the input boxes to obtain a running time of the form . We follow the common algorithmic approach of sampling. Specifically, we aim to sample a set from the union of boxes such that each point is selected with probability density . For appropriately defined , the value is a good estimate of the volume of the union. In the overview we assume that a proper is given and focus on the main difficulty: creating a sample set for the given .
We start by sorting the input boxes into classes by shape. Two boxes are in the same class if, in each dimension , their side lengths are both in for some . We call two classes similar if the corresponding side lengths are polynomially related (e.g., within a factor of ) in each dimension. We make two crucial observations:
-
1.
Each class is similar to only polylogarithmically many classes (Observation 11).
- 2.
Our algorithm continues as follows. We scan through the classes in an arbitrary order. For each class, we sample with density from the union of the boxes in this class, but we only keep a point if it is not contained in any class that comes later in the order. To efficiently test containment in a later class, we use an orthogonal range searching data structure (with an additional dimension for the index of the class). When the scan finishes, we obtain a desired sample set .
Let us elaborate why the algorithm is efficient.
- Sampling from a single class.
-
Our approach is simple yet powerful: (i) grid the space into cells of side lengths comparable to the boxes in this class; (ii) sample points from the relevant cells uniformly at random; (iii) discard points outside the union by querying an orthogonal range searching data structure. See Figure 2. All these steps exploit the geometry of boxes. Since the grid cells and boxes have roughly the same shape, a significant fraction of the points sampled in (ii) are contained in the union, i.e., not discarded in (iii). The orthogonal range searching data structure allows us to quickly decide which points to discard. These explain the efficiency of sampling from a class.
- Bounding the amount of wasted work.
-
Sampled points that appear in later classes also need to be discarded, and this is a potential (and only) source of inefficiency. So we need to bound the number of them. Such a point shows up later either (i) in a similar class, or (ii) in a dissimilar class. Our first observation stated that a class is similar to at most polylogarithmically many classes. So (i) can happen up to polylogarithmically many times from the perspective of a fixed point in the union. On the other hand, our second observation stated that the intersection of dissimilar classes is small. So on average (ii) rarely happens and does not affect the expected running time significantly.
2.2 Lower bound for union volume estimation
Our lower bound is shown by a reduction from the Query-Gap-Hamming problem: Given two vectors , distinguish whether their inner product is greater than or less than . It is known that any algorithm distinguishing these two cases with success probability at least must access bits of and .
We first sketch why queries are necessary to -approximate the volume of the union of two objects in the query model. Given a Query-Gap-Hamming instance , we construct two objects and . These are discrete point sets, but they can be inflated to polygons such that cardinality corresponds to volume. See Figure 3 for an example. Note that for all , we have
If an algorithm can -approximate , then it can approximate within additive error , thus it can also approximate within additive error . Setting , the additive error is at most , which suffices to distinguish from and thereby solves the Query-Gap-Hamming instance.
Note that , so a volume query does not disclose any information about and . Each sample or containment query accesses at most one bit of or . Therefore, algorithm has to make queries to .
In order to generalize this lower bound for the union of two objects to an lower bound for the union of objects, we need to ensure that each query gives away only bits of information about and . We apply two obfuscations that jointly slow down the exposure of bits; see Figure 4. Firstly, we introduce objects whose union is and objects whose union is . Imagine cutting each -induced rectangle in Figure 3 into side-by-side pieces and distributing them randomly among . Do the same for . The idea is that one needs to make containment queries on the rectangular region in order to hit the correct piece; only then is the corresponding bit revealed. Secondly, we introduce a large band shared by all and for . In Figure 4, this is the long dark-blue rectangle that spans from left to right. Intuitively it enforces sample queries to obtain a single point that contains any information about and .
3 Approximation algorithm for Klee’s measure problem
In this section, we present our approximation algorithm for Klee’s measure problem in constant dimension : See 2
Lemmas marked with are folklore or straightforward. Their proofs can be found in the full version of this paper [5].
3.1 Preliminaries
In Klee’s measure problem we are given boxes . Here, a box is an object of the form , and as input we are given the coordinates of each box. Based on these, it is easy to compute the side lengths and volume of each box.
Throughout we write for the volume of the union of boxes. The goal is to approximate up to a factor of . Our approach is based on sampling, so now let us introduce the relevant notions.
Recall the Poisson distribution with mean and variance : It captures the number of active points in a universe, under the assumption that active points occur uniformly and independently at random across the universe, and that points are active on average. The following definition is usually referred to as a homogeneous Poisson point process at rate . Intuitively, we activate each point in some universe independently with “probability density” , thus the number of activated points follows the Poisson distribution with mean .
Definition 3 (-sample).
Let be a measurable set, and let . We say that a random subset is a -sample of if for any measurable we have that .
In particular, if is a -sample of , then . Two more useful properties follow from the definition:
-
(i)
For any measurable subset , the restriction is a -sample of .
-
(ii)
The union of -samples of two disjoint sets is a -sample of .
Besides sampling, we will apply orthogonal range searching to handle the so-called query: Given and , is ?
Lemma 4 ().
We can build a data structure in time that answers queries in time.
3.2 Classifying boxes by shapes
As our first step in the algorithm, we classify boxes by their shapes.
Definition 6.
Let . We say that a box is of type if its side length in dimension is contained in , for all .
Using this definition, we partition the input boxes into classes such that each class corresponds to one type of boxes. The notation is fixed from now on. For each , we denote , namely the union of boxes in class .
Similar to appears, we may use orthogonal range searching to handle the so-called query: Is a given point contained in ?
Lemma 7 ().
We can build a data structure in time that answers queries in time.
Sampling from a class
The next lemma shows that we can -sample from any efficiently by rejection sampling.
Lemma 8.
Given , and the data structure from Lemma 7, one can generate a -sample of in expected time .
Proof.
Let be the type of class . We subdivide into the grid
We call each element of a cell. Let be the set of cells that intersect with . Let .
First we create a -sample of as follows. Generate , the number of points we are going to include. Then sample points uniformly at random from by repeating the following step times: Select a cell uniformly at random and then sample a point from uniformly at random. The sampled points constitute our set .
Next we compute : For each , we query ; if the answer is true then we keep , otherwise we discard it. The resulting set is a -sample of , since restricting to a fixed subset preserves the -sample property.
Before we analyze the running time, we show that makes up a decent proportion of . Recall that every box in class is of type . Consider the projection to any dimension . Each projected box from can intersect at most three projected cells from . So each box from intersects at most cells from , implying that . Moreover, since the volume of any cell is at most the volume of a box in , we have .
Recall that we assume to be constant and hence . The computation of takes time. The remaining time is dominated by the inClass queries. The expected size of is . As we query inClass once for each point of , the total expected time is by Lemma 7.
Classes do not overlap much
We show the following interesting property of classes, that the sum of their volumes is within a polylogarithmic factor of the total volume .
Lemma 9.
We have .
We later exploit it to draw -samples from efficiently. Towards proving the lemma, let us collect some simple observations.
Definition 10.
We say that a class of type is similar to a class of type if for all . Otherwise they are said to be dissimilar.
Observation 11.
Every class is similar to at most classes.
Proof.
Fix a type . For each , there are at most many integers such that .
Observation 12.
If two boxes are in dissimilar classes, then .
Proof.
Let be the type of , and be the type of . Since the boxes belong to dissimilar classes, there is a dimension such that . Without loss of generality, assume ; the other case is symmetric. Let and be the projections of and onto dimension , respectively. Note that and . So we have . In other words, at most a fraction of the interval intersects the interval . Hence,
Proof of Lemma 9.
Without loss of generality assume . We construct a set of indices by the following procedure:
-
Initialize .
-
For , if and are dissimilar for all , then add to .
For , we have only if there exists an such that are similar and ; we call a witness of . If multiple witnesses exist, then we pick an arbitrary one. Conversely, every can be a witness at most times by Observation 11. Hence,
(1) |
3.3 Joining the classes
Recall that are the classes of the input boxes and their respective unions. Assume without loss of generality that the boxes are ordered in accordance with the class ordering, that is, form the first class, form the second class, and so on. In general, we assume that for all , where .
Let be the points in that are not contained in later classes. Note that is a partition of . Hence, to generate a -sample of , it suffices to draw -samples from each and then take their union.444This idea has previously been used on objects, by considering the difference [18, 25], while we use this idea on classes. To this end, we draw a -sample from via Lemma 8. Then we remove all for which ; these are exactly the points that appear in a later class. What remains is a -sample of . The union of these sets thus is a -sample of , and we can use the size of this -sample to estimate the volume of . The complete algorithm is summarized in Algorithm 1.
Lemma 13.
Conditioned on , Algorithm 1 outputs a -approximation to with probability at least .
Proof.
Note that for all , the set is a -sample of . Since partition , their union is a -sample of . It follows that . The expectation and variance of are . So by Chebyshev,
Hence, with probability at least , the output is a -approximation to .
Lemma 14.
Conditioned on , Algorithm 1 runs in time in expectation.
Proof.
Step 1 takes time: we can compute the side lengths of each box, determine its class, then sort the boxes by class index. Step 2 takes time by Lemmas 4 and 7. Step 3 takes time by Lemma 5.
In step 4, iteration , sampling costs expected time by Lemma 8, and computing costs expected time by Lemma 4. Over all iterations, the expected running time is
Substituting and applying Lemma 9, we can bound
Hence the expected running time of step 4 is .
Finally, step 5 takes time.
Proof of Theorem 2.
Run Algorithm 1 with a time budget tenfold the bound in Lemma 14; abort the algorithm as soon as it spends more time than the budget allows. So the stated time bound is clearly satisfied. Now consider three bad events:
-
.
-
, but the algorithm is aborted.
-
and the algorithm is not aborted, but it does not output a -approximation to .
By Lemma 5, the first event happens with probability at most . By Lemma 14 and Markov’s inequality, the second event happens with probability at most . Lastly, by Lemma 13, the third event happens with probability at most . So the total error probability is at most . If none of the bad events happen, then the algorithm correctly outputs a -approximation to . The success probability of can be boosted to, say, by returning the median of a sufficiently large constant number of repetitions.
4 Lower bound for union volume estimation
In this section, any randomized algorithm or protocol is assumed to have success probability at least . We consider the discrete version of union volume estimation in two dimensions, where each object is a finite subset of the integer lattice . The goal is to -approximate the cardinality of their union. The algorithm does not have direct access to the objects, but it can ask three forms of queries:
-
: Return the cardinality .
-
: Draw a uniform random point from .
-
: Given a point , return whether or not.
We aim to prove that queries are necessary. Later in Section 5 we translate this to a lower bound in the continuous space , thereby proving Theorem 1.
Our starting point is the Query-Gap-Hamming problem: The inputs are two (hidden) vectors and we can access one bit of or of our choice at a time. The goal is to distinguish the cases and using as few accesses as possible. The following lemma is folklore; its proof can be found in the full version of this paper [5].
Lemma 15 ().
There exists a constant with the following property. For every , any randomized algorithm for Query-Gap-Hamming on vectors of length needs at least accesses in expectation.
From now on, let us fix the constant guaranteed by Lemma 15. Let , be arbitrary and define . We reduce Query-Gap-Hamming on vectors of length to estimating the cardinality of the union of objects in .
The reduction
From the hidden vectors we construct objects . Let . Take independent random permutations of and define
for . Then take another set of independent random permutations and define
for . Note that is a subset of all and .
How is related to the cardinality of the union ? Consider an arbitrary index . If then the point sets and are equal (regardless of the concrete permutations), so they together contribute to the cardinality of the union. On the other hand, if then they are disjoint and thus contribute . Furthermore, the point set is contained in all objects and contributes . Hence, the cardinality of the union is equal to
Suppose we get a -approximation to the cardinality of the union, then
Since , by computing we obtain a value in , that is, an additive approximation to . Our choice of allows to distinguish from .
The reduction is not yet complete. The catch is that an algorithm for union volume estimation can only learn about the objects via queries. It is left to us to answer those queries. In Algorithm 2 we describe how to answer queries on object . Queries on object can be handled similarly by replacing with and with . The correctness is easy to verify, and this completes the reduction. So far, the argument works for arbitrary permutations; randomness becomes relevant when we bound the number of bit accesses.
-
For query : Return .
-
For query : Draw a random block and a random shift . With probability return . Otherwise return .
-
For query : If then return true. Else, compute and . If or then return false. Otherwise, if then return true; else return false.
Bounding the number of accesses
Now for the sake of contradiction, suppose there exists a randomized algorithm that -approximates the cardinality of the union of any objects in , using less than queries. We run on the objects defined earlier (using Algorithm 2 to handle its queries). From the output, we can distinguish and . We claim that less than bits of are accessed from the side of Algorithm 2.
To this end, note that accesses to only happen in the “otherwise” clauses of and in Algorithm 2. For analysis we isolate the following mini game:
Lemma 16.
Consider a game of rounds. At the start, Merlin generates a secret random permutation of . In each round, Arthur adaptively sends one of the following two requests to Merlin:
-
: Answer “yes” with probability and “no” with the remaining probability.
-
: Answer “yes” if and “no” otherwise.
No matter what strategy Arthur employs, the probability that Merlin ever answers “yes” in the game is at most .
Proof.
Let be the event that Merlin ever answers “yes” and the first one was due to a request. Let be the event that Merlin ever answers “yes” and the first one was due to a request. We will bound and separately.
To bound , define events where indicates that Arthur’s -th request is and the answer is “yes”. Clearly , thus .
To bound we use an entropy argument. In particular, we will describe and analyze an encoding scheme for a random permutation of .
The encoder receives and acts as Merlin in the game, using as the secret permutation. It interacts with Arthur till the end of the game. If does not happen then the encoding is simply a bit 0 followed by . On the other hand, if happens, then consider the first round when “yes” appeared; denote that request by . The encoding starts with a bit 1, then the number , and finishes with the ordering of induced by .
The decoder receives the encoding and tries to reconstruct . First it reads the leading bit of the encoding. If the bit is 0 then immediately follows. If the bit is 1 then the decoder recovers the number . It pretends to be Merlin and starts a game with Arthur, using the same random tape as the encoder but without knowledge of . For the first requests it just answers “no” to Arthur. Now comes the -th request from Arthur, which must be a request with answer “yes”. This allows the decoder to deduce . The remaining entries in can be fully restored from .
Writing , the encoding length is at most
where we used in the second line. Since the encoding length must be at least the Shannon entropy of the encoded permutation , we conclude that .
Putting the two bounds together, we have .
Where does the game come into play? The reader might find it useful to recall Figure 4. The space is split into blocks horizontally, where each block holds information about the -th bit of the input vector. Let us focus on a particular block . As runs, some of its queries “hit” this block while others don’t. Precisely, we say that a query hits if the random block chosen in Algorithm 2 is . Likewise, we say that a query hits if . Now we restrict ourselves to the queries hitting and ignore all the others. Then, over time, (as Arthur) is exactly playing the game with us (as Merlin) on the secret permutation : queries correspond to requests; queries correspond to requests; Algorithm 2 branches to the “otherwise” clauses if and only if the answer is “yes”. Whatever information collects in the queries hitting other blocks , it has no effect on the decision in this game, as all other permutations are independent of , and as do not play a role in this game.
With this in mind, let random variable count the number of queries hitting . Define
where . By Lemma 16, for every , so . On the other hand, by definition of and our assumption that the number of queries is at most , we can bound
So Algorithm 2 accesses less than bits of in expectation. Symmetrically, the same bound holds for . Altogether it accesses less than bits of . But as we argued, the output can be used to distinguish and . This contradicts Lemma 15.
5 Reduction from discrete to continuous space
Union volume estimation and Klee’s measure problem were formulated in the continuous space . Their analogs in discrete space are only simpler due to the following reduction. Consider the natural embedding that blows up each point into a continuous cube . For any finite subset , denote its image by . We have:
-
.
-
To sample from uniformly, we can first sample uniformly and then sample from uniformly.
-
If is a discrete box then is a continuous box, and vice versa.
Given this correspondence, our algorithm for Klee’s measure problem can be made to handle discrete boxes in . On the other hand, the discrete objects in our lower bound construction can be realized as axis-aligned polygons in .
References
- [1] Pankaj K. Agarwal. An improved algorithm for computing the volume of the union of cubes. In David G. Kirkpatrick and Joseph S. B. Mitchell, editors, Proceedings of the 26th ACM Symposium on Computational Geometry, Snowbird, Utah, USA, June 13-16, 2010, pages 230–239. ACM, 2010. doi:10.1145/1810959.1811000.
- [2] Pankaj K. Agarwal, Haim Kaplan, and Micha Sharir. Computing the volume of the union of cubes. In Jeff Erickson, editor, Proceedings of the 23rd ACM Symposium on Computational Geometry, Gyeongju, South Korea, June 6-8, 2007, pages 294–301. ACM, 2007. doi:10.1145/1247069.1247121.
- [3] Karl Bringmann. An improved algorithm for Klee’s measure problem on fat boxes. Comput. Geom., 45(5-6):225–233, 2012. doi:10.1016/j.comgeo.2011.12.001.
- [4] Karl Bringmann and Tobias Friedrich. Approximating the volume of unions and intersections of high-dimensional geometric objects. Comput. Geom., 43(6-7):601–610, 2010. doi:10.1016/j.comgeo.2010.03.004.
- [5] Karl Bringmann, Kasper Green Larsen, André Nusser, Eva Rotenberg, and Yanheng Wang. Approximating klee’s measure problem and a lower bound for union volume estimation. CoRR, abs/2410.00996, 2024. doi:10.48550/arXiv.2410.00996.
- [6] Ran Canetti, Guy Even, and Oded Goldreich. Lower bounds for sampling algorithms for estimating the average. Inf. Process. Lett., 53(1):17–25, 1995. doi:10.1016/0020-0190(94)00171-T.
- [7] Nofar Carmeli, Shai Zeevi, Christoph Berkholz, Alessio Conte, Benny Kimelfeld, and Nicole Schweikardt. Answering (unions of) conjunctive queries using random access and random-order enumeration. ACM Trans. Database Syst., 47(3):9:1–9:49, 2022. doi:10.1145/3531055.
- [8] Ruoxu Cen, William He, Jason Li, and Debmalya Panigrahi. Beyond the quadratic time barrier for network unreliability. In David P. Woodruff, editor, Proceedings of the 2024 ACM-SIAM Symposium on Discrete Algorithms, SODA 2024, Alexandria, VA, USA, January 7-10, 2024, pages 1542–1567. SIAM, 2024. doi:10.1137/1.9781611977912.62.
- [9] Supratik Chakraborty, Kuldeep S. Meel, and Moshe Y. Vardi. A scalable approximate model counter. In Christian Schulte, editor, Principles and Practice of Constraint Programming - 19th International Conference, CP 2013, Uppsala, Sweden, September 16-20, 2013. Proceedings, volume 8124 of Lecture Notes in Computer Science, pages 200–216. Springer, 2013. doi:10.1007/978-3-642-40627-0_18.
- [10] Timothy M. Chan. Geometric applications of a randomized optimization technique. Discret. Comput. Geom., 22(4):547–567, 1999. doi:10.1007/PL00009478.
- [11] Timothy M. Chan. Semi-online maintenance of geometric optima and measures. SIAM J. Comput., 32(3):700–716, 2003. doi:10.1137/S0097539702404389.
- [12] Timothy M. Chan. A (slightly) faster algorithm for Klee’s measure problem. Comput. Geom., 43(3):243–250, 2010. doi:10.1016/j.comgeo.2009.01.007.
- [13] Timothy M. Chan. Klee’s measure problem made easy. In 54th Annual IEEE Symposium on Foundations of Computer Science, FOCS 2013, 26-29 October, 2013, Berkeley, CA, USA, pages 410–419. IEEE Computer Society, 2013. doi:10.1109/FOCS.2013.51.
- [14] Timothy M. Chan. Minimum Hausdorff distance of point sets under translation: Generalizing klee’s measure problem. In Erin W. Chambers and Joachim Gudmundsson, editors, 39th International Symposium on Computational Geometry, SoCG 2023, June 12-15, 2023, Dallas, Texas, USA, volume 258 of LIPIcs, pages 24:1–24:13. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2023. doi:10.4230/LIPICS.SOCG.2023.24.
- [15] Nilesh N. Dalvi and Dan Suciu. Efficient query evaluation on probabilistic databases. VLDB J., 16(4):523–544, 2007. doi:10.1007/s00778-006-0004-3.
- [16] David Eppstein and Jeff Erickson. Iterated nearest neighbors and finding minimal polytopes. Discret. Comput. Geom., 11:321–350, 1994. doi:10.1007/BF02574012.
- [17] David R. Karger. A randomized fully polynomial time approximation scheme for the all-terminal network reliability problem. SIAM J. Comput., 29(2):492–514, 1999. doi:10.1137/S0097539796298340.
- [18] Richard M. Karp and Michael Luby. Monte-Carlo algorithms for the planar multiterminal network reliability problem. J. Complex., 1(1):45–64, 1985. doi:10.1016/0885-064X(85)90021-4.
- [19] Richard M. Karp, Michael Luby, and Neal Madras. Monte-Carlo approximation algorithms for enumeration problems. J. Algorithms, 10(3):429–448, 1989. doi:10.1016/0196-6774(89)90038-2.
- [20] Benny Kimelfeld, Yuri Kosharovsky, and Yehoshua Sagiv. Query efficiency in probabilistic XML models. In Jason Tsong-Li Wang, editor, Proceedings of the ACM SIGMOD International Conference on Management of Data, SIGMOD 2008, Vancouver, BC, Canada, June 10-12, 2008, pages 701–714. ACM, 2008. doi:10.1145/1376616.1376687.
- [21] Victor Klee. Can the measure of be computed in less than steps? The American Mathematical Monthly, 84(4):284–285, 1977. doi:10.1080/00029890.1977.11994336.
- [22] Michael G. Luby. Monte-Carlo methods for estimating system reliability. Technical report, Report UCB/CSD 84/168, Computer Science Division, University of California, Berkeley, 1983.
- [23] Kuldeep S. Meel, Sourav Chakraborty, and N. V. Vinodchandran. Estimation of the size of union of delphic sets: Achieving independence from stream size. In Leonid Libkin and Pablo Barceló, editors, PODS ’22: International Conference on Management of Data, Philadelphia, PA, USA, June 12 - 17, 2022, pages 41–52. ACM, 2022. doi:10.1145/3517804.3526222.
- [24] Kuldeep S. Meel, Aditya A. Shrotri, and Moshe Y. Vardi. Not all FPRASs are equal: demystifying FPRASs for DNF-counting. Constraints An Int. J., 24(3-4):211–233, 2019. doi:10.1007/s10601-018-9301-x.
- [25] Kuldeep S. Meel, N. V. Vinodchandran, and Sourav Chakraborty. Estimating the size of union of sets in streaming models. In Leonid Libkin, Reinhard Pichler, and Paolo Guagliardo, editors, PODS’21: Proceedings of the 40th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems, Virtual Event, China, June 20-25, 2021, pages 126–137. ACM, 2021. doi:10.1145/3452021.3458333.
- [26] Mark H. Overmars and Chee-Keng Yap. New upper bounds in Klee’s measure problem. SIAM J. Comput., 20(6):1034–1045, 1991. doi:10.1137/0220065.
- [27] Aduri Pavan, N. V. Vinodchandran, Arnab Bhattacharyya, and Kuldeep S. Meel. Model counting meets estimation. In Leonid Libkin, Reinhard Pichler, and Paolo Guagliardo, editors, PODS’21: Proceedings of the 40th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems, Virtual Event, China, June 20-25, 2021, pages 299–311. ACM, 2021. doi:10.1145/3452021.3458311.
- [28] Christopher Ré, Nilesh N. Dalvi, and Dan Suciu. Efficient top- query evaluation on probabilistic data. In Rada Chirkova, Asuman Dogac, M. Tamer Özsu, and Timos K. Sellis, editors, Proceedings of the 23rd International Conference on Data Engineering, ICDE 2007, The Marmara Hotel, Istanbul, Turkey, April 15-20, 2007, pages 886–895. IEEE Computer Society, 2007. doi:10.1109/ICDE.2007.367934.
- [29] Qiaosheng Shi and Binay K. Bhattacharya. Application of computational geometry to network -center location problems. In Proceedings of the 20th Annual Canadian Conference on Computational Geometry, Montréal, Canada, August 13-15, 2008, 2008.
- [30] Srikanta Tirthapura and David P. Woodruff. Rectangle-efficient aggregation in spatial data streams. In Michael Benedikt, Markus Krötzsch, and Maurizio Lenzerini, editors, Proceedings of the 31st ACM SIGMOD-SIGACT-SIGART Symposium on Principles of Database Systems, PODS 2012, Scottsdale, AZ, USA, May 20-24, 2012, pages 283–294. ACM, 2012. doi:10.1145/2213556.2213595.
- [31] Jan van Leeuwen and Derick Wood. The measure problem for rectangular ranges in -space. J. Algorithms, 2(3):282–300, 1981. doi:10.1016/0196-6774(81)90027-4.
- [32] Hakan Yildiz and Subhash Suri. On Klee’s measure problem for grounded boxes. In Tamal K. Dey and Sue Whitesides, editors, Proceedings of the 28th ACM Symposium on Computational Geometry, Chapel Hill, NC, USA, June 17-20, 2012, pages 111–120. ACM, 2012. doi:10.1145/2261250.2261267.