Abstract 1 Introduction 2 Our Algorithm 3 Bounding Number of Phases 4 Controlling the Potential 5 Conclusions References Appendix A Proofs of Technical Claims

Online Disjoint Set Covers: Randomization Is Not Necessary

Marcin Bienkowski ORCID University of Wrocław, Poland Jarosław Byrka ORCID University of Wrocław, Poland Łukasz Jeż ORCID University of Wrocław, Poland
Abstract

In the online disjoint set covers problem, the edges of a hypergraph are revealed online, and the goal is to partition them into a maximum number of disjoint set covers. That is, n nodes of a hypergraph are given at the beginning, and then a sequence of hyperedges (subsets of [n]) is presented to an algorithm. For each hyperedge, an online algorithm must assign a color (an integer). Once an input terminates, the gain of the algorithm is the number of colors that correspond to valid set covers (i.e., the union of hyperedges that have that color contains all n nodes).

We present a deterministic online algorithm that is O(log2n)-competitive, exponentially improving on the previous bound of O(n) and matching the performance of the best randomized algorithm by Emek et al. [ESA 2019].

For color selection, our algorithm uses a novel potential function, which can be seen as an online counterpart of the derandomization method of conditional probabilities and pessimistic estimators. There are only a few cases where derandomization has been successfully used in the field of online algorithms. In contrast to previous approaches, our result extends to the following new challenges: (i) the potential function derandomizes not only the Chernoff bound, but also the coupon collector’s problem, (ii) the value of Opt of the maximization problem is not bounded a priori, and (iii) we do not produce a fractional solution first, but work directly on the input.

Keywords and phrases:
Disjoint Set Covers, Derandomization, pessimistic Estimator, potential Function, online Algorithms, competitive Analysis
Copyright and License:
[Uncaptioned image] © Marcin Bienkowski, Jarosław Byrka, and Łukasz Jeż; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Online algorithms
Funding:
Supported by Polish National Science Centre grants 2022/45/B/ST6/00559,
2020/39/B/ST6/01641, and 2020/39/B/ST6/01679.
Acknowledgements:
The authors are grateful to the anonymous reviewers for their insightful comments.
Editors:
Olaf Beyersdorff, Michał Pilipczuk, Elaine Pimentel, and Nguyễn Kim Thắng

1 Introduction

In this paper, we study online algorithms for maximizing the number of set covers of a set of nodes. We focus on a hypergraph (set system) G=(V,E) that has n=|V| nodes and where each hyperedge SE is a subset of nodes from V. A subset EE is called set cover if SES=V, i.e., every node is covered by at least one hyperedge of E. In the disjoint set covers (DSC) problem [15, 12, 20], the goal is to partition the set of hyperedges E into maximum number of mutually disjoint subsets E=E1E2Ek, where each Ej is a set cover. Note that E is a multi-set, i.e., it can contain multiple copies of the same hyperedge.

The problem has been studied in a theoretical setting, but as we discuss later, it also finds applications in sensor networks [12] or assignment tasks [20].

Coloring Perspective.

When constructing a solution to the DSC problem, it is convenient to think in terms of coloring hyperedges.111The DSC problem should not be confused with the hyperedge coloring problem, which requires that the hyperedges of the same color be disjoint. Each color then corresponds to a subset of hyperedges that have that color, and the color is fully used if the hyperedges colored with it form a set cover. The problem then becomes equivalent to coloring hyperedges, so that the number of fully used colors is maximized.

Online Variant.

In this paper, we focus on the online variant of the DSC problem, where the set V is given in advance, but the hyperedges of E arrive in an online fashion. Once a hyperedge SE arrives, it must be colored immediately and irrevocably. Again, the goal is to maximize the number of fully used colors. The performance of an online algorithm is measured by the competitive ratio, i.e., the ratio of the number of fully used colors produced by the optimal offline algorithm Opt to that of an online algorithm Alg.

Our Contribution.

We present a deterministic online O(log2n)-competitive algorithm Det for the DSC problem, which exponentially improves on the O(n)-competitive algorithm by Emek et al. [14] and matches the performance of their randomized algorithm [14]. We discuss the challenges and technical contribution in more detail in Subsection 1.3.

1.1 Offline Scenario: Previous Results

The disjoint set covers problem is a fundamental NP-complete problem [12] that can be approximated within a factor of (1+o(1))ln|V| [20] and cannot be approximated within a factor of (1ε)ln|V| for ε>0 unless NPDTIME(nloglogn) [15].222The authors of [15] call this problem set cover packing or one-sided domatic problem.

OPT vs. Min-degree.

We use Opt(E) to denote the maximum number of disjoint set covers of a hypergraph G=(V,E). This value is also called cover-decomposition number [6]. We denote the minimum degree of the hypergraph G=(V,E) as δ(E)miniV|SE:iS|. Note that trivially Opt(E)δ(E). While this bound may not be tight (cf. Subsection 1.4), δ(E) serves as a natural benchmark for approximation and online algorithms.

Random Coloring and its Straightforward Derandomization.

An O(logn)-approximation algorithm for the offline DSC problem [21] can be obtained by coloring each hyperedge with a color chosen uniformly at random from the palette of Θ(δ(E)/logn) colors. To analyze this approach, we focus on a single node iV. We say that node i gathers color r if i is contained in a hyperedge colored with r. Since node i is contained in at least δ(E) hyperedges, and there are Θ(δ(E)/logn) available colors, i gathers all palette colors with high probability. By the union bound, this property holds for all nodes, i.e., all Θ(δ(E)/logn) colors are fully used by an algorithm (also with high probability). Since Opt(E)δ(E), the approximation ratio of O(logn) follows.

The hyperedges can be processed in a fixed order, and the random choice of a color can be replaced by the deterministic one by a simple application of the method of conditional probabilities [21, 3].

1.2 Online Scenario: Previous Results

In the online case, an algorithm first learns the set of nodes V, and then the edges of E are revealed sequentially. For notational convenience, we use E to denote both the set and a sequence of hyperedges (the input). We use Alg(E) to denote the number of fully used colors in the solution of an online algorithm Alg.

It is important to note that no parameter of the hypergraph other than the number of nodes is known a priori. In particular, an online algorithm does not know the value of δ(E) in advance.

Competitive Ratio.

We say that Alg is (non-strictly) c-competitive if there exists β0 such that

Alg(E)Opt(E)/cβ. (1)

While β can be a function of n, it cannot depend on the input sequence E. If (1) holds with β=0, the algorithm is strictly c-competitive.

For randomized algorithms, we replace Alg(E) with its expected value taken over the random choices of the algorithm.

Randomized Algorithms.

The current best randomized algorithm was given by Emek et al. [14]; it achieves the (strict) competitive ratio of O(log2n). The general idea of their algorithm is as follows. To color a hyperedge S, their algorithm first computes the minimum degree δ of a node from S. By a combinatorial argument, they show that they can temporarily assume that δ(E)=O(nδ). Their algorithm then chooses uniformly at random from the set {δ,2δ,4δ,,2lognδ}; with probability Ω(1/logn) such is a 2-approximation of δ(E). Finally, to color S, it chooses a color uniformly at random from the palette of Θ(/log2n) colors, using the arguments similar to those in the offline case.

We refrain from discussing it further here, as we present a variant of their algorithm, called Rand, in Subsection 2.1 (along with a description of the differences from their algorithm).

The best lower bound for the (strict and non-strict) competitive ratio of a randomized algorithm is Ω(logn/loglogn) [14]; it improves on an earlier bound of Ω(logn) by Pananjady et al. [20].

Deterministic Algorithms.

The deterministic case is well understood if we restrict our attention to strict competitive ratios. In such a case the lower bound is Ω(n) [20]. The asymptotically matching upper bound O(n) is achieved by a simple greedy algorithm [14].

On the other hand, the current best lower bound for the non-strict deterministic competitive ratio is Ω(logn/loglogn) [14].333This discrepancy in achievable strict and non-strict competitive ratios is quite common for many maximization problems: the non-strict competitive ratio allows the algorithm to have zero gain on very short sequences, thus avoiding initial choices that would be bad in the long run. The O(n) upper bound of [14] clearly holds also in the non-strict setting, but no better algorithm has been known so far.

Table 1: Previous and new bounds on the strict and non-strict competitive ratios for the online DSC problem, for randomized and deterministic algorithms.
Upper bound Lower bound
randomized O(log2n) [14] Ω(logn/loglogn) [14]
strict and non-strict
deterministic O(n) [14] Ω(n) [20]
strict
deterministic O(n) [14] Ω(logn/loglogn) [14]
non-strict 𝐎(𝐥𝐨𝐠𝟐𝐧) (Theorem 5)

Lack of General Derandomization Tools.

Unlike approximation algorithms, derandomization is extremely rare in online algorithms.444Many online problems (e.g., caching [23, 1] or metrical task systems [7, 8, 9]) exhibit a provable exponential discrepancy between the performance of the best randomized and deterministic algorithms. For many other problems, the best known deterministic algorithms are quite different (and more complex) than randomized ones. To understand the difficulty, consider the standard (offline) method of conditional probabilities [3] applied to the offline DSC problem: the resulting algorithm considers all the random choices (color assignments) it could make for a given hyperedge S, and computes the probability that future random choices, conditioned on the current one, will lead to the desired solution. Choosing the color that maximizes this probability ensures that the probability of reaching the desired solution does not decrease as subsequent hyperedges are processed. In some cases, exact computations are not possible, but the algorithm can instead estimate this probability by computing a so-called pessimistic estimator [22]. This is not feasible in the online setting, since an algorithm does not know the future hyperedges, and thus cannot even estimate these probabilities.555As observed by Pananjady et al. [20], however, these probabilities can be estimated if an online algorithm knows the final min-degree δ(E) in advance, which would lead to an O(logn)-competitive algorithm for this semi-offline scenario.

1.3 Our Technical Contribution

In this paper, we present a deterministic online polynomial-time algorithm Det that is O(log2n)-competitive, which exponentially improves the previous bound of O(n). This resolves an open question posed by the authors of [14] who asked whether the method of conditional expectations could be used to derandomize their algorithm.

Our bound is obtained for the non-strict competitive ratio: as we discussed in the previous subsection, this is unavoidable for the DSC problem. Our result requires a relatively small (1/4) additive term β in the definition of the competitive ratio (1).

We begin by constructing a randomized solution Rand. It will be a variation of the approach by Emek et al. [14]; we present it in Subsection 2.1. As Rand is closely related to the previous randomized algorithm, it is quite plausible that it achieves the same competitive ratio of O(log2n). However, our analysis does not support this conjecture. Instead, we use Rand as a stepping stone to our deterministic algorithm in the following way.

  • We define a particular random event, denoted 𝒯E, of Rand executed on instance E.

  • We show that for each execution of Rand satisfying 𝒯E, it holds that Rand(E)Ω(1/log2n)Opt(E)1/4.

We will provide the exact definition of the event 𝒯E in Subsection 2.2. For now, we just note that 𝒯E is a property that certifies that, at each step t, we can relate the number of colors gathered so far by each node i to its current degree.

While a relaxed version of the property 𝒯E (requiring that the relation between the gathered colors and the current degree only at the end of the input) follows quite easily with a constant probability, it seems that 𝒯E itself is quite strong and does not hold with a reasonable probability. However, this is not an obstacle to creating a deterministic solution, as we show in the following claim.

  • It is possible to replace the random choices of Rand by deterministic ones (in an online manner), so that 𝒯E always holds.

This will immediately imply the competitive ratio of the resulting deterministic algorithm.

Our approach is based on a novel potential function that guides the choice of colors. As we discuss later in Subsection 2.2, our approach can be seen as an online-computable counterpart of the method of conditional probabilities that simultaneously controls the derandomization of the Chernoff bound and the coupon collector’s problem.

1.4 Related Work

Applications of the DSC problem include allocating servers to users in file systems and users to tasks in crowd-sourcing platforms [20].

Another application arises in a sensor network, where each node corresponds to a monitoring target and each hyperedge corresponds to a sensor that can monitor a set of targets. At any given time, all targets must be monitored. A possible battery-saving strategy is to partition the sensors into disjoint groups (each group covering all targets) and activate only one group at a time, while the other sensors remain in a low-power mode [12].

When the assumption that each sensor can only participate in a single group is dropped, this leads to more general sensor coverage problems. The goal is then to maximize the lifetime of the network of sensors while maintaining the coverage of all targets. The offline variants of this problem have been studied both in general graphs [4, 13, 21] and in geometric settings [11, 16].

Another line of work studied the relationship between the minimum degree δ(E) and the cover-decomposition number Opt(E). As pointed out in Subsection 1.1, δ(E)/Opt(E)1. This ratio is constant if G is a graph [17], and it is at most O(logn) for every hypergraph; the latter bound is asymptotically tight [6]. Interestingly enough, all the known papers on the DSC problem (including ours) relate the number of disjoint set covers to δ(E). It is an open question whether our estimate of the competitive ratio is tight: it could possibly be improved by relating the gain of an algorithm directly to Opt(E) instead of δ(E).

1.5 Preliminaries

An input to our problem is a gradually revealed hypergraph G=(V,E). V consists of n nodes numbered from 1 to n, i.e., V=[n]. The set E of hyperedges is presented one by one: in a step t, an algorithm learns StE, where St[n], and has to color it. We say that node i gathers color r if i is contained in a hyperedge colored with r. We say that a color r is fully used if all nodes have gathered it. The objective of an algorithm is to maximize the number of fully used colors.

At any point in time, the degree of a node j, denoted deg(j), is the current number of hyperedges containing j. Let δ(E)=mini[n]|{StE:viSt}|, i.e., δ(E) is the minimum degree of G at the end of the input E. Clearly, Opt(E)δ(E). It is important to note that δ(E) is not known in advance to an online algorithm.

2 Our Algorithm

2.1 Definition of RAND

We start with some notions, defined for each integer k0:

hlogn, qk(112n)2k, Rk{2k,,2k+11}.

Note that |Rk|=2k.

Algorithm 1 Definition of Rand for a hyperedge S in step t.

For each node i independently, Rand maintains its phase p(i), initially set to 0. We use Cik to denote the set of colors from the palette Rk that node i has gathered so far. A phase k for node i ends when it has gathered qk colors from palette Rk. In such a case, node i increments its phase number p(i) at the end of the step.

We now describe the behavior of Rand in a single step, when a hyperedge S appears. Let pS=miniSp(i), where p(i) is the phase of node i before S appears. Rand first chooses a random integer k uniformly from the set {pS,pS+1,,pS+h1}. Second, Rand chooses a color r uniformly at random from the set Rk and colors S with it; in effect all nodes in S gather color r.

The pseudocode of Rand is given in Algorithm 1.

Comparison with the Previous Randomized Algorithm.

Rand is closely related to the randomized algorithm by Emek et al. [14]. The main difference is that the phases of the nodes in their algorithm are of fixed lengths, being powers of 2.666The pseudocode of their algorithm does not use phases, but with some fiddling with constants, it can be transformed into one that does. They use probabilistic arguments to argue that with high probability each node gathers qk colors in a phase of length Θ(2klog2n). Instead, we treat the number of colors gathered in a phase as a hard constraint (we require that each node gathers qk of them), and instead the phase lengths of the nodes become random variables. As it turns out, this subtle difference allows us to derandomize the algorithm.

Another small difference concerns the color selection. While Rand chooses the color uniformly at random from the set Rk, the algorithm of [14] would choose it from the set j=0kRj. This difference affects only the constant factors of the analysis.

Gain of RAND.

As mentioned earlier, the gain of Rand is directly ensured by the algorithm definition provided that each node has completed some number of phases.

Note that qk is slightly less than |Rk|=2k; this gives a better bound on the the expected phase lengths, while still ensuring that if all nodes completed phase , they gathered at least 21 common colors.

Lemma 1.

If every node has completed phase , then Rand(E)21.

Proof.

In phase , each node gathers at least q colors from the palette R, i.e., all colors from R except at most |R|q2/(2n) colors. Thus, all nodes share at least |R|n(2/(2n))=21 common colors from R, i.e., Rand fully uses at least 21 colors.

Note that we could sum the above bound over all phases completed by all nodes, but this would not change the asymptotic gain.

The above lemma points us to the goal: to show that for an input E with a sufficiently large δ(E), each node completes an appropriately large number of phases. In Subsection 2.3, we show that if δ(E)=Ω(2log2n), then each node completes phases, provided a certain random event 𝒯E occurs.

2.2 Constructing the Potential: Insights and Definitions

In the field of online algorithms, the derandomization has been successfully conveyed several times by replacing the method of conditional probabilities by an online-computable potential function that guides the choice of the deterministic algorithm [2, 5, 10, 18].

This method is based on the following framework. First, introduce random expressions {Zi}i=1 to be controlled. Second, define a potential function Φ=i=1exp(Zi). Third, show that the random actions of an algorithm at each step decrease exp(Zi) (for each i) in expectation, which implies that Φ decreases as well. (By the probabilistic method, this implies the existence of a deterministic action of an algorithm in a step that does not increase Φ.) Finally, by the non-negativity of the exponential function, exp(Zi) is bounded by the initial value Φ0 of the potential (in each step), which shows that Zi can always be bounded by ln(Φ0). This process can be seen as a derandomization of the Chernoff bound / high probability argument.

In order to apply this very general framework, we must overcome several technical difficulties, which we explain below. Except for the definition of Φ (and related definitions of wik,cik, dk and Zi), the discussion in this section is informal and will not be used in formal proofs later.

Variables to be Controlled.

The first step is to identify the variables to control. Natural choices are the node degrees and the number of colors gathered so far by each node.

For this purpose, we define cik=|Cik| for each node i and each phase k0.

We also introduce counters wik, which are initially set to zero. Recall that whenever a new hyperedge S containing node i arrives, Rand computes pS=miniSp(i). For each node iS such that p(i)pS+h1, we increment the counter wi,p(i). Note that these wi,p(i) variables are incremented exactly for nodes i for which there is a non-zero probability of increasing their set of colors Ci,p(i) (as then a random integer k chosen by Rand has a non-zero chance of being equal to p(i)).

By the definition of wik, at each step k0wikdeg(i). While these quantities are not necessarily equal, we will treat k0wik as a good proxy for deg(i) and deal with the discrepancy between these two quantities later.

Linking the Variables.

Now we want to introduce an expression that links k0wik (the proxy for degree) with k0cik (the number of colors gathered) for a node i. A simple difference of these two terms does not make sense: the expected growth of cik varies over time, since it is easier to gather new colors when cik is small. This effect has been studied in the context of the coupon collector’s problem [19]. To overcome this issue, we introduce the following function, defined for each integer k0:

dk(m) hj=1m2k2kj+1 (defined for each m2k)

Note that in a process of choosing random colors from palette Rk of 2k colors, the expected number of steps till m different colors are gathered is dk(m)/h.

Now we focus on a single node i in phase p(i). We consider a sequence of hyperedges S containing node i, neglecting those hyperedges S for which p(i)pS+h. That is, all considered hyperedges increment the counter wi,p(i). Then, the value of dk(ci,p(i)) corresponds to the expected number of such hyperedges needed to gather ci,p(i) colors from the palette Rp(i). We can thus use the expression k0(wik2dk(cik)) to measure the progress of node i: small values of this expression indicate that it is gathering colors quite fast, while large values indicate that it is falling behind. Note that since cikqk2k, the value of dk(cik) is always well defined.

We note that the previous applications of the potential function method [2, 5, 10, 18] did not require such transformations of variables: in their case, the potential function was used to guide deterministic rounding: the function Φ directly compared the cost (or gain) of a deterministic algorithm with that of an online fractional solution. Instead, our solution operates directly on the input, without the need to generate a fractional solution first.

Scaling.

Using the random choices of Rand, we can argue that in expectation the value of Zik0(wik2dk(cik)) is decreasing in time. However, to argue that exp(Zi) is also decreasing in expectation, we would have to ensure that Zi is upper-bounded by a small constant (and use the fact that exp(x) can be approximated by a linear function for small x).

In the previous papers [2, 5, 10, 18], this property was achieved by scaling down Z by the value of Opt. An algorithm was then either given an upper bound on Opt (in the case of the throughput maximization of the virtual circuit routing [10]) or Opt was estimated by standard doubling techniques (in the case of the cost minimization for set cover variants [2, 5, 18]). In the latter case, the algorithm was restarted each time the estimate on Opt was doubled. Unfortunately, the DSC problem (which is an unbounded-gain maximization problem) does not fall into any of the above categories, and the doubling approach does not seem to work here.

Instead, we replace the scaling with a weighted average. That is, for each node i, we define

Zik0wik2dk(cik)4h2k

and the potential as

Φi[n]exp(Zi).

Note that all counters and variables defined above are random variables; they depend on particular random choices of Rand. We use pt(i), degt(i), wikt, cikt, Zit and Φt to denote the values of the corresponding variables at the end of step t of the algorithm (after Rand has processed the hyperedge presented in step t). The value of t=0 corresponds to the state of these variables at the beginning of the algorithm; note that p0(i)=deg0(i)=wik0=cik0=Zi0=0 for all i and k. Therefore,

Φ0=n. (2)

Random event 𝓣𝑬.

For an input instance E consisting of T steps, we define a random event 𝒯E that occurs if Φtn for each step t{0,,T} of Rand execution on input E.

2.3 Relating Potential to Algorithm Performance

We begin by presenting the usefulness of the event 𝒯E. We emphasize that the following lemma holds for all executions of Rand in which the event 𝒯E occurs. Its proof is deferred to Section 3.

Lemma 2.

Fix a sequence E such that δ(E)>24hln(4en)2 for some integer 0. If 𝒯E occurs, then each node has completed its phase .

Using the lemma above, we can relate the gain of Rand on an arbitrary input E to that of Opt, if only 𝒯E occurs.

Lemma 3.

Fix a sequence E. If 𝒯E occurs, then Rand(E)Opt(E)/(96hln(4en))1/4.

Proof.

Let r24hln(4en). We consider two cases.

  • First, we assume that δ(E)>r. Then we can find an integer 0 such that r2<δ(E)r2+1. By Lemma 2, each node then completes its phase , and so by Lemma 1, Rand(E)21δ(E)/(4r).

  • Second, we assume that δ(E)r. Trivially, Rand(E)0(δ(E)r)/(4r).

In both cases, Rand(E)(δ(E)r)/(4r)Opt(E)/(4r)1/4.

2.4 Derandomization of RAND

To analyze the evolution of Φ, we note that Φt (and also other variables wikt, cikt or Zit) depends only on the random choices of Rand till step t (inclusively). Moreover, {Φt}t0 is a supermartingale with respect to the random choices of Rand in consecutive steps. Specifically, the following lemma holds; its proof is deferred to Section 4.

Lemma 4.

Fix a step t and an outcome of random choices till step t1 inclusively. (In particular, this will fix the value of Φt1.) Then, 𝔼[Φt]Φt1, where the expectation is taken exclusively over random choices of Rand in step t.

The above lemma states that the value of Φ is non-increasing in expectation. In fact, an inductive application of this lemma shows that 𝔼[Φt]Φ0=n. However, this does not imply that 𝒯E occurs with a reasonable probability, especially when the input length is large.777By Markov’s inequality, for a chosen step t, Φt2n holds with probability at least 1/2. While such a relaxed bound on Φt would be sufficient for our needs, in our proof, we need such a bound to hold for all steps t simultaneously.

However, using Lemma 4, we can easily derandomize Rand using the method of conditional probabilities using potential Φ as an online-computable counterpart of a pessimistic estimator. To do this, we proceed iteratively, ensuring at each step t that Φtn. This is trivial at the beginning, since Φ0=n by (2).

Suppose we have already fixed the random choices of Rand till step t1 inclusively and have Φt1n. Consider a hyperedge S presented in step t. Note that all other variables indexed by t1, such as pt1(i), are also fixed. Then Lemma 4 states that 𝔼[Φt]Φt1n. That is, the random choice of a color at step t guarantees that 𝔼[Φt]n. This choice is made from a finite and well-defined set of colors pSkpS+h1Rk, where pS=miniSpt1(i).

By the probabilistic method, at each step t, there exists a deterministic choice of a color from  that ensures that Φtn. The resulting algorithm is called Det. (If more than one color leads to Φtn, Det chooses any of them.) Since || is bounded by a polynomial of n and |E|, Det can be implemented in polynomial time by simply checking all possible colors of .

Theorem 5.

Det is O(log2n)-competitive for the DSC problem.

Proof.

As described above, Det guarantees that Φtn for each step t{0,,T}, i.e., 𝒯E occurs. Note that Lemma 3 lower-bounds the gain of Rand in every execution conditioned on 𝒯E, and Det can be seen as such an execution. Therefore, the bound of Lemma 3 can be applied, yielding

Det(E)Opt(E)96hln(4en)14,

i.e., the competitive ratio of Det is at most 96hln(4en)=O(log2n).

Note that, by the lower bounds of [20, 14], an additive term (in our case equal to 1/4) is inevitable for obtaining a sub-linear competitive ratio.

3 Bounding Number of Phases

In this section, we fix an input sequence E consisting of T steps. Our goal is to estimate the number of phases of nodes in the execution of Rand, conditioned on the random event 𝒯E, i.e., to prove Lemma 2. To this end, we consider a node i, assume that it has completed  phases, and show an upper bound on the final degree of i as a function of .

Bounding Variables w Using Potential.

Recall that in some steps where the degree of a node i grows, the counter wi,p(i) is incremented. Thus, our first goal is to upper-bound values of these counters at the end of the execution of Rand.

Below, H(m) denotes the m-th harmonic number. The following technical claim is proved in Appendix A.

Claim 6.

For each k0 it holds that H(2k)H(2kqk)ln(4en).

Lemma 7.

Fix a step tT, a node i and a phase k0. Then, dk(cikt)hln(4en)2k.

Proof.

Note that at all times, ciktqk. Using the definition of dk,

dk(cik)dk(qk) =hj=1qk2k2kj+1
=h(H(2k)H(2kqk))2k
hln(4en)2k. (by Claim 6)

Lemma 8.

Fix a node i and a phase 0. If the event 𝒯E occurs, then k=0wikT8hln(4en)2.

Proof.

Fix the last step tT in which k=0wik increases. By the choice of t, we have wikt=cikt=0 for every phase k>.

Since 𝒯E occurs, nΦt=j[n]exp(Zjt). Due to the non-negativity and monotonicity of the exponential function, Zitlnn. Using the definition of Zit, we get

lnnZit =k0wikt2dk(cikt)4h2k
=k=0wikt2dk(cikt)4h2k
14h2(k=0wikt2k=0hln(4en)2k). (by Lemma 7)

Hence, again by the choice of t,

k=0wikT=k=0wikt 4h2lnn+4hln(4en)2<8hln(4en)2.

Bounding Node Degrees.

Recall that whenever a new hyperedge S containing node i arrives, pS=miniSp(i) is determined. Then, for each node iS, if p(i)pS+h1, the counter wi,p(i) is incremented. If p(i)pS+h, the degree of i grows, but wi,p(i) is not incremented. To estimate the degree of i, we therefore introduce the counters sik, which are incremented in the latter case. That is, for each node iS, we always have k0(wik+sik)=deg(i).

The growth of the variables sik is not controlled by the potential, but they grow only for nodes whose degree is very high compared to the current minimum degree. By constructing an appropriate charging argument, we can link their growth to the growth of other variables wik.

Lemma 9.

Fix a step tT, a node i, and a phase 0. Then, sitj[n]r=0hwjrt.

Proof.

We fix node i, phase 0, and show the lemma by induction on t. The inductive basis is trivial, as si0=0=j[n]r=0hwjr0.

Now suppose that the lemma statement holds for step t1. We look at how both sides of the inequality change as we increase the step superscripts from t1 to t, and argue that the increase of the right hand side is at least as large as the increase of the left hand side. If si does not change, the inductive claim follows trivially. Otherwise, si is incremented by 1. This happens only if =p(i), iS, and pS+h. By the definition of pS, this means that there exists at least one node jS such that p(j)=pS, and the corresponding counter wj,pS is also incremented. This means that the right hand side of the lemma inequality is incremented by at least j[n]r=0h(wjrtwjrt1)wj,pStwj,pSt1=1, and the inductive claim follows.

Lemma 10.

Fix a node i and a phase 0. If the event 𝒯E occurs, then k=0sikT16hln(4en)2.

Proof.

By Lemma 9,

sikTj[n]r=0khwjrT n8hln(4en)2kh (by Lemma 8)
8hln(4en)2k. (as h=logn)

Hence, k=0sikT<16hln(4en)2.

Finally, we can show Lemma 2, restated below.

Lemma 2. [Restated, see original statement.]

Fix a sequence E such that δ(E)>24hln(4en)2 for some integer 0. If 𝒯E occurs, then each node has completed its phase .

Proof.

Suppose for a contradiction that there exists a node i for which p(i) at the end of the input. Then,

δ(E)degT(i)=k0(wikT+sikT)=k=0(wikT+sikT)24hln(4en)2,

where the last inequality follows by Lemma 8 and Lemma 10. This would contradict the assumption of the lemma.

4 Controlling the Potential

In this section, we show that {Φt}t0 is a supermartingale with respect to the choices of Rand made in corresponding steps, i.e., we prove Lemma 4.

Throughout this section, we focus on a single step t, in which Rand processes a hyperedge S. Recall that Rand first chooses a random integer k uniformly from the set {pS,pS+1,,pS+h1}. Second, conditioned on the first choice, it chooses a random color uniformly from the set Rk.

By the definition of our variables, for each node i and each integer k0, it holds that wiktwikt1{0,1} and ciktcikt1{0,1}.

Lemma 11.

Fix a node iS and let p=pt1(i). If ppS+h1, then cipt=cipt1+1 with probability (2pcipt1)/(h2p).

Proof.

By the definition of pS, we have ppS. Combining this with the lemma assumption, we get p{pS,,pS+h1}.

Note that cipt=cipt1+1 when node i gathers a new color from Rp, and cipt=cipt1 otherwise. For a node i to gather a new color from Rp, first the integer k chosen randomly from the set {pS,,pS+h1} must be equal to p, which happens with probability 1/h. Second, conditioned on the former event, a color chosen randomly from the palette Rp must be different from all cipt1 colors from Rp gathered so far by node i, which happens with probability (|Rp|cipt1)/|Rp|=(2pcipt1)/2p. Hence, the probability of gathering a new color is (2pcipt1)/(h2p).

We emphasize that the relations involving random variables in the following lemma (e.g., the statements such as ZitZit1) hold for all random choices made by Rand.

Lemma 12.

Fix a node i. Let p=pt1(i). Then, either ZitZit1 or

ZitZit1+14h2p+{1/(2h2pα)with probability α,0otherwise.

where α=(2pcipt1)/(h2p). The probability is computed exclusively with respect to random choices of Rand in step t.

Proof.

For brevity, for an integer k0, we define

Δwik =wiktwikt1,
Δdk(cik) =dk(cikt)dk(cikt1).

As ciktcikt1 and dk is a non-decreasing function, we have Δdk(cik)0.

Let S be the hyperedge presented in step t. By the definition of the variables wik (cf. Subsection 2.2), we have

Δwik={1if iS and k=p and ppS+h1,0otherwise.

Now we consider two cases.

  • It holds that iS or ppS+h. Then,

    ZitZit1=k0Δwik2Δdk(cik)4h2kk0Δwik4h2k=0.
  • It holds that iS and ppS+h1. Then, Δwip=1 and Δwik=0 for kp. Therefore,

    ZitZit1 =k0Δwik2Δdk(cik)4h2k
    =Δwip2Δdp(cip)4h2p+kpΔwik2Δdk(cik)4h2k
    14h2pΔdp(cip)2h2p.

    To complete the proof, it now suffices to argue that Δdp(cipt)=1/α with probability α and 0 with the remaining probability. This follows immediately by Lemma 11: With probability α we have cipt=cipt1+1, and hence Δdp(cip)=dp(cipt)dp(cipt1)=h2p/(2pcipt+1)=h2p/(2pcipt1)=1/α. With the remaining probability cipt=cipt1 and thus Δdp(cip)=0.

For the final lemma, we need the following technical bound (proven in Appendix A). This can be seen as a reverse Jensen’s type inequality.

Claim 13.

Fix ε[0,1], α[ε,1] and a real x. Let X be a random variable such that

X={xε/αwith probability α,xotherwise.

Then, 𝔼[eX]exε/2.

Finally, we can prove Lemma 4, restated below.

Lemma 4. [Restated, see original statement.]

Fix a step t and an outcome of random choices till step t1 inclusively. (In particular, this will fix the value of Φt1.) Then, 𝔼[Φt]Φt1, where the expectation is taken exclusively over random choices of Rand in step t.

Proof.

Fix a node i. We will show that

𝔼[exp(Zit)]exp(Zit1). (3)

The lemma then follows by summing the above inequality over all nodes.

If ZitZit1, (3) follows trivially. Otherwise, Lemma 12 implies that

ZitZit1+14h2p+{1/(2h2pα)with probability α,0otherwise.

where p=pt1(i) and α=(2pcipt1)/(h2p). As the random choices are fixed until step t1, the variables Zit1, p and α are no longer random variables, but real numbers.

Note that cipt1qp12p1 as otherwise phase p of node i would have ended in an earlier step. Hence, α=(2pcipt1)/(h2p)1/(2h2p).

Thus, x=Zit1+1/(4h2p), ε=1/(2h2p), and α satisfy the conditions of Claim 13. (In particular, εα1.) This claim now yields

𝔼[exp(Zit)]exp(Zit1+14h2p1212h2p)=exp(Zit1),

which concludes the proof of (3), and thus also the lemma.

5 Conclusions

In this paper, we have constructed a deterministic O(log2n)-competitive algorithm for the disjoint set covers (DSC) problem. Closing the remaining logarithmic gap between the current upper and lower bounds is an interesting open problem that seems to require a new algorithm that goes beyond the phase-based approach.

We have developed new derandomization techniques that extend the existing potential function methods. We hope that these extensions will be useful for derandomizing other online randomized algorithms, and eventually for providing a coherent online derandomization toolbox.

References

  • [1] Anna Adamaszek, Artur Czumaj, Matthias Englert, and Harald Räcke. An O(log k)-competitive algorithm for generalized caching. ACM Transactions on Algorithms, 15(1):6:1–6:18, 2019. doi:10.1145/3280826.
  • [2] Noga Alon, Baruch Awerbuch, Yossi Azar, Niv Buchbinder, and Joseph Naor. The online set cover problem. SIAM Journal on Computing, 39(2):361–370, 2009. doi:10.1137/060661946.
  • [3] Noga Alon and Joel H. Spencer. The Probabilistic Method, Second Edition. John Wiley, 2000. doi:10.1002/0471722154.
  • [4] Piotr Berman, Gruia Călinescu, C. Shah, and Alexander Zelikovsky. Power efficient monitoring management in sensor networks. In 2004 IEEE Wireless Communications and Networking Conference, pages 2329–2334. IEEE, 2004. doi:10.1109/WCNC.2004.1311452.
  • [5] Marcin Bienkowski, Björn Feldkord, and Pawel Schmidt. A nearly optimal deterministic online algorithm for non-metric facility location. In Proc. 38th Symp. on Theoretical Aspects of Computer Science (STACS), LIPIcs, pages 14:1–14:17, 2021. doi:10.4230/LIPIcs.STACS.2021.14.
  • [6] Béla Bollobás, David Pritchard, Thomas Rothvoß, and Alex D. Scott. Cover-decomposition and polychromatic numbers. SIAM Journal on Discrete Mathematics, 27(1):240–256, 2013. doi:10.1137/110856332.
  • [7] Alan Borodin, Nati Linial, and Michael E. Saks. An optimal on-line algorithm for metrical task system. Journal of the ACM, 39(4):745–763, 1992. doi:10.1145/146585.146588.
  • [8] Sébastien Bubeck, Christian Coester, and Yuval Rabani. The randomized k-server conjecture is false! In Proc. 55th ACM Symp. on Theory of Computing (STOC), pages 581–594. ACM, 2023. doi:10.1145/3564246.3585132.
  • [9] Sébastien Bubeck, Michael B. Cohen, James R. Lee, and Yin Tat Lee. Metrical task systems on trees via mirror descent and unfair gluing. In Proc. 30th ACM-SIAM Symp. on Discrete Algorithms (SODA), pages 89–97. SIAM, 2019. doi:10.1137/1.9781611975482.6.
  • [10] Niv Buchbinder and Joseph (Seffi) Naor. Online primal-dual algorithms for covering and packing. Mathematics of Operations Research, 34(2):270–286, 2009. doi:10.1287/MOOR.1080.0363.
  • [11] Adam L. Buchsbaum, Alon Efrat, Shaili Jain, Suresh Venkatasubramanian, and Ke Yi. Restricted strip covering and the sensor cover problem. In Proc. 18th ACM-SIAM Symp. on Discrete Algorithms (SODA), pages 1056–1063. SIAM, 2007. URL: http://dl.acm.org/citation.cfm?id=1283383.1283497, doi:10.5555/1283383.1283497.
  • [12] Mihaela Cardei and Ding-Zhu Du. Improving wireless sensor network lifetime through power aware organization. Wireless Networks, 11(3):333–340, 2005. doi:10.1007/S11276-005-6615-6.
  • [13] Mihaela Cardei, My T. Thai, Yingshu Li, and Weili Wu. Energy-efficient target coverage in wireless sensor networks. In Proc. 2005 IEEE Int. Conf. on Computer Communications (INFOCOM), pages 1976–1984. IEEE, 2005. doi:10.1109/INFCOM.2005.1498475.
  • [14] Yuval Emek, Adam Goldbraikh, and Erez Kantor. Online disjoint set cover without prior knowledge. In Proc. 27th European Symp. on Algorithms (ESA), LIPIcs, pages 44:1–44:16, 2019. doi:10.4230/LIPICS.ESA.2019.44.
  • [15] Uriel Feige, Magnús M. Halldórsson, Guy Kortsarz, and Aravind Srinivasan. Approximating the domatic number. SIAM Journal on Computing, 32(1):172–195, 2002. doi:10.1137/S0097539700380754.
  • [16] Matt Gibson and Kasturi R. Varadarajan. Optimally decomposing coverings with translates of a convex polygon. Discret. Comput. Geom., 46(2):313–333, 2011. doi:10.1007/S00454-011-9353-9.
  • [17] Ram P. Gupta. On the chromatic index and the cover index of a multigraph. In Theory and Applications of Graphs, pages 204–215. Springer Berlin Heidelberg, 1978.
  • [18] Ngoc Mai Le, Seeun William Umboh, and Ningyuan Xie. The power of clairvoyance for multi-level aggregation and set cover with delay. In Proc. 2023 ACM-SIAM Symp. on Discrete Algorithms (SODA), pages 1594–1610. SIAM, 2023. doi:10.1137/1.9781611977554.CH59.
  • [19] Michael Mitzenmacher and Eli Upfal. Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis. Cambridge University Press, USA, 2nd edition, 2017.
  • [20] Ashwin Pananjady, Vivek Kumar Bagaria, and Rahul Vaze. The online disjoint set cover problem and its applications. In Proc. 2015 IEEE Int. Conf. on Computer Communications (INFOCOM), pages 1221–1229. IEEE, 2015. doi:10.1109/INFOCOM.2015.7218497.
  • [21] Ashwin Pananjady, Vivek Kumar Bagaria, and Rahul Vaze. Optimally approximating the coverage lifetime of wireless sensor networks. IEEE/ACM Transactions on Networking, 25(1):98–111, 2017. doi:10.1109/TNET.2016.2574563.
  • [22] Prabhakar Raghavan. Probabilistic construction of deterministic algorithms: Approximating packing integer programs. Journal of Computer and System Sciences, 37(2):130–143, 1988. doi:10.1016/0022-0000(88)90003-7.
  • [23] Neal E. Young. On-line file caching. Algorithmica, 33(3):371–383, 2002. doi:10.1007/S00453-001-0124-5.

Appendix A Proofs of Technical Claims

Claim 6. [Restated, see original statement.]

For each k0 it holds that H(2k)H(2kqk)ln(4en).

Proof.

Assume first that 2k<4n. As H(m)1+lnm=ln(em) for each m, we get H(2k)H(2kqk)<H(4n)ln(4en) and the lemma follows.

Hence, in the following, we assume that 2k4n, obtaining

2kqk =2k(11/(2n))2k (by the definition of qk)
2k(11/(2n))2k1
=2k/(2n)1
2k/(4n). (as 2k4n)

As qk2k, we can use the relation H(m)lnm holding for every m0, obtaining

H(2k)H(2kqk)ln(e2k)ln(2kqk)ln(e2k)ln(2k/(4n))=ln(4en).

Claim 13. [Restated, see original statement.]

Fix ε[0,1], α[ε,1] and a real x. Let X be a random variable such that

X={xε/αwith probability α,xotherwise.

Then, 𝔼[eX]exε/2.

Proof.

Using αε, we obtain

(1+ε/α)(1ε/(2α))=1+ε/(2α)ε2/(2α2)1. (4)

Next, we use the relation ey(1+y)1 (holding for every y>1) and (4) to argue that

exp(ε/α)(1+ε/α)11ε/(2α). (5)

Finally, this gives

𝔼[eX] =αexp(xε/α)+(1α)exp(x)
=ex(αexp(ε/α)+1α)
ex(αε/2+1α) (by (5))
exε/2. (as 1ε/2eε/2 for each ε)