Abstract 1 Introduction 2 Prize-collecting Ordered TSP 3 Analysis of Algorithm 1 4 Prize-collecting Multi-Path TSP 5 Derandomization 6 Discussion References

Approximating Prize-Collecting Variants of TSP

Morteza Alimi ORCID University of Augsburg, Germany Tobias Mömke ORCID University of Augsburg, Germany Michael Ruderer ORCID University of Augsburg, Germany
Abstract

We present an approximation algorithm for the Prize-collecting Ordered Traveling Salesman Problem (PCOTSP), which simultaneously generalizes the Prize-collecting TSP and the Ordered TSP. The Prize-collecting TSP is well-studied and has a long history, with the current best approximation factor slightly below 1.6, shown by Blauth, Klein and Nägele [IPCO 2024]. The best approximation ratio for Ordered TSP is 32+1e, presented by Böhm, Friggstad, Mömke, Spoerhase [SODA 2025] and Armbruster, Mnich, Nägele [Approx 2024]. The former also present a factor 2.2131 approximation algorithm for Multi-Path-TSP.

We present a 2.097-approximation algorithm for PCOTSP, which is, to the best of our knowledge, the first result for this problem. Key ideas in our approach are to sample a set of trees and then to probabilistically pick up some vertices, and to use the pruning ideas of Blauth, Klein, Nägele [IPCO 2024] on the sampled vertices. While the sampling probability of vertices for our problem is lower than for PCTSP, intuitively leaving less spare penalty to spend, we leverage the cycle structure induced by the sampled trees together with a simple combinatorial algorithm to bring the approximation factor below 2.1.

Our techniques extend to Prize-collecting Multi-Path TSP, building on results from Böhm, Friggstad, Mömke, Spoerhase [SODA 2025], leading to a 2.41-approximation.

Keywords and phrases:
Approximation Algorithms, TSP
Copyright and License:
[Uncaptioned image] © Morteza Alimi, Tobias Mömke, and Michael Ruderer; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Approximation algorithms analysis
Related Version:
Previous Version: https://arxiv.org/abs/2411.14994
Funding:
Partially supported by DFG Grant 439522729 (Heisenberg-Grant).
Editors:
Paweł Gawrychowski, Filip Mazowiecki, and Michał Skrzypczak

1 Introduction

The metric Traveling Salesman Problem (TSP), which asks for a shortest closed tour (i.e., a Hamiltonian cycle) in a metric space (V,c),c:V×V+ on an n element vertex set V visiting each vertex exactly once111Note that the problem can alternatively be defined as having a weighted graph as input, and seeking to find a shortest closed walk (or shortest Eulerian multi-subgraph) that visits each vertex at least once. is one of the most well-studied problems in combinatorial optimization in its various incarnations. Christofides [15] and Serdjukov [23] gave a simple 32 approximation algorithm for symmetric (undirected) TSP; an approximation factor slightly below 32 was provided by Karlin, Klein and Oveis Gharan [20].

The path versions of TSP have also been subject to intense study. This line of study led to a surprising outcome: Traub, Vygen, and Zenklusen [25] show that for any ϵ>0, there is a reduction from path-TSP to TSP which only loses ϵ in the approximation factor. In their book, Traub and Vygen [24] simplify the reduction and use Multi-Path-TSP (see below) as a building block. The book gives a comprehensive overview of the aforementioned results and more, including asymmetric (directed) variants of TSP.

A more general and more practical version of the problem can be defined by allowing the tour to omit some cities by paying some additional penalty. This prize collecting paradigm has been intensely studied for various combinatorial optimization problems; see, e.g., [1, 2] for some recent results regarding Prize-collecting Steiner Forest and Steiner Tree Problems. As regards the Prize-collecting Traveling Salesman Problem (PCTSP), Bienstock et al. [7] give a 2.5-approximation for this problem. Goemans and Williamson [18] give a factor 2 approximation. The first group to break the 2 barrier was Archer et al. [3], who achieved a factor of 1.979. Goemans [17] observed that by carefully combining the two aforementioned algorithms, one can achieve an approximation factor of 1.91. Blauth and Nägele [9] gave a factor 1.774-approximation. Blauth, Klein, and Nägele [8] achieved a factor of 1.599.

Another important generalization of TSP is attained by enforcing some ordering on a subset of vertices. Formally, in the metric Ordered Traveling Salesman Problem (OTSP), we are given a metric graph G=(V,E),c:E+ together with k terminals O={o1,,ok}, and the objective is to find a shortest tour that visits the terminals in order. OTSP is closely related to vehicle routing problems with pickup and delivery and the dial-a-ride problem; see [22, 6, 14, 16, 19]. Any α-approximation for TSP can be utilized to give an (α+1)-approximation algorithm for OTSP by finding an α-approximate tour on VO and combining it with the cycle on O [10]. Furthermore, there is a combinatorial (2.52/k)-approximation algorithm [11]. A substantial improvement in the approximation factor to 32+1e was achieved by [13] and [4].

In Multi-Path Traveling Salesman Problem (Multi-Path-TSP), in addition to a weighted graph, we are given a list of 2k terminals 𝒯={(s1,t1),,(sk,tk)}. The objective is to find k paths Pi, 1ik of minimum total length, covering all vertices of the graph (see also Section 16.4 of  [24]). Böhm, Friggstad, Mömke and Spoerhase [13] give a factor 2.2131 approximation algorithm for the problem.

1.1 Our Contribution and Overview of Techniques

As mentioned above, the prize-collecting setting is a standard setting for studying algorithmic problems. Theoretically, this is closely related to the Lagrangian relaxation paradigm, where constraints are allowed to be violated by paying some penalty. It is also natural from a practical perspective, where it is often possible to refuse to cover some entity if it incurs too high a cost, paying them some predefined penalty instead. In this article, we study OTSP and Multi-Path-TSP in this setting, which can also be viewed as generalizations of PCTSP.

Definition 1.

In the Prize-collecting Ordered Traveling Salesman Problem (PCOTSP) we are given a metric weighted complete graph with penalties for vertices G=(V,E),c:E+,π:V+, together with k terminals O={o1,,ok}. The objective is to find a tour C traversing o1,,ok in order, that minimizes c(C)+vV(C)π(v).

Definition 2.

In the Prize-collecting Multi-Path Traveling Salesman Problem (PC-Multi-Path-TSP), we are given a metric weighted complete graph with vertex penalties G=(V,E),c:E+,π:V+, and a set of k terminal pairs 𝒯={(s1,t1),,(sk,tk)}. The objective is to find k paths Pi,1ik, such that Pi is a path with endpoints si,ti, minimizing i=1kc(Pi)+v(V(Pi))π(v).222 Similar to Path-TSP, we consider the k terminal pairs in Multi-Path TSP as part of what defines the problem which cannot be relaxed. This is also consistent with standard practical applications where e.g. the start and finish depots cannot be dropped. Hence in PC-Multi-Path-TSP and its special case, PCOTSP, we assume the penalties for terminals are . The techniques in this paper do not seem to extend directly to the case where terminals are allowed to be dropped.

Theorem 3.

There is a 2.097-approximation algorithm for PCOTSP.

Theorem 4.

There is a 2.41-approximation algorithm for PC-Multi-Path-TSP.

In the remainder of this section we provide an overview of our algorithms and techniques for PCOTSP and PC-Multi-Path-TSP. The details of the algorithms and the proofs of correctness are covered in the following sections.

In Section 2, we specify the LP relaxation (OLP) of PCOTSP. Our algorithm first solves the LP, and then it samples a tree Ti for each of the k parts of the LP solution, in the manner of Lemma 2 in [8], and Theorem 4 in [13], which are both based on a result of Bang-Jensen, Frank and Jackson [5] and Post and Swamy [21]333Post and Swamy showed that the Bang-Jensen, Frank, Jackson decomposition can be computed in polynomial time. (see Lemma 6). The expected cost of each tree is at most the corresponding part of the LP solution, and its expected coverage is at least that of the corresponding y values. Now a key idea in our algorithm is to even out the penalty ratios paid by the algorithm. This is implemented in two ways.

  1. 1.

    Probabilistically picking up some of the vertices left out of the sampled trees (inevitably increasing the connection cost). This is achieved through a pick-up threshold, which determines the vertices which should be picked up (with a suitable probability), based on the LP solution.

  2. 2.

    Utilizing the spare penalty ratios of sampled vertices to bring down the cost of parity correction.

The second point is realized through an adaptation of the pruning idea from [8] to probabilistically prune away portions of the resultant structure with low connectivity. This allows us to assign weights to the edges of the tree, which, when combined with a suitable multiple of the LP solution, gives a point in the Q-join polytope, where Q is the set of odd degree vertices of the current structure. This allows for cheaper parity correction, because the tree edges with higher assigned weights, which are associated with lower value cuts, are pruned with higher probability, and thus the expected cost incurred by them remains low.

The prominent difference between our problem and the standard PCTSP, which necessitates finding new ideas for pruning, is the sampling probability for vertices. In PCTSP, the sampling probability for each vertex v is equal to yv, hence the expected penalty ratio for v is one, and intuitively, there is lots of spare penalty for v to utilize for pruning. But in PCOTSP, the sampling probability of a vertex v is lower (bounded from below by 1eyv); hence there is little spare penalty ratio left (and for fewer vertices) to utilize for pruning. We obtain an improved adjustment of the pruning by combining our algorithm with a version of the classical algorithm for OTSP [10], which combines the cycle on terminals with a 32-approximation for TSP on the remaining vertices. If the length of the cycle on terminals is low, this algorithm already gives a good approximation ratio. If the length of the cycle is large, this fact can be utilized to improve the analysis for parity correction, because we can show that there is no need to assign (nonzero) weights to the cycle edges.

We believe that this view of evening out the penalty ratio of vertices vis-a-vis the optimal LP solution is a useful conceptual tool for approaching prize collecting versions of TSP or other combinatorial optimization problems. In the case of PCOTSP, it is partly realized through the idea of probabilistically including in the solution those vertices which currently have a too high penalty ratio (based on the desired approximation factor). Squeezing out the slack penalty ratio (again, based on the target approximation factor), is achieved through pruning. Here, the specific structure of the OTSP, and the partially constructed solution proves useful; we can show that if the current cycle through the terminals is small, combining the cycle on terminals with the best PCTSP approximation for the rest of vertices gives a good factor, while a large cycle improves the cost of pruning.

In the case of Multi-Path-TSP, Böhm, Friggstad, Mömke, and Spoerhase [13] propose two algorithms, and show that a careful combination of them leads to a good approximation. The first one, which doubles all edges except those on si,ti paths (and is good when the sum of si,ti distances is large) can be adapted to our setting by probabilistically picking up vertices with high penalty ratio, as opposed to picking up all remaining vertices in [13]. We replace their second algorithm, which finds a minimum length forest in which each terminal appears in exactly one of the components and then adds direct si,ti edges with an algorithm that uses Lemma 6 to sample a tree from the related PCTSP, and then adds direct si,ti edges.

2 Prize-collecting Ordered TSP

In this section we describe our algorithm for PCOTSP. Some of the technical proofs will be presented in the following sections.

2.1 Preliminaries and Definitions

In the OTSP, a tour can be decomposed into k paths, between oi and oi+1.444For notational convenience, we identify ok+1 with o1. We can take the polyhedron determined by the following inequalities as the relaxation of one such path between two vertices s and t.555 Similar to the LP of [4]; see also [12, 8]. For each vertex v, the variable yv indicates its fractional degree in the stroll.

ys=yt=12x(δ(v))=2yvvVx(δ(S))1SV{t},sSx(δ(S))2yvvSV{s,t}x,y0(s-t-stroll relaxation)

Now, the PCOTSP can be modeled as the following linear program. For i=1,k we define xi and yi to be the vectors (xi,e)eE and (yi,v)vV, respectively. For each i, the vector (xi,yi) is constrained to be a feasible point in the oi-oi+1-stroll relaxation, the sum yv=i=1kyi,v over all fractional degrees of a vertex v is an indicator of to what degree v is used in the solution. When v is not fully used, i.e., when yv<1, the LP has to pay a fractional penalty of πv(1yv). We will usually refer to the pair (oi,oi+1) as (si,ti), to emphasize that (xi,yi) is a fractional tour/stroll from oi to oi+1.

minimizeeEi=1kcexi,e+vVπv(1yv)subject to yv=i=1kyi,vvV(xi,yi) lies in the si-ti-stroll relaxationi[k](OLP)

Note that every feasible solution to (OLP) has yo=1 for all terminals o. Similar to yv, we will use xe as a shorthand for i=1kxi,e. It follows immediately from the LP constraints that x(δ(S))2yv for any vSVO. By contracting all terminals oO into a single vertex r, we therefore obtain the relaxation of the normal PCTSP from [8], which also involves a root, which (without loss of generality; see, e.g., [3]) is required to be in the tour. Given a solution (x,y) to (OLP) and some threshold ρ[0,1], we define Vρ{vVyvρ}.

2.2 A simple algorithm

We consider the following simple algorithm for PCOTSP, inspired by [10]. First, directly connect the terminals o1,,ok in order, creating a simple cycle C^. Then, compute a solution to the PCTSP on the same instance. Since all terminals have infinite penalty, every tour T obtained in this way includes all terminals. We obtain an ordered tour of cost c(T)+c(C^) by following the original cycle C^ and grafting in the tour at an arbitrary terminal oO.

Using an α^-approximation algorithm to solve the PCTSP, we know that the sum of tour- and penalty costs for this solution is at most α^optPCTSPα^optPCOTSP. Since c(C^)optPCOTSP, this immediately implies a (1+α^)-approximation algorithm for PCOTSP. One can see that this algorithm performs even better if we can guarantee that C^ is small. To be precise, for any αα^ we obtain an α-approximation provided that c(C^)(αα^)optPCOTSP. We therefore may assume c(C^)(αα^)optPCOTSP in the analysis of our main algorithm. The currently best value for α^ is the approximation guarantee of (roughly) 1.599 obtained by [8].

2.3 Overview of our main algorithm

Fix α=2.097. We first solve (OLP)666The separation oracle for the LP boils down to separating subtour elimination constraints. Hence the LP can be solved in polynomial time using the ellipsoid algorithm., to get an optimal solution (x,y). Using the following Lemma 5, we then split off the vertices v for which yvθ, for a parameter θ to be determined later, to get a solution (x^,y^) for the LP were the remaining vertices have a certain minimum connectivity to the terminals. Vertices that have been split off will not be used in our final tour. Instead, we pay the full penalty for them. We state the following lemma for PCOTSP, as the proof is identical to the one for PCTSP.

Lemma 5 (Splitting off [8]).

Let (x,y) be a feasible solution to the PCOTSP relaxation and let SVO. Then we can efficiently compute another feasible solution (x,y) in which yv=0 for all vS, but yv=yv for all vS, and c(x)c(x).

Lemma 5 ensures that we can remove vertices from the support of x, without increasing the cost of x. Since we only split off vertices for which yv is relatively low, we can afford to pay the additional penalties if we set θ=11α (we will prove this in Section 3.1). We proceed by sampling a set of trees based on (x^,y^).

Lemma 6 ([8], following [5]).

Suppose (x,y) is a feasible point in the s-t-stroll relaxation. In polynomial time we can find a set of trees 𝒯 and weights (μ(T))T𝒯 such that (i) T𝒯μ(T)=1; and (ii) T𝒯:eE(T)μ(T)xeeE; and (iii) T𝒯:vV(T)μ(T)yv,vV{s,t}; and (iv) all trees span s and t.

Lemma 6 can be used to sample a random tree of expected cost at most c(x) which contains each vertex v with probability at least yv and is guaranteed to connect s to t. It is straightforward to see that this can be achieved by choosing each tree T𝒯 with probability μ(T).

For each component (x^i,y^i) of our solution, we apply Lemma 6 and sample a tree Ti as we have described. Define T:=˙i=1kTi. Note that T is no longer a tree, and it might even have repeated edges.

2.3.1 Pruning and Picking up

To get some intuition on the the following steps, suppose that we had to pay the penalty for a vertex v if and only if v has not been sampled in T. In Lemma 10, we will show that the probability for this is at most eyv, which gives us a bound of eyvπv on the expected penalty cost of v. The LP however pays a (1yv)-fraction of πv. We thus define the penalty ratio ρv of a vertex v as ρv=eyv1yv and define σ0 as unique solution to the equation α(1σ0)=eσ0. With this, if yv=σ0, we have ρv=α, i.e., our expected penalty for v is at most α times the LP penalty.

Note that in the PCTSP regime, e.g., in the result of [8], the probability that a vertex v is in the (one) sampled tree is exactly yv, which means that ρv=1<α for every v. But in our setting, the probability of a vertex not being sampled is [bounded from above by] eyv, which intuitively means that the spare penalties are much more constrained, and the gains would be more meager. Nonetheless, we show how to utilize the specific structure of our problem to gain almost as much from pruning as in the setting of PCTSP.

We will now describe how we deal with vertices whose y-value is smaller or larger than σ0. First, we consider those vertices for which yv<σ0 (and thus ρv<α). These vertices v have some spare penalty ratio, which we utilize to prune v with certain probability.

For our pruning step, we need the following definition from [8]:

Definition 7.

For a fixed LP solution (x,y), a tree T, and a threshold γ[0,1], we define core(T,γ) as the inclusion-wise minimal subtree of T that spans all vertices in V(T)Vγ.

Algorithmically, core(T,γ) can be obtained by iteratively removing leaves v with yv<γ. To prune a tree Ti simply means to replace it by Ti=core(Ti,γ). We emphasize that for pruning Ti, we do not consider the local penalty values yi,v, but the global values yv. Furthermore, our construction ensures that si as well as ti are part of the pruned tree Ti (remember that ysi=yti=1). We will draw the global pruning threshold γ according to a suitable distribution Dγ which is defined by the following cumulative probability function: Fγ(y)=Pr[γy]1α(1y)1ey. We will prove in Section 3.1 that this ensures that the slack in the penalty ratios is fully utilized.

We continue with those vertices for which yv>σ0 and describe our pickup step. For vertices v with yv>σ0, the penalty ratio is greater than α. In fact, their penalty ratio becomes arbitrarily high as yv approaches 1. Intuitively, this means that the probability of these vertices to not be sampled is too high. We call these vertices critical and define Vσ0={vVyvσ0}. We pick up these critical vertices with certain probability. That means, we will probabilistically select some unsampled critical vertices and connect them to our solution. We first describe how these vertices are selected and then give the details on how we connect them.

Note that by our observation on ρv, the chance for picking up a fixed unsampled critical vertex v should increase with the value of yv. Our strategy is therefore the following: As we did for our pruning threshold, we will draw a global pickup threshold σ[σ0,1] from a distribution Dσ. We then pick up all critical vertices whose y-value is above σ, i.e., all vertices in VσV(T).

The distribution Dσ is determined by the cumulative probability function Fσ which we state in the following. Choosing σ according to this function will ensure that for any vVσ0 the probability of being picked up is just high enough so that the expected penalty paid for v is at most α times the fractional penalty πv(1y) paid by the LP solution. We will formally prove this fact in Section 3.1. For now, we simply define Fσ(y)1α(1y)ey. We remark that by definition of σ0, we have α(1y)ey=ey(1y)eσ0(1σ0), from which it is easy to verify that Fσ is indeed a cumulative probability function, i.e., Fσ(σ0)=0, Fσ(1)=1 and Fσ is monotonously increasing on [σ0,1]. To properly describe how we connect VσT to T, we use the following definition from [13].

Definition 8.

Let XYV. An X-rooted spanning forest of Y is a spanning forest of Y such that each of its connected components contains a vertex of X.

A cheapest X-rooted spanning forest of Y can be efficiently computed by contracting X, computing a minimal spanning tree of Y in the contracted graph and then reversing the contraction.

For our pickup step, we buy the cheapest forest FP that spans Vσ and is rooted in VσV(T). This forest FP has two crucial properties: (i) after buying FP, each vertex vVσT will be connected to T; and (ii) FP only spans vertices with yvσ. Property (ii) follows immediately from the definition of FP, and property (i) follows from the fact that OVσT for any choice of σ.

Finally, we remark that our pickup and pruning steps target the disjoint vertex sets {vVyvσ0} and {vVθyv<σ0}, so neither step interferes with the other.

Algorithm 1 The Prize-collecting Ordered TSP Algorithm.

2.3.2 Obtaining an ordered tour

To turn T′′ into a feasible tour, we first need to correct parities, i.e., we ensure that every vertex has an even degree. Let H be the graph that is obtained by adding a cheapest odd(T′′)-join to T′′, where odd(T′′) is defined as the set of odd degree vertices of T′′. Observe that initially, each tree Ti contains a path Pi between oi and oi+1. Since all terminals oO have yo=1, all of these paths survive the pruning step. So the multigraph H still contains k edge-disjoint oioi+1–paths which, taken together, form a closed walk C. By removing C from H, we obtain a graph whose connected components have even degree and can thus be shortcut into cycles. Furthermore, since H was connected, each of these cycles has a common vertex with C. Hence we can obtain a feasible tour of no greater cost than H by following the walk C and grafting in the cycles obtained by shortcutting the components of HC on the way.

3 Analysis of Algorithm 1

In this section, we prove Theorem 3. We will compare the expected cost of our computed solution to the cost of the optimal LP solution (x,y). In particular, we compare the expected cost of our computed tour to the cost of x (Section 3.3) and the expected sum of our incurred penalties to the fractional penalty cost defined by y (Section 3.1). For a derandomized version of our algorithm, we refer to the appendix.

In the following, we use (x,y) to refer to the optimal LP solution computed by the algorithm, and (x,y) to refer to the LP solution after the splitting-off step.

3.1 Bounding the Penalty Ratios

In this subsection, we prove the following lemma:

Lemma 9.

The expected total penalty cost paid by Algorithm 1 is at most

αvVπv(1yv).

We note that we can express the expected total penalty cost paid by our algorithm as vVπvPr[vV(T′′)]. We can thus prove Lemma 9 by showing that for each vertex v, the ratio Pr[vV(T′′)]/(1yv) is at most α.

Due to Step 2 in Algorithm 1, i.e., due to the splitting-off step, T′′ spans only vertices whose y-value is at least θ. Vertices v with yv<θ are therefore included in V(T′′) with probability 0. However, our choice of θ=11α guarantees that for the vertices v with yv<θ, we have Pr[vV(T′′)]/(1yv)1/(1θ)=α. To continue our analysis for the remaining vertices, i.e., those with yvθ, we show the following lemma:

Lemma 10.

Let v be a vertex with yvθ. Then the probability that v is not in T is at most ey.

Proof.

By Lemma 6, the probability that v is not in Ti is at most 1yi,v for any fixed i[k]. The probability that this happens for all i[k] is at most

i=1kPr[vV(Ti)]i=1k(1yi,v)exp(i=1kyi,v)=eyv.

Note that the distributions from which σ and γ are chosen guarantee that θγ<σ0σ1, i.e., only vertices v with yv[θ,σ0) can be pruned and only vertices with yv[σ0,1] can be picked up.

Now consider a vertex v with y[θ,σ0). There are two cases in which we have to pay v’s penalty: if v is not sampled, and if v is sampled but immediately pruned afterwards.

By Lemma 10, the probability of the former is at most eyv, whereas the probability of being pruned – given that v was sampled in the first place – can be bounded from above by Pr[γ>yv]=1Fγ(yv).777The probability can be lower if one of the sampled trees has a vertex with higher y-value in the subtree rooted at v. Our choice of Fγ(y)=1α(1y)1ey guarantees that the expected penalty is just high enough:

Pr[vV(T′′)]1yv (1Pr[vT])+Pr[vT](1Fγ(yv))1yv
=1Pr[vT]Fγ(yv)1yv1(1eyy)Fγ(yv)1yv=α.

Finally, consider a vertex v with yv[σ0,1]. This time, there is only one case in which we have to pay the penalty for v: when v is neither sampled, nor picked up afterwards. Again, the probability of not being sampled is at most eyv whereas the probability of not being picked up – given that v was not sampled previously – is Pr[σ>yv]=1Fσ(yv). Again, our choice of Fσ(y)=1α(1y)ey guarantees that

Pr[vV(T′′)]1yv eyv(1Fσ(yv))1yv=α.

This concludes our proof of Lemma 9.

3.2 Parity Correction

In order to analyze the expected cost of both the parity correction step and of our final tour, we first need to further investigate the structure of our computed solution and introduce some additional notation. Recall that T denotes the union of all sampled trees and that it includes the closed walk C, which is obtained by joining all si-ti-paths PiTi. Let R be the graph that contains all remaining edges, i.e., the (multi-)graph induced by E(T)E(C).

While it is convenient to think of C as a cycle, note that the paths P1,,Pk are not necessarily vertex disjoint. However, we can guarantee that C is a Eulerian multigraph spanning all terminals. In the same sense, R can be thought of as a collection of small trees which are all rooted at the cycle C (in reality, some of those trees may intersect each other). In the following, we will call the edges of C cycle edges and the edges of R tree edges. A depiction of C and R can be seen in Fig. 1.

When we prune the trees Ti into Ti, the paths Pi are not affected (because the pruning step guarantees that oi and oi+1 stay connected in Ti). So our pruning step can only remove edges from R. We thus define core(R,γ) as the graph that is obtained by pruning all trees in T with pruning threshold γ and then removing the cycle C. We now partition the edge set of R into layers. Intuitively, the i-th layer of R contains those edges that are contained in core(R,γ) if and only if γ does not exceed some value ηi.

Let η1>>η be the y-values of all vertices that might be affected by our pruning step, i.e., {η1,,η}={yvvV and θyvσ0}{σ0}. By definition, we always have η1=σ0. Now let E1=core(T,η1) and Ei=core(T,ηi)Ei1 for i=2,,. One can see that E1E2E is a partitioning of E(R).

To be able to bound the cost of the cheapest odd(T′′)-join J , we define the following vector

zβx+i:ηiγ(12ηiβ)χEizγ+max{0,12βσ}χFPzσ (1)

where β=13σ0θ. To simplify future arguments, we also define the two components zγ and zσ of z, as specified in (1). We remark that the vector βx+zγ has been used (for a different value of β) in [8]. In Section 3.3, the cost of z will be used as an upper bound for c(J). To this end, we now prove the following lemma.

Lemma 11.

z lies in the dominant of the odd(T′′)-join polytope.

Proof.

First, observe that all coefficients ze are non-negative. We now consider an arbitrary SV for which |Sodd(H)| is odd and show that z(δ(S))1.

We begin with the case where S cuts the terminal set, i.e., where 0<|S{o1,,ok}|<k. In this case, S separates at least two terminal pairs (oi1,oi1+1) and (oi2,oi2+1), which implies x(δ(S))2 and therefore z(δ(S))βx(δ(S))1. In the following we can thus assume that SV{o1,,ok}.

Now suppose that δ(S) contains a pickup edge e={u,v}E(FP) s.t. uS and vS. Recall that FP only spans vertices w with ywσσ0. We therefore have uSV{o1,,ok} which (by the connectivity constraints of (OLP)) implies x(δ(S))2yu2σ0 and therefore z(δ(S))βx(δ(S))+max{0,12βσ0}1.

It remains to show the bound for the case when δ(S) only contains cycle and tree edges. By a simple counting argument, one can see that |δT′′(S)| has the same parity as |Sodd(T′′)| and therefore must be odd. Now observe that by our assumption, |δFP(S)|=0 and that because C is a Eulerian multigraph, |δC(S)| must be even. If follows that δ(S) must have an odd number of tree edges. We finish our proof by marginally adapting an argument from [8].

First, we consider the case where δ(S) contains exactly one tree edge e. This is only possible if S includes a whole subtree Te of one of the trees in T. Because this subtree Te has survived the pruning step, e must lie in some layer Ei for which ηiγ. Furthermore, if eEi, then Te must contain at least one vertex v with yvηi, which implies that x(δ(S))2ηi. So we have

z(δ(S))βx(δ(S))+(12ηi)2βηi+(12βηi)1.

If, however, δ(S) contains at least three tree edges, we know that all vertices vS have yvθ and that each tree edge contributes at least (12βσ0) to zγ. Therefore z(δ(S))2βθ+3(12βσ0)1.

(a)

(b)

Figure 1: (a) The graph T′′ after pruning and picking up critical vertices. The terminals in O are drawn as black rectangles. The cycle C is depicted in red, the surviving edges of R are drawn in black and edges in FP in green. The greyed out vertices and edges do not belong to T′′. They have either been pruned (the dashed vertices and edges), not sampled, or split off. (b) The same graph T′′ with the various cuts that are considered in the proof of Lemma 11 drawn in different colors. The dashed blue cut is an example for the case where 0<|SO|<k. The dashed pink cut shows the case where a pickup edge (drawn in green) is cut. Note that even though the vertex in S was not picked up itself, it is still part of FP and thus must have a high y-value. The remaining two cuts (drawn in brown and light blue) show the two cases where δ(S) contains an odd number of tree edges.

3.3 Bounding the Tour Cost

In this section, we show the following bound on the cost of our computed tour:

Lemma 12.

𝔼[c(T′′)]αc(x)

The total cost of our computed tour can be bounded from above by

𝔼[c(H)]𝔼[c(T)]+𝔼[c(FP)]+𝔼[c(J)].

Our bound on the cost of J is given by the cost of the parity correction vector z=βx+zγ+zσ. We start by analyzing the expected combined value of c(T) and c(zγ).

Lemma 13.

𝔼[c(T)+c(zγ)]c(x)(2+α^α(2+2α^)βσ0+2αβσ0) where α^ denotes the approximation guarantee of the current best PCTSP algorithm.

The main idea of the proof of Lemma 13 is that the weight which each layer Ei is assigned in zγ is large when ηi is small and vice versa. At the same time, the chance for layer Ei to be present after pruning and thus to contribute at all to both zγ and c(T) is an increasing function of ηi. By balancing out these two values, we obtain an upper bound or the contribution of all layers in R to c(T)+c(zγ). At the same time we take into account that the cycle C only contributes toward c(T) but crucially not towards the cost of the parity correction vector. This is where we utilize our assumption that C^ and therefore also C can be lower bounded by a constant fraction of opt. For formally proving Lemma 13, we need the following technical Lemma 14.

Lemma 14.

For 23σ01, the function g(y)=Pr[γy](22βy) attains its maximum value over the interval [θ,σ0] at y=σ0.

Proof.

First, observe that Pr[γy]=Fγ(y)=1α(1y)1ey and therefore g(y)=2Fγ(y)(1βy). The claim follows from proving that g is monotonously increasing on [θ,σ0] by showing that its derivative is strictly positive:

12ddyg(y) =fγ(y)(1βy)βFγ(y)fγ(y)(1βσ0)β
α(1βσ0)β (2)
>0, (3)

where fγ denotes the density function of γ. In (2) we used that for y[θ,σ0]

fγ(y) =α1ey(1α(1y))ey(1ey)2=α1eyFγ(y)ey1ey
αey1ey=α+α1ey1α.

Furthermore, (3) follows because α and β are determined by σ0:

α=eσ01σ0,θ=11α, andβ=13σ0θ.

By plugging in these dependencies we obtain the expression

α(1βσ0)β=αβ(ασ0+1)=eσ01σ013σ01+1eσ01σ0(eσ01σ0σ0+1),

which has a positive value if σ0(23,1).

Proof of Lemma 13.

Our analysis follows the basic approach from [8], but distinguishes between the cycle C and the remaining part R of the sampled solution T to leverage the lower bound on the the cost of the cycle, which we get by running the simple algorithm from Section 2 in parallel. Another difference to [8] is that T is not a single tree, but consists of k distinct sampled trees, which requires some additional notation.

Let 𝒯=𝒯1××𝒯k denote the set of all possible outcomes obtained by sampling the k trees as described in Section 2. One can see that the probability of a fixed outcome π=(T1,,Tk)𝒯 is μ(π)=i=1kμi(Ti) and that π𝒯μ(π)=1.

Recall that technically, the subgraphs C,T as well as the layers Ei depend on the combination of sampled trees π. We therefore write, e.g., Cπ to refer to the concrete cycle C that arises from sampling the trees Ti in π=(T1,,Tk). Now

𝔼[c(T)+c(zγ)] 𝔼[c(C)]+c(R)+𝔼[c(zγ)]
=π𝒯μ(π)(c(Cπ)+𝔼γ[c(Rπ)]+i=1πFγ(ηj)(12βηj)c(Ei,π))
=π𝒯μ(π)(c(Cπ)+i=1πPr[γηi](22βηj)g(ηi)c(Ei,π))
π𝒯μ(π)(c(Cπ)+g(σ0)c(Rπ)) (4)
=π𝒯μ(π)(c(Cπ)+g(σ0)(c(Tπ)c(Cπ)))
=π𝒯μ(π)(g(σ0)c(Tπ)(f(σ0)1)c(Cπ))
g(σ0)c(x)(g(σ0)1)c(C^)=2(1βσ0)c(x)(12βσ0)c(C^).

In (4) we used the fact from Lemma 14 that g(ηi) is maximized at ηi=σ0 and that the layers Ei,π partition the edge set of Rπ. Now recall that we may assume that c(C^)(αα^)optPCOTSP(αα^)c(x), which yields a bound of

𝔼[c(T)+c(zγ)] 2(1βσ0)c(x)(12βσ0)c(C^)
c(x)(22βσ0(12βσ0)(αα^))
=c(x)(22βσ0α+α^+2αβσ02α^βσ0)
=c(x)(2+α^α(2+2α^)βσ0+2αβσ0).

We will next analyze the cost of FP, in Lemma 16. The proof builds on the following theorem.

Theorem 15 (Böhm et al. [13]).

Let XUV. Furthermore, let SUX be a randomly chosen subset such that Pr[vS]ρ for each vUX. Let FX and FXS denote the cheapest X-rooted spanning forest and the cheapest (XS)-rooted spanning forest of U respectively. Then 𝔼[c(FXS)]ρc(FX).

Lemma 16.

For a fixed value of σ[σ0,1], the cost of FP is at most eσσc(x).

Proof.

We invoke Theorem 15 with U=Vσ, X=O and S=(VσO)V(T). Recall that our pickup forest FP is the cheapest VσV(T) rooted spanning forest of Vσ and observe that VσV(T)=XS. By Lemma 10, each vertex vUX is not sampled (and therefore not in S) with probability at most eσ. It remains to show that the cost of the cheapest O-rooted spanning forest of Vσ is at most c(x)σ.

As we have already observed in Section 2, the cost of the cheapest O-rooted spanning forest of Vσ is equal to the cost of a minimum spanning tree in the graph obtained by contracting all vertices of O into a single vertex r, i.e., on V′′=(VσO){r}. Furthermore, we have also observed that by applying the very same contraction to our solution (x,y), we obtain a feasible solution (x,y) to the (non-ordered) PCTSP relaxation. By splitting off all vertices in VVσ and scaling up by a factor of 1σ, we obtain a vector x′′ for which x(δ(s))2 for all SV′′, i.e., a feasible point in the dominant888It is possible to obtain a feasible point in the polytope itself by applying a sequence of splitting-off operations. of the subtours elimination polytope of V′′.

The cost of the minimum spanning tree on V′′ is therefore at most c(x′′)1σc(x), which concludes the proof. Equipped with this upper bound, we can now bound the expected cost of FP, utilizing the density function fσ(y)=ddyFσ(y)=αyey and integrating over the range [σ0,1] from which σ is drawn:

𝔼[c(FP)]=c(x)σ01fσ(y)eyy𝑑y=αc(x)σ011𝑑y=α(1σ0)c(x)=eσ0c(x). (5)

For the last equality we have used the definition of σ0. We emphasize that by randomizing the choice of σ instead of flatly using σ=σ0, we have gained a factor of σ0. In a similar way, we can use Lemma 16 to compute the expected cost of zσ.

Lemma 17.

𝔼[c(zσ)]=αc(x)(14βσ0+βσ02).

Proof.

Note that when σ>12β, then zσ is by definition the zero-vector. We compute

𝔼[c(zσ)] =𝔼[max{0,12βσ}c(χFP)]=c(x)σ012βfσ(y)ey(12βy)y𝑑y
=αc(x)σ012β(12βy)𝑑y=αc(x)(12βσ02βσ012βy𝑑y)
=αc(x)(12βσ02β(18β2σ022))=αc(x)(14βσ0+βσ02).

Finally, we are able to combine the bounds on the various parts of c(H), and obtain our final upper bound on the expected tour cost:

𝔼[c(H)] 𝔼[c(T)]+𝔼[c(FP)]+𝔼[c(J)]
=𝔼[c(T)+c(zγ)]+𝔼[c(FP)]+𝔼[c(zσ)]+βc(x)
c(x)(2+α^+βα(2+2α^)βσ0+2αβσ0+eσ0+α4βασ0+αβσ02).

Note that the values of β and σ0 are functions of α. We therefore can express our upper bound as 𝔼[c(H)]c(x)f(α,α^). By Lemma 9, Algorithm 1 pays at most α times the penalty incurred by (x,y). Running Algorithm 1 with θ and σ determined by the single parameter α thus yields an approximation factor of max(α,f(α,α^)). Recall that the value of α^ is currently slightly below 1.599. For α^=1.599, the term max(α,f(α,α^)) is minimized at α2.096896<2.097. In fact, if we set α=2.097, the term evaluates to 2.097.

For α=2.097 and α^=1.599, we have β0.548775 and σ00.781790. Thus we have proven Theorem 3.

4 Prize-collecting Multi-Path TSP

PC-Multi-Path-TSP can be modeled as the following linear program.

minimizeeEi=1kcexi,e+vV𝒯πv(1yv)subject to yv=i=1kyi,vvV(xi,yi) lies in the si-ti-stroll relaxationi[k](kLP)

Similar to [13], we describe two algorithms for the problem and show that an appropriate combination of the two algorithms gives an 2.41-approximation.

In one algorithm, (which we call Algorithm A), we first sample k trees using Lemma 6 and an optimal solution (x,y) of (kLP), where tree Ti connects terminals si,ti. The remaining vertices are picked up in a probabilistic fashion akin to Algorithm 1, i.e., we choose a random threshold σ[σ0,1] and buy a (Vσi=1kV(Ti))-rooted spanning forest FP of Vσ. Here, σ0 is a constant whose value will be determined later. Then we double every edge that does not, for any i, lie on the si-ti path in Ti. This gives a tour HA. Define Δ:=c(si,ti) and ηΔc(x). Then it is easy to see that

c(HA) 2c(x)+2𝔼[c(FP)]Δ.

One can already see that intuitively, this yields good results whenever Δ is large.

Now we define the distribution from which σ is drawn. We choose σ s.t. Pr[σy]=Fσ(y) where Fσ(y)=1ey(1y)eσ0(1σ0). Note that except for the constant σ0, this is exactly the same distribution that we used in Algorithm 1. In fact, if we define ρeσ01σ0, we obtain Fσ(y)=1ρ(1y)ey (compare this to Fσ(y)=1α(1y)ey), so any result about Fσ obtained in the previous setting carries over to Fσ if we replace α by ρ. We remark that we use the symbol ρ instead of α because unlike in the previous case, ρ will not be our final approximation factor.

We can thus bound the expected cost of FP in the same way as we did for PCOTSP. This is possible because the distribution Fσ as well as the (lower bound on the) probability of a vertex vVσ being sampled at least once are the same as in Section 3.3.

First, we prove an equivalent of Lemma 16, i.e., we show that c(FP)=eσσ for any fixed value of σ, and then we compute the expected cost 𝔼[c(FP)]=eσ0 by integrating over the range [σ0,1] from which σ is chosen as we have done in Equation 5. This gives us the following upper bound on the expected tour cost:

c(HA) (2+2eσ0η)c(x).

Furthermore, by similar reasoning as in Section 3.1, we know that the expected total penalty paid by this algorithm is at most ρ=eσ01σ0 times the fractional penalty cost incurred by (x,y).

Now we give a simple Algorithm B that works well when η is small. Contract all the 2k terminals into one mega-vertex w with penalty , solve the PCTSP LP for this instance, and sample a single tree T from the solution, using Lemma 6. In the original graph, T corresponds to a 𝒯-rooted forest of cost at most c(x) that contains each vertex v with probability at least yv. Now double every edge of T (obtained in the original graph), and then add the k edges {si,ti} to get a final solution HB. It is easy to see that c(HB)(2+η)c(x) and that the incurred penalty is no higher than the fractional penalty cost of (x,y).

Running both algorithms A and B and returning each solution with probability 12, yields a tour of expected cost c(HA)+c(HB)2(2+eσ0)c(x) and an expected total penalty cost of at most

12(eσ01σ0+1)vV𝒯(1yv).

The approximation ratio that we get from combining both algorithms is

max{2+eσ0,12(eσ01σ0+1)},

which is minimized at σ00.892769. This yields an approximation guarantee slightly below 2.41, proving Theorem 4. We remark that both algorithm A and algorithm B can be derandomized in the same fashion as Algorithm 1.

5 Derandomization

In this section we discuss how our algorithm can be derandomized. Note that we used randomization for the choice of the thresholds σ and γ, as well as for sampling the trees T1,,Tk.

For the thresholds, we follow the approach of [8]: since the thresholds are used to prune (pick up) vertices whose y-value is below (above) the respective threshold, we may generate all possible outcomes by running our algorithm once for each pair of values γ,σ{0,1}{yvvV}.

The sampling of the trees T1,,Tk can be derandomized using the method of conditional expectations, as it is done in [4].

The basic idea is to iteratively fix the trees TiTi for i=1,,k, while minimizing the conditional expectation

𝔼σ,γ,Ti+1,,Tk[c(T)+c(Fp)+c(J)T1=T1,,Ti=Ti]

at each step. The expected conditional costs of T and Fp can be computed in a similar manner as it is done in [4], whereas the expected conditional cost of J can be bounded by a linear combination of the expected conditional costs of T and Fp, and the costs of x and C^. Thus, the conditional expectation can be computed efficiently each round.

6 Discussion

The PCOTSP, as a generalization of both PCTSP and OTSP, poses the challenges of each of the individual problems plus new challenges. The best approximation algorithms for OTSP ([13, 4]) both pick up vertices which have been left out of sampling, which imposes a cost of 1e over the solution. In the latest PCTSP result ([8]), the cost of parity correction is slightly below 0.6. So even if we simply add up the overheads of the two problems for parity correction and vertex pickup, the approximation factor would be close to two. But a straightforward application of the techniques from the latest results on Ordered TSP and Prize-Collecting TSP to PCOTSP actually leads to an approximation factor much higher than 2. This is because in PCOTSP the sampling probability for each vertex is lower than both PCTSP and OTSP; this makes pickup more expensive. Together with the need for parity correction for the picked up vertices, this also makes parity correction more costly than PCTSP.

It is not difficult to see that for the special cases of PCTSP, OTSP, our algorithms produce the best previously-known approximation factors for these problems. For example, when k=1, PCOTSP is simply PCTSP. In this case, the cycle length over the one vertex is zero, and the output of our algorithm would be no worse than the output PCTSP algorithm of [8] on the remaining vertices. Likewise, setting all penalties to turns PCOTSP into OTSP (and thus all y values are 1). Thus all vertices that have not been sampled will be picked up (i.e., σ=1), and no vertex is split off. This is equivalent to the algorithms of [13] and [4]. In a similar vein, setting all penalties to turns PC-Multi-Path-TSP into Multi-Path-TSP, and here our algorithm would do the same as the factor 2.367 algorithm of [13]. Their factor 2.2131 algorithm, however, does not directly carry over to our setting. The issue is that in the prize collecting setting, we require a picking-up step which leads to additional costs exceeding the additional gains as soon as we have to sample more than one tree.

The distributions used in this article have been carefully balanced to achieve the target approximation factor; it seems unlikely that their further tuning leads to better factors. It is an intriguing open question whether the problem admits an approximation factor of 2 or below, which we believe requires improvements of at least one of PCTSP or OTSP.

References

  • [1] Ali Ahmadi, Iman Gholami, MohammadTaghi Hajiaghayi, Peyman Jabbarzade, and Mohammad Mahdavi. 2-approximation for prize-collecting steiner forest. In Proceedings of the 2024 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 669–693. SIAM, 2024. doi:10.1137/1.9781611977912.25.
  • [2] Ali Ahmadi, Iman Gholami, MohammadTaghi Hajiaghayi, Peyman Jabbarzade, and Mohammad Mahdavi. Prize-collecting steiner tree: A 1.79 approximation. In Proceedings of the 56th Annual ACM Symposium on Theory of Computing, pages 1641–1652, 2024. doi:10.1145/3618260.3649789.
  • [3] Aaron Archer, MohammadHossein Bateni, MohammadTaghi Hajiaghayi, and Howard Karloff. Improved approximation algorithms for prize-collecting Steiner tree and TSP. SIAM journal on computing, 40(2):309–332, 2011. doi:10.1137/090771429.
  • [4] Susanne Armbruster, Matthias Mnich, and Martin Nägele. A (3/2+ 1/e)-approximation algorithm for ordered TSP. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2024). Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024.
  • [5] Jørgen Bang-Jensen, András Frank, and Bill Jackson. Preserving and increasing local edge-connectivity in mixed graphs. SIAM Journal on Discrete Mathematics, 8(2):155–178, 1995. doi:10.1137/S0036142993226983.
  • [6] Gerardo Berbeglia, Jean-François Cordeau, Irina Gribkovskaia, and Gilbert Laporte. Static pickup and delivery problems: a classification scheme and survey. Top, 15:1–31, 2007.
  • [7] Daniel Bienstock, Michel X Goemans, David Simchi-Levi, and David Williamson. A note on the prize collecting traveling salesman problem. Mathematical programming, 59(1):413–420, 1993. doi:10.1007/BF01581256.
  • [8] Jannis Blauth, Nathan Klein, and Martin Nägele. A better-than-1.6-approximation for prize-collecting TSP. In International Conference on Integer Programming and Combinatorial Optimization, pages 28–42. Springer, 2024. doi:10.1007/978-3-031-59835-7_3.
  • [9] Jannis Blauth and Martin Nägele. An improved approximation guarantee for prize-collecting TSP. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing, pages 1848–1861, 2023. doi:10.1145/3564246.3585159.
  • [10] Hans-Joachim Böckenhauer, Juraj Hromkovič, Joachim Kneis, and Joachim Kupke. On the approximation hardness of some generalizations of TSP. In Algorithm Theory–SWAT 2006: 10th Scandinavian Workshop on Algorithm Theory, Riga, Latvia, July 6-8, 2006. Proceedings 10, pages 184–195. Springer, 2006. doi:10.1007/11785293_19.
  • [11] Hans-Joachim Böckenhauer, Tobias Mömke, and Monika Steinová. Improved approximations for TSP with simple precedence constraints. J. Discrete Algorithms, 21:32–40, 2013. doi:10.1016/J.JDA.2013.04.002.
  • [12] Martin Böhm, Zachary Friggstad, Tobias Mömke, and Joachim Spoerhase. Approximating TSP variants using a bridge lemma. arXiv preprint arXiv:2405.12876, 2024. doi:10.48550/arXiv.2405.12876.
  • [13] Martin Böhm, Zachary Friggstad, Tobias Mömke, and Joachim Spoerhase. Approximating traveling salesman problems using a bridge lemma. In Proceedings of the 2025 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA). SIAM, 2025.
  • [14] Moses Charikar and Balaji Raghavachari. The finite capacity dial-a-ride problem. In Proceedings 39th Annual Symposium on Foundations of Computer Science (Cat. No. 98CB36280), pages 458–467. IEEE, 1998. doi:10.1109/SFCS.1998.743496.
  • [15] Nicos Christofides. Worst-case analysis of a new heuristic for the travelling salesman problem. Operations Research Forum, 3(1):20, 2022. doi:10.1007/S43069-021-00101-Z.
  • [16] Jean-François Cordeau and Gilbert Laporte. The dial-a-ride problem: models and algorithms. Annals of operations research, 153:29–46, 2007. doi:10.1007/S10479-007-0170-8.
  • [17] Michel X Goemans. Combining approximation algorithms for the prize-collecting TSP. arXiv preprint arXiv:0910.0553, 2009. arXiv:0910.0553.
  • [18] Michel X Goemans and David P Williamson. A general approximation technique for constrained forest problems. SIAM Journal on Computing, 24(2):296–317, 1995. doi:10.1137/S0097539793242618.
  • [19] Inge Li Gørtz, Viswanath Nagarajan, and Ramamoorthi Ravi. Minimum makespan multi-vehicle dial-a-ride. ACM Transactions on Algorithms (TALG), 11(3):1–29, 2015. doi:10.1145/2629653.
  • [20] Anna R Karlin, Nathan Klein, and Shayan Oveis Gharan. A (slightly) improved approximation algorithm for metric TSP. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, pages 32–45, 2021. doi:10.1145/3406325.3451009.
  • [21] Ian Post and Chaitanya Swamy. Linear programming-based approximation algorithms for multi-vehicle minimum latency problems (extended abstract). In Piotr Indyk, editor, Proceedings of the Twenty-Sixth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2015, San Diego, CA, USA, January 4-6, 2015, pages 512–531. SIAM, 2015. doi:10.1137/1.9781611973730.35.
  • [22] Martin WP Savelsbergh and Marc Sol. The general pickup and delivery problem. Transportation science, 29(1):17–29, 1995. doi:10.1287/TRSC.29.1.17.
  • [23] AI Serdjukov. Some extremal bypasses in graphs [in russian]. Upravlyaemye Sistemy, 17(89):76–79, 1978.
  • [24] Vera Traub and Jens Vygen. Approximation Algorithms for Traveling Salesman Problems. Cambridge University Press, 2024. URL: https://books.google.de/books?id=o5jV0AEACAAJ.
  • [25] Vera Traub, Jens Vygen, and Rico Zenklusen. Reducing path TSP to TSP. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 14–27, 2020. doi:10.1145/3357713.3384256.