posted on 1971-01-01, 00:00authored byAvrim Blum, Shuchi Chawla, David R. Karger, Terran Lane, Adam Meyerson, Maria Minkoff
In this paper, we give the first constant-factor approximation algorithm for the rooted ORIENTEERING
problem, as well as a new problem that we call the DISCOUNTED-REWARD-TSP, motivated by
robot navigation. In both problems, we are given a graph with lengths on edges and rewards on nodes,
and a start node s. In the ORIENTEERING problem, the goal is to find a path starting at s that maximizes
the reward collected, subject to a hard limit on the total length of the path. In the DISCOUNTED-REWARD-TSP, instead of a length limit we are given a discount factor γ, and the goal is to maximize
total discounted reward collected, where reward for a node reached at time t is discounted by γt. This
problem is motivated by an approximation to a planning problem in the Markov decision process (MDP)
framework under the commonly employed infinite horizon discounted reward optimality criterion. The
approximation arises from a need to deal with exponentially large state spaces that emerge when trying
to model one-time events and non-repeatable rewards (such as for package deliveries). We also consider
tree and multiple-path variants of these problems and provide approximations for those as well.
Although the unrooted ORIENTEERING problem, where there is no fixed start node s, has been known
to be approximable using algorithms for related problems such as k-TSP (in which the amount of reward
to be collected is fixed and the total length is approximately minimized), ours is the first to approximate
the rooted question, solving an open problem [3, 1]. We complement our approximation result for
ORIENTEERING by showing that the problem is APX-hard.