41 research outputs found

    Ramsey-type theorems for metric spaces with applications to online problems

    Get PDF
    A nearly logarithmic lower bound on the randomized competitive ratio for the metrical task systems problem is presented. This implies a similar lower bound for the extensively studied k-server problem. The proof is based on Ramsey-type theorems for metric spaces, that state that every metric space contains a large subspace which is approximately a hierarchically well-separated tree (and in particular an ultrametric). These Ramsey-type theorems may be of independent interest.Comment: Fix an error in the metadata. 31 pages, 0 figures. Preliminary version in FOCS '01. To be published in J. Comput. System Sc

    Parametrized Metrical Task Systems

    Get PDF
    We consider parametrized versions of metrical task systems and metrical service systems, two fundamental models of online computing, where the constrained parameter is the number of possible distinct requests m. Such parametrization occurs naturally in a wide range of applications. Striking examples are certain power management problems, which are modeled as metrical task systems with m = 2. We characterize the competitive ratio in terms of the parameter m for both deterministic and randomized algorithms on hierarchically separated trees. Our findings uncover a rich and unexpected picture that differs substantially from what is known or conjectured about the unparametrized versions of these problems. For metrical task systems, we show that deterministic algorithms do not exhibit any asymptotic gain beyond one-level trees (namely, uniform metric spaces), whereas randomized algorithms do not exhibit any asymptotic gain even for one-level trees. In contrast, the special case of metrical service systems (subset chasing) behaves very differently. Both deterministic and randomized algorithms exhibit gain, for m sufficiently small compared to n, for any number of levels. Most significantly, they exhibit a large gain for uniform metric spaces and a smaller gain for two-level trees. Moreover, it turns out that in these cases (as well as in the case of metrical task systems for uniform metric spaces with m being an absolute constant), deterministic algorithms are essentially as powerful as randomized algorithms. This is surprising and runs counter to the ubiquitous intuition/conjecture that, for most problems that can be modeled as metrical task systems, the randomized competitive ratio is polylogarithmic in the deterministic competitive ratio

    k-server via multiscale entropic regularization

    Full text link
    We present an O((logk)2)O((\log k)^2)-competitive randomized algorithm for the kk-server problem on hierarchically separated trees (HSTs). This is the first o(k)o(k)-competitive randomized algorithm for which the competitive ratio is independent of the size of the underlying HST. Our algorithm is designed in the framework of online mirror descent where the mirror map is a multiscale entropy. When combined with Bartal's static HST embedding reduction, this leads to an O((logk)2logn)O((\log k)^2 \log n)-competitive algorithm on any nn-point metric space. We give a new dynamic HST embedding that yields an O((logk)3logΔ)O((\log k)^3 \log \Delta)-competitive algorithm on any metric space where the ratio of the largest to smallest non-zero distance is at most Δ\Delta

    Online Metric Allocation and Time-Varying Regularization

    Get PDF
    We introduce a general online allocation problem that connects several of the most fundamental problems in online optimization. Let be an -point metric space. Consider a resource that can be allocated in arbitrary fractions to the points of . At each time , a convex monotone cost function : [0, 1] → ℝ+ appears at some point ∈ . In response, an algorithm may change the allocation of the resource, paying movement cost as determined by the metric and service cost ( ), where is the fraction of the resource at at the end of time . For example, when the cost functions are () = , this is equivalent to randomized MTS, and when the cost functions are () = ∞·<1/, this is equivalent to fractional -server. Because of an inherent scale-freeness property of the problem, existing techniques for MTS and -server fail to achieve similar guarantees for metric allocation. To handle this, we consider a generalization of the online multiplicative update method where we decouple the rate at which a variable is updated from its value, resulting in interesting new dynamics. We use this to give an (log)-competitive algorithm for weighted star metrics. We then show how this corresponds to an extension of the online mirror descent framework to a setting where the regularizer is time-varying. Using this perspective, we further refine the guarantees of our algorithm. We also consider the case of non-convex cost functions. Using a simple ₂²-regularizer, we give tight bounds of Θ() on tree metrics, which imply deterministic and randomized competitive ratios of (2) and ( log ) respectively on arbitrary metrics

    Stochastic k-Server: How Should Uber Work?

    Get PDF
    In this paper we study a stochastic variant of the celebrated kk-server problem. In the k-server problem, we are required to minimize the total movement of k servers that are serving an online sequence of tt requests in a metric. In the stochastic setting we are given t independent distributions in advance, and at every time step i a request is drawn from P_i. Designing the optimal online algorithm in such setting is NP-hard, therefore the emphasis of our work is on designing an approximately optimal online algorithm. We first show a structural characterization for a certain class of non-adaptive online algorithms. We prove that in general metrics, the best of such algorithms has a cost of no worse than three times that of the optimal online algorithm. Next, we present an integer program that finds the optimal algorithm of this class for any arbitrary metric. Finally by rounding the solution of the linear relaxation of this program, we present an online algorithm for the stochastic k-server problem with an approximation factor of 33 in the line and circle metrics and factor of O(log n) in general metrics. In this way, we achieve an approximation factor that is independent of k, the number of servers. Moreover, we define the Uber problem, motivated by extraordinary growth of online network transportation services. In the Uber problem, each demand consists of two points -a source and a destination- in the metric. Serving a demand is to move a server to its source and then to its destination. The objective is again minimizing the total movement of the k given servers. It is not hard to show that given an alpha-approximation algorithm for the k-server problem, we can obtain a max{3,alpha}-approximation algorithm for the Uber problem. Motivated by the fact that demands are usually highly correlated with the time (e.g. what day of the week or what time of the day the demand is arrived), we study the stochastic Uber problem. Using our results for stochastic k-server we can obtain a 3-approximation algorithm for the stochastic Uber problem in line and circle metrics, and a O(log n)-approximation algorithm for a general metric of size n. Furthermore, we extend our results to the correlated setting where the probability of a request arriving at a certain point depends not only on the time step but also on the previously arrived requests

    Multiscale Entropic Regularization for MTS on General Metric Spaces

    Get PDF
    We present an O((logn)2)O((\log n)^2)-competitive algorithm for metrical task systems (MTS) on any nn-point metric space that is also 11-competitive for service costs. This matches the competitive ratio achieved by Bubeck, Cohen, Lee, and Lee (2019) and the refined competitive ratios obtained by Coester and Lee (2019). Those algorithms work by first randomly embedding the metric space into an ultrametric and then solving MTS there. In contrast, our algorithm is cast as regularized gradient descent where the regularizer is a multiscale metric entropy defined directly on the metric space. This answers an open question of Bubeck (Highlights of Algorithms, 2019).Comment: 23 pages, 1 figure, to appear in ITCS '2
    corecore