306 research outputs found

    A Bicriteria Approximation for the Reordering Buffer Problem

    Full text link
    In the reordering buffer problem (RBP), a server is asked to process a sequence of requests lying in a metric space. To process a request the server must move to the corresponding point in the metric. The requests can be processed slightly out of order; in particular, the server has a buffer of capacity k which can store up to k requests as it reads in the sequence. The goal is to reorder the requests in such a manner that the buffer constraint is satisfied and the total travel cost of the server is minimized. The RBP arises in many applications that require scheduling with a limited buffer capacity, such as scheduling a disk arm in storage systems, switching colors in paint shops of a car manufacturing plant, and rendering 3D images in computer graphics. We study the offline version of RBP and develop bicriteria approximations. When the underlying metric is a tree, we obtain a solution of cost no more than 9OPT using a buffer of capacity 4k + 1 where OPT is the cost of an optimal solution with buffer capacity k. Constant factor approximations were known previously only for the uniform metric (Avigdor-Elgrabli et al., 2012). Via randomized tree embeddings, this implies an O(log n) approximation to cost and O(1) approximation to buffer size for general metrics. Previously the best known algorithm for arbitrary metrics by Englert et al. (2007) provided an O(log^2 k log n) approximation without violating the buffer constraint.Comment: 13 page

    Selfish traffic allocation for server farms

    Get PDF
    We study the price of selfish routing in noncooperative networks like the Internet. In particular, we investigate the price of selfish routing using the price of anarchy (a.k.a. the coordination ratio) and other (e.g., bicriteria) measures in the recently introduced game theoretic parallel links network model of Koutsoupias and Papadimitriou. We generalize this model toward general, monotone families of cost functions and cost functions from queueing theory. A summary of our main results for general, monotone cost functions is as follows: 1. We give an exact characterization of all cost functions having a bounded/unbounded price of anarchy. For example, the price of anarchy for cost functions describing the expected delay in queueing systems is unbounded. 2. We show that an unbounded price of anarchy implies an extremely high performance degradation under bicriteria measures. In fact, the price of selfish routing can be as high as a bandwidth degradation by a factor that is linear in the network size. 3. We separate the game theoretic (integral) allocation model from the (fractional) flow model by demonstrating that even a very small or negligible amount of integrality can lead to a dramatic performance degradation. 4. We unify recent results on selfish routing under different objectives by showing that an unbounded price of anarchy under the min-max objective implies an unbounded price of anarchy under the average cost objective and vice versa. Our special focus lies on cost functions describing the behavior of Web servers that can open only a limited number of Transmission Control Protocol (TCP) connections. In particular, we compare the performance of queueing systems that serve all incoming requests with servers that reject requests in case of overload. Our analysis indicates that all queueing systems without rejection cannot give any reasonable guarantee on the expected delay of requests under selfish routing even when the injected load is far away from the capacity of the system. In contrast, Web server farms that are allowed to reject requests can guarantee a high quality of service for every individual request stream even under relatively high injection rates

    Serve or Skip: The Power of Rejection in Online Bottleneck Matching

    Get PDF
    We consider the online matching problem, where n server-vertices lie in a metric space and n request-vertices that arrive over time each must immediately be permanently assigned to a server-vertex.We focus on the egalitarian bottleneck objective, where the goal is to minimize the maximum distance between any request and its server. It has been demonstrated that while there are effective algorithms for the utilitarian objective (minimizing total cost) in the resource augmentation setting where the offline adversary has half the resources, these are not effective for the egalitarian objective. Thus, we propose a new Serve-or-Skip bicriteria analysis model, where the online algorithm may reject or skip up to a specified number of requests, and propose two greedy algorithms: GRI NN(t) and GRIN(t) . We show that the Serve-or-Skip model of resource augmentation analysis can essentially simulate the doubled-server capacity model, and then examine the performance of GRI NN(t) and GRIN(t)

    Incremental Medians via Online Bidding

    Full text link
    In the k-median problem we are given sets of facilities and customers, and distances between them. For a given set F of facilities, the cost of serving a customer u is the minimum distance between u and a facility in F. The goal is to find a set F of k facilities that minimizes the sum, over all customers, of their service costs. Following Mettu and Plaxton, we study the incremental medians problem, where k is not known in advance, and the algorithm produces a nested sequence of facility sets where the kth set has size k. The algorithm is c-cost-competitive if the cost of each set is at most c times the cost of the optimum set of size k. We give improved incremental algorithms for the metric version: an 8-cost-competitive deterministic algorithm, a 2e ~ 5.44-cost-competitive randomized algorithm, a (24+epsilon)-cost-competitive, poly-time deterministic algorithm, and a (6e+epsilon ~ .31)-cost-competitive, poly-time randomized algorithm. The algorithm is s-size-competitive if the cost of the kth set is at most the minimum cost of any set of size k, and has size at most s k. The optimal size-competitive ratios for this problem are 4 (deterministic) and e (randomized). We present the first poly-time O(log m)-size-approximation algorithm for the offline problem and first poly-time O(log m)-size-competitive algorithm for the incremental problem. Our proofs reduce incremental medians to the following online bidding problem: faced with an unknown threshold T, an algorithm submits "bids" until it submits a bid that is at least the threshold. It pays the sum of all its bids. We prove that folklore algorithms for online bidding are optimally competitive.Comment: conference version appeared in LATIN 2006 as "Oblivious Medians via Online Bidding

    Enhanced genetic algorithm-based fuzzy multiobjective strategy to multiproduct batch plant design

    Get PDF
    This paper addresses the problem of the optimal design of batch plants with imprecise demands in product amounts. The design of such plants necessary involves how equipment may be utilized, which means that plant scheduling and production must constitute a basic part of the design problem. Rather than resorting to a traditional probabilistic approach for modeling the imprecision on product demands, this work proposes an alternative treatment by using fuzzy concepts. The design problem is tackled by introducing a new approach based on a multiobjective genetic algorithm, combined wit the fuzzy set theory for computing the objectives as fuzzy quantities. The problem takes into account simultaneous maximization of the fuzzy net present value and of two other performance criteria, i.e. the production delay/advance and a flexibility index. The delay/advance objective is computed by comparing the fuzzy production time for the products to a given fuzzy time horizon, and the flexibility index represents the additional fuzzy production that the plant would be able to produce. The multiobjective optimization provides the Pareto's front which is a set of scenarios that are helpful for guiding the decision's maker in its final choices. About the solution procedure, a genetic algorithm was implemented since it is particularly well-suited to take into account the arithmetic of fuzzy numbers. Furthermore because a genetic algorithm is working on populations of potential solutions, this type of procedure is well adapted for multiobjective optimization

    Online Bicriteria Load Balancing using Object Reallocation

    Get PDF
    Cataloged from PDF version of article.We study the bicriteria load balancing problem on two independent parameters under the allowance of object reallocation. The scenario is a system of M distributed file servers located in a cluster, and we propose three online approximate algorithms for balancing their loads and required storage spaces during document placement. The first algorithm is for heterogeneous servers. Each server has its individual trade-off of load and storage space under the same rule of selection. The other two algorithms are for homogeneous servers. The second algorithm combines the idea of the first one and the best existing solution for homogeneous servers. Using document reallocation, we obtain a smooth trade-off curve of the upper bounds of load and storage space. The last one bounds the load and storage space of each server by less than three times of their trivial lower bounds, respectively; and more importantly, for each server, the value of at least one parameter is far from its worst case. The time complexities of these three algorithms are O(log M) plus the cost of document reallocation

    Multiobjective genetic algorithm strategies for electricity production from generation IV nuclear technology

    Get PDF
    Development of a technico-economic optimization strategy of cogeneration systems of electricity/hydrogen, consists in finding an optimal efficiency of the generating cycle and heat delivery system, maximizing the energy production and minimizing the production costs. The first part of the paper is related to the development of a multiobjective optimization library (MULTIGEN) to tackle all types of problems arising from cogeneration. After a literature review for identifying the most efficient methods, the MULTIGEN library is described, and the innovative points are listed. A new stopping criterion, based on the stagnation of the Pareto front, may lead to significant decrease of computational times, particularly in the case of problems involving only integer variables. Two practical examples are presented in the last section. The former is devoted to a bicriteria optimization of both exergy destruction and total cost of the plant, for a generating cycle coupled with a Very High Temperature Reactor (VHTR). The second example consists in designing the heat exchanger of the generating turbomachine. Three criteria are optimized: the exchange surface, the exergy destruction and the number of exchange modules
    corecore