13,926 research outputs found

    Location models in the public sector

    Get PDF
    The past four decades have witnessed an explosive growth in the field of networkbased facility location modeling. This is not at all surprising since location policy is one of the most profitable areas of applied systems analysis in regional science and ample theoretical and applied challenges are offered. Location-allocation models seek the location of facilities and/or services (e.g., schools, hospitals, and warehouses) so as to optimize one or several objectives generally related to the efficiency of the system or to the allocation of resources. This paper concerns the location of facilities or services in discrete space or networks, that are related to the public sector, such as emergency services (ambulances, fire stations, and police units), school systems and postal facilities. The paper is structured as follows: first, we will focus on public facility location models that use some type of coverage criterion, with special emphasis in emergency services. The second section will examine models based on the P-Median problem and some of the issues faced by planners when implementing this formulation in real world locational decisions. Finally, the last section will examine new trends in public sector facility location modeling.Location analysis, public facilities, covering models

    Datacenter Traffic Control: Understanding Techniques and Trade-offs

    Get PDF
    Datacenters provide cost-effective and flexible access to scalable compute and storage resources necessary for today's cloud computing needs. A typical datacenter is made up of thousands of servers connected with a large network and usually managed by one operator. To provide quality access to the variety of applications and services hosted on datacenters and maximize performance, it deems necessary to use datacenter networks effectively and efficiently. Datacenter traffic is often a mix of several classes with different priorities and requirements. This includes user-generated interactive traffic, traffic with deadlines, and long-running traffic. To this end, custom transport protocols and traffic management techniques have been developed to improve datacenter network performance. In this tutorial paper, we review the general architecture of datacenter networks, various topologies proposed for them, their traffic properties, general traffic control challenges in datacenters and general traffic control objectives. The purpose of this paper is to bring out the important characteristics of traffic control in datacenters and not to survey all existing solutions (as it is virtually impossible due to massive body of existing research). We hope to provide readers with a wide range of options and factors while considering a variety of traffic control mechanisms. We discuss various characteristics of datacenter traffic control including management schemes, transmission control, traffic shaping, prioritization, load balancing, multipathing, and traffic scheduling. Next, we point to several open challenges as well as new and interesting networking paradigms. At the end of this paper, we briefly review inter-datacenter networks that connect geographically dispersed datacenters which have been receiving increasing attention recently and pose interesting and novel research problems.Comment: Accepted for Publication in IEEE Communications Surveys and Tutorial

    Rail-Road terminal locations: aggregation errors and best potential locations on large networks

    Get PDF
    In network location problems, the number of potential locations is often too large in order to find a solution in a reasonable computing time. That is why aggregation techniques are often used to reduce the number of nodes. This reduction of the size of the location problems makes them more computationally tractable, but aggregation introduces errors into the solutions. Some of these errors will be estimated in this paper. A method that helps to isolate the best potential locations for rail-road terminals embedded in a hub-and-spoke network will further be outlined. Hub location problems arise when it is desirable to consolidate flows at certain locations called hubs. The basic idea is to use the flows of commodities and their geographic spreading as input to determine a set of potential locations for hub terminals. The exercise will be done for the trans-European networks. These potential locations can then further be used as input by an optimal location method

    An Integrated Contraflow Strategy for Multimodal Evacuation

    Get PDF
    To improve the efficiency of multimodal evacuation, a network aggregation method and an integrated contraflow strategy are proposed in this paper. The network aggregation method indicates the uncertain evacuation demand on the arterial subnetwork and balances accuracy and efficiency by refining the local road subnetworks. The integrated contraflow strategy contains three arterial configurations: noncontraflow to shorten the strategy setup time, full-lane contraflow to maximize the evacuation network capacity, and bus contraflow to realize the transit cycle operation. The application of this strategy takes two steps to provide transit priority during evacuation: solve the transit-based evacuation problem with a minimum-cost flow model, firstly, and then address the auto-based evacuation problem with a bilevel network flow model. The numerical results from optimizing an evacuation network for a super typhoon justify the validness and usefulness of the network aggregation method and the integrated contraflow strategy

    Boosting the Basic Counting on Distributed Streams

    Get PDF
    We revisit the classic basic counting problem in the distributed streaming model that was studied by Gibbons and Tirthapura (GT). In the solution for maintaining an (ϵ,δ)(\epsilon,\delta)-estimate, as what GT's method does, we make the following new contributions: (1) For a bit stream of size nn, where each bit has a probability at least γ\gamma to be 1, we exponentially reduced the average total processing time from GT's Θ(nlog(1/δ))\Theta(n \log(1/\delta)) to O((1/(γϵ2))(log2n)log(1/δ))O((1/(\gamma\epsilon^2))(\log^2 n) \log(1/\delta)), thus providing the first sublinear-time streaming algorithm for this problem. (2) In addition to an overall much faster processing speed, our method provides a new tradeoff that a lower accuracy demand (a larger value for ϵ\epsilon) promises a faster processing speed, whereas GT's processing speed is Θ(nlog(1/δ))\Theta(n \log(1/\delta)) in any case and for any ϵ\epsilon. (3) The worst-case total time cost of our method matches GT's Θ(nlog(1/δ))\Theta(n\log(1/\delta)), which is necessary but rarely occurs in our method. (4) The space usage overhead in our method is a lower order term compared with GT's space usage and occurs only O(logn)O(\log n) times during the stream processing and is too negligible to be detected by the operating system in practice. We further validate these solid theoretical results with experiments on both real-world and synthetic data, showing that our method is faster than GT's by a factor of several to several thousands depending on the stream size and accuracy demands, without any detectable space usage overhead. Our method is based on a faster sampling technique that we design for boosting GT's method and we believe this technique can be of other interest.Comment: 32 page

    p-Median problems in a fuzzy environment

    Get PDF
    In this paper a formulation for the fuzzy p-median model in a fuzzy environment is presented. The model allows to find optimal locations of p facilities and their related cost when data related to the node demands and the edge distances are imprecise and uncertain and also to know the degree of certainty of the solution. For the sake of illustration, the proposed model is applied in a reduced map of Kinshasa (Democratic Republic of Congo) obtaining results which are rather than realistic one

    Improved Algorithms for Time Decay Streams

    Get PDF
    In the time-decay model for data streams, elements of an underlying data set arrive sequentially with the recently arrived elements being more important. A common approach for handling large data sets is to maintain a coreset, a succinct summary of the processed data that allows approximate recovery of a predetermined query. We provide a general framework that takes any offline-coreset and gives a time-decay coreset for polynomial time decay functions. We also consider the exponential time decay model for k-median clustering, where we provide a constant factor approximation algorithm that utilizes the online facility location algorithm. Our algorithm stores O(k log(h Delta)+h) points where h is the half-life of the decay function and Delta is the aspect ratio of the dataset. Our techniques extend to k-means clustering and M-estimators as well
    corecore