13,926 research outputs found
Location models in the public sector
The past four decades have witnessed an explosive growth in the field of networkbased facility location modeling. This is not at all surprising since location policy is one of the most profitable areas of applied systems analysis in regional science and ample theoretical and applied challenges are offered. Location-allocation models seek the location of facilities and/or services (e.g., schools, hospitals, and warehouses) so as to optimize one or several objectives generally related to the efficiency of the system or to the allocation of resources. This paper concerns the location of facilities or services in discrete space or networks, that are related to the public sector, such as emergency services (ambulances, fire stations, and police units), school systems and postal facilities. The paper is structured as follows: first, we will focus on public facility location models that use some type of coverage criterion, with special emphasis in emergency services. The second section will examine models based on the P-Median problem and some of the issues faced by planners when implementing this formulation in real world locational decisions. Finally, the last section will examine new trends in public sector facility location modeling.Location analysis, public facilities, covering models
Datacenter Traffic Control: Understanding Techniques and Trade-offs
Datacenters provide cost-effective and flexible access to scalable compute
and storage resources necessary for today's cloud computing needs. A typical
datacenter is made up of thousands of servers connected with a large network
and usually managed by one operator. To provide quality access to the variety
of applications and services hosted on datacenters and maximize performance, it
deems necessary to use datacenter networks effectively and efficiently.
Datacenter traffic is often a mix of several classes with different priorities
and requirements. This includes user-generated interactive traffic, traffic
with deadlines, and long-running traffic. To this end, custom transport
protocols and traffic management techniques have been developed to improve
datacenter network performance.
In this tutorial paper, we review the general architecture of datacenter
networks, various topologies proposed for them, their traffic properties,
general traffic control challenges in datacenters and general traffic control
objectives. The purpose of this paper is to bring out the important
characteristics of traffic control in datacenters and not to survey all
existing solutions (as it is virtually impossible due to massive body of
existing research). We hope to provide readers with a wide range of options and
factors while considering a variety of traffic control mechanisms. We discuss
various characteristics of datacenter traffic control including management
schemes, transmission control, traffic shaping, prioritization, load balancing,
multipathing, and traffic scheduling. Next, we point to several open challenges
as well as new and interesting networking paradigms. At the end of this paper,
we briefly review inter-datacenter networks that connect geographically
dispersed datacenters which have been receiving increasing attention recently
and pose interesting and novel research problems.Comment: Accepted for Publication in IEEE Communications Surveys and Tutorial
Rail-Road terminal locations: aggregation errors and best potential locations on large networks
In network location problems, the number of potential locations is often too large in order to find a solution in a reasonable computing time. That is why aggregation techniques are often used to reduce the number of nodes. This reduction of the size of the location problems makes them more computationally tractable, but aggregation introduces errors into the solutions. Some of these errors will be estimated in this paper. A method that helps to isolate the best potential locations for rail-road terminals embedded in a hub-and-spoke network will further be outlined. Hub location problems arise when it is desirable to consolidate flows at certain locations called hubs. The basic idea is to use the flows of commodities and their geographic spreading as input to determine a set of potential locations for hub terminals. The exercise will be done for the trans-European networks. These potential locations can then further be used as input by an optimal location method
An Integrated Contraflow Strategy for Multimodal Evacuation
To improve the efficiency of multimodal evacuation, a network aggregation method and an integrated contraflow strategy are proposed in this paper. The network aggregation method indicates the uncertain evacuation demand on the arterial subnetwork and balances accuracy and efficiency by refining the local road subnetworks. The integrated contraflow strategy contains three arterial configurations: noncontraflow to shorten the strategy setup time, full-lane contraflow to maximize the evacuation network capacity, and bus contraflow to realize the transit cycle operation. The application of this strategy takes two steps to provide transit priority during evacuation: solve the transit-based evacuation problem with a minimum-cost flow model, firstly, and then address the auto-based evacuation problem with a bilevel network flow model. The numerical results from optimizing an evacuation network for a super typhoon justify the validness and usefulness of the network aggregation method and the integrated contraflow strategy
Boosting the Basic Counting on Distributed Streams
We revisit the classic basic counting problem in the distributed streaming
model that was studied by Gibbons and Tirthapura (GT). In the solution for
maintaining an -estimate, as what GT's method does, we make
the following new contributions: (1) For a bit stream of size , where each
bit has a probability at least to be 1, we exponentially reduced the
average total processing time from GT's to
, thus providing the first
sublinear-time streaming algorithm for this problem. (2) In addition to an
overall much faster processing speed, our method provides a new tradeoff that a
lower accuracy demand (a larger value for ) promises a faster
processing speed, whereas GT's processing speed is
in any case and for any . (3) The worst-case total time cost of our
method matches GT's , which is necessary but rarely
occurs in our method. (4) The space usage overhead in our method is a lower
order term compared with GT's space usage and occurs only times
during the stream processing and is too negligible to be detected by the
operating system in practice. We further validate these solid theoretical
results with experiments on both real-world and synthetic data, showing that
our method is faster than GT's by a factor of several to several thousands
depending on the stream size and accuracy demands, without any detectable space
usage overhead. Our method is based on a faster sampling technique that we
design for boosting GT's method and we believe this technique can be of other
interest.Comment: 32 page
p-Median problems in a fuzzy environment
In this paper a formulation for the fuzzy p-median model in a fuzzy environment
is presented. The model allows to find optimal locations of p facilities and their
related cost when data related to the node demands and the edge distances are
imprecise and uncertain and also to know the degree of certainty of the solution. For
the sake of illustration, the proposed model is applied in a reduced map of Kinshasa
(Democratic Republic of Congo) obtaining results which are rather than realistic
one
Improved Algorithms for Time Decay Streams
In the time-decay model for data streams, elements of an underlying data set arrive sequentially with the recently arrived elements being more important. A common approach for handling large data sets is to maintain a coreset, a succinct summary of the processed data that allows approximate recovery of a predetermined query. We provide a general framework that takes any offline-coreset and gives a time-decay coreset for polynomial time decay functions.
We also consider the exponential time decay model for k-median clustering, where we provide a constant factor approximation algorithm that utilizes the online facility location algorithm. Our algorithm stores O(k log(h Delta)+h) points where h is the half-life of the decay function and Delta is the aspect ratio of the dataset. Our techniques extend to k-means clustering and M-estimators as well
- …