9,322 research outputs found
Separation of timescales in a two-layered network
We investigate a computer network consisting of two layers occurring in, for
example, application servers. The first layer incorporates the arrival of jobs
at a network of multi-server nodes, which we model as a many-server Jackson
network. At the second layer, active servers at these nodes act now as
customers who are served by a common CPU. Our main result shows a separation of
time scales in heavy traffic: the main source of randomness occurs at the
(aggregate) CPU layer; the interactions between different types of nodes at the
other layer is shown to converge to a fixed point at a faster time scale; this
also yields a state-space collapse property. Apart from these fundamental
insights, we also obtain an explicit approximation for the joint law of the
number of jobs in the system, which is provably accurate for heavily loaded
systems and performs numerically well for moderately loaded systems. The
obtained results for the model under consideration can be applied to
thread-pool dimensioning in application servers, while the technique seems
applicable to other layered systems too.Comment: 8 pages, 2 figures, 1 table, ITC 24 (2012
Measuring and Understanding Throughput of Network Topologies
High throughput is of particular interest in data center and HPC networks.
Although myriad network topologies have been proposed, a broad head-to-head
comparison across topologies and across traffic patterns is absent, and the
right way to compare worst-case throughput performance is a subtle problem.
In this paper, we develop a framework to benchmark the throughput of network
topologies, using a two-pronged approach. First, we study performance on a
variety of synthetic and experimentally-measured traffic matrices (TMs).
Second, we show how to measure worst-case throughput by generating a
near-worst-case TM for any given topology. We apply the framework to study the
performance of these TMs in a wide range of network topologies, revealing
insights into the performance of topologies with scaling, robustness of
performance across TMs, and the effect of scattered workload placement. Our
evaluation code is freely available
Multiobjective Tactical Planning under Uncertainty for Air Traffic Flow and Capacity Management
We investigate a method to deal with congestion of sectors and delays in the
tactical phase of air traffic flow and capacity management. It relies on
temporal objectives given for every point of the flight plans and shared among
the controllers in order to create a collaborative environment. This would
enhance the transition from the network view of the flow management to the
local view of air traffic control. Uncertainty is modeled at the trajectory
level with temporal information on the boundary points of the crossed sectors
and then, we infer the probabilistic occupancy count. Therefore, we can model
the accuracy of the trajectory prediction in the optimization process in order
to fix some safety margins. On the one hand, more accurate is our prediction;
more efficient will be the proposed solutions, because of the tighter safety
margins. On the other hand, when uncertainty is not negligible, the proposed
solutions will be more robust to disruptions. Furthermore, a multiobjective
algorithm is used to find the tradeoff between the delays and congestion, which
are antagonist in airspace with high traffic density. The flow management
position can choose manually, or automatically with a preference-based
algorithm, the adequate solution. This method is tested against two instances,
one with 10 flights and 5 sectors and one with 300 flights and 16 sectors.Comment: IEEE Congress on Evolutionary Computation (2013). arXiv admin note:
substantial text overlap with arXiv:1309.391
Power efficient job scheduling by predicting the impact of processor manufacturing variability
Modern CPUs suffer from performance and power consumption variability due to the manufacturing process. As a result, systems that do not consider such variability caused by manufacturing issues lead to performance degradations and wasted power. In order to avoid such negative impact, users and system administrators must actively counteract any manufacturing variability.
In this work we show that parallel systems benefit from taking into account the consequences of manufacturing variability when making scheduling decisions at the job scheduler level. We also show that it is possible to predict the impact of this variability on specific applications by using variability-aware power prediction models. Based on these power models, we propose two job scheduling policies that consider the effects of manufacturing variability for each application and that ensure that power consumption stays under a system-wide power budget. We evaluate our policies under different power budgets and traffic scenarios, consisting of both single- and multi-node parallel applications, utilizing up to 4096 cores in total. We demonstrate that they decrease job turnaround time, compared to contemporary scheduling policies used on production clusters, up to 31% while saving up to 5.5% energy.Postprint (author's final draft
- …