17,870 research outputs found
Control-Based Resource Management Procedures for Satellite Networks
This paper describes the resource management of a DVBRCS
geostationary satellite network. The functional modules
of the access layer aim at efficiently exploiting the link
resources while assuring the contracted Quality of Service
(QoS) to the traffic entering the satellite network. The main
novelty is the integration between the Connection Admission
Control and the Congestion Control procedures. Both them
exploit the estimation of the traffic load, performed by a
Kalman filter. The proposed solution has been analysed via
computer simulations, which confirmed their effectiveness
Low-Latency Millimeter-Wave Communications: Traffic Dispersion or Network Densification?
This paper investigates two strategies to reduce the communication delay in
future wireless networks: traffic dispersion and network densification. A
hybrid scheme that combines these two strategies is also considered. The
probabilistic delay and effective capacity are used to evaluate performance.
For probabilistic delay, the violation probability of delay, i.e., the
probability that the delay exceeds a given tolerance level, is characterized in
terms of upper bounds, which are derived by applying stochastic network
calculus theory. In addition, to characterize the maximum affordable arrival
traffic for mmWave systems, the effective capacity, i.e., the service
capability with a given quality-of-service (QoS) requirement, is studied. The
derived bounds on the probabilistic delay and effective capacity are validated
through simulations. These numerical results show that, for a given average
system gain, traffic dispersion, network densification, and the hybrid scheme
exhibit different potentials to reduce the end-to-end communication delay. For
instance, traffic dispersion outperforms network densification, given high
average system gain and arrival rate, while it could be the worst option,
otherwise. Furthermore, it is revealed that, increasing the number of
independent paths and/or relay density is always beneficial, while the
performance gain is related to the arrival rate and average system gain,
jointly. Therefore, a proper transmission scheme should be selected to optimize
the delay performance, according to the given conditions on arrival traffic and
system service capability
Resource dimensioning through buffer sampling
Link dimensioning, i.e., selecting a (minimal) link capacity such that the users’ performance requirements are met, is a crucial component of network design. It requires insight into the interrelationship among the traffic offered (in terms of the mean offered load , but also its fluctuation around the mean, i.e., ‘burstiness’), the envisioned performance level, and the capacity needed. We first derive, for different performance criteria, theoretical dimensioning formulas that estimate the required capacity as a function of the input traffic and the performance target. For the special case of Gaussian input traffic, these formulas reduce to , where directly relates to the performance requirement (as agreed upon in a service level agreement) and reflects the burstiness (at the timescale of interest). We also observe that Gaussianity applies for virtually all realistic scenarios; notably, already for a relatively low aggregation level, the Gaussianity assumption is justified.\ud
As estimating is relatively straightforward, the remaining open issue concerns the estimation of . We argue that particularly if corresponds to small time-scales, it may be inaccurate to estimate it directly from the traffic traces. Therefore, we propose an indirect method that samples the buffer content, estimates the buffer content distribution, and ‘inverts’ this to the variance. We validate the inversion through extensive numerical experiments (using a sizeable collection of traffic traces from various representative locations); the resulting estimate of is then inserted in the dimensioning formula. These experiments show that both the inversion and the dimensioning formula are remarkably accurate
Dynamic Resource Management in Clouds: A Probabilistic Approach
Dynamic resource management has become an active area of research in the
Cloud Computing paradigm. Cost of resources varies significantly depending on
configuration for using them. Hence efficient management of resources is of
prime interest to both Cloud Providers and Cloud Users. In this work we suggest
a probabilistic resource provisioning approach that can be exploited as the
input of a dynamic resource management scheme. Using a Video on Demand use case
to justify our claims, we propose an analytical model inspired from standard
models developed for epidemiology spreading, to represent sudden and intense
workload variations. We show that the resulting model verifies a Large
Deviation Principle that statistically characterizes extreme rare events, such
as the ones produced by "buzz/flash crowd effects" that may cause workload
overflow in the VoD context. This analysis provides valuable insight on
expectable abnormal behaviors of systems. We exploit the information obtained
using the Large Deviation Principle for the proposed Video on Demand use-case
for defining policies (Service Level Agreements). We believe these policies for
elastic resource provisioning and usage may be of some interest to all
stakeholders in the emerging context of cloud networkingComment: IEICE Transactions on Communications (2012). arXiv admin note:
substantial text overlap with arXiv:1209.515
Resource dimensioning through buffer sampling
Link dimensioning, i.e., selecting a (minimal) link capacity such that the users’ performance requirements are met, is a crucial component of network design. It requires insight into the interrelationship between the traffic offered (in terms of the mean offered load M, but also its fluctuation around the mean, i.e., ‘burstiness’), the envisioned performance level, and the capacity needed. We first derive, for different performance criteria, theoretical dimensioning formulae that estimate the required capacity C as a function of the input traffic and the performance target. For the special case of Gaussian input traffic these formulae reduce to C = M+V , where directly relates to the performance requirement (as agreed upon in a service level agreement) and V reflects the burstiness (at the timescale of interest). We also observe that Gaussianity applies for virtually all realistic scenarios; notably, already for a relatively low aggregation level the Gaussianity assumption is justified.\ud
As estimating M is relatively straightforward, the remaining open issue concerns the estimation of V . We argue that, particularly if V corresponds to small time-scales, it may be inaccurate to estimate it directly from the traffic traces. Therefore, we propose an indirect method that samples the buffer content, estimates the buffer content distribution, and ‘inverts’ this to the variance. We validate the inversion through extensive numerical experiments (using a sizeable collection of traffic traces from various representative locations); the resulting estimate of V is then inserted in the dimensioning formula. These experiments show that both the inversion and the dimensioning formula are remarkably accurate
- …