7,341 research outputs found
Quality of service support in differentiated services packet networks
During the past few years, new types of Internet applications which require performance beyond the best-effort service that is provided by the current Internet have emerged. These applications include the transmission of voice and video, which require a fixed end-to-end delay bound in order for the end-user to perceive an acceptable level of service quality. The Differentiated Services (Diffserv) model has been proposed recently to enhance the traditional best-effort service, and provide certain Quality of Serviee (QoS) guarantees to these applications. Its current definition, however, does not allow for a high level of flexibility or assurance and, therefore, it can not be widely deployed. In this paper, we introduce a new protocol for a Diffserv architecture which provides a simple and efficient solution to the above problem. It is a complete protocol, in the sense that it deals with the issues of packet scheduling, admission control, and congestion control. We will show, through experimental results, that our proposed protocol can improve the flexibility and assurance provided by current solutions, while maintaining a high level of network utilization.published_or_final_versio
A Taxonomy for Management and Optimization of Multiple Resources in Edge Computing
Edge computing is promoted to meet increasing performance needs of
data-driven services using computational and storage resources close to the end
devices, at the edge of the current network. To achieve higher performance in
this new paradigm one has to consider how to combine the efficiency of resource
usage at all three layers of architecture: end devices, edge devices, and the
cloud. While cloud capacity is elastically extendable, end devices and edge
devices are to various degrees resource-constrained. Hence, an efficient
resource management is essential to make edge computing a reality. In this
work, we first present terminology and architectures to characterize current
works within the field of edge computing. Then, we review a wide range of
recent articles and categorize relevant aspects in terms of 4 perspectives:
resource type, resource management objective, resource location, and resource
use. This taxonomy and the ensuing analysis is used to identify some gaps in
the existing research. Among several research gaps, we found that research is
less prevalent on data, storage, and energy as a resource, and less extensive
towards the estimation, discovery and sharing objectives. As for resource
types, the most well-studied resources are computation and communication
resources. Our analysis shows that resource management at the edge requires a
deeper understanding of how methods applied at different levels and geared
towards different resource types interact. Specifically, the impact of mobility
and collaboration schemes requiring incentives are expected to be different in
edge architectures compared to the classic cloud solutions. Finally, we find
that fewer works are dedicated to the study of non-functional properties or to
quantifying the footprint of resource management techniques, including
edge-specific means of migrating data and services.Comment: Accepted in the Special Issue Mobile Edge Computing of the Wireless
Communications and Mobile Computing journa
Meta-QoS performance of earliest-deadline-first and rate-monotonic scheduling of smoothed video data in a client-server environment
In this paper we present an extensive performance study of two modified EDF and RM scheduling algorithms which are enhanced to provide quality of service (QoS) guarantees for smoothed video data. With a probabilistic definition of QoS, we incorporate admission control conditions into the two algorithms. Furthermore, we also include a counter-based scheduling module as the core scheduling mechanism which adaptively adjusts the actual QoS levels assigned to requests. Our theoretical analysis of the two enhanced algorithms, called QEDF and QRM, shows that the QRM algorithm is more robust than the QEDF algorithm for different workload and utilization conditions. We also propose to use a new metric called meta-QoS to quantify the overall performance of a packet scheduler given a set of simultaneous requests. In our experiments, we find that the QRM algorithm can sustain a rather stable level of meta-QoS even when the workload and utilization levels are increased. On the other hand, the QEDF algorithm is found to be less desirable for a high level of utilization and a large number of requests.published_or_final_versio
Wired and Wireless Reliable Real-Time Communication in Industrial Systems
In modern factory automation systems, data communication plays a vital role. Different nodes like control systems, sensors and actuators can communicate over a wireless or wired industrial network. The data traffic generated is often scheduled for periodic transmission, where each single message or packet must arrive in time. For this real-time communication, methods have been developed to support communication services with a guaranteed throughput and delay bound for such periodic traffic, but merely under the assumption of error-free communication. However, the possibility for errors in the transmission still exists due to, e.g. noise or interference. A node receiving sensor values from a sensor in the system might then be forced to rely upon an older sensor value from the latest period, possibly leading to inaccuracies in control loops which can compromise the functioning of the system. In safety-critical systems, redundant networks or communication channels are frequently added to cope with errors, leading to more expensive systems. In this chapter, we will describe an alternative approach where erroneous data packets are retransmitted in a way that does not jeopardise any earlier stated real-time guarantees for ordinary transmissions. Using our framework, the reliability of real-time communication can be increased in a more cost-efficient way. We describe in this chapter an overview of our framework for reliable real-time communication, while details of our approach can be found in our earlier publications. In the light of the emerging use of wireless communication, the framework proves to be especially useful due to the high bit error rate inherent to the wireless medium. However, the framework is naturally also attractive for wired communication systems
Mixed-criticality real-time task scheduling with graceful degradation
”The mixed-criticality real-time systems implement functionalities of different degrees of importance (or criticalities) upon a shared platform. In traditional mixed-criticality systems, under a hi mode switch, no guaranteed service is provided to lo-criticality tasks. After a mode switch, only hi-criticality tasks are considered for execution while no guarantee is made to the lo-criticality tasks. However, with careful optimistic design, a certain degree of service guarantee can be provided to lo-criticality tasks upon a mode switch. This concept is broadly known as graceful degradation. Guaranteed graceful degradation provides a better quality of service as well as it utilizes the system resource more efficiently. In this thesis, we study two efficient techniques of graceful degradation.
First, we study a mixed-criticality scheduling technique where graceful degradation is provided in the form of minimum cumulative completion rates. We present two easy-to-implement admission-control algorithms to determine which lo-criticality jobs to complete in hi mode. The scheduling is done by following deadline virtualization, and two heuristics are shown for virtual deadline settings. We further study the schedulability analysis and the backward mode switch conditions, which are proposed and proved in (Guo et al., 2018).
Next, we present a probabilistic scheduling technique for mixed-criticality tasks on multiprocessor systems where a system-wide permitted failure probability is known. The schedulability conditions are derived along with the processor allocation scheme. The work is extended from (Guo et al., 2015), where the probabilistic model is first introduced for independent task scheduling on a uniprocessor platform. We further consider the failure dependency between tasks while scheduling on multiprocessor platforms.
We provide related theoretical analysis to show the correctness of our work. To show the effectiveness of our proposed techniques, we conduct a detailed experimental evaluation under different circumstances”--Abstract, page iii
- …