14,037 research outputs found
Optimizing Service Differentiation Scheme with Sized-based Queue Management in DiffServ Networks
In this paper we introduced Modified Sized-based Queue Management as a
dropping scheme that aims to fairly prioritize and allocate more service to
VoIP traffic over bulk data like FTP as the former one usually has small packet
size with less impact to the network congestion. In the same time, we want to
guarantee that this prioritization is fair enough for both traffic types. On
the other hand we study the total link delay over the congestive link with the
attempt to alleviate this congestion as much as possible at the by function of
early congestion notification. Our M-SQM scheme has been evaluated with NS2
experiments to measure the packets received from both and total link-delay for
different traffic. The performance evaluation results of M-SQM have been
validated and graphically compared with the performance of other three legacy
AQMs (RED, RIO, and PI). It is depicted that our M-SQM outperformed these AQMs
in providing QoS level of service differentiation.Comment: 10 pages, 9 figures, 1 table, Submitted to Journal of
Telecommunication
Comparative Study Of Congestion Control Techniques In High Speed Networks
Congestion in network occurs due to exceed in aggregate demand as compared to
the accessible capacity of the resources. Network congestion will increase as
network speed increases and new effective congestion control methods are
needed, especially to handle bursty traffic of todays very high speed networks.
Since late 90s numerous schemes i.e. [1]...[10] etc. have been proposed. This
paper concentrates on comparative study of the different congestion control
schemes based on some key performance metrics. An effort has been made to judge
the performance of Maximum Entropy (ME) based solution for a steady state
GE/GE/1/N censored queues with partial buffer sharing scheme against these key
performance metrics.Comment: 10 pages IEEE format, International Journal of Computer Science and
Information Security, IJCSIS November 2009, ISSN 1947 5500,
http://sites.google.com/site/ijcsis
Recommended from our members
Performance modelling of a multiple threshold RED mechanism for bursty and correlated Internet traffic with MMPP arrival process
Access to the large web content hosted all over the world by users of the Internet engage
many hosts, routers/switches and faster links. They challenge the internet backbone to operate at
its capacity to assure e±cient content access. This may result in congestion and raises concerns over
various Quality of Service (QoS) issues like high delays, high packet loss and low throughput of the
system for various Internet applications. Thus, there is a need to develop effective congestion control
mechanisms in order to meet various Quality of Service (QoS) related performance parameters. In this
paper, our emphasis is on the Active Queue Management (AQM) mechanisms, particularly Random
Early Detection (RED). We propose a threshold based novel analytical model based on standard RED
mechanism. Various numerical examples are presented for Internet traffic scenarios containing both the
burstiness and correlation properties of the network traffic
A Priority-based Fair Queuing (PFQ) Model for Wireless Healthcare System
Healthcare is a very active research area, primarily due to the increase in the elderly population that leads to increasing number of emergency situations that require urgent actions. In recent years some of wireless networked medical devices were equipped with different sensors to measure and report on vital signs of patient remotely. The most important sensors are Heart Beat Rate (ECG), Pressure and Glucose sensors. However, the strict requirements and real-time nature of medical applications dictate the extreme importance and need for appropriate Quality of Service (QoS), fast and accurate delivery of a patient’s measurements in reliable e-Health ecosystem.
As the elderly age and older adult population is increasing (65 years and above) due to the advancement in medicine and medical care in the last two decades; high QoS and reliable e-health ecosystem has become a major challenge in Healthcare especially for patients who require continuous monitoring and attention. Nevertheless, predictions have indicated that elderly population will be approximately 2 billion in developing countries by 2050 where availability of medical staff shall be unable to cope with this growth and emergency cases that need immediate intervention. On the other side, limitations in communication networks capacity, congestions and the humongous increase of devices, applications and IOT using the available communication networks add extra layer of challenges on E-health ecosystem such as time constraints, quality of measurements and signals reaching healthcare centres.
Hence this research has tackled the delay and jitter parameters in E-health M2M wireless communication and succeeded in reducing them in comparison to current available models. The novelty of this research has succeeded in developing a new Priority Queuing model ‘’Priority Based-Fair Queuing’’ (PFQ) where a new priority level and concept of ‘’Patient’s Health Record’’ (PHR) has been developed and
integrated with the Priority Parameters (PP) values of each sensor to add a second level of priority. The results and data analysis performed on the PFQ model under different scenarios simulating real M2M E-health environment have revealed that the PFQ has outperformed the results obtained from simulating the widely used current models such as First in First Out (FIFO) and Weight Fair Queuing (WFQ).
PFQ model has improved transmission of ECG sensor data by decreasing delay and jitter in emergency cases by 83.32% and 75.88% respectively in comparison to FIFO and 46.65% and 60.13% with respect to WFQ model. Similarly, in pressure sensor the improvements were 82.41% and 71.5% and 68.43% and 73.36% in comparison to FIFO and WFQ respectively. Data transmission were also improved in the Glucose sensor by 80.85% and 64.7% and 92.1% and 83.17% in comparison to FIFO and WFQ respectively. However, non-emergency cases data transmission using PFQ model was negatively impacted and scored higher rates than FIFO and WFQ since PFQ tends to give higher priority to emergency cases.
Thus, a derivative from the PFQ model has been developed to create a new version namely “Priority Based-Fair Queuing-Tolerated Delay” (PFQ-TD) to balance the data transmission between emergency and non-emergency cases where tolerated delay in emergency cases has been considered. PFQ-TD has succeeded in balancing fairly this issue and reducing the total average delay and jitter of emergency and non-emergency cases in all sensors and keep them within the acceptable allowable standards. PFQ-TD has improved the overall average delay and jitter in emergency and non-emergency cases among all sensors by 41% and 84% respectively in comparison to PFQ model
Heart-like fair queuing algorithms (HLFQA)
We propose a new family of fair, work conserving traffic scheduling mechanisms that imitate the behavior of the human heart in the cardiovascular system. The algorithms have MAX (where MAX is the maximum packet size) fairness and O(log N) complexity and thus compare favorably with existing algorithms. The algorithms are simple enough to be implemented in hardwar
Multiplexing regulated traffic streams: design and performance
The main network solutions for supporting QoS rely on traf- fic policing (conditioning, shaping). In particular, for IP networks the IETF has developed Intserv (individual flows regulated) and Diffserv (only ag- gregates regulated). The regulator proposed could be based on the (dual) leaky-bucket mechanism. This explains the interest in network element per- formance (loss, delay) for leaky-bucket regulated traffic. This paper describes a novel approach to the above problem. Explicitly using the correlation structure of the sources’ traffic, we derive approxi- mations for both small and large buffers. Importantly, for small (large) buffers the short-term (long-term) correlations are dominant. The large buffer result decomposes the traffic stream in a stream of constant rate and a periodic impulse stream, allowing direct application of the Brownian bridge approximation. Combining the small and large buffer results by a concave majorization, we propose a simple, fast and accurate technique to statistically multiplex homogeneous regulated sources. To address heterogeneous inputs, we present similarly efficient tech- niques to evaluate the performance of multiple classes of traffic, each with distinct characteristics and QoS requirements. These techniques, applica- ble under more general conditions, are based on optimal resource (band- width and buffer) partitioning. They can also be directly applied to set GPS (Generalized Processor Sharing) weights and buffer thresholds in a shared resource system
- …