111 research outputs found

    Performance analysis of buffers with train arrivals and correlated output interruptions

    Get PDF
    In this paper, we study a discrete-time buffer system with a timecorrelated packet arrival process and one unreliable output line. In particular, packets arrive to the buffer in the form of variable-length packet trains at a fixed rate of exactly one packet per slot. The packet trains are assumed to have a geometric length, such that each packet has a fixed probability of being the last of its corresponding train. The output line is governed by a Markovian process, such that the probability that the line is available during a slot depends on the state of the underlying J-state Markov process during that slot. First, we provide a general analysis of the state of the buffer system based on a matrix generating functions approach. This also leads to an expression for the mean buffer content. Additionally, we take a closer look at the distributions of the packet delay and the train delay. In order to make matters more concrete, we next present a detailed and explicit analysis of the buffer system in case the output line is governed by a 2-state Markov process. Some numerical examples help to visualise the influence of the various model parameters

    Analysis of generic discrete-time buffer models with irregular packet arrival patterns

    Get PDF
    De kwaliteit van de multimediadiensten die worden aangeboden over de huidige breedband-communicatienetwerken, wordt in hoge mate bepaald door de performantie van de buffers die zich in de diverse netwerkele-menten (zoals schakelknooppunten, routers, modems, toegangsmultiplexers, netwerkinter- faces, ...) bevinden. In dit proefschrift bestuderen we de performantie van een dergelijke buffer met behulp van een geschikt stochastisch discrete-tijd wachtlijnmodel, waarbij we het geval van meerdere uitgangskanalen en (niet noodzakelijk identieke) pakketbronnen beschouwen, en de pakkettransmissietijden in eerste instantie één slot bedragen. De grillige, of gecorreleerde, aard van een pakketstroom die door een bron wordt gegenereerd, wordt gekarakteriseerd aan de hand van een algemeen D-BMAP (discrete-batch Markovian arrival process), wat een generiek kader creëert voor het beschrijven van een superpositie van dergelijke informatiestromen. In een later stadium breiden we onze studie uit tot het geval van transmissietijden met een algemene verdeling, waarbij we ons beperken tot een buffer met één enkel uitgangskanaal. De analyse van deze wachtlijnmodellen gebeurt hoofdzakelijk aan de hand van een particuliere wiskundig-analytische aanpak waarbij uitvoerig gebruik gemaakt wordt van probabiliteitsgenererende functies, die er toe leidt dat de diverse performantiematen (min of meer expliciet) kunnen worden uitgedrukt als functie van de systeemparameters. Dit resul-teert op zijn beurt in efficiënte en accurate berekeningsalgoritmen voor deze grootheden, die op relatief eenvoudige wijze geïmplementeerd kunnen worden

    ATM virtual connection performance modeling

    Get PDF

    Analysis of discrete-time queueing systems with multidimensional state space

    Get PDF

    Discrete-time queueing models: generalized service mechanisms and correlation effects

    Get PDF

    Analysis of queueing models with batch service

    Get PDF
    This dissertation is the result of my research work at the SMACS research group (Department of Telecommunications and Information Processing, Ghent University) and it concerns the analysis of queueing models with batch service. A queueing model basically is a mathematical abstraction of a situation where customers arrive and queue up until they receive some kind of service. These phenomena are omnipresent in real life: people waiting at a counter of a post office or bank, people in the waiting room of a doctor, airplanes waiting to take off, people waiting until they get connected with the call center, data packets which are temporarily stored into a buffer until the transmisssion channel is available, et cetera. The analysis of queueing models constitutes the subject of the applied mathematical discipline called queueing theory and amounts to answering questions such as “How many customers are waiting on average?”, “How long do customers have to wait?”, “Is there a large variation on the waiting time?”, “What is the probability that data packets are lost due to a full buffer?”, “What is the probability that a customer suffers a lengthy delay?”, et cetera. In queueing theory, the number of customers and their waiting time are often denominated by respectively buffer content and customer delay. In addition, the probability that a quantity, such as the buffer content or customer delay, is very large or lengthy, is generally called a tail probability. The models we investigate throughout this dissertation have in common that customers can be served in batches, meaning that several customers can be served simultaneously. An elevator can be viewed as a classic example, as several people can be transported simultaneously to another floor. Also, in a variety of production or transport processes several goods can be processed together. Furthermore, in quality control, classification of items as good or bad can often be achieved more economically by examining the items in groups rather than individually. If the result of a group test is good, all items within it can then be classified as good, whereas one or more items are bad in the opposite case, where the items can then be retested by considering smaller groups. Group testing is especially of importance when the percentage of bad items is small. In addition, in telecommunications networks, packets with the same destination and quality of service (QoS) requirements are often aggregated into so-called bursts and these bursts are transmitted over the network. This is mainly done for efficiency reasons, since only one header per aggregated burst has to be constructed, instead of one header per single information unit, thus leading to an increased throughput. Technologies using packet aggregation include for instance Optical burst switched (OBS) networks and IEEE 802.11n wireless local area networks (WLANs). An inherent aspect of batch service is that newly arriving customers cannot join the ongoing service, even if there is free capacity (we denominate the maximum number of customers that can be served simultaneously by server capacity). For instance, an arriving person cannot enter an elevator that has just left, even if space is available. This person has to wait until the elevator has transported its occupants to their requested floors and has returned, which might take a long time in high buildings. In view of this, it is of importance to take a well-considered decision when the server becomes available and finds less customers than it can serve in theory. This decision is called the service policy. A whole spectrum of service policies exist. The server could, for instance, start serving the already present customers immediately. Although the present customers benefit from this approach, capacity is wasted: customers that arrive later cannot join the ongoing service. An alternative for this so-called immediate-batch service policy is the full-batch service policy. In this case, the available server postpones service until the number of present customers reaches or exceeds the server capacity, which, in turn, has a negative effect on the delay of the customers waiting to form a full batch (postponing delay). The threshold-based policy is a kind of compromise between immediate-batch service policy and full-batch service policy. When the number of present customers is below some service threshold, service is postponed, whereas service is initiated when the number of present customers reaches or exceeds this threshold. It is important to realize that even with this compromise, long postponing delays are possible. Therefore, in this dissertation, we combine a thresholdbased policy with a timer mechanism that avoids excessive postponing delays. The purpose of this dissertation is to calculate a large spectrum of performance measures, which enable to evaluate a broad set of situations with batch service and aid in selecting an efficient service policy. The studied performance measures are moments, such as the mean value and variance, and tail probabilities of the buffer content and the customer delay. This dissertation is structured as follows. In chapter 1, we motivate our work and we introduce crucial concepts such as probability generating functions (PGFs), whose useful properties are frequently relied upon throughout the analysis. Then we deduce moments and tail probabilities of the buffer content in chapter 2. The resulting formulas still contain unknown probabilities that have to be calculated numerically. As this might become unfeasible in some cases, we compute in chapter 3 approximations for the buffer content. Next, moments and tail probabilities of the customer delay are covered in respectively chapters 4 and 5. In order to analyze the moments, we conceive the customer delay as the sum of two non-overlapping parts, whereas for the tail probabilities, it turns out to be more convenient to interpret the delay as the maximum of two time periods. Further, in real life the customer arrival process often exhibits some kind of dependency. For instance, if a large amount of customers have recently arrived, it is likely that many customers arrive in the near future, as it might be an indication of a peak moment. Therefore, we investigate in chapter 6 the influence of dependency in the arrival process on the behaviour of batch-service phenomena and on the selection of an efficient service policy. Finally, the main contributions are summarized in chapter 7

    Study on the Performance of TCP over 10Gbps High Speed Networks

    Get PDF
    Internet traffic is expected to grow phenomenally over the next five to ten years. To cope with such large traffic volumes, high-speed networks are expected to scale to capacities of terabits-per-second and beyond. Increasing the role of optics for packet forwarding and transmission inside the high-speed networks seems to be the most promising way to accomplish this capacity scaling. Unfortunately, unlike electronic memory, it remains a formidable challenge to build even a few dozen packets of integrated all-optical buffers. On the other hand, many high-speed networks depend on the TCP/IP protocol for reliability which is typically implemented in software and is sensitive to buffer size. For example, TCP requires a buffer size of bandwidth delay product in switches/routers to maintain nearly 100\% link utilization. Otherwise, the performance will be much downgraded. But such large buffer will challenge hardware design and power consumption, and will generate queuing delay and jitter which again cause problems. Therefore, improve TCP performance over tiny buffered high-speed networks is a top priority. This dissertation studies the TCP performance in 10Gbps high-speed networks. First, a 10Gbps reconfigurable optical networking testbed is developed as a research environment. Second, a 10Gbps traffic sniffing tool is developed for measuring and analyzing TCP performance. New expressions for evaluating TCP loss synchronization are presented by carefully examining the congestion events of TCP. Based on observation, two basic reasons that cause performance problems are studied. We find that minimize TCP loss synchronization and reduce flow burstiness impact are critical keys to improve TCP performance in tiny buffered networks. Finally, we present a new TCP protocol called Multi-Channel TCP and a new congestion control algorithm called Desynchronized Multi-Channel TCP (DMCTCP). Our algorithm implementation takes advantage of a potential parallelism from the Multi-Path TCP in Linux. Over an emulated 10Gbps network ruled by routers with only a few dozen packets of buffers, our experimental results confirm that bottleneck link utilization can be much better improved by DMCTCP than by many other TCP variants. Our study is a new step towards the deployment of optical packet switching/routing networks

    Stochastic Dynamic Programming and Stochastic Fluid-Flow Models in the Design and Analysis of Web-Server Farms

    Get PDF
    A Web-server farm is a specialized facility designed specifically for housing Web servers catering to one or more Internet facing Web sites. In this dissertation, stochastic dynamic programming technique is used to obtain the optimal admission control policy with different classes of customers, and stochastic uid- ow models are used to compute the performance measures in the network. The two types of network traffic considered in this research are streaming (guaranteed bandwidth per connection) and elastic (shares available bandwidth equally among connections). We first obtain the optimal admission control policy using stochastic dynamic programming, in which, based on the number of requests of each type being served, a decision is made whether to allow or deny service to an incoming request. In this subproblem, we consider a xed bandwidth capacity server, which allocates the requested bandwidth to the streaming requests and divides all of the remaining bandwidth equally among all of the elastic requests. The performance metric of interest in this case will be the blocking probability of streaming traffic, which will be computed in order to be able to provide Quality of Service (QoS) guarantees. Next, we obtain bounds on the expected waiting time in the system for elastic requests that enter the system. This will be done at the server level in such a way that the total available bandwidth for the requests is constant. Trace data will be converted to an ON-OFF source and fluid- flow models will be used for this analysis. The results are compared with both the mean waiting time obtained by simulating real data, and the expected waiting time obtained using traditional queueing models. Finally, we consider the network of servers and routers within the Web farm where data from servers flows and merges before getting transmitted to the requesting users via the Internet. We compute the waiting time of the elastic requests at intermediate and edge nodes by obtaining the distribution of the out ow of the upstream node. This out ow distribution is obtained by using a methodology based on minimizing the deviations from the constituent in flows. This analysis also helps us to compute waiting times at different bandwidth capacities, and hence obtain a suitable bandwidth to promise or satisfy the QoS guarantees. This research helps in obtaining performance measures for different traffic classes at a Web-server farm so as to be able to promise or provide QoS guarantees; while at the same time helping in utilizing the resources of the server farms efficiently, thereby reducing the operational costs and increasing energy savings
    corecore