174 research outputs found
Modelling of self-similar teletraffic for simulation
Recent studies of real teletraffic data in modern computer networks have shown that teletraffic exhibits self-similar (or fractal) properties over a wide range of time scales. The properties of self-similar teletraffic are very different from the traditional models of teletraffic based on Poisson, Markov-modulated Poisson, and related processes. The use of traditional models in networks characterised by self-similar processes can lead to incorrect conclusions about the performance of analysed networks. These include serious over-estimations of the performance of computer networks, insufficient allocation of communication and data processing resources, and difficulties ensuring the quality of service expected by network users. Thus, full understanding of the self-similar nature in teletraffic is an important issue.
Due to the growing complexity of modern telecommunication networks, simulation has become the only feasible paradigm for their performance evaluation. In this thesis, we make some contributions to discrete-event simulation of networks with strongly-dependent, self-similar teletraffic.
First, we have evaluated the most commonly used methods for estimating the self-similarity parameter H using appropriately long sequences of data. After assessing properties of available H estimators, we identified the most efficient estimators for practical studies of self-similarity.
Next, the generation of arbitrarily long sequences of pseudo-random numbers possessing specific stochastic properties was considered. Various generators of pseudo-random self-similar sequences have been proposed. They differ in computational complexity and accuracy of the self-similar sequences they generate. In this thesis, we propose two new generators of self-similar teletraffic: (i) a generator based on Fractional Gaussian Noise and Daubechies Wavelets (FGN-DW), that is one of the fastest and the most accurate generators so far proposed; and (ii) a generator based on the Successive Random Addition (SRA) algorithm. Our comparative study of sequential and fixed-length self-similar pseudo-random teletraffic generators showed that the FFT, FGN-DW and SRP-FGN generators are the most efficient, both in the sense of accuracy and speed.
To conduct simulation studies of telecommunication networks, self-similar processes often need to be transformed into suitable self-similar processes with arbitrary marginal distributions. Thus, the next problem addressed was how well the self-similarity and autocorrelation function of an original self-similar process are preserved when the self-similar sequences are converted into suitable self-similar processes with arbitrary marginal distributions. We also show how pseudo-random self-similar sequences can be applied to produce a model of teletraffic associated with the transmission of VBR JPEG /MPEG video. A combined gamma/Pareto model based on the application of the FGN-DW generator was used to synthesise VBR JPEG /MPEG video traffic.
Finally, effects of self-similarity on the behaviour of queueing systems have been investigated. Using M/M/1/∞ as a reference queueing system with no long-range dependence, we have investigated how self-similarity and long-range dependence in arrival processes affect the length of sequential simulations being executed for obtaining steady-state results with the required level of statistical error. Our results show that the finite buffer overflow probability of a queueing system with self-similar input is much greater than the equivalent queueing system with Poisson or a short-range dependent input process, and that the overflow probability increases as the self-similarity parameter approaches one
Renegotiation based dynamic bandwidth allocation for selfsimilar VBR traffic
The provision of QoS to applications traffic depends heavily on how different traffic types are categorized and classified, and how the prioritization of these applications are managed. Bandwidth is the most scarce network resource. Therefore, there is a need for a method or system that distributes an available bandwidth in a network among different applications in such a way that each class or type of traffic receives their constraint QoS requirements.
In this dissertation, a new renegotiation based dynamic resource allocation method for variable bit rate (VBR) traffic is presented. First, pros and cons of available off-line methods that are used to estimate selfsimilarity level (represented by Hurst parameter) of a VBR traffic trace are empirically investigated, and criteria to select measurement parameters for online resource management are developed. It is shown that wavelet analysis based methods are the strongest tools in estimation of Hurst parameter with their low computational complexities, compared to the variance-time method and R/S pox plot. Therefore, a temporal energy distribution of a traffic data arrival counting process among different frequency sub-bands is considered as a traffic descriptor, and then a robust traffic rate predictor is developed by using the Haar wavelet analysis. The empirical results show that the new on-line dynamic bandwidth allocation scheme for VBR traffic is superior to traditional dynamic bandwidth allocation methods that are based on adaptive algorithms such as Least Mean Square, Recursive Least Square, and Mean Square Error etc. in terms of high utilization and low queuing delay. Also a method is developed to minimize the number of bandwidth renegotiations to decrease signaling costs on traffic schedulers (e.g. WFQ) and networks (e.g. ATM). It is also quantified that the introduced renegotiation based bandwidth management scheme decreases heavytailedness of queue size distributions, which is an inherent impact of traffic self similarity.
The new design increases the achieved utilization levels in the literature, provisions given queue size constraints and minimizes the number of renegotiations simultaneously. This renegotiation -based design is online and practically embeddable into QoS management blocks, edge routers and Digital Subscriber Lines Access Multiplexers (DSLAM) and rate adaptive DSL modems
Video traffic : characterization, modelling and transmission
EThOS - Electronic Theses Online ServiceGBUnited Kingdo
Rate-distortion analysis and traffic modeling of scalable video coders
In this work, we focus on two important goals of the transmission of scalable video over the Internet. The first goal is to provide high quality video to end users and the second one is to properly design networks and predict network performance for video transmission based on the characteristics of existing video traffic. Rate-distortion (R-D) based schemes are often applied to improve and stabilize video quality; however, the lack of R-D modeling of scalable coders limits their applications in scalable streaming.
Thus, in the first part of this work, we analyze R-D curves of scalable video coders and propose a novel operational R-D model. We evaluate and demonstrate the accuracy of our R-D function in various scalable coders, such as Fine Granular Scalable (FGS) and Progressive FGS coders. Furthermore, due to the time-constraint nature of Internet streaming, we propose another operational R-D model, which is accurate yet with low computational cost, and apply it to streaming applications for quality control purposes.
The Internet is a changing environment; however, most quality control approaches only consider constant bit rate (CBR) channels and no specific studies have been conducted for quality control in variable bit rate (VBR) channels. To fill this void, we examine an asymptotically stable congestion control mechanism and combine it with our R-D model to present smooth visual quality to end users under various network conditions.
Our second focus in this work concerns the modeling and analysis of video traffic, which is crucial to protocol design and efficient network utilization for video transmission. Although scalable video traffic is expected to be an important source for the Internet, we find that little work has been done on analyzing or modeling it. In this regard, we develop a frame-level hybrid framework for modeling multi-layer VBR video traffic. In the proposed framework, the base layer is modeled using a combination of wavelet and time-domain methods and the enhancement layer is linearly predicted from the base layer using the cross-layer correlation
Recommended from our members
Estimation of LRD present in H.264 video traces using wavelet analysis and proving the paramount of H.264 using OPF technique in wi-fi environment.
While there has always been a tremendous demand for streaming video over
Wireless networks, the nature of the application still presents some challenging
issues. These applications that transmit coded video sequence data over best-effort
networks like the Internet, the application must cope with the changing network
behaviour; especially, the source encoder rate should be controlled based on
feedback from a channel estimator that explores the network intermittently. The
arrival of powerful video compression techniques such as H.264, which advance in
networking and telecommunications, opened up a whole new frontier for multimedia
communications. The aim of this research is to transmit the H.264 coded video
frames in the wireless network with maximum reliability and in a very efficient
manner. When the H.264 encoded video sequences are to be transmitted through
wireless network, it faces major difficulties in reaching the destination. The
characteristics of H.264 video coded sequences are studied fully and their capability
of transmitting in wireless networks are examined and a new approach called
Optimal Packet Fragmentation (OPF) is framed and the H.264 coded sequences are
tested in the wireless simulated environment. This research has three major studies
involved in it. First part of the research has the study about Long Range Dependence
(LRD) and the ways by which the self-similarity can be estimated. For estimating the
LRD a few studies are carried out and Wavelet-based estimator is selected for the
research because Wavelets incarcerate both time and frequency features in the data
and regularly provides a more affluent picture than the classical Fourier analysis.
The Wavelet used to estimate the self-similarity by using the variable called Hurst
Parameter. Hurst Parameter tells the researcher about how a data can behave inside the transmitted network. This Hurst Parameter should be calculated for a more
reliable transmission in the wireless network. The second part of the research deals
with MPEG-4 and H.264 encoder. The study is carried out to prove which encoder is
superior to the other. We need to know which encoder can provide excellent Quality
of Service (QoS) and reliability. This study proves with the help of Hurst parameter
that H.264 is superior to MPEG-4. The third part of the study is the vital part in this
research; it deals with the H.264 video coded frames that are segmented into optimal
packet size in the MAC Layer for an efficient and more reliable transfer in the
wireless network. Finally the H.264 encoded video frames incorporated with the
Optimal Packet Fragmentation are tested in the NS-2 wireless simulated network.
The research proves the superiority of H.264 video encoder and OPF¿s master class
Modelling packet departure times using a known PDF
This paper deals with IPTV traffic source modelling and describes a packet generator based on a known probability density function which is measured and formed from a histogram. Histogram based probability density functions destroy an amount of information, because classes used to form the histogram often cover significantly more events than one. In this work, we propose an algorithm to generate far more output states of random variable X than the input probability distribution function is made from. In this generator is assumed that all IPTV packets of the same video stream are the same length. Therefore, only packet times are generated. These times are generated using the measured normalized histogram that is converted to a cumulative distribution function which acts as a finite number of states that can be addressed. To address these states we use an ON/OFF model that is driven by an uniform random number generator in (0, 1). When a state is chosen then the resulting value is equal to a histogram class. To raise the number of possible output states of the random variable X, we propose to use an uniform random number generator that generates numbers within the range of the chosen histogram class. This second uniform random number generator assures that the number of output states is far more larger than the number of histogram classes
Modeling And Dynamic Resource Allocation For High Definition And Mobile Video Streams
Video streaming traffic has been surging in the last few years, which has resulted in an increase of its Internet traffic share on a daily basis. The importance of video streaming management has been emphasized with the advent of High Definition: HD) video streaming, as it requires by its nature more network resources. In this dissertation, we provide a better support for managing HD video traffic over both wireless and wired networks through several contributions. We present a simple, general and accurate video source model: Simplified Seasonal ARIMA Model: SAM). SAM is capable of capturing the statistical characteristics of video traces with less than 5% difference from their calculated optimal models. SAM is shown to be capable of modeling video traces encoded with MPEG-4 Part2, MPEG-4 Part10, and Scalable Video Codec: SVC) standards, using various encoding settings. We also provide a large and publicly-available collection of HD video traces along with their analyses results. These analyses include a full statistical analysis of HD videos, in addition to modeling, factor and cluster analyses. These results show that by using SAM, we can achieve up to 50% improvement in video traffic prediction accuracy. In addition, we developed several video tools, including an HD video traffic generator based on our model. Finally, to improve HD video streaming resource management, we present a SAM-based delay-guaranteed dynamic resource allocation: DRA) scheme that can provide up to 32.4% improvement in bandwidth utilization
Combined Industry, Space and Earth Science Data Compression Workshop
The sixth annual Space and Earth Science Data Compression Workshop and the third annual Data Compression Industry Workshop were held as a single combined workshop. The workshop was held April 4, 1996 in Snowbird, Utah in conjunction with the 1996 IEEE Data Compression Conference, which was held at the same location March 31 - April 3, 1996. The Space and Earth Science Data Compression sessions seek to explore opportunities for data compression to enhance the collection, analysis, and retrieval of space and earth science data. Of particular interest is data compression research that is integrated into, or has the potential to be integrated into, a particular space or earth science data information system. Preference is given to data compression research that takes into account the scien- tist's data requirements, and the constraints imposed by the data collection, transmission, distribution and archival systems
Efficient compression of motion compensated residuals
EThOS - Electronic Theses Online ServiceGBUnited Kingdo
- …