228 research outputs found

    Comparison of HDR quality metrics in Per-Clip Lagrangian multiplier optimisation with AV1

    Get PDF
    The complexity of modern codecs along with the increased need of delivering high-quality videos at low bitrates has reinforced the idea of a per-clip tailoring of parameters for optimised rate-distortion performance. While the objective quality metrics used for Standard Dynamic Range (SDR) videos have been well studied, the transitioning of consumer displays to support High Dynamic Range (HDR) videos, poses a new challenge to rate-distortion optimisation. In this paper, we review the popular HDR metrics DeltaE100 (DE100), PSNRL100, wPSNR, and HDR-VQM. We measure the impact of employing these metrics in per-clip direct search optimisation of the rate-distortion Lagrange multiplier in AV1. We report, on 35 HDR videos, average Bjontegaard Delta Rate (BD-Rate) gains of 4.675%, 2.226%, and 7.253% in terms of DE100, PSNRL100, and HDR-VQM. We also show that the inclusion of chroma in the quality metrics has a significant impact on optimisation, which can only be partially addressed by the use of chroma offsets.Comment: Accepted version for ICME 2023 Special Session, "Optimised Media Delivery

    A tabu search algorithm for dynamic routing in ATM cell-switching networks

    Get PDF
    This paper deals with the dynamic routing problem in ATM cell-switching networks. We present a mathematical programming model based on cell loss and a Tabu Search algorithm with short-term memory that is reinforced with a long-term memory procedure. The estimation of the quality of the solutions is fast, due to the specific encoding of the feasible solutions. The Tabu Search algorithm reaches good quality solutions, outperforming other approaches such as Genetic Algorithms and the Minimum Switching Path heuristic, regarding both cell loss and the CPU time consumption. The best results were found for the more complex networks with a high number of switches and links

    Reactive traffic control mechanisms for communication networks with self-similar bandwidth demands

    Get PDF
    Communication network architectures are in the process of being redesigned so that many different services are integrated within the same network. Due to this integration, traffic management algorithms need to balance the requirements of the traffic which the algorithms are directly controlling with Quality of Service (QoS) requirements of other classes of traffic which will be encountered in the network. Of particular interest is one class of traffic, termed elastic traffic, that responds to dynamic feedback from the network regarding the amount of available resources within the network. Examples of this type of traffic include the Available Bit Rate (ABR) service in Asynchronous Transfer Mode (ATM) networks and connections using Transmission Control Protocol (TCP) in the Internet. Both examples aim to utilise available bandwidth within a network. Reactive traffic management, like that which occurs in the ABR service and TCP, depends explicitly on the dynamic bandwidth requirements of other traffic which is currently using the network. In particular, there is significant evidence that a wide range of network traffic, including Ethernet, World Wide Web, Varible Bit Rate video and signalling traffic, is self-similar. The term self-similar refers to the particular characteristic of network traffic to remain bursty over a wide range of time scales. A closely associated characteristic of self-similar traffic is its long-range dependence (LRD), which refers to the significant correlations that occur with the traffic. By utilising these correlations, greater predictability of network traffic can be achieved, and hence the performance of reactive traffic management algorithms can be enhanced. A predictive rate control algorithm, called PERC (Predictive Explicit Rate Control), is proposed in this thesis which is targeted to the ABR service in ATM networks. By incorporating the LRD stochastic structure of background traffic, measurements of the bandwidth requirements of background traffic, and the delay associated with a particular ABR connection, a predictive algorithm is defined which provides explicit rate information that is conveyed to ABR sources. An enhancement to PERC is also described. This algorithm, called PERC+, uses previous control information to correct prediction errors that occur for connections with larger round-trip delay. These algorithms have been extensively analysed with regards to their network performance, and simulation results show that queue lengths and cell loss rates are significantly reduced when these algorithms are deployed. An adaptive version of PERC has also been developed using real-time parameter estimates of self-similar traffic. This has excellent performance compared with standard ABR rate control algorithms such as ERICA. Since PERC and its enhancement PERC+ have explicitly utilised the index of self-similarity, known as the Hurst parameter, the sensitivity of these algorithms to this parameter can be determined analytically. Research work described in this thesis shows that the algorithms have an asymmetric sensitivity to the Hurst parameter, with significant sensitivity in the region where the parameter is underestimated as being close to 0.5. Simulation results reveal the same bias in the performance of the algorithm with regards to the Hurst parameter. In contrast, PERC is insensitive to estimates of the mean, using the sample mean estimator, and estimates of the traffic variance, which is due to the algorithm primarily utilising the correlation structure of the traffic to predict future bandwidth requirements. Sensitivity analysis falls into the area of investigative research, but it naturally leads to the area of robust control, where algorithms are designed so that uncertainty in traffic parameter estimation or modelling can be accommodated. An alternative robust design approach, to the standard maximum entropy approach, is proposed in this thesis that uses the maximum likelihood function to develop the predictive rate controller. The likelihood function defines the proximity of a specific traffic model to the traffic data, and hence gives a measure of the performance of a chosen model. Maximising the likelihood function leads to optimising robust performance, and it is shown, through simulations, that the system performance is close to the optimal performance as compared with maximising the spectral entropy. There is still debate regarding the influence of LRD on network performance. This thesis also considers the question of the influence of LRD on traffic predictability, and demonstrates that predictive rate control algorithms that only use short-term correlations have close performance to algorithms that utilise long-term correlations. It is noted that predictors based on LRD still out-perform ones which use short-term correlations, but that there is Potential simplification in the design of predictors, since traffic predictability can be achieved using short-term correlations. This thesis forms a substantial contribution to the understanding of control in the case where self-similar processes form part of the overall system. Rather than doggedly pursuing self-similar control, a broader view has been taken where the performance of algorithms have been considered from a number of perspectives. A number of different research avenues lead on from this work, and these are outlined

    A new charging scheme for ATM based on QoS

    Get PDF
    PhDNew services are emerging rapidly within the world of telecommunications. Charging strategies that were appropriate for individual transfer capabilities are no longer appropriate for an integrated broadband communications network. There is currently a range of technologies (such as cable television, telephony and narrow band ISDN) for the different services in use and a limited number of charging schemes are applicable for each of the underlying technologies irrespective of the services used over it. Difficulties arise when a wide range of services has to be supported on the same integrated technology such as asynchronous transfer mode (ATM); in such cases the type of service in use and the impact it has on the network becomes much more important. The subject of this thesis, therefore, is the charging strategies for integrated broadband communications networks. That is, the identification of the requirements associated with ATM charging schemes and the proposal of a new approach to charging for ATM called the “quality of service based charging scheme”. Charging for ATM is influenced by three important components: the type and content of a service being offered; the type of customer using the services; and the traffic characteristics belonging to the application supporting the services. The first two issues will largely be dependent on the business and regulatory requirements of the operators. The last item, and an essential one for ATM, is the bridge between technology and business; how are the resources used by a service quantified? Charging that is based on resource usage at the network level was the prime focus of the research reported here. With the proposed charging scheme, a distinction is first made between the four different ATM transfer capabilities that will support various services and the different quality of service requirements that may be applicable to each of them. Then, resources are distributed among buffers set-up to support the combination of these transfer capabilities and quality of services. The buffers are dimensioned according to the M/D/1/K and the ND/D/1 queuing analysis to determine the buffer efficiency and quality of service requirements. This dimensioning provides the basis for fixing the price per unit of resource and time. The actual resource used by a connection is based on the volume of cells transmitted or peak cell rate allocation in combination with traffic shapers if appropriate. Shapers are also dimensioned using the quality of service parameters. Since the buffer 4 efficiency is dependent on the quality of service requirements, users (customers) of ATM networks buy quality of service. The actual price of a connection is further subjected to a number of transformations based on the size of the resource purchased, the time of the day at which a connection is made, and the geographical locality of the destination switch. It is demonstrated that the proposed charging scheme meets all the requirements of customers and of network operators. In addition the result of the comparison of the new scheme with a number of existing, prominent, ATM charging schemes is presented, showing that the performance of the proposed scheme is better in terms of meeting the expectations of both the customers and the network operators

    Direct Optimisation of λ\boldsymbol\lambda for HDR Content Adaptive Transcoding in AV1

    Get PDF
    Since the adoption of VP9 by Netflix in 2016, royalty-free coding standards continued to gain prominence through the activities of the AOMedia consortium. AV1, the latest open source standard, is now widely supported. In the early years after standardisation, HDR video tends to be under served in open source encoders for a variety of reasons including the relatively small amount of true HDR content being broadcast and the challenges in RD optimisation with that material. AV1 codec optimisation has been ongoing since 2020 including consideration of the computational load. In this paper, we explore the idea of direct optimisation of the Lagrangian λ\lambda parameter used in the rate control of the encoders to estimate the optimal Rate-Distortion trade-off achievable for a High Dynamic Range signalled video clip. We show that by adjusting the Lagrange multiplier in the RD optimisation process on a frame-hierarchy basis, we are able to increase the Bjontegaard difference rate gains by more than 3.98×\times on average without visually affecting the quality.Comment: SPIE2022:Applications of Digital Image Processing XLV accepted manuscrip

    Cross-layer signalling and middleware: a survey for inelastic soft real-time applications in MANETs

    Get PDF
    This paper provides a review of the different cross-layer design and protocol tuning approaches that may be used to meet a growing need to support inelastic soft real-time streams in MANETs. These streams are characterised by critical timing and throughput requirements and low packet loss tolerance levels. Many cross-layer approaches exist either for provision of QoS to soft real-time streams in static wireless networks or to improve the performance of real and non-real-time transmissions in MANETs. The common ground and lessons learned from these approaches, with a view to the potential provision of much needed support to real-time applications in MANETs, is therefore discussed

    Mechanisms for QoE optimisation of video traffic: a review paper

    Get PDF
    Transmission of video traffic over the Internet has grown exponentially in the past few years with no sign of waning. This increasing demand for video services has changed user expectation of quality. Various mechanisms have been proposed to optimise the Quality of Experience (QoE) of end users’ video. Studying these approaches are necessary for new methods to be proposed or combination of existing ones to be tailored. We discuss challenges facing the optimisation of QoE for video traffic in this paper. It surveys and classifies these mechanisms based on their functions. The limitation of each of them is identified and future directions are highlighted

    Quality of service over ATM networks

    Get PDF
    PhDAbstract not availabl
    corecore