75 research outputs found

    A policy-based framework towards smooth adaptive playback for dynamic video streaming over HTTP

    Get PDF
    The growth of video streaming in the Internet in the last few years has been highly significant and promises to continue in the future. This fact is related to the growth of Internet users and especially with the diversification of the end-user devices that happens nowadays. Earlier video streaming solutions didn´t consider adequately the Quality of Experience from the user’s perspective. This weakness has been since overcame with the DASH video streaming. The main feature of this protocol is to provide different versions, in terms of quality, of the same content. This way, depending on the status of the network infrastructure between the video server and the user device, the DASH protocol automatically selects the more adequate content version. Thus, it provides to the user the best possible quality for the consumption of that content. The main issue with the DASH protocol is associated to the loop, between each client and video server, which controls the rate of the video stream. In fact, as the network congestion increases, the client requests to the server a video stream with a lower rate. Nevertheless, due to the network latency, the DASH protocol in a standalone way may not be able to stabilize the video stream rate at a level that can guarantee a satisfactory QoE to the end-users. Network programming is a very active and popular topic in the field of network infrastructures management. In this area, the Software Defined Networking paradigm is an approach where a network controller, with a relatively abstracted view of the physical network infrastructure, tries to perform a more efficient management of the data path. The current work studies the combination of the DASH protocol and the Software Defined Networking paradigm in order to achieve a more adequate sharing of the network resources that could benefit both the users’ QoE and network management.O streaming de vídeo na Internet é um fenómeno que tem vindo a crescer de forma significativa nos últimos anos e que promete continuar a crescer no futuro. Este facto está associado ao aumento do número de utilizadores na Internet e, sobretudo, à crescente diversificação de dispositivos que se verifica atualmente. As primeiras soluções utilizadas no streaming de vídeo não acomodavam adequadamente o ponto de vista do utilizador na avaliação da qualidade do vídeo, i.e., a Qualidade de Experiência (QoE) do utilizador. Esta debilidade foi suplantada com o protocolo de streaming de vídeo adaptativo DASH. A principal funcionalidade deste protocolo é fornecer diferente versões, em termos de qualidade, para o mesmo conteúdo. Desta forma, dependendo do estado da infraestrutura de rede entre o servidor de vídeo e o dispositivo do utilizador, o protocolo DASH seleciona automaticamente a versão do conteúdo mais adequada a essas condições. Tal permite fornecer ao utilizador a melhor qualidade possível para o consumo deste conteúdo. O principal problema com o protocolo DASH está associado com o ciclo, entre cada cliente e o servidor de vídeo, que controla o débito de cada fluxo de vídeo. De facto, à medida que a rede fica congestionada, o cliente irá começar a requerer ao servidor um fluxo de vídeo com um débito menor. Ainda assim, devido à latência da rede, o protocolo DASH pode não ser capaz por si só de estabilizar o débito do fluxo de vídeo num nível que consiga garantir uma QoE satisfatória para os utilizadores. A programação de redes é uma área muito popular e ativa na gestão de infraestruturas de redes. Nesta área, o paradigma de Software Defined Networking é uma abordagem onde um controlador da rede, com um ponto de vista relativamente abstrato da infraestrutura física da rede, tenta desempenhar uma gestão mais eficiente do encaminhamento de rede. Neste trabalho estuda-se a junção do protocolo DASH e do paradigma de Software Defined Networking, de forma a atingir uma partilha mais adequada dos recursos da rede. O objetivo é implementar uma solução que seja benéfica tanto para a qualidade de experiência dos utilizadores como para a gestão da rede

    Machine Learning for Multimedia Communications

    Get PDF
    Machine learning is revolutionizing the way multimedia information is processed and transmitted to users. After intensive and powerful training, some impressive efficiency/accuracy improvements have been made all over the transmission pipeline. For example, the high model capacity of the learning-based architectures enables us to accurately model the image and video behavior such that tremendous compression gains can be achieved. Similarly, error concealment, streaming strategy or even user perception modeling have widely benefited from the recent learningoriented developments. However, learning-based algorithms often imply drastic changes to the way data are represented or consumed, meaning that the overall pipeline can be affected even though a subpart of it is optimized. In this paper, we review the recent major advances that have been proposed all across the transmission chain, and we discuss their potential impact and the research challenges that they raise

    QoE-Aware Resource Allocation For Crowdsourced Live Streaming: A Machine Learning Approach

    Get PDF
    In the last decade, empowered by the technological advancements of mobile devices and the revolution of wireless mobile network access, the world has witnessed an explosion in crowdsourced live streaming. Ensuring a stable high-quality playback experience is compulsory to maximize the viewers’ Quality of Experience and the content providers’ profits. This can be achieved by advocating a geo-distributed cloud infrastructure to allocate the multimedia resources as close as possible to viewers, in order to minimize the access delay and video stalls. Additionally, because of the instability of network condition and the heterogeneity of the end-users capabilities, transcoding the original video into multiple bitrates is required. Video transcoding is a computationally expensive process, where generally a single cloud instance needs to be reserved to produce one single video bitrate representation. On demand renting of resources or inadequate resources reservation may cause delay of the video playback or serving the viewers with a lower quality. On the other hand, if resources provisioning is much higher than the required, the extra resources will be wasted. In this thesis, we introduce a prediction-driven resource allocation framework, to maximize the QoE of viewers and minimize the resources allocation cost. First, by exploiting the viewers’ locations available in our unique dataset, we implement a machine learning model to predict the viewers’ number near each geo-distributed cloud site. Second, based on the predicted results that showed to be close to the actual values, we formulate an optimization problem to proactively allocate resources at the viewers’ proximity. Additionally, we will present a trade-off between the video access delay and the cost of resource allocation. Considering the complexity and infeasibility of our offline optimization to respond to the volume of viewing requests in real-time, we further extend our work, by introducing a resources forecasting and reservation framework for geo-distributed cloud sites. First, we formulate an offline optimization problem to allocate transcoding resources at the viewers’ proximity, while creating a tradeoff between the network cost and viewers QoE. Second, based on the optimizer resource allocation decisions on historical live videos, we create our time series datasets containing historical records of the optimal resources needed at each geo-distributed cloud site. Finally, we adopt machine learning to build our distributed time series forecasting models to proactively forecast the exact needed transcoding resources ahead of time at each geo-distributed cloud site. The results showed that the predicted number of transcoding resources needed in each cloud site is close to the optimal number of transcoding resources

    Deep Reinforcement Learning with Importance Weighted A3C for QoE enhancement in Video Delivery Services

    Full text link
    Adaptive bitrate (ABR) algorithms are used to adapt the video bitrate based on the network conditions to improve the overall video quality of experience (QoE). Recently, reinforcement learning (RL) and asynchronous advantage actor-critic (A3C) methods have been used to generate adaptive bit rate algorithms and they have been shown to improve the overall QoE as compared to fixed rule ABR algorithms. However, a common issue in the A3C methods is the lag between behaviour policy and target policy. As a result, the behaviour and the target policies are no longer synchronized which results in suboptimal updates. In this work, we present ALISA: An Actor-Learner Architecture with Importance Sampling for efficient learning in ABR algorithms. ALISA incorporates importance sampling weights to give more weightage to relevant experience to address the lag issues with the existing A3C methods. We present the design and implementation of ALISA, and compare its performance to state-of-the-art video rate adaptation algorithms including vanilla A3C implemented in the Pensieve framework and other fixed-rule schedulers like BB, BOLA, and RB. Our results show that ALISA improves average QoE by up to 25%-48% higher average QoE than Pensieve, and even more when compared to fixed-rule schedulers.Comment: Number of pages: 10, Number of figures: 9, Conference name: 24th IEEE International Symposium on a World of Wireless, Mobile and Multimedia Networks (WoWMoM
    • …
    corecore