18 research outputs found

    Video streaming with quality adaption using collaborative active grid networks

    Get PDF
    Due to the services and demands of the end users, Distributed Computing (Grid Technology, Web Services, and Peer-to-Peer) has been developedrapidJy in thelastyears. Theconvergence of these architectures has been possible using mechanisms such as Collaborative work and Resources Sharing. Grid computing is a platform to enable flexible, secure, controlled, scalable, ubiquitous and heterogeneous services. On the other hand, Video Streaming applications demand a greater deployment over connected Internet users. The present work uses the Acti ve Grid technology as a fundamental platform to give a solution of multimediacontentrecovery. This solution takes into account the following key concepts: collaborative work, multi-source recovery and adapti ve quality. A new archi tecture is designed to deliver video content over a Grid Network. The acti ve and passi ve roles of the nodes are important to guarantee a high quality and efficiency for the video streaming system. The acti ve sender nodes are the content suppliers, while the passive sender nodes wiU perform the backup functions, based on global resource control policies. The aim of the backup node is minirnize the time to restore the systemin caseoffailures. In this way, all participant peers work in a collaborati ve manner following a mul ti -source recovery scheme. Furthermore, Video La yered Encoding is used to manage the video data in a high scalable way, di viding the video in multiple layers. This video codification scheme enables thequality adaptation according to the availability of system resources. In addition, a buffer by sender peer and by layer is needed for an effecti ve control ofthe video retrieve. The QoS will fit considering the state of each buffer and the measurement tools provide by the Acti ve Grid on the network nodes. Ke ywords: Peer -to-Peer Grid Architecture, Services for Active Grids, Streaming Media, Layered Coding, Quality Adaptation, CoUaborative Work.Peer Reviewe

    Video streaming with quality adaption using collaborative active grid networks

    Get PDF
    Due to the services and demands of the end users, Distributed Computing (Grid Technology, Web Services, and Peer-to-Peer) has been developedrapidJy in thelastyears. Theconvergence of these architectures has been possible using mechanisms such as Collaborative work and Resources Sharing. Grid computing is a platform to enable flexible, secure, controlled, scalable, ubiquitous and heterogeneous services. On the other hand, Video Streaming applications demand a greater deployment over connected Internet users. The present work uses the Acti ve Grid technology as a fundamental platform to give a solution of multimediacontentrecovery. This solution takes into account the following key concepts: collaborative work, multi-source recovery and adapti ve quality. A new archi tecture is designed to deliver video content over a Grid Network. The acti ve and passi ve roles of the nodes are important to guarantee a high quality and efficiency for the video streaming system. The acti ve sender nodes are the content suppliers, while the passive sender nodes wiU perform the backup functions, based on global resource control policies. The aim of the backup node is minirnize the time to restore the systemin caseoffailures. In this way, all participant peers work in a collaborati ve manner following a mul ti -source recovery scheme. Furthermore, Video La yered Encoding is used to manage the video data in a high scalable way, di viding the video in multiple layers. This video codification scheme enables thequality adaptation according to the availability of system resources. In addition, a buffer by sender peer and by layer is needed for an effecti ve control ofthe video retrieve. The QoS will fit considering the state of each buffer and the measurement tools provide by the Acti ve Grid on the network nodes. Ke ywords: Peer -to-Peer Grid Architecture, Services for Active Grids, Streaming Media, Layered Coding, Quality Adaptation, CoUaborative Work.Peer Reviewe

    Codage réseau pour des applications multimédias avancées

    Get PDF
    Network coding is a paradigm that allows an efficient use of the capacity of communication networks. It maximizes the throughput in a multi-hop multicast communication and reduces the delay. In this thesis, we focus our attention to the integration of the network coding framework to multimedia applications, and in particular to advanced systems that provide enhanced video services to the users. Our contributions concern several instances of advanced multimedia communications: an efficient framework for transmission of a live stream making joint use of network coding and multiple description coding; a novel transmission strategy for lossy wireless networks that guarantees a trade-off between loss resilience and short delay based on a rate-distortion optimized scheduling of the video frames, that we also extended to the case of interactive multi-view streaming; a distributed social caching system that, using network coding in conjunction with the knowledge of the users' preferences in terms of views, is able to select a replication scheme such that to provide a high video quality by accessing only other members of the social group without incurring the access cost associated with a connection to a central server and without exchanging large tables of metadata to keep track of the replicated parts; and, finally, a study on using blind source separation techniques to reduce the overhead incurred by network coding schemes based on error-detecting techniques such as parity coding and message digest generation. All our contributions are aimed at using network coding to enhance the quality of video transmission in terms of distortion and delay perceivedLe codage réseau est un paradigme qui permet une utilisation efficace du réseau. Il maximise le débit dans un réseau multi-saut en multicast et réduit le retard. Dans cette thèse, nous concentrons notre attention sur l’intégration du codage réseau aux applications multimédias, et en particulier aux systèmes avancès qui fournissent un service vidéo amélioré pour les utilisateurs. Nos contributions concernent plusieurs scénarios : un cadre de fonctions efficace pour la transmission de flux en directe qui utilise à la fois le codage réseau et le codage par description multiple, une nouvelle stratégie de transmission pour les réseaux sans fil avec perte qui garantit un compromis entre la résilience vis-à-vis des perte et la reduction du retard sur la base d’une optimisation débit-distorsion de l'ordonnancement des images vidéo, que nous avons également étendu au cas du streaming multi-vue interactive, un système replication sociale distribuée qui, en utilisant le réseau codage en relation et la connaissance des préférences des utilisateurs en termes de vue, est en mesure de sélectionner un schéma de réplication capable de fournir une vidéo de haute qualité en accédant seulement aux autres membres du groupe social, sans encourir le coût d’accès associé à une connexion à un serveur central et sans échanger des larges tables de métadonnées pour tenir trace des éléments répliqués, et, finalement, une étude sur l’utilisation de techniques de séparation aveugle de source -pour réduire l’overhead encouru par les schémas de codage réseau- basé sur des techniques de détection d’erreur telles que le codage de parité et la génération de message digest

    Methods to enhance content distribution for very large scale online communities

    Get PDF
    The Internet has experienced an exponential growth in the last years, and its number of users far from decay keeps on growing. Popular Web 2.0 services such as Facebook, YouTube or Twitter among others sum millions of users and employ vast infrastructures deployed worldwide. The size of these infrastructures is getting huge in order to support such a massive number of users. This increment of the infrastructure size has brought new problems regarding scalability, power consumption, cooling, hardware lifetime, underutilization, investment recovery, etc. Owning this kind of infrastructures is not always affordable nor convenient. This could be a major handicap for starting projects with a humble budget whose success is based on reaching a large audience. However, current technologies might permit to deploy vast infrastructures reducing their cost. We refer to peer-to-peer networks and cloud computing. Peer-to-peer systems permit users to yield their own resources to distributed infrastructures. These systems have demonstrated to be a valuable choice capable of distributing vast amounts of data to large audiences with a minimal starting infrastructure. Nevertheless, aspects such as content availability cannot be controlled in these systems, whereas classic server infrastructures can improve this aspect. In the recent time, the cloud has been revealed as a promising paradigm for hosting horizontally scalable Web systems. The cloud offers elastic capabilities that permit to save costs by adapting the number of resources to the incoming demand. Additionally, the cloud makes accessible a vast amount of resources that may be employed on peak workloads. However, how to determine the amount of resources to use remains a challenge. In this thesis, we describe a hierarchical architecture that combines both: peer-to-peer and elastic server infrastructures in order to enhance content distribution. The peer-topeer infrastructure brings a scalable solution that reduces the workload in the servers, while the server infrastructure assures availability and reduces costs varying its size when necessary. We propose a distributed collaborative caching infrastructure that employs a clusterbased locality-aware self-organizing P2P system. This system, leverages collaborative data classification in order to improve content locality. Our evaluation demonstrates that incrementing data locality permits to improve data search while reducing traffic. We explore the utilization of elastic server infrastructures addressing three issues: system sizing, data grouping and content distribution. We propose novel multi-model techniques for hierarchical workload prediction. These predictions are employed to determine the system size and request distribution policies. Additionally, we propose novel techniques for adaptive control that permit to identify inaccurate models and redefine them. Our evaluation using traces extracted from real systems indicate that the utilization of a hierarchy of multiple models increases prediction accuracy. This hierarchy in conjunction with our adaptive control techniques increments the accuracy during unexpected workload variations. Finally, we demonstrate that locality-aware request distribution policies can take advantage of prediction models to adequate content distribution independently of the system size

    User-centric power-friendly quality-based network selection strategy for heterogeneous wireless environments

    Get PDF
    The ‘Always Best Connected’ vision is built around the scenario of a mobile user seamlessly roaming within a multi-operator multi-technology multi-terminal multi-application multi-user environment supported by the next generation of wireless networks. In this heterogeneous environment, users equipped with multi-mode wireless mobile devices will access rich media services via one or more access networks. All these access networks may differ in terms of technology, coverage range, available bandwidth, operator, monetary cost, energy usage etc. In this context, there is a need for a smart network selection decision to be made, to choose the best available network option to cater for the user’s current application and requirements. The decision is a difficult one, especially given the number and dynamics of the possible input parameters. What parameters are used and how those parameters model the application requirements and user needs is important. Also, game theory approaches can be used to model and analyze the cooperative or competitive interaction between the rational decision makers involved, which are users, seeking to get good service quality at good value prices, and/or the network operators, trying to increase their revenue. This thesis presents the roadmap towards an ‘Always Best Connected’ environment. The proposed solution includes an Adapt-or-Handover solution which makes use of a Signal Strength-based Adaptive Multimedia Delivery mechanism (SAMMy) and a Power-Friendly Access Network Selection Strategy (PoFANS) in order to help the user in taking decisions, and to improve the energy efficiency at the end-user mobile device. A Reputation-based System is proposed, which models the user-network interaction as a repeated cooperative game following the repeated Prisoner’s Dilemma game from Game Theory. It combines reputation-based systems, game theory and a network selection mechanism in order to create a reputation-based heterogeneous environment. In this environment, the users keep track of their individual history with the visited networks. Every time, a user connects to a network the user-network interaction game is played. The outcome of the game is a network reputation factor which reflects the network’s previous behavior in assuring service guarantees to the user. The network reputation factor will impact the decision taken by the user next time, when he/she will have to decide whether to connect or not to that specific network. The performance of the proposed solutions was evaluated through in-depth analysis and both simulation-based and experimental-oriented testing. The results clearly show improved performance of the proposed solutions in comparison with other similar state-of-the-art solutions. An energy consumption study for a Google Nexus One streaming adaptive multimedia was performed, and a comprehensive survey on related Game Theory research are provided as part of the work

    Sixth Goddard Conference on Mass Storage Systems and Technologies Held in Cooperation with the Fifteenth IEEE Symposium on Mass Storage Systems

    Get PDF
    This document contains copies of those technical papers received in time for publication prior to the Sixth Goddard Conference on Mass Storage Systems and Technologies which is being held in cooperation with the Fifteenth IEEE Symposium on Mass Storage Systems at the University of Maryland-University College Inn and Conference Center March 23-26, 1998. As one of an ongoing series, this Conference continues to provide a forum for discussion of issues relevant to the management of large volumes of data. The Conference encourages all interested organizations to discuss long term mass storage requirements and experiences in fielding solutions. Emphasis is on current and future practical solutions addressing issues in data management, storage systems and media, data acquisition, long term retention of data, and data distribution. This year's discussion topics include architecture, tape optimization, new technology, performance, standards, site reports, vendor solutions. Tutorials will be available on shared file systems, file system backups, data mining, and the dynamics of obsolescence

    Creation of value with open source software in the telecommunications field

    Get PDF
    Tese de doutoramento. Engenharia Electrotécnica e de Computadores. Faculdade de Engenharia. Universidade do Porto. 200

    Net Neutrality

    Get PDF
    This book is available as open access through the Bloomsbury Open Access programme and is available on www.bloomsburycollections.com. Chris Marsden maneuvers through the hype articulated by Netwrok Neutrality advocates and opponents. He offers a clear-headed analysis of the high stakes in this debate about the Internet's future, and fearlessly refutes the misinformation and misconceptions that about' Professor Rob Freiden, Penn State University Net Neutrality is a very heated and contested policy principle regarding access for content providers to the Internet end-user, and potential discrimination in that access where the end-user's ISP (or another ISP) blocks that access in part or whole. The suggestion has been that the problem can be resolved by either introducing greater competition, or closely policing conditions for vertically integrated service, such as VOIP. However, that is not the whole story, and ISPs as a whole have incentives to discriminate between content for matters such as network management of spam, to secure and maintain customer experience at current levels, and for economic benefit from new Quality of Service standards. This includes offering a ‘priority lane' on the network for premium content types such as video and voice service. The author considers market developments and policy responses in Europe and the United States, draws conclusions and proposes regulatory recommendations

    Resource Management schemes for mobile ad hoc networks

    Get PDF
    Ph.DDOCTOR OF PHILOSOPH

    Net Neutrality

    Get PDF
    This book is available as open access through the Bloomsbury Open Access programme and is available on www.bloomsburycollections.com. Chris Marsden maneuvers through the hype articulated by Netwrok Neutrality advocates and opponents. He offers a clear-headed analysis of the high stakes in this debate about the Internet's future, and fearlessly refutes the misinformation and misconceptions that about' Professor Rob Freiden, Penn State University Net Neutrality is a very heated and contested policy principle regarding access for content providers to the Internet end-user, and potential discrimination in that access where the end-user's ISP (or another ISP) blocks that access in part or whole. The suggestion has been that the problem can be resolved by either introducing greater competition, or closely policing conditions for vertically integrated service, such as VOIP. However, that is not the whole story, and ISPs as a whole have incentives to discriminate between content for matters such as network management of spam, to secure and maintain customer experience at current levels, and for economic benefit from new Quality of Service standards. This includes offering a ‘priority lane' on the network for premium content types such as video and voice service. The author considers market developments and policy responses in Europe and the United States, draws conclusions and proposes regulatory recommendations
    corecore