149 research outputs found
The Virtue of Patience when Scheduling Media in Presence of Feedback
We consider streaming of pre-encoded and packetized media over best-effort networks in presence of acknowledgment feedback. Given an estimation of future transmission resources and knowing about past transmissions and received acknowledgments, a scheduling algorithm is defined as a mechanism that selects the data to send over the network at any given time, so as to minimize the end-to-end distortion. Our work first reveals the suboptimality of popular greedy schedulers, which might be strongly penalized by anticipated retransmissions. It then proposes an original scheduling algorithm that avoids premature retransmissions, while preserving the simplicity of the greedy paradigm. The proposed patient greedy (PG) scheduler appears to save up to 50% of rate in comparison with the conventional greedy approach
MediaSync: Handbook on Multimedia Synchronization
This book provides an approachable overview of the most recent advances in the fascinating field of media synchronization (mediasync), gathering contributions from the most representative and influential experts. Understanding the challenges of this field in the current multi-sensory, multi-device, and multi-protocol world is not an easy task. The book revisits the foundations of mediasync, including theoretical frameworks and models, highlights ongoing research efforts, like hybrid broadband broadcast (HBB) delivery and users' perception modeling (i.e., Quality of Experience or QoE), and paves the way for the future (e.g., towards the deployment of multi-sensory and ultra-realistic experiences). Although many advances around mediasync have been devised and deployed, this area of research is getting renewed attention to overcome remaining challenges in the next-generation (heterogeneous and ubiquitous) media ecosystem. Given the significant advances in this research area, its current relevance and the multiple disciplines it involves, the availability of a reference book on mediasync becomes necessary. This book fills the gap in this context. In particular, it addresses key aspects and reviews the most relevant contributions within the mediasync research space, from different perspectives. Mediasync: Handbook on Multimedia Synchronization is the perfect companion for scholars and practitioners that want to acquire strong knowledge about this research area, and also approach the challenges behind ensuring the best mediated experiences, by providing the adequate synchronization between the media elements that constitute these experiences
Content-Aware Multimedia Communications
The demands for fast, economic and reliable dissemination of multimedia
information are steadily growing within our society. While people and
economy increasingly rely on communication technologies, engineers still
struggle with their growing complexity.
Complexity in multimedia communication originates from several sources. The
most prominent is the unreliability of packet networks like the Internet.
Recent advances in scheduling and error control mechanisms for streaming
protocols have shown that the quality and robustness of multimedia delivery
can be improved significantly when protocols are aware of the content they
deliver. However, the proposed mechanisms require close cooperation between
transport systems and application layers which increases the overall system
complexity. Current approaches also require expensive metrics and focus on
special encoding formats only. A general and efficient model is missing so
far.
This thesis presents efficient and format-independent solutions to support
cross-layer coordination in system architectures. In particular, the first
contribution of this work is a generic dependency model that enables
transport layers to access content-specific properties of media streams,
such as dependencies between data units and their importance. The second
contribution is the design of a programming model for streaming
communication and its implementation as a middleware architecture. The
programming model hides the complexity of protocol stacks behind simple
programming abstractions, but exposes cross-layer control and monitoring
options to application programmers. For example, our interfaces allow
programmers to choose appropriate failure semantics at design time while
they can refine error protection and visibility of low-level errors at
run-time.
Based on some examples we show how our middleware simplifies the
integration of stream-based communication into large-scale application
architectures. An important result of this work is that despite cross-layer
cooperation, neither application nor transport protocol designers
experience an increase in complexity. Application programmers can even
reuse existing streaming protocols which effectively increases system
robustness.Der Bedarf unsere Gesellschaft nach kostengünstiger und
zuverlässiger
Kommunikation wächst stetig. Während wir uns selbst immer mehr von modernen
Kommunikationstechnologien abhängig machen, müssen die Ingenieure dieser
Technologien sowohl den Bedarf nach schneller Einführung neuer Produkte
befriedigen als auch die wachsende Komplexität der Systeme beherrschen.
Gerade die Übertragung multimedialer Inhalte wie Video und Audiodaten ist
nicht trivial. Einer der prominentesten Gründe dafür ist die
Unzuverlässigkeit heutiger Netzwerke, wie z.B.~dem Internet. Paketverluste
und schwankende Laufzeiten können die Darstellungsqualität massiv
beeinträchtigen. Wie jüngste Entwicklungen im Bereich der
Streaming-Protokolle zeigen, sind jedoch Qualität und Robustheit der
Übertragung effizient kontrollierbar, wenn Streamingprotokolle
Informationen über den Inhalt der transportierten Daten ausnutzen.
Existierende Ansätze, die den Inhalt von Multimediadatenströmen
beschreiben, sind allerdings meist auf einzelne Kompressionsverfahren
spezialisiert und verwenden berechnungsintensive Metriken. Das reduziert
ihren praktischen Nutzen deutlich. Außerdem erfordert der
Informationsaustausch eine enge Kooperation zwischen Applikationen und
Transportschichten. Da allerdings die Schnittstellen aktueller
Systemarchitekturen nicht darauf vorbereitet sind, müssen entweder die
Schnittstellen erweitert oder alternative Architekturkonzepte geschaffen
werden. Die Gefahr beider Varianten ist jedoch, dass sich die Komplexität
eines Systems dadurch weiter erhöhen kann.
Das zentrale Ziel dieser Dissertation ist es deshalb,
schichtenübergreifende Koordination bei gleichzeitiger Reduzierung der
Komplexität zu erreichen. Hier leistet die Arbeit zwei Beträge zum
aktuellen Stand der Forschung. Erstens definiert sie ein universelles
Modell zur Beschreibung von Inhaltsattributen, wie Wichtigkeiten und
Abhängigkeitsbeziehungen innerhalb eines Datenstroms. Transportschichten
können dieses Wissen zur effizienten Fehlerkontrolle verwenden. Zweitens
beschreibt die Arbeit das Noja Programmiermodell für multimediale
Middleware. Noja definiert Abstraktionen zur Übertragung und Kontrolle
multimedialer Ströme, die die Koordination von Streamingprotokollen mit
Applikationen ermöglichen. Zum Beispiel können Programmierer geeignete
Fehlersemantiken und Kommunikationstopologien auswählen und den konkreten
Fehlerschutz dann zur Laufzeit verfeinern und kontrolliere
Packetized Media Streaming with Comprehensive Exploitation of Feedback Information
This paper addresses the problem of streaming packetized media over a lossy packet network, with sender-driven (re)transmission using acknowledgement feedback. The different transmission scenarios associated to a group of interdependent media data units are abstracted in terms of a finite alphabet of policies, for each single data unit. A rate-distortion optimized markovian framework is proposed, which supports the use of comprehensive feedback information. Contrarily to previous works in rate-distortion optimized streaming, whose transmission policies definitions do not take into account the feedback expected for other data units, our framework considers all the acknowledgment packets in defining the streaming policy of a single data unit. More specifically, the notion of master and slave data unit is introduced, to define dependent streaming policies between media packets; the policy adopted to transmit a slave data unit becomes dependent on the acknowledgments received about its masters. One of the main contributions of our work is to propose a methodology that limits the space of dependent policies for the RD optimized streaming strategy. A number of rules are formulated to select a set of relevant master/slave relationships, defined as the dependencies that are likely to bring RD performance gain in the streaming system. These rules provide a limited complexity solution to the rate-distortion optimized streaming problem, with comprehensive use of feedback information. Based on extensive simulations, we conclude that (i) the proposed set of relevant dependent policies achieves close to optimal performance, while being computationally tractable, and (ii) the benefit of dependent policies is driven by the relative sizes and importance of interdependent data units. Our simulations demonstrate that dependent streaming policies can perform significantly better than independent streaming strategies, especially for cases where some media data units bring a relatively large gain in distortion, in comparison with other data units they depend on for correct decoding. We observe however that the benefit becomes marginal when the gain in distortion per unit of rate decreases along the media decoding dependency path. Since such a trend characterizes most conventional scalable coders, the implementation of dependent policies can reasonably be ruled out in these specific cases
Flexi-WVSNP-DASH: A Wireless Video Sensor Network Platform for the Internet of Things
abstract: Video capture, storage, and distribution in wireless video sensor networks
(WVSNs) critically depends on the resources of the nodes forming the sensor
networks. In the era of big data, Internet of Things (IoT), and distributed
demand and solutions, there is a need for multi-dimensional data to be part of
the Sensor Network data that is easily accessible and consumable by humanity as
well as machinery. Images and video are expected to become as ubiquitous as is
the scalar data in traditional sensor networks. The inception of video-streaming
over the Internet, heralded a relentless research for effective ways of
distributing video in a scalable and cost effective way. There has been novel
implementation attempts across several network layers. Due to the inherent
complications of backward compatibility and need for standardization across
network layers, there has been a refocused attention to address most of the
video distribution over the application layer. As a result, a few video
streaming solutions over the Hypertext Transfer Protocol (HTTP) have been
proposed. Most notable are Apple’s HTTP Live Streaming (HLS) and the Motion
Picture Experts Groups Dynamic Adaptive Streaming over HTTP (MPEG-DASH). These
frameworks, do not address the typical and future WVSN use cases. A highly
flexible Wireless Video Sensor Network Platform and compatible DASH (WVSNP-DASH)
are introduced. The platform's goal is to usher video as a data element that
can be integrated into traditional and non-Internet networks. A low cost,
scalable node is built from the ground up to be fully compatible with the
Internet of Things Machine to Machine (M2M) concept, as well as the ability to
be easily re-targeted to new applications in a short time. Flexi-WVSNP design
includes a multi-radio node, a middle-ware for sensor operation and
communication, a cross platform client facing data retriever/player framework,
scalable security as well as a cohesive but decoupled hardware and software
design.Dissertation/ThesisDoctoral Dissertation Electrical Engineering 201
Recommended from our members
Interoperability of wireless communication technologies in hybrid networks: Evaluation of end-to-end interoperability issues and quality of service requirements
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Hybrid Networks employing wireless communication technologies have nowadays brought closer the vision of communication “anywhere, any time with anyone”. Such communication technologies consist of various standards, protocols, architectures, characteristics, models, devices, modulation and coding techniques. All these different technologies naturally may share some common characteristics, but there are also many important differences. New advances in these technologies are emerging very rapidly, with the advent of new models, characteristics, protocols and architectures. This rapid evolution imposes many challenges and issues to be addressed, and of particular importance are the interoperability issues of the following wireless technologies: Wireless Fidelity (Wi-Fi) IEEE802.11, Worldwide Interoperability for Microwave Access (WiMAX) IEEE 802.16, Single Channel per Carrier (SCPC), Digital Video Broadcasting of Satellite (DVB-S/DVB-S2), and Digital Video Broadcasting Return Channel through Satellite (DVB-RCS). Due to the differences amongst wireless technologies, these technologies do not generally interoperate easily with each other because of various interoperability and Quality of Service (QoS) issues.
The aim of this study is to assess and investigate end-to-end interoperability issues and QoS requirements, such as bandwidth, delays, jitter, latency, packet loss, throughput, TCP performance, UDP performance, unicast and multicast services and availability, on hybrid wireless communication networks (employing both satellite broadband and terrestrial wireless technologies).
The thesis provides an introduction to wireless communication technologies followed by a review of previous research studies on Hybrid Networks (both satellite and terrestrial wireless technologies, particularly Wi-Fi, WiMAX, DVB-RCS, and SCPC). Previous studies have discussed Wi-Fi, WiMAX, DVB-RCS, SCPC and 3G technologies and their standards as well as their properties and characteristics, such as operating frequency, bandwidth, data rate, basic configuration, coverage, power, interference, social issues, security problems, physical and MAC layer design and development issues. Although some previous studies provide valuable contributions to this area of research, they are limited to link layer characteristics, TCP performance, delay, bandwidth, capacity, data rate, and throughput. None of the studies cover all aspects of end-to-end interoperability issues and QoS requirements; such as bandwidth, delay, jitter, latency, packet loss, link performance, TCP and UDP performance, unicast and multicast performance, at end-to-end level, on Hybrid wireless networks.
Interoperability issues are discussed in detail and a comparison of the different technologies and protocols was done using appropriate testing tools, assessing various performance measures including: bandwidth, delay, jitter, latency, packet loss, throughput and availability testing. The standards, protocol suite/ models and architectures for Wi-Fi, WiMAX, DVB-RCS, SCPC, alongside with different platforms and applications, are discussed and compared. Using a robust approach, which includes a new testing methodology and a generic test plan, the testing was conducted using various realistic test scenarios on real networks, comprising variable numbers and types of nodes. The data, traces, packets, and files were captured from various live scenarios and sites. The test results were analysed in order to measure and compare the characteristics of wireless technologies, devices, protocols and applications.
The motivation of this research is to study all the end-to-end interoperability issues and Quality of Service requirements for rapidly growing Hybrid Networks in a comprehensive and systematic way.
The significance of this research is that it is based on a comprehensive and systematic investigation of issues and facts, instead of hypothetical ideas/scenarios or simulations, which informed the design of a test methodology for empirical data gathering by real network testing, suitable for the measurement of hybrid network single-link or end-to-end issues using proven test tools.
This systematic investigation of the issues encompasses an extensive series of tests measuring delay, jitter, packet loss, bandwidth, throughput, availability, performance of audio and video session, multicast and unicast performance, and stress testing. This testing covers most common test scenarios in hybrid networks and gives recommendations in achieving good end-to-end interoperability and QoS in hybrid networks.
Contributions of study include the identification of gaps in the research, a description of interoperability issues, a comparison of most common test tools, the development of a generic test plan, a new testing process and methodology, analysis and network design recommendations for end-to-end interoperability issues and QoS requirements. This covers the complete cycle of this research.
It is found that UDP is more suitable for hybrid wireless network as compared to TCP, particularly for the demanding applications considered, since TCP presents significant problems for multimedia and live traffic which requires strict QoS requirements on delay, jitter, packet loss and bandwidth. The main bottleneck for satellite communication is the delay of approximately 600 to 680 ms due to the long distance factor (and the finite speed of light) when communicating over geostationary satellites.
The delay and packet loss can be controlled using various methods, such as traffic classification, traffic prioritization, congestion control, buffer management, using delay compensator, protocol compensator, developing automatic request technique, flow scheduling, and bandwidth allocation
Recommended from our members
Network coding for sensor networks, distributed storage and video streaming
The classical store-and-forward routing has and will continue to be the most important routing architecture in many modern packet-switched communication networks. In a packet-switched network, data is sent in the form of discrete packets that traverse hop-by-hop from a source to a destination. At each intermediate hop, the router stores and examines the packets it receives then forwards them to the next hop until they reach the correct destinations according to some pre-defined routing algorithms. Importantly, the intermediate routers do not modify but simply store and forward the contents of the packets. In contrast, a new generalized approach to routing called Network Coding (NC) allows the intermediate routers to modify and combine packets from different sources and destinations in such a way that increases the overall throughput. The core idea of NC allowing the intermediate nodes in a network to perform data processing has a wide range of applications well beyond its initial application to routing, impacting different disciplines from distributed data storage and security to energy efficient sensor networks and Internet media streaming. To that end, this dissertation aims to develop the theories and applications of NC via four main thrusts:
1) Energy efficient NC techniques for sensor networks,
2) Novel NC techniques and protocols for Internet video streaming,
3) Stochastic data replenishment for large scale NC-based distributed storage
systems,
4) Real-world implementation of NC-based distributed video streaming system.
In thrust one, we describe a novel cross-sensor coding technique that combines
network topology and coding techniques to maximize the life-time of a sensor network,
by addressing the uneven energy consumption problem in data gathering
sensor networks where the nodes closer to the sink tend to consume more energy
than those of the farther nodes. Our approach is based on the following observation
from the sensor networks using On-Off Keying and digital transmission:
transmitting bit "1" consumes much more energy than bit "0". Our proposed
coding technique exploits this difference to reduce the communication energy by
limiting the number of bits "1" in the output codeword (low-weight codeword) and
to use NC-based cross-sensor coding technique to equalize the communication energy
among the nodes. This cross-sensor coding scheme can significantly extend
the network lifetime as compared with traditional (binary) coding by solving the
energy-consumption unfairness problem. The theoretical and experimental results
confirm that transmission energy can be reduced substantially (e.g., a factor of 15)
and the unequal energy consumption among nodes can be practically eliminated.
In thrust two, we describe a rate distortion aware hierarchical NC technique
and transport protocol for Internet video streaming. We begin by proposing
a NC-based multi-sender streaming framework that reduces the overall storage,
eliminates the complexity of sender synchronization, and enables TCP streaming.
Furthermore, we propose a Hierarchical Network Coding (HNC) technique that
facilitates scalable video streaming to combat bandwidth fluctuation on the Internet.
This HNC technique enables receiver to recover the important data gracefully
in the presence of limited bandwidth which causes an increase in decoding delay.
Simulations demonstrate that under certain scenarios, our proposed NC techniques
can result in bandwidth saving up to 60% over the traditional schemes.
In thrust three, we present a theory of NC-based data replenishment to automate
the process of data maintenance for large scale distributed storage systems.
The data replenishment mechanism is the core of these systems that promises to
reduce the coordination complexity and increases performance scalability. The
data replenishment automates the process of maintaining a sufficient level of data
redundancy to ensure the availability of data in presence of peer departures and
failures. The dynamics of peers entering and leaving the network is modeled as
a stochastic process. We propose a novel analytical time-backward technique to
bound the expected time, the longer the better, for a piece of data to remain in
P2P systems. Both theoretical and simulation results are in agreement, indicating
that our proposed data replenishment via random linear network coding (RLNC)
outperforms other popular strategies that employ repetition and channel coding
techniques. Specifically, we show that the expected time for a piece of data to
remain in a P2P system is exponential in the number of peers used to store the
data for the RLNC-based strategy, while they are quadratic for other strategies.
Furthermore, the time-backward technique can be applied to problems in other
disciplines such as gene population modeling in theoretical biology.
Finally in thrust four, we present the architecture, design, and experimental
results of an actual NC-based distributed video streaming system. We first implement
random linear network coding (RLNC) library and show the feasibility of
using RLNC in P2P video streaming applications. Then we design, implement and
analyze RESnc - a resilient P2P video storage and streaming over the Internet using
network coding. RESnc increases the streaming throughput and data resiliency
against peer departures and failures using peer diversity. These improvements are
based on three architectural elements:
1) The RLNC scheme that breaks a video stream into multiple smaller pieces,
codes, and disperses them throughout peers in the network, in such a way to
maximize the probability of recovering the original video under peer departures
and failures;
2) The scalable mechanism for automating the data replenishment process using
RLNC to maintain a sufficient level of redundancy for video stored in the system;
3) The path-diversity streaming protocol for a client to simultaneously stream
a video from multiple peers with minimal coordination.
Experimental results demonstrated that our system adapts well with bandwidth
fluctuation, provides significant playback quality improvement and bandwidth saving
Smart Wireless Sensor Networks
The recent development of communication and sensor technology results in the growth of a new attractive and challenging area - wireless sensor networks (WSNs). A wireless sensor network which consists of a large number of sensor nodes is deployed in environmental fields to serve various applications. Facilitated with the ability of wireless communication and intelligent computation, these nodes become smart sensors which do not only perceive ambient physical parameters but also be able to process information, cooperate with each other and self-organize into the network. These new features assist the sensor nodes as well as the network to operate more efficiently in terms of both data acquisition and energy consumption. Special purposes of the applications require design and operation of WSNs different from conventional networks such as the internet. The network design must take into account of the objectives of specific applications. The nature of deployed environment must be considered. The limited of sensor nodes� resources such as memory, computational ability, communication bandwidth and energy source are the challenges in network design. A smart wireless sensor network must be able to deal with these constraints as well as to guarantee the connectivity, coverage, reliability and security of network's operation for a maximized lifetime. This book discusses various aspects of designing such smart wireless sensor networks. Main topics includes: design methodologies, network protocols and algorithms, quality of service management, coverage optimization, time synchronization and security techniques for sensor networks
Interoperability of wireless communication technologies in hybrid networks : evaluation of end-to-end interoperability issues and quality of service requirements
Hybrid Networks employing wireless communication technologies have nowadays brought closer the vision of communication “anywhere, any time with anyone”. Such communication technologies consist of various standards, protocols, architectures, characteristics, models, devices, modulation and coding techniques. All these different technologies naturally may share some common characteristics, but there are also many important differences. New advances in these technologies are emerging very rapidly, with the advent of new models, characteristics, protocols and architectures. This rapid evolution imposes many challenges and issues to be addressed, and of particular importance are the interoperability issues of the following wireless technologies: Wireless Fidelity (Wi-Fi) IEEE802.11, Worldwide Interoperability for Microwave Access (WiMAX) IEEE 802.16, Single Channel per Carrier (SCPC), Digital Video Broadcasting of Satellite (DVB-S/DVB-S2), and Digital Video Broadcasting Return Channel through Satellite (DVB-RCS). Due to the differences amongst wireless technologies, these technologies do not generally interoperate easily with each other because of various interoperability and Quality of Service (QoS) issues. The aim of this study is to assess and investigate end-to-end interoperability issues and QoS requirements, such as bandwidth, delays, jitter, latency, packet loss, throughput, TCP performance, UDP performance, unicast and multicast services and availability, on hybrid wireless communication networks (employing both satellite broadband and terrestrial wireless technologies). The thesis provides an introduction to wireless communication technologies followed by a review of previous research studies on Hybrid Networks (both satellite and terrestrial wireless technologies, particularly Wi-Fi, WiMAX, DVB-RCS, and SCPC). Previous studies have discussed Wi-Fi, WiMAX, DVB-RCS, SCPC and 3G technologies and their standards as well as their properties and characteristics, such as operating frequency, bandwidth, data rate, basic configuration, coverage, power, interference, social issues, security problems, physical and MAC layer design and development issues. Although some previous studies provide valuable contributions to this area of research, they are limited to link layer characteristics, TCP performance, delay, bandwidth, capacity, data rate, and throughput. None of the studies cover all aspects of end-to-end interoperability issues and QoS requirements; such as bandwidth, delay, jitter, latency, packet loss, link performance, TCP and UDP performance, unicast and multicast performance, at end-to-end level, on Hybrid wireless networks. Interoperability issues are discussed in detail and a comparison of the different technologies and protocols was done using appropriate testing tools, assessing various performance measures including: bandwidth, delay, jitter, latency, packet loss, throughput and availability testing. The standards, protocol suite/ models and architectures for Wi-Fi, WiMAX, DVB-RCS, SCPC, alongside with different platforms and applications, are discussed and compared. Using a robust approach, which includes a new testing methodology and a generic test plan, the testing was conducted using various realistic test scenarios on real networks, comprising variable numbers and types of nodes. The data, traces, packets, and files were captured from various live scenarios and sites. The test results were analysed in order to measure and compare the characteristics of wireless technologies, devices, protocols and applications. The motivation of this research is to study all the end-to-end interoperability issues and Quality of Service requirements for rapidly growing Hybrid Networks in a comprehensive and systematic way. The significance of this research is that it is based on a comprehensive and systematic investigation of issues and facts, instead of hypothetical ideas/scenarios or simulations, which informed the design of a test methodology for empirical data gathering by real network testing, suitable for the measurement of hybrid network single-link or end-to-end issues using proven test tools. This systematic investigation of the issues encompasses an extensive series of tests measuring delay, jitter, packet loss, bandwidth, throughput, availability, performance of audio and video session, multicast and unicast performance, and stress testing. This testing covers most common test scenarios in hybrid networks and gives recommendations in achieving good end-to-end interoperability and QoS in hybrid networks. Contributions of study include the identification of gaps in the research, a description of interoperability issues, a comparison of most common test tools, the development of a generic test plan, a new testing process and methodology, analysis and network design recommendations for end-to-end interoperability issues and QoS requirements. This covers the complete cycle of this research. It is found that UDP is more suitable for hybrid wireless network as compared to TCP, particularly for the demanding applications considered, since TCP presents significant problems for multimedia and live traffic which requires strict QoS requirements on delay, jitter, packet loss and bandwidth. The main bottleneck for satellite communication is the delay of approximately 600 to 680 ms due to the long distance factor (and the finite speed of light) when communicating over geostationary satellites. The delay and packet loss can be controlled using various methods, such as traffic classification, traffic prioritization, congestion control, buffer management, using delay compensator, protocol compensator, developing automatic request technique, flow scheduling, and bandwidth allocation.EThOS - Electronic Theses Online ServiceGBUnited Kingdo
Multimedia-Streaming in Benutzergruppen
At the time being, multimedia services using IP technology like IPTV or video on-demand are a hot topic. Technically, they can be classified under the notion of streaming. A server sends media data in a continuous fashion to one or several clients, which consume and display data portions as soon as they arrive. Using a feedback channel customers may influence the play-back, watching programs time-shifted or pausing the program. An enhancement of such streaming services is to watch those movies with a group of people on several devices in parallel. Similar approaches have been developed using IP multicast. However, users cannot control the presentation: pausing or skipping of more unimportant parts is impossible. Moreover, members cannot be added to the session directly within the application. The costream architecture developed in this works offers a collaborative streaming service without these limitations: People may join others watching a movie or invite others to such a collaborative streaming session. Dependent on the desired course of the session the participants' control operations are executed for all users, or the group is split into subgroups to let watchers follow their own time-lines. A group management controls this by means of user roles. Separate from the group management, the so-called association service provides for streaming session control and synchronization among participants. This separation of duties is advantageous in the sense that standard components can be used: For group management, SIP conferencing servers are suitable, whereas session control can best be handled using RTSP proxies as already used for caching of media data. Eventually, the evaluation of this architecture shows that such a service offers both low latency for clients and an acceptable synchronization of media streams to different client devices. Moreover, the communication overhead compared to usual conferencing or streaming systems is very low.Mit Hilfe der IP-Technologie erbrachte Multimedia-Dienste wie IPTV oder Video-on-Demand sind zur Zeit ein gefragtes Thema. Technisch werden solche Dienste unter dem Begriff "Streaming" eingeordnet. Ein Server sendet Mediendaten kontinuierlich an Empfänger, welche die Daten sofort weiterverarbeiten und anzeigen. Über einen Rückkanal hat der Kunde die Möglichkeit der Einflussnahme auf die Wiedergabe. Eine Weiterentwicklung dieser Streaming-Dienste ist die Möglichkeit, gemeinsam mit anderen denselben Film auf mehreren Geräten anzusehen. Ähnliche Ansätze gibt es im Internet bereits durch IP-Multicast. Allerdings können Benutzer hierbei keinen Einfluss auf die Übertragung nehmen - das Überspringen von Teilen ist zum Beispiel nicht möglich. Andere Benutzer können nicht direkt zur Streaming-Sitzung eingeladen werden. Collaborative Streaming ohne solche Einschränkungen bietet die in dieser Arbeit entwickelte costream-Architektur: Sie erlaubt es, andere zum gemeinsamen Betrachten eines Filmes einzuladen oder sich selbst in eine Benutzergruppe einzuklinken. Abhängig vom gewünschten Ablauf der Sitzung wird die Steuerung für alle Teilnehmer durchgeführt oder die Gruppe aufgeteilt. Eine Gruppenverwaltung regelt dies mit Hilfe von Rollenzuweisungen. Davon getrennt sorgt eine weitere Komponente für die Steuerung der Streaming-Sitzungen und die Synchronisation zwischen Teilnehmern. Diese Aufteilung hat den Vorteil, dass von der IETF entwickelte Standardprotokolle eingesetzt werden können. Für die Gruppenverwaltung sind SIP-Konferenzsysteme geeignet, während für die Sitzungssteuerung ein RTSP-Zwischensystem benutzt wurde. Die Evaluierung dieser Architektur zeigt schließlich, dass ein solcher Dienst nicht nur geringe Wartezeiten aufweist, sondern eine akzeptable Synchronisation der Datenströme auf die verschiedenen Ausgabegeräte der Benutzer erreicht wird. Zudem ist der Zusatzaufwand verglichen mit üblichen Konferenz- oder Streaming-Systemen sehr gering
- …