1,245 research outputs found
3D video coding and transmission
The capture, transmission, and display of
3D content has gained a lot of attention in the last few
years. 3D multimedia content is no longer con fined to
cinema theatres but is being transmitted using stereoscopic
video over satellite, shared on Blu-RayTMdisks,
or sent over Internet technologies. Stereoscopic displays
are needed at the receiving end and the viewer needs to
wear special glasses to present the two versions of the
video to the human vision system that then generates
the 3D illusion. To be more e ffective and improve the
immersive experience, more views are acquired from a
larger number of cameras and presented on di fferent displays,
such as autostereoscopic and light field displays.
These multiple views, combined with depth data, also
allow enhanced user experiences and new forms of interaction
with the 3D content from virtual viewpoints.
This type of audiovisual information is represented by a
huge amount of data that needs to be compressed and
transmitted over bandwidth-limited channels. Part of
the COST Action IC1105 \3D Content Creation, Coding
and Transmission over Future Media Networks" (3DConTourNet)
focuses on this research challenge.peer-reviewe
Recommended from our members
Multimedia delivery in the future internet
The term âNetworked Mediaâ implies that all kinds of media including text, image, 3D graphics, audio
and video are produced, distributed, shared, managed and consumed on-line through various networks,
like the Internet, Fiber, WiFi, WiMAX, GPRS, 3G and so on, in a convergent manner [1]. This white
paper is the contribution of the Media Delivery Platform (MDP) cluster and aims to cover the Networked
challenges of the Networked Media in the transition to the Future of the Internet.
Internet has evolved and changed the way we work and live. End users of the Internet have been confronted
with a bewildering range of media, services and applications and of technological innovations concerning
media formats, wireless networks, terminal types and capabilities. And there is little evidence that the pace
of this innovation is slowing. Today, over one billion of users access the Internet on regular basis, more
than 100 million users have downloaded at least one (multi)media file and over 47 millions of them do so
regularly, searching in more than 160 Exabytes1 of content. In the near future these numbers are expected
to exponentially rise. It is expected that the Internet content will be increased by at least a factor of 6, rising
to more than 990 Exabytes before 2012, fuelled mainly by the users themselves. Moreover, it is envisaged
that in a near- to mid-term future, the Internet will provide the means to share and distribute (new)
multimedia content and services with superior quality and striking flexibility, in a trusted and personalized
way, improving citizensâ quality of life, working conditions, edutainment and safety.
In this evolving environment, new transport protocols, new multimedia encoding schemes, cross-layer inthe
network adaptation, machine-to-machine communication (including RFIDs), rich 3D content as well as
community networks and the use of peer-to-peer (P2P) overlays are expected to generate new models of
interaction and cooperation, and be able to support enhanced perceived quality-of-experience (PQoE) and
innovative applications âon the moveâ, like virtual collaboration environments, personalised services/
media, virtual sport groups, on-line gaming, edutainment. In this context, the interaction with content
combined with interactive/multimedia search capabilities across distributed repositories, opportunistic P2P
networks and the dynamic adaptation to the characteristics of diverse mobile terminals are expected to
contribute towards such a vision.
Based on work that has taken place in a number of EC co-funded projects, in Framework Program 6 (FP6)
and Framework Program 7 (FP7), a group of experts and technology visionaries have voluntarily
contributed in this white paper aiming to describe the status, the state-of-the art, the challenges and the way
ahead in the area of Content Aware media delivery platforms
Adaptive delivery of immersive 3D multi-view video over the Internet
The increase in Internet bandwidth and the developments in 3D video technology have paved the way for the delivery of 3D Multi-View Video (MVV) over the Internet. However, large amounts of data and dynamic network conditions result in frequent network congestion, which may prevent video packets from being delivered on time. As a consequence, the 3D video experience may well be degraded unless content-aware precautionary mechanisms and adaptation methods are deployed. In this work, a novel adaptive MVV streaming method is introduced which addresses the future generation 3D immersive MVV experiences with multi-view displays. When the user experiences network congestion, making it necessary to perform adaptation, the rate-distortion optimum set of views that are pre-determined by the server, are truncated from the delivered MVV streams. In order to maintain high Quality of Experience (QoE) service during the frequent network congestion, the proposed method involves the calculation of low-overhead additional metadata that is delivered to the client. The proposed adaptive 3D MVV streaming solution is tested using the MPEG Dynamic Adaptive Streaming over HTTP (MPEG-DASH) standard. Both extensive objective and subjective evaluations are presented, showing that the proposed method provides significant quality enhancement under the adverse network conditions
Quality-aware adaptive delivery of multi-view video
Advances in video coding and networking technologies have
paved the way for the Multi-View Video (MVV) streaming.
However, large amounts of data and dynamic network conditions
result in frequent network congestion, which may prevent
video packets from being delivered on time. As a consequence,
the 3D viewing experience may be degraded signifi-
cantly, unless quality-aware adaptation methods are deployed.
There is no research work to discuss the MVV adaptation of
decision strategy or provide a detailed analysis of a dynamic
network environment. This work addresses the mentioned issues
for MVV streaming over HTTP for emerging multi-view
displays. In this research work, the effect of various adaptations
of decision strategies are evaluated and, as a result, a
new quality-aware adaptation method is designed. The proposed
method is benefiting from layer based video coding in
such a way that high Quality of Experience (QoE) is maintained
in a cost-effective manner. The conducted experimental
results on MVV streaming using the proposed strategy are
showing that the perceptual 3D video quality, under adverse
network conditions, is enhanced significantly as a result of the
proposed quality-aware adaptation
Survey on QoE/QoS Correlation Models for Video Streaming over Vehicular Ad-hoc Networks
Vehicular Ad-hoc Networks (VANETs) are a new emerging technology which has attracted enormous interest over the last few years. It enables vehicles to communicate with each other and with roadside infrastructures for many applications. One of the promising applications is multimedia services for traffic safety or infotainment. The video service requires a good quality to satisfy the end-user known as the Quality of Experience (QoE). Several models have been suggested in the literature to measure or predict this metric. In this paper, we present an overview of interesting researches, which propose QoE models for video streaming over VANETs. The limits and deficiencies of these models are identified, which shed light on the challenges and real problems to overcome in the future
DĂŒnaamiline kiiruse jaotamine interaktiivses mitmevaatelises video vaatevahetuse ennustamineses
In Interactive Multi-View Video (IMVV), the video has been captured by numbers of
cameras positioned in array and transmitted those camera views to users. The user can
interact with the transmitted video content by choosing viewpoints (views from different
cameras in the array) with the expectation of minimum transmission delay while
changing among various views. View switching delay is one of the primary concern that
is dealt in this thesis work, where the contribution is to minimize the transmission delay
of new view switch frame through a novel process of selection of the predicted view
and compression considering the transmission efficiency. Mainly considered a realtime
IMVV streaming, and the view switch is mapped as discrete Markov chain, where
the transition probability is derived using Zipf distribution, which provides information
regarding view switch prediction. To eliminate Round-Trip Time (RTT) transmission
delay, Quantization Parameters (QP) are adaptively allocated to the remaining redundant
transmitted frames to maintain view switching time minimum, trading off with
the quality of the video till RTT time-span. The experimental results of the proposed
method show superior performance on PSNR and view switching delay for better viewing quality over the existing methods
Multiple description image and video coding for P2P transmissions
Peer-to-Peer (P2P) media streaming is, nowadays, a very attractive topic due to the bandwidth available to serve demanding content scales. A key challenge, however, is making content distribution robust to peer transience. Multiple description coding (MDC) has, indeed, proven to be very effective with problems concerning the packetsâ losses, since it generates several descriptions and may reconstruct the original information with any number of descriptions that may reach the decoder. Therefore multiple descriptions may be effective for robust peer-to-peer media streaming. In this dissertation, it will not only be showed that, but also that varying the redundancy level of description on the fly may lead to a better performance than the one obtained without varying this parameter. Besides that, it is shown, as well, that varying the Bitrate on the fly outperforms the redundancy on it. Furthermore, the redundancy and the Bitrate were varied simultaneously. Thus, it is shown that this variation is more efficient when the packet loss is high.
The experiments reported above were done using an experimental test bed developed for this purpose at the NMCG lab of the University of Beira Interior. It was also used the REGPROT, a video encoder developed by our research team, to splitted the video into multiple descriptions, which were, later, distributed among the peers in the test bed. After the request of the client, the referred encoder decoded the descriptions as they were being received.Fundação para a CiĂȘncia e a Tecnologia (FCT
- âŠ