3,023 research outputs found
Recommended from our members
Multimedia delivery in the future internet
The term “Networked Media” implies that all kinds of media including text, image, 3D graphics, audio
and video are produced, distributed, shared, managed and consumed on-line through various networks,
like the Internet, Fiber, WiFi, WiMAX, GPRS, 3G and so on, in a convergent manner [1]. This white
paper is the contribution of the Media Delivery Platform (MDP) cluster and aims to cover the Networked
challenges of the Networked Media in the transition to the Future of the Internet.
Internet has evolved and changed the way we work and live. End users of the Internet have been confronted
with a bewildering range of media, services and applications and of technological innovations concerning
media formats, wireless networks, terminal types and capabilities. And there is little evidence that the pace
of this innovation is slowing. Today, over one billion of users access the Internet on regular basis, more
than 100 million users have downloaded at least one (multi)media file and over 47 millions of them do so
regularly, searching in more than 160 Exabytes1 of content. In the near future these numbers are expected
to exponentially rise. It is expected that the Internet content will be increased by at least a factor of 6, rising
to more than 990 Exabytes before 2012, fuelled mainly by the users themselves. Moreover, it is envisaged
that in a near- to mid-term future, the Internet will provide the means to share and distribute (new)
multimedia content and services with superior quality and striking flexibility, in a trusted and personalized
way, improving citizens’ quality of life, working conditions, edutainment and safety.
In this evolving environment, new transport protocols, new multimedia encoding schemes, cross-layer inthe
network adaptation, machine-to-machine communication (including RFIDs), rich 3D content as well as
community networks and the use of peer-to-peer (P2P) overlays are expected to generate new models of
interaction and cooperation, and be able to support enhanced perceived quality-of-experience (PQoE) and
innovative applications “on the move”, like virtual collaboration environments, personalised services/
media, virtual sport groups, on-line gaming, edutainment. In this context, the interaction with content
combined with interactive/multimedia search capabilities across distributed repositories, opportunistic P2P
networks and the dynamic adaptation to the characteristics of diverse mobile terminals are expected to
contribute towards such a vision.
Based on work that has taken place in a number of EC co-funded projects, in Framework Program 6 (FP6)
and Framework Program 7 (FP7), a group of experts and technology visionaries have voluntarily
contributed in this white paper aiming to describe the status, the state-of-the art, the challenges and the way
ahead in the area of Content Aware media delivery platforms
Recommended from our members
Error resilient video transcoding for robust inter-network communications using GPRS
A novel fully comprehensive mobile video communications
system is proposed in this paper. This system exploits
the useful rate management features of the video transcoders and
combines them with error resilience for transmissions of coded
video streams over general packet radio service (GPRS) mobileaccess
networks. The error-resilient video transcoding operation
takes place at a centralized point, referred to as a video proxy,
which provides the necessary output transmission rates with the
required amount of robustness. With the use of this proposed
algorithm, error resilience can be added to an already compressed
video stream at an intermediate stage at the edge of two or more
different networks through two resilience schemes, namely the
adaptive intra refresh (AIR) and feedback control signaling (FCS)
methods. Both resilience tools impose an output rate increase
which can also be prevented with the proposed novel technique in
this paper. Thus, an error-resilient video transcoding scheme is
presented to give robust video outputs at near target transmission
rates that only require the same number of GPRS timeslots as
the nonresilient schemes. Moreover, an ultimate robustness is
also accomplished with the combination of the two resilience
algorithms at the video proxy. Extensive computer simulations
demonstrate the effectiveness of the proposed system
The aceToolbox: low-level audiovisual feature extraction for retrieval and classification
In this paper we present an overview of a software platform
that has been developed within the aceMedia project,
termed the aceToolbox, that provides global and local lowlevel feature extraction from audio-visual content. The toolbox is based on the MPEG-7 eXperimental Model (XM),
with extensions to provide descriptor extraction from arbitrarily shaped image segments, thereby supporting local descriptors reflecting real image content. We describe the architecture of the toolbox as well as providing an overview of the descriptors supported to date. We also briefly describe the segmentation algorithm provided. We then demonstrate the usefulness of the toolbox in the context of two different content processing scenarios: similarity-based retrieval in large collections and scene-level classification of still images
Optimal packetisation of MPEG-4 using RTP over mobile networks
The introduction of third-generation wireless networks should result in real-time mobile
video communications becoming a reality. Delivery of such video is likely to be facilitated by the realtime
transport protocol (RTP). Careful packetisation of the video data is necessary to ensure the
optimal trade-off between channel utilisation and error robustness. Theoretical analyses for two basic
schemes of MPEG-4 data encapsulation within RTP packets are presented. Simulations over a GPRS
(general packet radio service) network are used to validate the analysis of the most efficient scheme.
Finally, a motion adaptive system for deriving MPEG-4 video packet sizes is presented. Further
simulations demonstrate the benefits of the adaptive system
Semantic multimedia remote display for mobile thin clients
Current remote display technologies for mobile thin clients convert practically all types of graphical content into sequences of images rendered by the client. Consequently, important information concerning the content semantics is lost. The present paper goes beyond this bottleneck by developing a semantic multimedia remote display. The principle consists of representing the graphical content as a real-time interactive multimedia scene graph. The underlying architecture features novel components for scene-graph creation and management, as well as for user interactivity handling. The experimental setup considers the Linux X windows system and BiFS/LASeR multimedia scene technologies on the server and client sides, respectively. The implemented solution was benchmarked against currently deployed solutions (VNC and Microsoft-RDP), by considering text editing and WWW browsing applications. The quantitative assessments demonstrate: (1) visual quality expressed by seven objective metrics, e.g., PSNR values between 30 and 42 dB or SSIM values larger than 0.9999; (2) downlink bandwidth gain factors ranging from 2 to 60; (3) real-time user event management expressed by network round-trip time reduction by factors of 4-6 and by uplink bandwidth gain factors from 3 to 10; (4) feasible CPU activity, larger than in the RDP case but reduced by a factor of 1.5 with respect to the VNC-HEXTILE
A video object generation tool allowing friendly user interaction
In this paper we describe an interactive video object segmentation tool developed in the framework of the ACTS-AC098 MOMUSYS project. The Video Object Generator with User Environment (VOGUE) combines three different sets of automatic and semi-automatic-tool (spatial segmentation, object tracking and temporal segmentation) with general purpose tools for user interaction. The result is an integrated environment allowing the user-assisted segmentation of any sort of video sequences in a friendly and efficient manner.Peer ReviewedPostprint (published version
A Turbo-Detection Aided Serially Concatenated MPEG-4/TCM Videophone Transceiver
A Turbo-detection aided serially concatenated inner Trellis Coded Modulation (TCM) scheme is combined with four different outer codes, namely with a Reversible Variable Length Code (RVLC), a Non-Systematic Convolutional (NSC) code a Recursive Systematic Convolutional (RSC) code or a Low Density Parity Check (LDPC) code. These four outer constituent codes are comparatively studied in the context of an MPEG4 videophone transceiver. These serially concatenated schemes are also compared to a stand-alone LDPC coded MPEG4 videophone system at the same effective overall coding rate. The performance of the proposed schemes is evaluated when communicating over uncorrelated Rayleigh fading channels. It was found that the serially concatenated TCM-NSC scheme was the most attractive one in terms of coding gain and decoding complexity among all the schemes considered in the context of the MPEG4 videophone transceiver. By contrast, the serially concatenated TCM-RSC scheme was found to attain the highest iteration gain among the schemes considered
- …