868 research outputs found

    In-Network View Synthesis for Interactive Multiview Video Systems

    Get PDF
    To enable Interactive multiview video systems with a minimum view-switching delay, multiple camera views are sent to the users, which are used as reference images to synthesize additional virtual views via depth-image-based rendering. In practice, bandwidth constraints may however restrict the number of reference views sent to clients per time unit, which may in turn limit the quality of the synthesized viewpoints. We argue that the reference view selection should ideally be performed close to the users, and we study the problem of in-network reference view synthesis such that the navigation quality is maximized at the clients. We consider a distributed cloud network architecture where data stored in a main cloud is delivered to end users with the help of cloudlets, i.e., resource-rich proxies close to the users. In order to satisfy last-hop bandwidth constraints from the cloudlet to the users, a cloudlet re-samples viewpoints of the 3D scene into a discrete set of views (combination of received camera views and virtual views synthesized) to be used as reference for the synthesis of additional virtual views at the client. This in-network synthesis leads to better viewpoint sampling given a bandwidth constraint compared to simple selection of camera views, but it may however carry a distortion penalty in the cloudlet-synthesized reference views. We therefore cast a new reference view selection problem where the best subset of views is defined as the one minimizing the distortion over a view navigation window defined by the user under some transmission bandwidth constraints. We show that the view selection problem is NP-hard, and propose an effective polynomial time algorithm using dynamic programming to solve the optimization problem. Simulation results finally confirm the performance gain offered by virtual view synthesis in the network

    PlaceRaider: Virtual Theft in Physical Spaces with Smartphones

    Full text link
    As smartphones become more pervasive, they are increasingly targeted by malware. At the same time, each new generation of smartphone features increasingly powerful onboard sensor suites. A new strain of sensor malware has been developing that leverages these sensors to steal information from the physical environment (e.g., researchers have recently demonstrated how malware can listen for spoken credit card numbers through the microphone, or feel keystroke vibrations using the accelerometer). Yet the possibilities of what malware can see through a camera have been understudied. This paper introduces a novel visual malware called PlaceRaider, which allows remote attackers to engage in remote reconnaissance and what we call virtual theft. Through completely opportunistic use of the camera on the phone and other sensors, PlaceRaider constructs rich, three dimensional models of indoor environments. Remote burglars can thus download the physical space, study the environment carefully, and steal virtual objects from the environment (such as financial documents, information on computer monitors, and personally identifiable information). Through two human subject studies we demonstrate the effectiveness of using mobile devices as powerful surveillance and virtual theft platforms, and we suggest several possible defenses against visual malware

    Three-dimensional media for mobile devices

    Get PDF
    Cataloged from PDF version of article.This paper aims at providing an overview of the core technologies enabling the delivery of 3-D Media to next-generation mobile devices. To succeed in the design of the corresponding system, a profound knowledge about the human visual system and the visual cues that form the perception of depth, combined with understanding of the user requirements for designing user experience for mobile 3-D media, are required. These aspects are addressed first and related with the critical parts of the generic system within a novel user-centered research framework. Next-generation mobile devices are characterized through their portable 3-D displays, as those are considered critical for enabling a genuine 3-D experience on mobiles. Quality of 3-D content is emphasized as the most important factor for the adoption of the new technology. Quality is characterized through the most typical, 3-D-specific visual artifacts on portable 3-D displays and through subjective tests addressing the acceptance and satisfaction of different 3-D video representation, coding, and transmission methods. An emphasis is put on 3-D video broadcast over digital video broadcasting-handheld (DVB-H) in order to illustrate the importance of the joint source-channel optimization of 3-D video for its efficient compression and robust transmission over error-prone channels. The comparative results obtained identify the best coding and transmission approaches and enlighten the interaction between video quality and depth perception along with the influence of the context of media use. Finally, the paper speculates on the role and place of 3-D multimedia mobile devices in the future internet continuum involving the users in cocreation and refining of rich 3-D media content

    Advanced Free Viewpoint Video Streaming Techniques

    Get PDF
    Free-viewpoint video is a new type of interactive multimedia service allowing users to control their viewpoint and generate new views of a dynamic scene from any perspective. The uniquely generated and displayed views are composed from two or more high bitrate camera streams that must be delivered to the users depending on their continuously changing perspective. Due to significant network and computational resource requirements, we proposed scalable viewpoint generation and delivery schemes based on multicast forwarding and distributed approach. Our aim was to find the optimal deployment locations of the distributed viewpoint synthesis processes in the network topology by allowing network nodes to act as proxy servers with caching and viewpoint synthesis functionalities. Moreover, a predictive multicast group management scheme was introduced in order to provide all camera views that may be requested in the near future and prevent the viewpoint synthesizer algorithm from remaining without camera streams. The obtained results showed that even 42% traffic decrease can be realized using distributed viewpoint synthesis and the probability of viewpoint synthesis starvation can be also significantly reduced in future free viewpoint video services

    Appearance-Based Gaze Estimation in the Wild

    Full text link
    Appearance-based gaze estimation is believed to work well in real-world settings, but existing datasets have been collected under controlled laboratory conditions and methods have been not evaluated across multiple datasets. In this work we study appearance-based gaze estimation in the wild. We present the MPIIGaze dataset that contains 213,659 images we collected from 15 participants during natural everyday laptop use over more than three months. Our dataset is significantly more variable than existing ones with respect to appearance and illumination. We also present a method for in-the-wild appearance-based gaze estimation using multimodal convolutional neural networks that significantly outperforms state-of-the art methods in the most challenging cross-dataset evaluation. We present an extensive evaluation of several state-of-the-art image-based gaze estimation algorithms on three current datasets, including our own. This evaluation provides clear insights and allows us to identify key research challenges of gaze estimation in the wild

    A comparison of the performance of prediction techniques in curtailing uplink transmission and energy requirements in mobile free-viewpoint video applications

    Get PDF
    The rapid deployment of multimedia services on mobile networks together with the increase in consumer demand for immersive entertainment have paved the way for innovative video representations. Amongst these new applications is free-viewpoint video (FVV), whereby a scene is captured by an array of cameras distributed around a site to allow the user to alter the viewing perspective on demand, creating a three-dimensional (3D) effect. The implementation on mobile infrastructures is however still hindered by intrinsic wireless limitations, such as bandwidth constraints and limited battery power. To this effect, this paper presents a solution that reduces the number of uplink requests performed by the mobile terminal through view prediction techniques. The implementation and performance of four distinct prediction algorithms in anticipating the next viewpoint request by a mobile user in a typical FVV system are compared and contrasted. Additionally, each solution removes the jitter experienced by the user whilst moving from a view pattern to another by allowing some hysterisis in the convergence signal. Thus, this technique enhances the performance of all the algorithms by taking into consideration the fact that the user adapts to the presented views and will react accordingly. Simulation results illustrate that an uplink transmission reduction of up to 96.7% can be achieved in a conventional FVV simulation scenario. Therefore, the application of prediction schemes can drastically reduce the mobile terminal’s power consumption and bandwidth resource requirements on the uplink channel.peer-reviewe

    Mixed-Resolution HEVC based multiview video codec for low bitrate transmission

    Get PDF

    Applying prediction techniques to reduce uplink transmission and energy requirements in mobile free-viewpoint video applications

    Get PDF
    The increased efficiency of video compression algorithms, the improvements registered in reliability, data rates and quality of service of wireless transmission technologies, and the development of mobile multimedia terminals have made possible the implementation of Free-Viewpoint Video (FVV) technology on mobile platforms. The mobile environment however presents several restrictions. Two of these limiting factors being bandwidth constraints and energy availability in battery-operated mobile terminals. This paper looks at the possibility of employing prediction algorithms at the FVV server to anticipate the next viewpoint expected by the user. In doing so, the number of uplink requests is reduced to situations where the estimated view defers from the requested one and the associated transmissions required in retraining the algorithm once this occurs. Simulation results on two different prediction techniques demonstrate that the uplink transmission rate is reduced by up to 96.7% when emulating a conventional FVV usage scenario. Both prediction algorithms infer a substantial decrease in the mobile terminal’s power consumption and reduce the network’s uplink bandwidth utilization.peer-reviewe
    • 

    corecore