215 research outputs found

    Survey of Error Concealment techniques: Research directions and open issues

    Full text link
    © 2015 IEEE. Error Concealment (EC) techniques use either spatial, temporal or a combination of both types of information to recover the data lost in transmitted video. In this paper, existing EC techniques are reviewed, which are divided into three categories, namely Intra-frame EC, Inter-frame EC, and Hybrid EC techniques. We first focus on the EC techniques developed for the H.264/AVC standard. The advantages and disadvantages of these EC techniques are summarized with respect to the features in H.264. Then, the EC algorithms are also analyzed. These EC algorithms have been recently adopted in the newly introduced H.265/HEVC standard. A performance comparison between the classic EC techniques developed for H.264 and H.265 is performed in terms of the average PSNR. Lastly, open issues in the EC domain are addressed for future research consideration

    Video modeling via implicit motion representations

    Get PDF
    Video modeling refers to the development of analytical representations for explaining the intensity distribution in video signals. Based on the analytical representation, we can develop algorithms for accomplishing particular video-related tasks. Therefore video modeling provides us a foundation to bridge video data and related-tasks. Although there are many video models proposed in the past decades, the rise of new applications calls for more efficient and accurate video modeling approaches.;Most existing video modeling approaches are based on explicit motion representations, where motion information is explicitly expressed by correspondence-based representations (i.e., motion velocity or displacement). Although it is conceptually simple, the limitations of those representations and the suboptimum of motion estimation techniques can degrade such video modeling approaches, especially for handling complex motion or non-ideal observation video data. In this thesis, we propose to investigate video modeling without explicit motion representation. Motion information is implicitly embedded into the spatio-temporal dependency among pixels or patches instead of being explicitly described by motion vectors.;Firstly, we propose a parametric model based on a spatio-temporal adaptive localized learning (STALL). We formulate video modeling as a linear regression problem, in which motion information is embedded within the regression coefficients. The coefficients are adaptively learned within a local space-time window based on LMMSE criterion. Incorporating a spatio-temporal resampling and a Bayesian fusion scheme, we can enhance the modeling capability of STALL on more general videos. Under the framework of STALL, we can develop video processing algorithms for a variety of applications by adjusting model parameters (i.e., the size and topology of model support and training window). We apply STALL on three video processing problems. The simulation results show that motion information can be efficiently exploited by our implicit motion representation and the resampling and fusion do help to enhance the modeling capability of STALL.;Secondly, we propose a nonparametric video modeling approach, which is not dependent on explicit motion estimation. Assuming the video sequence is composed of many overlapping space-time patches, we propose to embed motion-related information into the relationships among video patches and develop a generic sparsity-based prior for typical video sequences. First, we extend block matching to more general kNN-based patch clustering, which provides an implicit and distributed representation for motion information. We propose to enforce the sparsity constraint on a higher-dimensional data array signal, which is generated by packing the patches in the similar patch set. Then we solve the inference problem by updating the kNN array and the wanted signal iteratively. Finally, we present a Bayesian fusion approach to fuse multiple-hypothesis inferences. Simulation results in video error concealment, denoising, and deartifacting are reported to demonstrate its modeling capability.;Finally, we summarize the proposed two video modeling approaches. We also point out the perspectives of implicit motion representations in applications ranging from low to high level problems

    Optimization of Coding of AR Sources for Transmission Across Channels with Loss

    Get PDF

    Transforming the Chinese Pole Circus Apparatus into an Interactive Musical Instrument

    Get PDF
    The objective of this project is to create a modified version of the Chinese pole circus apparatus in order to artistically musicalize and visualize a circus performer’s movement in real time. Wireless, wearable inertial measurement units (IMUs) allow for tracking the position of the performer’s hands and feet. The vertical height of the performer is then used to play a corresponding pitch on a musical scale, while the position with respect to the other two dimensions is used to produce a bird’s-eye-view visualization. Radio-frequency identification (RFID) tags added to the pole improve the accuracy of the IMU position tracking by providing anchor points with which to recalibrate an IMU\u27s position in space. This works to reduce the effect of drift, the result of small inaccuracies in the acceleration data collected by the IMUs which are compounded when integrating over time to determine velocity and position. This project allows an audience to experience the movement of a circus performer from a new perspective. In addition to the choreography, the audience is left with a unique musical composition and visual art pieces they can remember. More importantly, the audience can experience the movement from a perspective that cannot be experienced without the use of this technology

    Transforming the Chinese Pole Circus Apparatus into an Interactive Musical Instrument

    Get PDF
    The objective of this project is to create a modified version of the Chinese pole circus apparatus in order to artistically musicalize and visualize a circus performer’s movement in real time. Wireless, wearable inertial measurement units (IMUs) allow for tracking the position of the performer’s hands and feet. The vertical height of the performer is then used to play a corresponding pitch on a musical scale, while the position with respect to the other two dimensions is used to produce a bird’s-eye-view visualization. Radio-frequency identification (RFID) tags added to the pole improve the accuracy of the IMU position tracking by providing anchor points with which to recalibrate an IMU\u27s position in space. This works to reduce the effect of drift, the result of small inaccuracies in the acceleration data collected by the IMUs which are compounded when integrating over time to determine velocity and position. This project allows an audience to experience the movement of a circus performer from a new perspective. In addition to the choreography, the audience is left with a unique musical composition and visual art pieces they can remember. More importantly, the audience can experience the movement from a perspective that cannot be experienced without the use of this technology

    Vision Science and Technology at NASA: Results of a Workshop

    Get PDF
    A broad review is given of vision science and technology within NASA. The subject is defined and its applications in both NASA and the nation at large are noted. A survey of current NASA efforts is given, noting strengths and weaknesses of the NASA program

    Context Aided Tracking with Adaptive Hyperspectral Imagery

    Get PDF
    A methodology for the context-aided tracking of ground vehicles in remote airborne imagery is developed in which a background model is inferred from hyperspectral imagery. The materials comprising the background of a scene are remotely identified and lead to this model. Two model formation processes are developed: a manual method, and method that exploits an emerging adaptive, multiple-object-spectrometer instrument. A semi-automated background modeling approach is shown to arrive at a reasonable background model with minimal operator intervention. A novel, adaptive, and autonomous approach uses a new type of adaptive hyperspectral sensor, and converges to a 66% correct background model in 5% the time of the baseline {a 95% reduction in sensor acquisition time. A multiple-hypothesis-tracker is incorporated, which utilizes background statistics to form track costs and associated track maintenance thresholds. The context-aided system is demonstrated in a high- fidelity tracking testbed, and reduces track identity error by 30%

    Error Concealment for Frame Losses in MDC

    Full text link
    corecore