567 research outputs found

    Slow and steady feature analysis: higher order temporal coherence in video

    Full text link
    How can unlabeled video augment visual learning? Existing methods perform "slow" feature analysis, encouraging the representations of temporally close frames to exhibit only small differences. While this standard approach captures the fact that high-level visual signals change slowly over time, it fails to capture *how* the visual content changes. We propose to generalize slow feature analysis to "steady" feature analysis. The key idea is to impose a prior that higher order derivatives in the learned feature space must be small. To this end, we train a convolutional neural network with a regularizer on tuples of sequential frames from unlabeled video. It encourages feature changes over time to be smooth, i.e., similar to the most recent changes. Using five diverse datasets, including unlabeled YouTube and KITTI videos, we demonstrate our method's impact on object, scene, and action recognition tasks. We further show that our features learned from unlabeled video can even surpass a standard heavily supervised pretraining approach.Comment: in Computer Vision and Pattern Recognition (CVPR) 2016, Las Vegas, NV, June 201

    Real-time content-aware video retargeting on the Android platform for tunnel vision assistance

    Get PDF
    As mobile devices continue to rise in popularity, advances in overall mobile device processing power lead to further expansion of their capabilities. This, coupled with the fact that many people suffer from low vision, leaves substantial room for advancing mobile development for low vision assistance. Computer vision is capable of assisting and accommodating individuals with blind spots or tunnel vision by extracting the necessary information and presenting it to the user in a manner they are able to visualize. Such a system would enable individuals with low vision to function with greater ease. Additionally, offering assistance on a mobile platform allows greater access. The objective of this thesis is to develop a computer vision application for low vision assistance on the Android mobile device platform. Specifically, the goal of the application is to reduce the effects tunnel vision inflicts on individuals. This is accomplished by providing an in-depth real-time video retargeting model that builds upon previous works and applications. Seam carving is a content-aware retargeting operator which defines 8-connected paths, or seams, of pixels. The optimality of these seams is based on a specific energy function. Discrete removal of these seams permits changes in the aspect ratio while simultaneously preserving important regions. The video retargeting model incorporates spatial and temporal considerations to provide effective image and video retargeting. Data reduction techniques are utilized in order to generate an efficient model. Additionally, a minimalistic multi-operator approach is constructed to diminish the disadvantages experienced by individual operators. In the event automated techniques fail, interactive options are provided that allow for user intervention. Evaluation of the application and its video retargeting model is based on its comparison to existing standard algorithms and its ability to extend itself to real-time. Performance metrics are obtained for both PC environments and mobile device platforms for comparison

    Content Aware Video Retargeting using Seam Carving

    Get PDF
    Video retargeting method achieves high - quality resizing to arbitrary aspect ratios for complex videos containing diverse camera and dynamic motions . Video retargeting from a full - resolution video to a lower resolution display will inevitably cause information loss. While retargeting the video the important contents must also be preserved. Seam carving works well for images without straight lines or regular patterns like landscape images but may cause distortions if used for images with straight lines. Our approach combines Seam Carving method along with Hough transform to preserve the origi nality of the video

    Stereoscopic Seam Carving With Temporal Consistency

    Full text link
    In this paper, we present a novel technique for seam carving of stereoscopic video. It removes seams of pixels in areas that are most likely not noticed by the viewer. When applying seam carving to stereoscopic video rather than monoscopic still images, new challenges arise. The detected seams must be consistent between the left and the right view, so that no depth information is destroyed. When removing seams in two consecutive frames, temporal consistency between the removed seams must be established to avoid flicker in the resulting video. By making certain assumptions, the available depth information can be harnessed to improve the quality achieved by seam carving. Assuming that closer pixels are more important, the algorithm can focus on removing distant pixels first. Furthermore, we assume that coherent pixels belonging to the same object have similar depth. By avoiding to cut through edges in the depth map, we can thus avoid cutting through object boundaries

    Image resizing with minimum distortion

    Get PDF
    Displays became cheap and were combined with many devices, like camera, mobile, and so on…, so there has been an increased interest on resizing methods to make the image suitable and fill any screen size. Common and known methods like cropping or resampling can cause undesirable effects such as: losses in information or distortion in perception. Recently, content-aware image resizing methods have been proposed to get rid of these problems and produce exceptional results. Seam-carving produced by Avidan and Shamir has gained attention as an effective solution. This paper discussed about this method and used it to resize (minimize and maximize) four colored images vertically and horizontally respectively, and maintained the main features of the images by deleting or repeating only the uninfluenced features. The energy map was calculated that described the basic and influential details of the image using energy function. But instead of gradient function (as in Avidan and Shamir) entropy function was used to compute the energy of the images. A vertical or a horizontal seam of pixels with minimum energy values was either deleted or inserted to resize the image. Good results were obtained especially when the image contains spaces within its details. The work was programmed using Matlab2018a

    Motion-based video retargeting with optimized crop-and-warp

    Full text link

    Content-Preserving Warps for 3D Video Stabilization

    Get PDF
    We describe a technique that transforms a video from a hand-held video camera so that it appears as if it were taken with a directed camera motion. Our method adjusts the video to appear as if it were taken from nearby viewpoints, allowing 3D camera movements to be simulated. By aiming only for perceptual plausibility, rather than accurate reconstruction, we are able to develop algorithms that can effectively recreate dynamic scenes from a single source video. Our technique first recovers the original 3D camera motion and a sparse set of 3D, static scene points using an off-the-shelf structure-frommotion system. Then, a desired camera path is computed either automatically (e.g., by fitting a linear or quadratic path) or interactively. Finally, our technique performs a least-squares optimization that computes a spatially-varying warp from each input video frame into an output frame. The warp is computed to both follow the sparse displacements suggested by the recovered 3D structure, and avoid deforming the content in the video frame. Our experiments on stabilizing challenging videos of dynamic scenes demonstrate the effectiveness of our technique
    • …
    corecore