233 research outputs found

    High-speed Video from Asynchronous Camera Array

    Get PDF
    This paper presents a method for capturing high-speed video using an asynchronous camera array. Our method sequentially fires each sensor in a camera array with a small time offset and assembles captured frames into a high-speed video according to the time stamps. The resulting video, however, suffers from parallax jittering caused by the viewpoint difference among sensors in the camera array. To address this problem, we develop a dedicated novel view synthesis algorithm that transforms the video frames as if they were captured by a single reference sensor. Specifically, for any frame from a non-reference sensor, we find the two temporally neighboring frames captured by the reference sensor. Using these three frames, we render a new frame with the same time stamp as the non-reference frame but from the viewpoint of the reference sensor. Specifically, we segment these frames into super-pixels and then apply local content-preserving warping to warp them to form the new frame. We employ a multi-label Markov Random Field method to blend these warped frames. Our experiments show that our method can produce high-quality and high-speed video of a wide variety of scenes with large parallax, scene dynamics, and camera motion and outperforms several baseline and state-of-the-art approaches.Comment: 10 pages, 82 figures, Published at IEEE WACV 201

    Interactive videos: Plausible video editing using sparse structure points

    Get PDF
    Video remains the method of choice for capturing temporal events. However, without access to the underlying 3D scene models, it remains difficult to make object level edits in a single video or across multiple videos. While it may be possible to explicitly reconstruct the 3D geometries to facilitate these edits, such a workflow is cumbersome, expensive, and tedious. In this work, we present a much simpler workflow to create plausible editing and mixing of raw video footage using only sparse structure points (SSP) directly recovered from the raw sequences. First, we utilize user-scribbles to structure the point representations obtained using structure-from-motion on the input videos. The resultant structure points, even when noisy and sparse, are then used to enable various video edits in 3D, including view perturbation, keyframe animation, object duplication and transfer across videos, etc. Specifically, we describe how to synthesize object images from new views adopting a novel image-based rendering technique using the SSPs as proxy for the missing 3D scene information. We propose a structure-preserving image warping on multiple input frames adaptively selected from object video, followed by a spatio-temporally coherent image stitching to compose the final object image. Simple planar shadows and depth maps are synthesized for objects to generate plausible video sequence mimicking real-world interactions. We demonstrate our system on a variety of input videos to produce complex edits, which are otherwise difficult to achieve

    Stitching for multi-view videos with large parallax based on adaptive pixel warping

    Get PDF
    Conventional stitching techniques for images and videos are based on smooth warping models, and therefore, they often fail to work on multi-view images and videos with large parallax captured by cameras with wide baselines. In this paper, we propose a novel video stitching algorithm for such challenging multi-view videos. We estimate the parameters of ground plane homography, fundamental matrix, and vertical vanishing points reliably, using both of the appearance and activity based feature matches validated by geometric constraints. We alleviate the parallax artifacts in stitching by adaptively warping the off-plane pixels into geometrically accurate matching positions through their ground plane pixels based on the epipolar geometry. We also exploit the inter-view and inter-frame correspondence matching information together to estimate the ground plane pixels reliably, which are then refined by energy minimization. Experimental results show that the proposed algorithm provides geometrically accurate stitching results of multi-view videos with large parallax and outperforms the state-of-the-art stitching methods qualitatively and quantitatively

    Software simulation of depth of field effects in video from small aperture cameras

    Get PDF
    Thesis (M. Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2010.Cataloged from PDF version of thesis.Includes bibliographical references (p. 71-73).This thesis proposes a technique for post processing digital video to introduce a simulated depth of field effect. Because the technique is implemented in software, it affords the user greater control over the parameters of the effect (such as the amount of defocus, aperture shape, and defocus plane) and allows the effect to be used even on hardware which would not typically allow for depth of field. In addition, because it is a completely post processing technique and requires no change in capture method or hardware, it can be used on any video and introduces no new costs. This thesis describes the technique, evaluates its performance on example videos, and proposes further work to improve the technique's reliability.by Jordan Sorensen.M.Eng

    Video Upright Adjustment and Stabilization

    Get PDF
    Upright adjustment, Video stabilization, Camera pathWe propose a novel video upright adjustment method that can reliably correct slanted video contents that are often found in casual videos. Our approach combines deep learning and Bayesian inference to estimate accurate rotation angles from video frames. We train a convolutional neural network to obtain initial estimates of the rotation angles of input video frames. The initial estimates from the network are temporally inconsistent and inaccurate. To resolve this, we use Bayesian inference. We analyze estimation errors of the network, and derive an error model. We then use the error model to formulate video upright adjustment as a maximum a posteriori problem where we estimate consistent rotation angles from the initial estimates, while respecting relative rotations between consecutive frames. Finally, we propose a joint approach to video stabilization and upright adjustment, which minimizes information loss caused by separately handling stabilization and upright adjustment. Experimental results show that our video upright adjustment method can effectively correct slanted video contents, and its combination with video stabilization can achieve visually pleasing results from shaky and slanted videos.openI. INTRODUCTION 1.1. Related work II. ROTATION ESTIMATION NETWORK III. ERROR ANALYSIS IV. VIDEO UPRIGHT ADJUSTMENT 4.1. Initial angle estimation 4.2. Robust angle estimation 4.3. Optimization 4.4. Warping V. JOINT UPRIGHT ADJUSTMENT AND STABILIZATION 5.1. Bundled camera paths for video stabilization 5.2. Joint approach VI. EXPERIMENTS VII. CONCLUSION ReferencesCNN)을 ν›ˆλ ¨μ‹œν‚¨λ‹€. μ‹ κ²½λ§μ˜ 초기 μΆ”μ •μΉ˜λŠ” μ™„μ „νžˆ μ •ν™•ν•˜μ§€ μ•ŠμœΌλ©° μ‹œκ°„μ μœΌλ‘œλ„ μΌκ΄€λ˜μ§€ μ•ŠλŠ”λ‹€. 이λ₯Ό ν•΄κ²°ν•˜κΈ° μœ„ν•΄ λ² μ΄μ§€μ•ˆ 인퍼런슀λ₯Ό μ‚¬μš©ν•œλ‹€. λ³Έ 논문은 μ‹ κ²½λ§μ˜ μΆ”μ • 였λ₯˜λ₯Ό λΆ„μ„ν•˜κ³  였λ₯˜ λͺ¨λΈμ„ λ„μΆœν•œλ‹€. 그런 λ‹€μŒ 였λ₯˜ λͺ¨λΈμ„ μ‚¬μš©ν•˜μ—¬ 연속 ν”„λ ˆμž„ κ°„μ˜ μƒλŒ€ νšŒμ „ 각도(Relative rotation angle)λ₯Ό λ°˜μ˜ν•˜λ©΄μ„œ 초기 μΆ”μ •μΉ˜λ‘œλΆ€ν„° μ‹œκ°„μ μœΌλ‘œ μΌκ΄€λœ νšŒμ „ 각도λ₯Ό μΆ”μ •ν•˜λŠ” μ΅œλŒ€ 사후 문제(Maximum a posteriori problem)둜 λ™μ˜μƒ μˆ˜ν‰ 보정을 κ³΅μ‹ν™”ν•œλ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ, λ™μ˜μƒ μˆ˜ν‰ 보정 및 λ™μ˜μƒ μ•ˆμ •ν™”(Video stabilization)에 λŒ€ν•œ λ™μ‹œ μ ‘κ·Ό 방법을 μ œμ•ˆν•˜μ—¬ μˆ˜ν‰ 보정과 μ•ˆμ •ν™”λ₯Ό λ³„λ„λ‘œ μˆ˜ν–‰ν•  λ•Œ λ°œμƒν•˜λŠ” 곡간 정보 손싀과 μ—°μ‚°λŸ‰μ„ μ΅œμ†Œν™”ν•˜λ©° μ•ˆμ •ν™”μ˜ μ„±λŠ₯을 μ΅œλŒ€ν™”ν•œλ‹€. μ‹€ν—˜ 결과에 λ”°λ₯΄λ©΄ λ™μ˜μƒ μˆ˜ν‰ λ³΄μ •μœΌλ‘œ κΈ°μšΈμ–΄μ§„ λ™μ˜μƒμ„ 효과적으둜 보정할 수 있으며 λ™μ˜μƒ μ•ˆμ •ν™” 방법과 κ²°ν•©ν•˜μ—¬ 흔듀리고 κΈ°μšΈμ–΄μ§„ λ™μ˜μƒμœΌλ‘œλΆ€ν„° μ‹œκ°μ μœΌλ‘œ 만쑱슀러운 μƒˆλ‘œμš΄ λ™μ˜μƒμ„ νšλ“ν•  수 μžˆλ‹€.λ³Έ 논문은 μΌλ°˜μΈλ“€μ΄ μ΄¬μ˜ν•œ λ™μ˜μƒμ—μ„œ ν”νžˆ λ°œμƒν•˜λŠ” 문제인 κΈ°μšΈμ–΄μ§μ„ μ œκ±°ν•˜μ—¬ μˆ˜ν‰μ΄ μ˜¬λ°”λ₯Έ λ™μ˜μƒμ„ νšλ“ν•  수 있게 ν•˜λŠ” λ™μ˜μƒ μˆ˜ν‰ 보정(Video upright adjustment) 방법을 μ œμ•ˆν•œλ‹€. λ³Έ λ…Όλ¬Έμ˜ μ ‘κ·Ό 방식은 λ”₯ λŸ¬λ‹(Deep learning)κ³Ό λ² μ΄μ§€μ•ˆ 인퍼런슀(Bayesian inference)λ₯Ό κ²°ν•©ν•˜μ—¬ λ™μ˜μƒ ν”„λ ˆμž„(Frame)μ—μ„œ μ •ν™•ν•œ 각도λ₯Ό μΆ”μ •ν•œλ‹€. λ¨Όμ € μž…λ ₯ λ™μ˜μƒ ν”„λ ˆμž„μ˜ νšŒμ „ κ°λ„μ˜ 초기 μΆ”μ •μΉ˜λ₯Ό μ–»κΈ° μœ„ν•΄ νšŒμ„  신경망(Convolutional neural networkMasterdCollectio
    • …
    corecore