233 research outputs found
High-speed Video from Asynchronous Camera Array
This paper presents a method for capturing high-speed video using an
asynchronous camera array. Our method sequentially fires each sensor in a
camera array with a small time offset and assembles captured frames into a
high-speed video according to the time stamps. The resulting video, however,
suffers from parallax jittering caused by the viewpoint difference among
sensors in the camera array. To address this problem, we develop a dedicated
novel view synthesis algorithm that transforms the video frames as if they were
captured by a single reference sensor. Specifically, for any frame from a
non-reference sensor, we find the two temporally neighboring frames captured by
the reference sensor. Using these three frames, we render a new frame with the
same time stamp as the non-reference frame but from the viewpoint of the
reference sensor. Specifically, we segment these frames into super-pixels and
then apply local content-preserving warping to warp them to form the new frame.
We employ a multi-label Markov Random Field method to blend these warped
frames. Our experiments show that our method can produce high-quality and
high-speed video of a wide variety of scenes with large parallax, scene
dynamics, and camera motion and outperforms several baseline and
state-of-the-art approaches.Comment: 10 pages, 82 figures, Published at IEEE WACV 201
Interactive videos: Plausible video editing using sparse structure points
Video remains the method of choice for capturing temporal events. However, without access to the underlying 3D scene models, it remains difficult to make object level edits in a single video or across multiple videos. While it may be possible to explicitly reconstruct the 3D geometries to facilitate these edits, such a workflow is cumbersome, expensive, and tedious. In this work, we present a much simpler workflow to create plausible editing and mixing of raw video footage using only sparse structure points (SSP) directly recovered from the raw sequences. First, we utilize user-scribbles to structure the point representations obtained using structure-from-motion on the input videos. The resultant structure points, even when noisy and sparse, are then used to enable various video edits in 3D, including view perturbation, keyframe animation, object duplication and transfer across videos, etc. Specifically, we describe how to synthesize object images from new views adopting a novel image-based rendering technique using the SSPs as proxy for the missing 3D scene information. We propose a structure-preserving image warping on multiple input frames adaptively selected from object video, followed by a spatio-temporally coherent image stitching to compose the final object image. Simple planar shadows and depth maps are synthesized for objects to generate plausible video sequence mimicking real-world interactions. We demonstrate our system on a variety of input videos to produce complex edits, which are otherwise difficult to achieve
Stitching for multi-view videos with large parallax based on adaptive pixel warping
Conventional stitching techniques for images and videos are based on smooth warping models, and therefore, they often fail to work on multi-view images and videos with large parallax captured by cameras with wide baselines. In this paper, we propose a novel video stitching algorithm for such challenging multi-view videos. We estimate the parameters of ground plane homography, fundamental matrix, and vertical vanishing points reliably, using both of the appearance and activity based feature matches validated by geometric constraints. We alleviate the parallax artifacts in stitching by adaptively warping the off-plane pixels into geometrically accurate matching positions through their ground plane pixels based on the epipolar geometry. We also exploit the inter-view and inter-frame correspondence matching information together to estimate the ground plane pixels reliably, which are then refined by energy minimization. Experimental results show that the proposed algorithm provides geometrically accurate stitching results of multi-view videos with large parallax and outperforms the state-of-the-art stitching methods qualitatively and quantitatively
Software simulation of depth of field effects in video from small aperture cameras
Thesis (M. Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2010.Cataloged from PDF version of thesis.Includes bibliographical references (p. 71-73).This thesis proposes a technique for post processing digital video to introduce a simulated depth of field effect. Because the technique is implemented in software, it affords the user greater control over the parameters of the effect (such as the amount of defocus, aperture shape, and defocus plane) and allows the effect to be used even on hardware which would not typically allow for depth of field. In addition, because it is a completely post processing technique and requires no change in capture method or hardware, it can be used on any video and introduces no new costs. This thesis describes the technique, evaluates its performance on example videos, and proposes further work to improve the technique's reliability.by Jordan Sorensen.M.Eng
Video Upright Adjustment and Stabilization
Upright adjustment, Video stabilization, Camera pathWe propose a novel video upright adjustment method that can reliably correct slanted video contents that are often found in casual videos. Our approach combines deep learning and Bayesian inference to estimate accurate rotation angles from video frames. We train a convolutional neural network to obtain initial estimates of the rotation angles of input video frames. The initial estimates from the network are temporally inconsistent and inaccurate. To resolve this, we use Bayesian inference. We analyze estimation errors of the network, and derive an error model. We then use the error model to formulate video upright adjustment as a maximum a posteriori problem where we estimate consistent rotation angles from the initial estimates, while respecting relative rotations between consecutive frames. Finally, we propose a joint approach to video stabilization and upright adjustment, which minimizes information loss caused by separately handling stabilization and upright adjustment. Experimental results show that our video upright adjustment method can effectively correct slanted video contents, and its combination with video stabilization can achieve visually pleasing results from shaky and slanted videos.openI. INTRODUCTION
1.1. Related work
II. ROTATION ESTIMATION NETWORK
III. ERROR ANALYSIS
IV. VIDEO UPRIGHT ADJUSTMENT
4.1. Initial angle estimation
4.2. Robust angle estimation
4.3. Optimization
4.4. Warping
V. JOINT UPRIGHT ADJUSTMENT AND STABILIZATION
5.1. Bundled camera paths for video stabilization
5.2. Joint approach
VI. EXPERIMENTS
VII. CONCLUSION
ReferencesCNN)μ νλ ¨μν¨λ€. μ κ²½λ§μ μ΄κΈ° μΆμ μΉλ μμ ν μ ννμ§ μμΌλ©° μκ°μ μΌλ‘λ μΌκ΄λμ§ μλλ€. μ΄λ₯Ό ν΄κ²°νκΈ° μν΄ λ² μ΄μ§μ μΈνΌλ°μ€λ₯Ό μ¬μ©νλ€. λ³Έ λ
Όλ¬Έμ μ κ²½λ§μ μΆμ μ€λ₯λ₯Ό λΆμνκ³ μ€λ₯ λͺ¨λΈμ λμΆνλ€. κ·Έλ° λ€μ μ€λ₯ λͺ¨λΈμ μ¬μ©νμ¬ μ°μ νλ μ κ°μ μλ νμ κ°λ(Relative rotation angle)λ₯Ό λ°μνλ©΄μ μ΄κΈ° μΆμ μΉλ‘λΆν° μκ°μ μΌλ‘ μΌκ΄λ νμ κ°λλ₯Ό μΆμ νλ μ΅λ μ¬ν λ¬Έμ (Maximum a posteriori problem)λ‘ λμμ μν 보μ μ 곡μννλ€. λ§μ§λ§μΌλ‘, λμμ μν 보μ λ° λμμ μμ ν(Video stabilization)μ λν λμ μ κ·Ό λ°©λ²μ μ μνμ¬ μν 보μ κ³Ό μμ νλ₯Ό λ³λλ‘ μνν λ λ°μνλ κ³΅κ° μ 보 μμ€κ³Ό μ°μ°λμ μ΅μννλ©° μμ νμ μ±λ₯μ μ΅λννλ€. μ€ν κ²°κ³Όμ λ°λ₯΄λ©΄ λμμ μν 보μ μΌλ‘ κΈ°μΈμ΄μ§ λμμμ ν¨κ³Όμ μΌλ‘ 보μ ν μ μμΌλ©° λμμ μμ ν λ°©λ²κ³Ό κ²°ν©νμ¬ νλ€λ¦¬κ³ κΈ°μΈμ΄μ§ λμμμΌλ‘λΆν° μκ°μ μΌλ‘ λ§μ‘±μ€λ¬μ΄ μλ‘μ΄ λμμμ νλν μ μλ€.λ³Έ λ
Όλ¬Έμ μΌλ°μΈλ€μ΄ 촬μν λμμμμ νν λ°μνλ λ¬Έμ μΈ κΈ°μΈμ΄μ§μ μ κ±°νμ¬ μνμ΄ μ¬λ°λ₯Έ λμμμ νλν μ μκ² νλ λμμ μν 보μ (Video upright adjustment) λ°©λ²μ μ μνλ€. λ³Έ λ
Όλ¬Έμ μ κ·Ό λ°©μμ λ₯ λ¬λ(Deep learning)κ³Ό λ² μ΄μ§μ μΈνΌλ°μ€(Bayesian inference)λ₯Ό κ²°ν©νμ¬ λμμ νλ μ(Frame)μμ μ νν κ°λλ₯Ό μΆμ νλ€. λ¨Όμ μ
λ ₯ λμμ νλ μμ νμ κ°λμ μ΄κΈ° μΆμ μΉλ₯Ό μ»κΈ° μν΄ νμ μ κ²½λ§(Convolutional neural networkMasterdCollectio
- β¦