7 research outputs found

    Video alignment to a common reference

    Get PDF
    2015 Spring.Includes bibliographical references.Handheld videos often include unintentional motion (jitter) and intentional motion (pan and/or zoom). Human viewers prefer to see jitter removed, creating a smoothly moving camera. For video analysis, in contrast, aligning to a fixed stable background is sometimes preferable. This paper presents an algorithm that removes both forms of motion using a novel and efficient way of tracking background points while ignoring moving foreground points. The approach is related to image mosaicing, but the result is a video rather than an enlarged still image. It is also related to multiple object tracking approaches, but simpler since moving objects need not be explicitly tracked. The algorithm presented takes as input a video and returns one or several stabilized videos. Videos are broken into parts when the algorithm detects background change and it becomes necessary to fix upon a new background. We present two techniques in this thesis. One technique stabilizes the video with respect to the first available frame. Another technique stabilizes the videos with respect to a best frame. Our approach assumes the person holding the camera is standing in one place and that objects in motion do not dominate the image. Our algorithm performs better than previously published approaches when compared on 1,401 handheld videos from the recently released Point-and-Shoot Face Recognition Challenge (PASC)

    동적 ν™˜κ²½μ— κ°•μΈν•œ λͺ¨μ…˜ λΆ„λ₯˜ 기반의 μ˜μƒ 항법 μ•Œκ³ λ¦¬μ¦˜ 개발

    Get PDF
    ν•™μœ„λ…Όλ¬Έ (석사)-- μ„œμšΈλŒ€ν•™κ΅ λŒ€ν•™μ› κ³΅κ³ΌλŒ€ν•™ 기계항곡곡학뢀, 2017. 8. κΉ€ν˜„μ§„.In the paper, we propose a robust visual odometry algorithm for dynamic environments via rigid motion segmentation using a grid-based optical flow. The algorithm first divides image frame by a fixed-size grid, then calculates the three-dimensional motion of grids for light computational load and uniformly distributed optical flow vectors. Next, it selects several adjacent points among grid-based optical flow vectors based on a so-called entropy and generates motion hypotheses formed by three-dimensional rigid transformation. These processes for a spatial motion segmentation utilizes the principle of randomized hypothesis generation and the existing clustering algorithm, thus separating objects that move independently of each other. Moreover, we use a dual-mode simple Gaussian model in order to differentiate static and dynamic parts persistently. The model measures the output of the spatial motion segmentation algorithm and updates a probability vector consisting of the likelihood of representing specific label. For the evaluation of the proposed algorithm, we use a self-made dataset captured by ASUS Xtion Pro live RGB-D camera and Vicon motion capture system. We compare our algorithm with the existing motion segmentation algorithm and the current state-of-the-art visual odometry algorithm respectively, and the proposed algorithm estimates the ego-motion robustly and accurately in dynamic environments while showing the competitive performance of the motion segmentation.κΈ°μ‘΄ λŒ€λ‹€μˆ˜μ˜ μ˜μƒ 항법 μ•Œκ³ λ¦¬μ¦˜μ€ 정적인 ν™˜κ²½μ„ κ°€μ •ν•˜μ—¬ κ°œλ°œλ˜μ–΄ μ™”μœΌλ©°, 잘 μ •μ˜λœ λ°μ΄ν„°μ…‹μ—μ„œ μ„±λŠ₯이 κ²€μ¦λ˜μ–΄ μ™”λ‹€. ν•˜μ§€λ§Œ 무인 λ‘œλ΄‡μ΄ μ˜μƒ 항법을 ν™œμš©ν•˜μ—¬ μž„λ¬΄λ₯Ό μˆ˜ν–‰ν•˜μ—¬μ•Ό ν•˜λŠ” μž₯μ†ŒλŠ”, μ‹€μ œ μ‚¬λžŒμ΄λ‚˜ μ°¨λŸ‰μ΄ μ™•λž˜ν•˜λŠ” λ“± 동적인 ν™˜κ²½μΌ κ°€λŠ₯성이 크닀. 비둝 RANSAC을 ν™œμš©ν•˜μ—¬ μ˜μƒ 항법을 μˆ˜ν–‰ν•˜λŠ” 일뢀 μ•Œκ³ λ¦¬μ¦˜λ“€μ€ ν”„λ ˆμž„ λ‚΄μ˜ 비정상적인 μ›€μ§μž„μ„ μœ„μΉ˜ μΆ”μ • κ³Όμ •μ—μ„œ λ°°μ œν•  수 μžˆμ§€λ§Œ, μ΄λŠ” 동적 물체가 μ˜μƒ ν”„λ ˆμž„μ˜ μž‘μ€ 뢀뢄을 μ°¨μ§€ν•˜λŠ” κ²½μš°μ—λ§Œ 적용이 κ°€λŠ₯ν•˜λ‹€. λ”°λΌμ„œ λΆˆν™•μ‹€μ„±μ΄ μ‘΄μž¬ν•˜λŠ” 동적 ν™˜κ²½μ—μ„œ 자기 μœ„μΉ˜λ₯Ό κ°•μΈν•˜κ²Œ μΆ”μ •ν•˜κΈ° μœ„ν•΄, λ³Έ λ…Όλ¬Έμ—μ„œλŠ” 동적 ν™˜κ²½μ— κ°•μΈν•œ μ˜μƒ 기반 μ£Όν–‰ 기둝계 μ•Œκ³ λ¦¬μ¦˜μ„ μ œμ•ˆν•œλ‹€. μ œμ•ˆν•œ μ•Œκ³ λ¦¬μ¦˜μ€ μ›ν™œν•œ μˆ˜ν–‰ 속도와 이미지 내에 κ· μΌν•˜κ²Œ λΆ„ν¬λœ λͺ¨μ…˜μ„ κ³„μ‚°ν•˜κΈ° μœ„ν•΄, 격자 기반 μ˜΅ν‹°μ»¬ ν”Œλ‘œμš°λ₯Ό μ΄μš©ν•œλ‹€. 그리고 격자 λ‹¨μœ„ κ·Έλ¦¬λ“œμ˜ λͺ¨μ…˜μ„ 톡해 단일 ν”„λ ˆμž„ λ‚΄μ—μ„œ 3차원 곡간 λͺ¨μ…˜ 뢄할을 μˆ˜ν–‰ν•˜κ³ , λ‹€μˆ˜μ˜ 동적 물체 및 정적 μš”μ†Œλ₯Ό μ§€μ†μ μœΌλ‘œ ꡬ뢄 및 κ΅¬λ³„ν•˜κΈ° μœ„ν•΄ μ‹œκ°„μ  λͺ¨μ…˜ 뢄할을 μˆ˜ν–‰ν•œλ‹€. 특히 μ§€μ†μ μœΌλ‘œ 동적 및 정적 μš”μ†Œλ₯Ό κ΅¬λ³„ν•˜κΈ° μœ„ν•΄, μš°λ¦¬λŠ” 이미지 λ‚΄μ˜ 각 κ·Έλ¦¬λ“œμ— 이쀑 λͺ¨λ“œ κ°€μš°μ‹œμ•ˆ λͺ¨λΈμ„ μ μš©ν•˜μ—¬ μ•Œκ³ λ¦¬μ¦˜μ΄ 곡간적 λͺ¨μ…˜ λΆ„ν• μ˜ μΌμ‹œμ  λ…Έμ΄μ¦ˆμ— κ°•μΈν•˜κ²Œ ν•˜κ³ , ν™•λ₯  벑터λ₯Ό κ΅¬μ„±ν•˜μ—¬ κ·Έλ¦¬λ“œκ°€ μ„œλ‘œ κ΅¬λ³„λ˜λŠ” 각각의 μš”μ†Œλ‘œ λ°œν˜„ν•  ν™•λ₯ μ„ κ³„μ‚°ν•˜κ²Œ ν•œλ‹€. κ°œλ°œν•œ μ•Œκ³ λ¦¬μ¦˜μ˜ μ„±λŠ₯ 검증을 μœ„ν•΄ ASUS Xtion RGB-D 카메라와 Vicon λͺ¨μ…˜ 캑쳐 μ‹œμŠ€ν…œμ„ 톡해 κ΅¬μ„±ν•œ 데이터셋을 μ΄μš©ν•˜μ˜€μœΌλ©°, κΈ°μ‘΄ λͺ¨μ…˜ λΆ„ν•  μ•Œκ³ λ¦¬μ¦˜κ³Όμ˜ μž¬ν˜„μœ¨ (recall), 정밀도 (precision) 비ꡐ 및 κΈ°μ‘΄ μ˜μƒ 기반 μ£Όν–‰ 기둝계 μ•Œκ³ λ¦¬μ¦˜κ³Όμ˜ μΆ”μ • 였차 비ꡐλ₯Ό 톡해 타 μ•Œκ³ λ¦¬μ¦˜ λŒ€λΉ„ μš°μˆ˜ν•œ λͺ¨μ…˜ κ²€μΆœ 및 μœ„μΉ˜ μΆ”μ • μ„±λŠ₯을 ν™•μΈν•˜μ˜€λ‹€.Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iii Table of Contents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv List of Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v List of Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii Chapter 1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.1 Literature review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 1.2 Thesis contribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 1.3 Thesis outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 2 Background Knowledge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 2.1 Rigid transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 2.2 Grid-based optical flow . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 3 Motion Spatial Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 3.1 Motion hypothesis search . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 3.2 Motion hypothesis refinement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 3.3 Motion hypothesis clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 4 Motion Temporal Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 4.1 Label matching . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 4.2 Dual-mode simple Gaussian model . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 4.2.1 Update model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 4.2.2 Compensate model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 5 Evaluation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 5.1 Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 5.2 Motion segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 5.3 Visual odometry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34Maste

    Robust motion segmentation with subspace constraints

    No full text
    Motion segmentation is an important task in computer vision with many applications such as dynamic scene understanding and multi-body structure from motion. When the point correspondences across frames are given, motion segmentation can be addressed as a subspace clustering problem under an affine camera model. In the first two parts of this thesis, we target the general subspace clustering problem and propose two novel methods, namely Efficient Dense Subspace Clustering (EDSC) and the Robust Shape Interaction Matrix (RSIM) method. Instead of following the standard compressive sensing approach, in EDSC we formulate subspace clustering as a Frobenius norm minimization problem, which inherently yields denser connections between data points. While in the noise-free case we rely on the self-expressiveness of the observations, in the presence of noise we recover a clean dictionary to represent the data. Our formulation lets us solve the subspace clustering problem efficiently. More specifically, for outlier-free observations, the solution can be obtained in closed-form, and in the presence of outliers, we solve the problem by performing a series of linear operations. Furthermore, we show that our Frobenius norm formulation shares the same solution as the popular nuclear norm minimization approach when the data is free of any noise. In RSIM, we revisit the Shape Interaction Matrix (SIM) method, one of the earliest approaches for motion segmentation (or subspace clustering), and reveal its connections to several recent subspace clustering methods. We derive a simple, yet effective algorithm to robustify the SIM method and make it applicable to real-world scenarios where the data is corrupted by noise. We validate the proposed method by intuitive examples and justify it with the matrix perturbation theory. Moreover, we show that RSIM can be extended to handle missing data with a Grassmannian gradient descent method. The above subspace clustering methods work well for motion segmentation, yet they require that point trajectories across frames are known {\it a priori}. However, finding point correspondences is in itself a challenging task. Existing approaches tackle the correspondence estimation and motion segmentation problems separately. In the third part of this thesis, given a set of feature points detected in each frame of the sequence, we develop an approach which simultaneously performs motion segmentation and finds point correspondences across the frames. We formulate this problem in terms of Partial Permutation Matrices (PPMs) and aim to match feature descriptors while simultaneously encouraging point trajectories to satisfy subspace constraints. This lets us handle outliers in both point locations and feature appearance. The resulting optimization problem is solved via the Alternating Direction Method of Multipliers (ADMM), where each subproblem has an efficient solution. In particular, we show that most of the subproblems can be solved in closed-form, and one binary assignment subproblem can be solved by the Hungarian algorithm. Obtaining reliable feature tracks in a frame-by-frame manner is desirable in applications such as online motion segmentation. In the final part of the thesis, we introduce a novel multi-body feature tracker that exploits a multi-body rigidity assumption to improve tracking robustness under a general perspective camera model. A conventional approach to addressing this problem would consist of alternating between solving two subtasks: motion segmentation and feature tracking under rigidity constraints for each segment. This approach, however, requires knowing the number of motions, as well as assigning points to motion groups, which is typically sensitive to motion estimates. By contrast, we introduce a segmentation-free solution to multi-body feature tracking that bypasses the motion assignment step and reduces to solving a series of subproblems with closed-form solutions. In summary, in this thesis, we exploit the powerful subspace constraints and develop robust motion segmentation methods in different challenging scenarios where the trajectories are either given as input, or unknown beforehand. We also present a general robust multi-body feature tracker which can be used as the first step of motion segmentation to get reliable trajectories
    corecore