6,985 research outputs found

    Video Upright Adjustment and Stabilization

    Get PDF
    Upright adjustment, Video stabilization, Camera pathWe propose a novel video upright adjustment method that can reliably correct slanted video contents that are often found in casual videos. Our approach combines deep learning and Bayesian inference to estimate accurate rotation angles from video frames. We train a convolutional neural network to obtain initial estimates of the rotation angles of input video frames. The initial estimates from the network are temporally inconsistent and inaccurate. To resolve this, we use Bayesian inference. We analyze estimation errors of the network, and derive an error model. We then use the error model to formulate video upright adjustment as a maximum a posteriori problem where we estimate consistent rotation angles from the initial estimates, while respecting relative rotations between consecutive frames. Finally, we propose a joint approach to video stabilization and upright adjustment, which minimizes information loss caused by separately handling stabilization and upright adjustment. Experimental results show that our video upright adjustment method can effectively correct slanted video contents, and its combination with video stabilization can achieve visually pleasing results from shaky and slanted videos.openI. INTRODUCTION 1.1. Related work II. ROTATION ESTIMATION NETWORK III. ERROR ANALYSIS IV. VIDEO UPRIGHT ADJUSTMENT 4.1. Initial angle estimation 4.2. Robust angle estimation 4.3. Optimization 4.4. Warping V. JOINT UPRIGHT ADJUSTMENT AND STABILIZATION 5.1. Bundled camera paths for video stabilization 5.2. Joint approach VI. EXPERIMENTS VII. CONCLUSION ReferencesCNN)을 ν›ˆλ ¨μ‹œν‚¨λ‹€. μ‹ κ²½λ§μ˜ 초기 μΆ”μ •μΉ˜λŠ” μ™„μ „νžˆ μ •ν™•ν•˜μ§€ μ•ŠμœΌλ©° μ‹œκ°„μ μœΌλ‘œλ„ μΌκ΄€λ˜μ§€ μ•ŠλŠ”λ‹€. 이λ₯Ό ν•΄κ²°ν•˜κΈ° μœ„ν•΄ λ² μ΄μ§€μ•ˆ 인퍼런슀λ₯Ό μ‚¬μš©ν•œλ‹€. λ³Έ 논문은 μ‹ κ²½λ§μ˜ μΆ”μ • 였λ₯˜λ₯Ό λΆ„μ„ν•˜κ³  였λ₯˜ λͺ¨λΈμ„ λ„μΆœν•œλ‹€. 그런 λ‹€μŒ 였λ₯˜ λͺ¨λΈμ„ μ‚¬μš©ν•˜μ—¬ 연속 ν”„λ ˆμž„ κ°„μ˜ μƒλŒ€ νšŒμ „ 각도(Relative rotation angle)λ₯Ό λ°˜μ˜ν•˜λ©΄μ„œ 초기 μΆ”μ •μΉ˜λ‘œλΆ€ν„° μ‹œκ°„μ μœΌλ‘œ μΌκ΄€λœ νšŒμ „ 각도λ₯Ό μΆ”μ •ν•˜λŠ” μ΅œλŒ€ 사후 문제(Maximum a posteriori problem)둜 λ™μ˜μƒ μˆ˜ν‰ 보정을 κ³΅μ‹ν™”ν•œλ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ, λ™μ˜μƒ μˆ˜ν‰ 보정 및 λ™μ˜μƒ μ•ˆμ •ν™”(Video stabilization)에 λŒ€ν•œ λ™μ‹œ μ ‘κ·Ό 방법을 μ œμ•ˆν•˜μ—¬ μˆ˜ν‰ 보정과 μ•ˆμ •ν™”λ₯Ό λ³„λ„λ‘œ μˆ˜ν–‰ν•  λ•Œ λ°œμƒν•˜λŠ” 곡간 정보 손싀과 μ—°μ‚°λŸ‰μ„ μ΅œμ†Œν™”ν•˜λ©° μ•ˆμ •ν™”μ˜ μ„±λŠ₯을 μ΅œλŒ€ν™”ν•œλ‹€. μ‹€ν—˜ 결과에 λ”°λ₯΄λ©΄ λ™μ˜μƒ μˆ˜ν‰ λ³΄μ •μœΌλ‘œ κΈ°μšΈμ–΄μ§„ λ™μ˜μƒμ„ 효과적으둜 보정할 수 있으며 λ™μ˜μƒ μ•ˆμ •ν™” 방법과 κ²°ν•©ν•˜μ—¬ 흔듀리고 κΈ°μšΈμ–΄μ§„ λ™μ˜μƒμœΌλ‘œλΆ€ν„° μ‹œκ°μ μœΌλ‘œ 만쑱슀러운 μƒˆλ‘œμš΄ λ™μ˜μƒμ„ νšλ“ν•  수 μžˆλ‹€.λ³Έ 논문은 μΌλ°˜μΈλ“€μ΄ μ΄¬μ˜ν•œ λ™μ˜μƒμ—μ„œ ν”νžˆ λ°œμƒν•˜λŠ” 문제인 κΈ°μšΈμ–΄μ§μ„ μ œκ±°ν•˜μ—¬ μˆ˜ν‰μ΄ μ˜¬λ°”λ₯Έ λ™μ˜μƒμ„ νšλ“ν•  수 있게 ν•˜λŠ” λ™μ˜μƒ μˆ˜ν‰ 보정(Video upright adjustment) 방법을 μ œμ•ˆν•œλ‹€. λ³Έ λ…Όλ¬Έμ˜ μ ‘κ·Ό 방식은 λ”₯ λŸ¬λ‹(Deep learning)κ³Ό λ² μ΄μ§€μ•ˆ 인퍼런슀(Bayesian inference)λ₯Ό κ²°ν•©ν•˜μ—¬ λ™μ˜μƒ ν”„λ ˆμž„(Frame)μ—μ„œ μ •ν™•ν•œ 각도λ₯Ό μΆ”μ •ν•œλ‹€. λ¨Όμ € μž…λ ₯ λ™μ˜μƒ ν”„λ ˆμž„μ˜ νšŒμ „ κ°λ„μ˜ 초기 μΆ”μ •μΉ˜λ₯Ό μ–»κΈ° μœ„ν•΄ νšŒμ„  신경망(Convolutional neural networkMasterdCollectio

    Transitioning360: Content-aware NFoV Virtual Camera Paths for 360Β° Video Playback

    Get PDF

    The highD Dataset: A Drone Dataset of Naturalistic Vehicle Trajectories on German Highways for Validation of Highly Automated Driving Systems

    Full text link
    Scenario-based testing for the safety validation of highly automated vehicles is a promising approach that is being examined in research and industry. This approach heavily relies on data from real-world scenarios to derive the necessary scenario information for testing. Measurement data should be collected at a reasonable effort, contain naturalistic behavior of road users and include all data relevant for a description of the identified scenarios in sufficient quality. However, the current measurement methods fail to meet at least one of the requirements. Thus, we propose a novel method to measure data from an aerial perspective for scenario-based validation fulfilling the mentioned requirements. Furthermore, we provide a large-scale naturalistic vehicle trajectory dataset from German highways called highD. We evaluate the data in terms of quantity, variety and contained scenarios. Our dataset consists of 16.5 hours of measurements from six locations with 110 000 vehicles, a total driven distance of 45 000 km and 5600 recorded complete lane changes. The highD dataset is available online at: http://www.highD-dataset.comComment: IEEE International Conference on Intelligent Transportation Systems (ITSC) 201

    Long-Term Visual Object Tracking Benchmark

    Full text link
    We propose a new long video dataset (called Track Long and Prosper - TLP) and benchmark for single object tracking. The dataset consists of 50 HD videos from real world scenarios, encompassing a duration of over 400 minutes (676K frames), making it more than 20 folds larger in average duration per sequence and more than 8 folds larger in terms of total covered duration, as compared to existing generic datasets for visual tracking. The proposed dataset paves a way to suitably assess long term tracking performance and train better deep learning architectures (avoiding/reducing augmentation, which may not reflect real world behaviour). We benchmark the dataset on 17 state of the art trackers and rank them according to tracking accuracy and run time speeds. We further present thorough qualitative and quantitative evaluation highlighting the importance of long term aspect of tracking. Our most interesting observations are (a) existing short sequence benchmarks fail to bring out the inherent differences in tracking algorithms which widen up while tracking on long sequences and (b) the accuracy of trackers abruptly drops on challenging long sequences, suggesting the potential need of research efforts in the direction of long-term tracking.Comment: ACCV 2018 (Oral

    VNect: Real-time 3D Human Pose Estimation with a Single RGB Camera

    Full text link
    We present the first real-time method to capture the full global 3D skeletal pose of a human in a stable, temporally consistent manner using a single RGB camera. Our method combines a new convolutional neural network (CNN) based pose regressor with kinematic skeleton fitting. Our novel fully-convolutional pose formulation regresses 2D and 3D joint positions jointly in real time and does not require tightly cropped input frames. A real-time kinematic skeleton fitting method uses the CNN output to yield temporally stable 3D global pose reconstructions on the basis of a coherent kinematic skeleton. This makes our approach the first monocular RGB method usable in real-time applications such as 3D character control---thus far, the only monocular methods for such applications employed specialized RGB-D cameras. Our method's accuracy is quantitatively on par with the best offline 3D monocular RGB pose estimation methods. Our results are qualitatively comparable to, and sometimes better than, results from monocular RGB-D approaches, such as the Kinect. However, we show that our approach is more broadly applicable than RGB-D solutions, i.e. it works for outdoor scenes, community videos, and low quality commodity RGB cameras.Comment: Accepted to SIGGRAPH 201
    • …
    corecore