5 research outputs found

    Assembling convolution neural networks for automatic viewing transformation

    Get PDF
    Images taken under different camera poses are rotated or distorted, which leads to poor perception experiences. This paper proposes a new framework to automatically transform the images to the conformable view setting by assembling different convolution neural networks. Specifically, a referential 3D ground plane is firstly derived from the RGB image and a novel projection mapping algorithm is developed to achieve automatic viewing transformation. Extensive experimental results demonstrate that the proposed method outperforms the state-ofthe-art vanishing points based methods by a large margin in terms of accuracy and robustness

    Video Upright Adjustment and Stabilization

    Get PDF
    Upright adjustment, Video stabilization, Camera pathWe propose a novel video upright adjustment method that can reliably correct slanted video contents that are often found in casual videos. Our approach combines deep learning and Bayesian inference to estimate accurate rotation angles from video frames. We train a convolutional neural network to obtain initial estimates of the rotation angles of input video frames. The initial estimates from the network are temporally inconsistent and inaccurate. To resolve this, we use Bayesian inference. We analyze estimation errors of the network, and derive an error model. We then use the error model to formulate video upright adjustment as a maximum a posteriori problem where we estimate consistent rotation angles from the initial estimates, while respecting relative rotations between consecutive frames. Finally, we propose a joint approach to video stabilization and upright adjustment, which minimizes information loss caused by separately handling stabilization and upright adjustment. Experimental results show that our video upright adjustment method can effectively correct slanted video contents, and its combination with video stabilization can achieve visually pleasing results from shaky and slanted videos.openI. INTRODUCTION 1.1. Related work II. ROTATION ESTIMATION NETWORK III. ERROR ANALYSIS IV. VIDEO UPRIGHT ADJUSTMENT 4.1. Initial angle estimation 4.2. Robust angle estimation 4.3. Optimization 4.4. Warping V. JOINT UPRIGHT ADJUSTMENT AND STABILIZATION 5.1. Bundled camera paths for video stabilization 5.2. Joint approach VI. EXPERIMENTS VII. CONCLUSION ReferencesCNN)을 ν›ˆλ ¨μ‹œν‚¨λ‹€. μ‹ κ²½λ§μ˜ 초기 μΆ”μ •μΉ˜λŠ” μ™„μ „νžˆ μ •ν™•ν•˜μ§€ μ•ŠμœΌλ©° μ‹œκ°„μ μœΌλ‘œλ„ μΌκ΄€λ˜μ§€ μ•ŠλŠ”λ‹€. 이λ₯Ό ν•΄κ²°ν•˜κΈ° μœ„ν•΄ λ² μ΄μ§€μ•ˆ 인퍼런슀λ₯Ό μ‚¬μš©ν•œλ‹€. λ³Έ 논문은 μ‹ κ²½λ§μ˜ μΆ”μ • 였λ₯˜λ₯Ό λΆ„μ„ν•˜κ³  였λ₯˜ λͺ¨λΈμ„ λ„μΆœν•œλ‹€. 그런 λ‹€μŒ 였λ₯˜ λͺ¨λΈμ„ μ‚¬μš©ν•˜μ—¬ 연속 ν”„λ ˆμž„ κ°„μ˜ μƒλŒ€ νšŒμ „ 각도(Relative rotation angle)λ₯Ό λ°˜μ˜ν•˜λ©΄μ„œ 초기 μΆ”μ •μΉ˜λ‘œλΆ€ν„° μ‹œκ°„μ μœΌλ‘œ μΌκ΄€λœ νšŒμ „ 각도λ₯Ό μΆ”μ •ν•˜λŠ” μ΅œλŒ€ 사후 문제(Maximum a posteriori problem)둜 λ™μ˜μƒ μˆ˜ν‰ 보정을 κ³΅μ‹ν™”ν•œλ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ, λ™μ˜μƒ μˆ˜ν‰ 보정 및 λ™μ˜μƒ μ•ˆμ •ν™”(Video stabilization)에 λŒ€ν•œ λ™μ‹œ μ ‘κ·Ό 방법을 μ œμ•ˆν•˜μ—¬ μˆ˜ν‰ 보정과 μ•ˆμ •ν™”λ₯Ό λ³„λ„λ‘œ μˆ˜ν–‰ν•  λ•Œ λ°œμƒν•˜λŠ” 곡간 정보 손싀과 μ—°μ‚°λŸ‰μ„ μ΅œμ†Œν™”ν•˜λ©° μ•ˆμ •ν™”μ˜ μ„±λŠ₯을 μ΅œλŒ€ν™”ν•œλ‹€. μ‹€ν—˜ 결과에 λ”°λ₯΄λ©΄ λ™μ˜μƒ μˆ˜ν‰ λ³΄μ •μœΌλ‘œ κΈ°μšΈμ–΄μ§„ λ™μ˜μƒμ„ 효과적으둜 보정할 수 있으며 λ™μ˜μƒ μ•ˆμ •ν™” 방법과 κ²°ν•©ν•˜μ—¬ 흔듀리고 κΈ°μšΈμ–΄μ§„ λ™μ˜μƒμœΌλ‘œλΆ€ν„° μ‹œκ°μ μœΌλ‘œ 만쑱슀러운 μƒˆλ‘œμš΄ λ™μ˜μƒμ„ νšλ“ν•  수 μžˆλ‹€.λ³Έ 논문은 μΌλ°˜μΈλ“€μ΄ μ΄¬μ˜ν•œ λ™μ˜μƒμ—μ„œ ν”νžˆ λ°œμƒν•˜λŠ” 문제인 κΈ°μšΈμ–΄μ§μ„ μ œκ±°ν•˜μ—¬ μˆ˜ν‰μ΄ μ˜¬λ°”λ₯Έ λ™μ˜μƒμ„ νšλ“ν•  수 있게 ν•˜λŠ” λ™μ˜μƒ μˆ˜ν‰ 보정(Video upright adjustment) 방법을 μ œμ•ˆν•œλ‹€. λ³Έ λ…Όλ¬Έμ˜ μ ‘κ·Ό 방식은 λ”₯ λŸ¬λ‹(Deep learning)κ³Ό λ² μ΄μ§€μ•ˆ 인퍼런슀(Bayesian inference)λ₯Ό κ²°ν•©ν•˜μ—¬ λ™μ˜μƒ ν”„λ ˆμž„(Frame)μ—μ„œ μ •ν™•ν•œ 각도λ₯Ό μΆ”μ •ν•œλ‹€. λ¨Όμ € μž…λ ₯ λ™μ˜μƒ ν”„λ ˆμž„μ˜ νšŒμ „ κ°λ„μ˜ 초기 μΆ”μ •μΉ˜λ₯Ό μ–»κΈ° μœ„ν•΄ νšŒμ„  신경망(Convolutional neural networkMasterdCollectio

    Beyond Geo-localization: Fine-grained Orientation of Street-view Images by Cross-view Matching with Satellite Imagery

    Full text link
    Street-view imagery provides us with novel experiences to explore different places remotely. Carefully calibrated street-view images (e.g. Google Street View) can be used for different downstream tasks, e.g. navigation, map features extraction. As personal high-quality cameras have become much more affordable and portable, an enormous amount of crowdsourced street-view images are uploaded to the internet, but commonly with missing or noisy sensor information. To prepare this hidden treasure for "ready-to-use" status, determining missing location information and camera orientation angles are two equally important tasks. Recent methods have achieved high performance on geo-localization of street-view images by cross-view matching with a pool of geo-referenced satellite imagery. However, most of the existing works focus more on geo-localization than estimating the image orientation. In this work, we re-state the importance of finding fine-grained orientation for street-view images, formally define the problem and provide a set of evaluation metrics to assess the quality of the orientation estimation. We propose two methods to improve the granularity of the orientation estimation, achieving 82.4% and 72.3% accuracy for images with estimated angle errors below 2 degrees for CVUSA and CVACT datasets, corresponding to 34.9% and 28.2% absolute improvement compared to previous works. Integrating fine-grained orientation estimation in training also improves the performance on geo-localization, giving top 1 recall 95.5%/85.5% and 86.8%/80.4% for orientation known/unknown tests on the two datasets.Comment: This paper has been accepted by ACM Multimedia 2022. The version contains additional supplementary material

    Robust upright adjustment of 360 spherical panoramas

    No full text
    With the recent advent of 360 cameras, spherical panorama images are becoming more popular and widely available. In a spherical panorama, alignment of the scene orientation to the image axes is important for providing comfortable and pleasant viewing experiences using VR headsets and traditional displays. This paper presents an automatic method for upright adjustment of 360 spherical panorama images without any prior information, such as depths and gyro sensor data. We take the Atlanta world assumption and use the horizontal and vertical lines in the scene to formulate a cost function for upright adjustment. In addition to fast optimization of the cost function, our method includes outlier handling to improve the robustness and accuracy of upright adjustment. Our method produces visually pleasing results for a variety of real-world spherical panoramas in less than a second, and the accuracy is verified using ground-truth data. ? 2017, Springer-Verlag Berlin Heidelberg.112sciescopu
    corecore