3,034 research outputs found

    Guidance for benthic habitat mapping: an aerial photographic approach

    Get PDF
    This document, Guidance for Benthic Habitat Mapping: An Aerial Photographic Approach, describes proven technology that can be applied in an operational manner by state-level scientists and resource managers. This information is based on the experience gained by NOAA Coastal Services Center staff and state-level cooperators in the production of a series of benthic habitat data sets in Delaware, Florida, Maine, Massachusetts, New York, Rhode Island, the Virgin Islands, and Washington, as well as during Center-sponsored workshops on coral remote sensing and seagrass and aquatic habitat assessment. (PDF contains 39 pages) The original benthic habitat document, NOAA Coastal Change Analysis Program (C-CAP): Guidance for Regional Implementation (Dobson et al.), was published by the Department of Commerce in 1995. That document summarized procedures that were to be used by scientists throughout the United States to develop consistent and reliable coastal land cover and benthic habitat information. Advances in technology and new methodologies for generating these data created the need for this updated report, which builds upon the foundation of its predecessor

    ๋ฌธ์„œ ๊ฒฝ๊ณ„์™€ 3์ฐจ์› ์žฌ๊ตฌ์„ฑ์— ๊ธฐ๋ฐ˜ํ•œ ๋ฌธ์„œ ์ด๋ฏธ์ง€ ํ‰ํŒํ™”

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(์„์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ์ž์—ฐ๊ณผํ•™๋Œ€ํ•™ ์ˆ˜๋ฆฌ๊ณผํ•™๋ถ€, 2022. 8. ํ˜„๋™ํ›ˆ.In recent days, most of the scanned images are obtained from mobile devices such as cameras, smartphones, and tablets rather than traditional flatbed scanners. Contrary to the scanning process of the traditional scanners, capturing process of mobile devices might be accompanied by distortions in various forms such as perspective distortion, fold distortion, and page curls. In this thesis, we propose robust dewarping methods which correct such distortions based on the document boundary and 3D reconstruction. In the first method, we construct a curvilinear grid on the document image using the document boundary and reconstruct the document surface in the three dimensional space. Then we rectify the image using a family of local homographies computed from the reconstructed document surface. Although some of the steps of the proposed method have been proposed separately in other research, our approach exploited and combined their advantages to propose a robust dewarping process in addition to improving the stability in the overall process. Moreover, we refined the process by correcting the distorted text region boundary and developed this process into an independent dewarping method which is concise, straight-forward, and robust while still producing a well-rectified document image.์ตœ๊ทผ์—๋Š” ๋Œ€๋ถ€๋ถ„์˜ ์Šค์บ”๋œ ์ด๋ฏธ์ง€๋“ค์ด ์ „ํ†ต์ ์ธ ํ‰ํŒ์Šค์บ๋„ˆ๊ฐ€ ์•„๋‹Œ ์นด๋ฉ”๋ผ, ์Šค๋งˆํŠธํฐ, ํƒœ๋ธ”๋ฆฟ PC ๋“ฑ์˜ ํœด๋Œ€๊ธฐ๊ธฐ๋“ค๋กœ๋ถ€ํ„ฐ ์–ป์–ด์ง„๋‹ค. ์ด์ „ ์Šค์บ๋„ˆ๋“ค์˜ ์Šค์บ๋‹ ๊ณผ์ •๊ณผ๋Š” ๋‹ค๋ฅด๊ฒŒ ํœด๋Œ€๊ธฐ๊ธฐ๋“ค์„ ์ด์šฉํ•œ ์ด๋ฏธ์ง€ ์บก์ณ๋ง ๊ณผ์ •์€ ์›๊ทผ์™œ๊ณก, ์ข…์ด์˜ ์ ‘ํž˜์œผ๋กœ ์ธํ•œ ์™œ๊ณก, ๊ทธ๋ฆฌ๊ณ  ์ข…์ด์˜ ํœ˜์–ด์ง์œผ๋กœ ์ธํ•œ ์™œ๊ณก ๋“ฑ ๋‹ค์–‘ํ•œ ์™œ๊ณก๋“ค์„ ์ˆ˜๋ฐ˜ํ•  ์ˆ˜ ์žˆ๋‹ค. ์ด ๋…ผ๋ฌธ์—์„œ๋Š” ์ด๋Ÿฌํ•œ ์™œ๊ณก๋“ค์„ ์ œ๊ฑฐํ•  ์ˆ˜ ์žˆ๋Š” ๋ฌธ์„œ ๊ฒฝ๊ณ„์™€ 3์ฐจ์› ์žฌ๊ตฌ์„ฑ์— ๊ธฐ๋ฐ˜ํ•œ ๊ฐ•๋ ฅํ•œ ๋””์›Œํ•‘ ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•˜๊ณ ์ž ํ•œ๋‹ค. ์ฒซ๋ฒˆ์งธ ๋ฐฉ๋ฒ•์—์„œ๋Š”, ๋ฌธ์„œ ๊ฒฝ๊ณ„๋ฅผ ์ด์šฉํ•˜์—ฌ ๋ฌธ์„œ ์ด๋ฏธ์ง€ ์œ„์— ๊ณก์„ ์œผ๋กœ ์ด๋ฃจ์–ด์ง„ ๊ทธ๋ฆฌ๋“œ๋ฅผ ๋งŒ๋“ค๊ณ , 3์ฐจ์› ๊ณต๊ฐ„ ์ƒ์˜ ๋ฌธ์„œ ๊ณก๋ฉด์„ ์žฌ๊ตฌ์„ฑํ•œ๋‹ค. ๊ทธ๋ฆฌ๊ณ  ์žฌ๊ตฌ์„ฑ๋œ ๋ฌธ์„œ ๊ณก๋ฉด์œผ๋กœ๋ถ€ํ„ฐ ๊ณ„์‚ฐ๋œ ๊ตญ์†Œ์  ํ˜ธ๋ชจ๊ทธ๋ž˜ํ”ผ๋“ค์„ ์ด์šฉํ•˜์—ฌ ์ด๋ฏธ์ง€๋ฅผ ์ˆ˜์ •ํ•œ๋‹ค. ์šฐ๋ฆฌ๊ฐ€ ์ œ์•ˆํ•˜๋Š” ๋ฐฉ๋ฒ•์˜ ๋ช‡๋ช‡ ๋‹จ๊ณ„๋Š” ๋‹ค๋ฅธ ์—ฐ๊ตฌ์—์„œ ๊ฐœ๋ณ„์ ์œผ๋กœ ์‚ฌ์šฉ๋œ ๊ฒฝ์šฐ๋„ ์žˆ์ง€๋งŒ, ์šฐ๋ฆฌ๋Š” ์ „์ฒด์ ์ธ ๊ณผ์ •์—์„œ ์•ˆ์ •์„ฑ์„ ๋†’์ด๋Š” ๋™์‹œ์— ๊ฐ ๋ฐฉ๋ฒ•์˜ ์žฅ์ ๋“ค์„ ์ด์šฉํ•˜๊ณ  ์กฐํ•ฉํ•˜์—ฌ ๊ฐ•๋ ฅํ•œ ๋””์›Œํ•‘ ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•œ๋‹ค. ์ด์— ๋”ํ•˜์—ฌ, ์šฐ๋ฆฌ๋Š” ์™œ๊ณก๋œ ํ…์ŠคํŠธ ์˜์—ญ์˜ ๊ฒฝ๊ณ„๋ฅผ ์ˆ˜์ •ํ•˜์—ฌ ์ „์ฒด์ ์ธ ๊ณผ์ •์„ ๋ณด์™„ํ•˜์˜€๊ณ , ์ด ์ ˆ์ฐจ๋ฅผ ๊ฐ„๊ฒฐํ•˜๊ณ , ์ง๊ด€์ ์ด๋ฉฐ, ๊ฐ•๋ ฅํ•˜๋ฉด์„œ๋„ ์ข‹์€ ๊ฒฐ๊ณผ๋ฅผ ๋‚ด๋Š” ๋…๋ฆฝ์ ์ธ ๋””์›Œํ•‘ ๋ฐฉ๋ฒ•์œผ๋กœ ๊ฐœ๋ฐœํ•˜์˜€๋‹ค.1. Introduction 1 2. Review on Camera Geometry 6 2.1. Basic Camera Model 6 2.2. 3D Reconstruction Problem 8 3. Related Works 10 3.1. Dewarping Methods based on the Text-lines 10 3.2. Dewarping Methods based on the Document Boundary 11 3.3. Dewarping Methods based on the Grid Construction 12 3.4. Dewarping Methods based on the Document Surface Model in 3D Space 13 4. Document Image Dewarping based on the Document Boundary and 3D Reconstruction 15 4.1. Input Document Image Processing 17 4.1.1. Binarization of the Input Document Image 17 4.1.2. Perspective Distortion Removal using the Document Boundary 19 4.2. Grid Construction on the Document Image 21 4.3. 3D Reconstruction of the Document Surface 23 4.3.1. Geometric Model 23 4.3.2. Normalization of the Grid Corners 24 4.3.3. 3D Reconstruction of the Document Surface 26 4.4. Rectification of the Document Image under a Family of Local Homographies 27 4.5. Global Rectification of the Document Image 29 5. Document Image Dewarping by Straightening Document Boundary Curves 33 6. Conclusion 37 Appendix A. 38 A.1. 4-point Algorithm 38 A.2. Optimization of the Cost Function 40 Bibliography 42 Abstract (in Korean) 47 Acknowledgement (in Korean) 48์„

    HoughNet: neural network architecture for vanishing points detection

    Full text link
    In this paper we introduce a novel neural network architecture based on Fast Hough Transform layer. The layer of this type allows our neural network to accumulate features from linear areas across the entire image instead of local areas. We demonstrate its potential by solving the problem of vanishing points detection in the images of documents. Such problem occurs when dealing with camera shots of the documents in uncontrolled conditions. In this case, the document image can suffer several specific distortions including projective transform. To train our model, we use MIDV-500 dataset and provide testing results. The strong generalization ability of the suggested method is proven with its applying to a completely different ICDAR 2011 dewarping contest. In previously published papers considering these dataset authors measured the quality of vanishing point detection by counting correctly recognized words with open OCR engine Tesseract. To compare with them, we reproduce this experiment and show that our method outperforms the state-of-the-art result.Comment: 6 pages, 6 figures, 2 tables, 28 references, conferenc

    Procedures for condition mapping using 360ยฐ images

    Get PDF
    The identification of deterioration mechanisms and their monitoring over time is an essential phase for conservation. This work aimed at developing a novel approach for deterioration mapping and monitoring based on 360ยฐ images, which allows for simple and rapid data collection. The opportunity to capture the whole scene around a 360ยฐ camera reduces the number of images needed in a condition mapping project, resulting in a powerful solution to document small and narrow spaces. The paper will describe the implemented workflow for deterioration mapping based on 360ยฐ images, which highlights pathologies on surfaces and quantitatively measures their extension. Such a result will be available as standard outputs as well as an innovative virtual environment for immersive visualization. The case of multi-temporal data acquisition will be considered and discussed as well. Multiple 360ยฐ images acquired at different epochs from slightly different points are co-registered to obtain pixel-to-pixel correspondence, providing a solution to quantify and track deterioration effects

    Deep Learning for Vanishing Point Detection Using an Inverse Gnomonic Projection

    Full text link
    We present a novel approach for vanishing point detection from uncalibrated monocular images. In contrast to state-of-the-art, we make no a priori assumptions about the observed scene. Our method is based on a convolutional neural network (CNN) which does not use natural images, but a Gaussian sphere representation arising from an inverse gnomonic projection of lines detected in an image. This allows us to rely on synthetic data for training, eliminating the need for labelled images. Our method achieves competitive performance on three horizon estimation benchmark datasets. We further highlight some additional use cases for which our vanishing point detection algorithm can be used.Comment: Accepted for publication at German Conference on Pattern Recognition (GCPR) 2017. This research was supported by German Research Foundation DFG within Priority Research Programme 1894 "Volunteered Geographic Information: Interpretation, Visualisation and Social Computing

    Development of a text reading system on video images

    Get PDF
    Since the early days of computer science researchers sought to devise a machine which could automatically read text to help people with visual impairments. The problem of extracting and recognising text on document images has been largely resolved, but reading text from images of natural scenes remains a challenge. Scene text can present uneven lighting, complex backgrounds or perspective and lens distortion; it usually appears as short sentences or isolated words and shows a very diverse set of typefaces. However, video sequences of natural scenes provide a temporal redundancy that can be exploited to compensate for some of these deficiencies. Here we present a complete end-to-end, real-time scene text reading system on video images based on perspective aware text tracking. The main contribution of this work is a system that automatically detects, recognises and tracks text in videos of natural scenes in real-time. The focus of our method is on large text found in outdoor environments, such as shop signs, street names and billboards. We introduce novel efficient techniques for text detection, text aggregation and text perspective estimation. Furthermore, we propose using a set of Unscented Kalman Filters (UKF) to maintain each text regionยฟs identity and to continuously track the homography transformation of the text into a fronto-parallel view, thereby being resilient to erratic camera motion and wide baseline changes in orientation. The orientation of each text line is estimated using a method that relies on the geometry of the characters themselves to estimate a rectifying homography. This is done irrespective of the view of the text over a large range of orientations. We also demonstrate a wearable head-mounted device for text reading that encases a camera for image acquisition and a pair of headphones for synthesized speech output. Our system is designed for continuous and unsupervised operation over long periods of time. It is completely automatic and features quick failure recovery and interactive text reading. It is also highly parallelised in order to maximize the usage of available processing power and to achieve real-time operation. We show comparative results that improve the current state-of-the-art when correcting perspective deformation of scene text. The end-to-end system performance is demonstrated on sequences recorded in outdoor scenarios. Finally, we also release a dataset of text tracking videos along with the annotated ground-truth of text regions
    • โ€ฆ
    corecore