2,956 research outputs found
Improved Depth Map Estimation from Stereo Images based on Hybrid Method
In this paper, a stereo matching algorithm based on image segments is presented. We propose the hybrid segmentation algorithm that is based on a combination of the Belief Propagation and Mean Shift algorithms with aim to refine the disparity and depth map by using a stereo pair of images. This algorithm utilizes image filtering and modified SAD (Sum of Absolute Differences) stereo matching method. Firstly, a color based segmentation method is applied for segmenting the left image of the input stereo pair (reference image) into regions. The aim of the segmentation is to simplify representation of the image into the form that is easier to analyze and is able to locate objects in images. Secondly, results of the segmentation are used as an input of the local window-based matching method to determine the disparity estimate of each image pixel. The obtained experimental results demonstrate that the final depth map can be obtained by application of segment disparities to the original images. Experimental results with the stereo testing images show that our proposed Hybrid algorithm HSAD gives a good performance
Guidance for benthic habitat mapping: an aerial photographic approach
This document, Guidance for Benthic Habitat Mapping: An Aerial Photographic Approach, describes proven technology that can be applied in an operational manner by state-level scientists and resource managers. This information is based on the experience gained by NOAA Coastal Services Center staff and state-level cooperators in the production of a series of benthic habitat data sets in Delaware, Florida, Maine, Massachusetts, New York, Rhode Island, the Virgin Islands, and Washington, as well as during Center-sponsored workshops on coral remote sensing and seagrass and aquatic habitat assessment. (PDF contains 39 pages)
The original benthic habitat document, NOAA Coastal Change Analysis Program (C-CAP): Guidance for Regional Implementation (Dobson et al.), was published by the
Department of Commerce in 1995. That document summarized procedures that were to be used by scientists throughout the United States to develop consistent and reliable
coastal land cover and benthic habitat information. Advances in technology and new methodologies for generating these data created the need for this updated report,
which builds upon the foundation of its predecessor
Restoring Warped Document Image Based on Text Line Correction
[[abstract]]Document images captured by camera often suffer from warping and distortions because of the bounded volumes and complex environment light source. These effects not only reduce the document readability but also the OCR recognition performance. In this paper, we propose a method to combine non-linear and linear compensation for correcting distortions of document images. First, due to the broken text result of Otsu binarization, an image preprocessing is used to remove the effect of background light. Second, the dewarping method using the cubic polynomial fitting equation is proposed to find out the optimal approximate text line for vertical direction rectification. Finally, we use linear compensation for horizontal direction rectification. Experimental results demonstrate the robustness of the proposed methodology and improve the accuracy rate of OCR recognition.[[conferencetype]]國際[[conferencedate]]20130618~20130620[[booktype]]紙本[[iscallforpapers]]Y[[conferencelocation]]Jeju Island, Republic of Kore
DeepMatching: Hierarchical Deformable Dense Matching
We introduce a novel matching algorithm, called DeepMatching, to compute
dense correspondences between images. DeepMatching relies on a hierarchical,
multi-layer, correlational architecture designed for matching images and was
inspired by deep convolutional approaches. The proposed matching algorithm can
handle non-rigid deformations and repetitive textures and efficiently
determines dense correspondences in the presence of significant changes between
images. We evaluate the performance of DeepMatching, in comparison with
state-of-the-art matching algorithms, on the Mikolajczyk (Mikolajczyk et al
2005), the MPI-Sintel (Butler et al 2012) and the Kitti (Geiger et al 2013)
datasets. DeepMatching outperforms the state-of-the-art algorithms and shows
excellent results in particular for repetitive textures.We also propose a
method for estimating optical flow, called DeepFlow, by integrating
DeepMatching in the large displacement optical flow (LDOF) approach of Brox and
Malik (2011). Compared to existing matching algorithms, additional robustness
to large displacements and complex motion is obtained thanks to our matching
approach. DeepFlow obtains competitive performance on public benchmarks for
optical flow estimation
Guidelines for Best Practice and Quality Checking of Ortho Imagery
For almost 10 years JRC's ¿Guidelines for Best Practice and Quality Control of Ortho Imagery¿ has served as a reference document for the production of orthoimagery not only for the purposes of CAP but also for many medium-to-large scale photogrammetric applications. The aim is to provide the European Commission and the remote sensing user community with a general framework of the best approaches for quality checking of orthorectified remotely sensed imagery, and the expected best practice, required to achieve good results.
Since the last major revision (2003) the document was regularly updated in order to include state-of-the-art technologies. The major revision of the document was initiated last year in order to consolidate the information that was introduced to the document in the last five years. Following the internal discussion and the outcomes of the meeting with an expert panel it was decided to adopt as possible a process-based structure instead of a more sensor-based used before and also to keep the document as much generic as possible by focusing on the core aspects of the photogrammetric process. Additionally to any structural changes in the document new information was introduced mainly concerned with image resolution and radiometry, digital airborne sensors, data fusion, mosaicking and data compression.
The Guidelines of best practice is used as the base for our work on the definition of technical specifications for the orthoimagery. The scope is to establish a core set of measures to ensure sufficient image quality for the purposes of CAP and particularly for the Land Parcel Identification System (PLIS), and also to define the set of metadata necessary for data documentation and overall job tracking.JRC.G.3-Agricultur
Development of a text reading system on video images
Since the early days of computer science researchers sought to devise a machine which could automatically read text to help people with visual impairments. The problem of extracting and recognising text on document images has been largely resolved, but reading text from images of natural scenes remains a challenge. Scene text can present uneven lighting, complex backgrounds or perspective and lens distortion; it usually appears as short sentences or isolated words and shows a very diverse set of typefaces. However, video sequences of natural scenes provide a temporal redundancy that can be exploited to compensate for some of these deficiencies. Here we present a complete end-to-end, real-time scene text reading system on video images based on perspective aware text tracking.
The main contribution of this work is a system that automatically detects, recognises and tracks text in videos of natural scenes in real-time. The focus of our method is on large text found in outdoor environments, such as shop signs, street names and billboards. We introduce novel efficient techniques for text detection, text aggregation and text perspective estimation. Furthermore, we propose using a set of Unscented Kalman Filters (UKF) to maintain each text region¿s identity and to continuously track the homography transformation of the text into a fronto-parallel view, thereby being resilient to erratic camera motion and wide baseline changes in orientation. The orientation of each text line is estimated using a method that relies on the geometry of the characters themselves to estimate a rectifying homography. This is done irrespective of the view of the text over a large range of orientations. We also demonstrate a wearable head-mounted device for text reading that encases a camera for image acquisition and a pair of headphones for synthesized speech output.
Our system is designed for continuous and unsupervised operation over long periods of time. It is completely automatic and features quick failure recovery and interactive text reading. It is also highly parallelised in order to maximize the usage of available processing power and to achieve real-time operation. We show comparative results that improve the current state-of-the-art when correcting perspective deformation of scene text. The end-to-end system performance is demonstrated on sequences recorded in outdoor scenarios. Finally, we also release a dataset of text tracking videos along with the annotated ground-truth of text regions
Enhancing text recognition on Tor Darknet images
[Abstract] Text Spotting can be used as an approach to retrieve
information found in images that cannot be
obtained otherwise, by performing text detection
rst and then recognizing the located text. Examples
of images to apply this task on can be
found in Tor network images, which contain information
that may not be found in plain text. When
comparing both stages, the latter performs worse
due to the low resolution of the cropped areas
among other problems. Focusing on the recognition
part of the pipeline, we study the performance
of ve recognition approaches, based on state-ofthe-
art neural network models, standalone OCR,
and OCR enhancements. We complement them
using string-matching techniques with two lexicons
and compare computational time on ve
di erent datasets, including Tor network images.
Our nal proposal achieved 39,70% precision of
text recognition in a custom dataset of images
taken from Tor domain
- …