1,717 research outputs found
Object Contour and Edge Detection with RefineContourNet
A ResNet-based multi-path refinement CNN is used for object contour
detection. For this task, we prioritise the effective utilization of the
high-level abstraction capability of a ResNet, which leads to state-of-the-art
results for edge detection. Keeping our focus in mind, we fuse the high, mid
and low-level features in that specific order, which differs from many other
approaches. It uses the tensor with the highest-levelled features as the
starting point to combine it layer-by-layer with features of a lower
abstraction level until it reaches the lowest level. We train this network on a
modified PASCAL VOC 2012 dataset for object contour detection and evaluate on a
refined PASCAL-val dataset reaching an excellent performance and an Optimal
Dataset Scale (ODS) of 0.752. Furthermore, by fine-training on the BSDS500
dataset we reach state-of-the-art results for edge-detection with an ODS of
0.824.Comment: Keywords: Object Contour Detection, Edge Detection, Multi-Path
Refinement CN
New techniques for the automatic registration of microwave and optical remotely sensed images
Remote sensing is a remarkable tool for monitoring and mapping the land and ocean surfaces of the Earth. Recently, with the launch of many new Earth observation satellites, there has been an increase in the amount of data that is being acquired, and the potential for mapping is greater than ever before. Furthermore, sensors which are currently operational are acquiring data in many different parts of the electromagnetic spectrum. It has long been known that by combining images that have been acquired at different wavelengths, or at different times, the ability to detect and recognise features on the ground is greatly increased. This thesis investigates the possibilities for automatically combining radar and optical remotely sensed images. The process of combining images, known as data integration, is a two step procedure: geometric integration (image registration) and radiometric integration (data fusion). Data fusion is essentially an automatic procedure, but the problems associated with automatic registration of multisource images have not, in general, been resolved. This thesis proposes a method of automatic image registration based on the extraction and matching of common features which are visible in both images. The first stage of the registration procedure uses patches as the matching primitives in order to determine the approximate alignment of the images. The second stage refines the registration results by matching edge features. Throughout the development of the proposed registration algorithm, reliability, robustness and automation were always considered priorities. Tests with both small images (512x512 pixels) and full scene images showed that the algorithm could successfully register images to an acceptable level of accuracy
Recommended from our members
Geophysical data registration using modified plane-wave destruction filters
I propose a method to efficiently measure local shifts, slopes, and scaling functions between seismic traces using modified plane-wave destruction filters.
Plane-wave destruction can efficiently measure shifts of less than a few samples, making this algorithm particularly effective for detecting small shifts.
When shifts are large, amplitude-adjusted plane-wave destruction can also be used to refine shift estimates obtained by other methods.
Amplitude-adjusted plane-wave destruction separates estimation of local shifts and amplitude weights, allowing the time-shift to be measured more accurately.
This algorithm has clear applications to geophysical data registration problems, including time-lapse image registration, multicomponent image registration, automatic gather flattening, automatic seismic-well ties, and image merging.
The effectiveness of this algorithm in predicting shifts associated with fluid migration, wave mode conversions, and anisotropy and amplitude gradients associated with amplitude variations with offset or angle is demonstrated by applying the algorithm to a synthetic trace, a time-lapse field data example from the Cranfield COâ‚‚ sequestration project, a multicomponent field data example from West Texas, and the Mobil AVO prestack seismic data.
Finding correspondence between different parts of the same dataset falls into the same category of problems as local shift estimation.
Computation of structure-oriented amplitude gradients for attribute-assisted interpretation requires the estimation of local slopes by correlating reflections between neighboring seismic traces in an image.
One of the major challenges of interpreting seismic images is the delineation of reflection discontinuities that are related to geologic features, such as faults, channels, salt boundaries, and unconformities.
Visually prominent reflection features often overshadow these subtle discontinuous features which are critical to understanding the structural and depositional environment of the subsurface.
For this reason, precise manual interpretation of these reflection discontinuities in seismic images can be tedious and time-consuming, especially when data quality is poor.
Discontinuity enhancement attributes are commonly used to facilitate the interpretation process by enhancing edges in seismic images and providing a quantitative measure of the significance of discontinuous features.
These attributes require careful pre-processing to maintain geologic features and suppress acquisition and processing artifacts which may be artificially detected as a geologic edge.
The plane-wave Sobel filter cascades plane-wave destruction filters with plane-wave shaping in the transverse direction to compute an enhanced discontinuity attribute.
The plane-wave Sobel attribute can be applied directly to a seismic image to efficiently and effectively enhance discontinuous features, or to a coherence image to create a sharper and more detailed image.
I demonstrate the effectiveness of this method by applying it to two field data sets from offshore New Zealand and offshore Nova Scotia with several faults and channel features and compare the results to other coherence attributes.Geological Science
Vision based strategies for implementing Sense and Avoid capabilities onboard Unmanned Aerial Systems
Current research activities are worked out to develop fully autonomous unmanned platform systems, provided with Sense and Avoid technologies in order to achieve the access to the National Airspace System (NAS), flying with manned airplanes. The TECVOl project is set in this framework, aiming at developing an autonomous prototypal Unmanned Aerial Vehicle which performs Detect Sense and Avoid functionalities, by means of an integrated sensors package, composed by a pulsed radar and four electro-optical cameras, two visible and two Infra-Red. This project is carried out by the Italian Aerospace Research Center in collaboration with the Department of Aerospace Engineering of the University of Naples “Federico II”, which has been involved in the developing of the Obstacle Detection and IDentification system.
Thus, this thesis concerns the image processing technique customized for the Sense and Avoid applications in the TECVOL project, where the EO system has an auxiliary role to radar, which is the main sensor. In particular, the panchromatic camera performs the aiding function of object detection, in order to increase accuracy and data rate performance of radar system. Therefore, the thesis describes the implemented steps to evaluate the most suitable panchromatic camera image processing technique for our applications, the test strategies adopted to study its performance and the analysis conducted to optimize it in terms of false alarms, missed detections and detection range. Finally, results from the tests will be explained, and they will demonstrate that the Electro-Optical sensor is beneficial to the overall Detect Sense and Avoid system; in fact it is able to improve upon it, in terms of object detection and tracking performance
City Scene Super-Resolution via Geometric Error Minimization
Super-resolution techniques are crucial in improving image granularity,
particularly in complex urban scenes, where preserving geometric structures is
vital for data-informed cultural heritage applications. In this paper, we
propose a city scene super-resolution method via geometric error minimization.
The geometric-consistent mechanism leverages the Hough Transform to extract
regular geometric features in city scenes, enabling the computation of
geometric errors between low-resolution and high-resolution images. By
minimizing mixed mean square error and geometric align error during the
super-resolution process, the proposed method efficiently restores details and
geometric regularities. Extensive validations on the SET14, BSD300, Cityscapes
and GSV-Cities datasets demonstrate that the proposed method outperforms
existing state-of-the-art methods, especially in urban scenes.Comment: 26 pages, 10 figure
Improving Dynamic HDR Imaging with Fusion Transformer
Reconstructing a High Dynamic Range (HDR) image from several Low Dynamic Range (LDR) images with different exposures is a challenging task, especially in the presence of camera and object motion. Though existing models using convolutional neural networks (CNNs) have made great progress, challenges still exist, e.g., ghosting artifacts. Transformers, originating from the field of natural language processing, have shown success in computer vision tasks, due to their ability to address a large receptive field even within a single layer. In this paper, we propose a transformer model for HDR imaging. Our pipeline includes three steps: alignment, fusion, and reconstruction. The key component is the HDR transformer module. Through experiments and ablation studies, we demonstrate that our model outperforms the state-of-the-art by large margins on several popular public datasets
- …