1,677 research outputs found

    FULL-WAVEFORM AND DISCRETE-RETURN LIDAR IN SALT MARSH ENVIRONMENTS: AN ASSESSMENT OF BIOPHYSICAL PARAMETERS, VERTICAL UNCERTATINTY, AND NONPARAMETRIC DEM CORRECTION

    Get PDF
    High-resolution and high-accuracy elevation data sets of coastal salt marsh environments are necessary to support restoration and other management initiatives, such as adaptation to sea level rise. Lidar (light detection and ranging) data may serve this need by enabling efficient acquisition of detailed elevation data from an airborne platform. However, previous research has revealed that lidar data tend to have lower vertical accuracy (i.e., greater uncertainty) in salt marshes than in other environments. The increase in vertical uncertainty in lidar data of salt marshes can be attributed primarily to low, dense-growing salt marsh vegetation. Unfortunately, this increased vertical uncertainty often renders lidar-derived digital elevation models (DEM) ineffective for analysis of topographic features controlling tidal inundation frequency and ecology. This study aims to address these challenges by providing a detailed assessment of the factors influencing lidar-derived elevation uncertainty in marshes. The information gained from this assessment is then used to: 1) test the ability to predict marsh vegetation biophysical parameters from lidar-derived metrics, and 2) develop a method for improving salt marsh DEM accuracy. Discrete-return and full-waveform lidar, along with RTK GNSS (Real-time Kinematic Global Navigation Satellite System) reference data, were acquired for four salt marsh systems characterized by four major taxa (Spartina alterniflora, Spartina patens, Distichlis spicata, and Salicornia spp.) on Cape Cod, Massachusetts. These data were used to: 1) develop an innovative combination of full-waveform lidar and field methods to assess the vertical distribution of aboveground biomass as well as its light blocking properties; 2) investigate lidar elevation bias and standard deviation using varying interpolation and filtering methods; 3) evaluate the effects of seasonality (temporal differences between peak growth and senescent conditions) using lidar data flown in summer and spring; 4) create new products, called Relative Uncertainty Surfaces (RUS), from lidar waveform-derived metrics and determine their utility; and 5) develop and test five nonparametric regression model algorithms (MARS - Multivariate Adaptive Regression, CART - Classification and Regression Trees, TreeNet, Random Forests, and GPSM - Generalized Path Seeker) with 13 predictor variables derived from both discrete and full waveform lidar sources in order to develop a method of improving lidar DEM quality. Results of this study indicate strong correlations for Spartina alterniflora (r \u3e 0.9) between vertical biomass (VB), the distribution of vegetation biomass by height, and vertical obscuration (VO), the measure of the vertical distribution of the ratio of vegetation to airspace. It was determined that simple, feature-based lidar waveform metrics, such as waveform width, can provide new information to estimate salt marsh vegetation biophysical parameters such as vegetation height. The results also clearly illustrate the importance of seasonality, species, and lidar interpolation and filtering methods on elevation uncertainty in salt marshes. Relative uncertainty surfaces generated from lidar waveform features were determined useful in qualitative/visual assessment of lidar elevation uncertainty and correlate well with vegetation height and presence of Spartina alterniflora. Finally, DEMs generated using full-waveform predictor models produced corrections (compared to ground based RTK GNSS elevations) with R2 values of up to 0.98 and slopes within 4% of a perfect 1:1 correlation. The findings from this research have strong potential to advance tidal marsh mapping, research and management initiatives

    Design of a High-Speed Architecture for Stabilization of Video Captured Under Non-Uniform Lighting Conditions

    Get PDF
    Video captured in shaky conditions may lead to vibrations. A robust algorithm to immobilize the video by compensating for the vibrations from physical settings of the camera is presented in this dissertation. A very high performance hardware architecture on Field Programmable Gate Array (FPGA) technology is also developed for the implementation of the stabilization system. Stabilization of video sequences captured under non-uniform lighting conditions begins with a nonlinear enhancement process. This improves the visibility of the scene captured from physical sensing devices which have limited dynamic range. This physical limitation causes the saturated region of the image to shadow out the rest of the scene. It is therefore desirable to bring back a more uniform scene which eliminates the shadows to a certain extent. Stabilization of video requires the estimation of global motion parameters. By obtaining reliable background motion, the video can be spatially transformed to the reference sequence thereby eliminating the unintended motion of the camera. A reflectance-illuminance model for video enhancement is used in this research work to improve the visibility and quality of the scene. With fast color space conversion, the computational complexity is reduced to a minimum. The basic video stabilization model is formulated and configured for hardware implementation. Such a model involves evaluation of reliable features for tracking, motion estimation, and affine transformation to map the display coordinates of a stabilized sequence. The multiplications, divisions and exponentiations are replaced by simple arithmetic and logic operations using improved log-domain computations in the hardware modules. On Xilinx\u27s Virtex II 2V8000-5 FPGA platform, the prototype system consumes 59% logic slices, 30% flip-flops, 34% lookup tables, 35% embedded RAMs and two ZBT frame buffers. The system is capable of rendering 180.9 million pixels per second (mpps) and consumes approximately 30.6 watts of power at 1.5 volts. With a 1024×1024 frame, the throughput is equivalent to 172 frames per second (fps). Future work will optimize the performance-resource trade-off to meet the specific needs of the applications. It further extends the model for extraction and tracking of moving objects as our model inherently encapsulates the attributes of spatial distortion and motion prediction to reduce complexity. With these parameters to narrow down the processing range, it is possible to achieve a minimum of 20 fps on desktop computers with Intel Core 2 Duo or Quad Core CPUs and 2GB DDR2 memory without a dedicated hardware

    Combined Learned and Classical Methods for Real-Time Visual Perception in Autonomous Driving

    Full text link
    Autonomy, robotics, and Artificial Intelligence (AI) are among the main defining themes of next-generation societies. Of the most important applications of said technologies is driving automation which spans from different Advanced Driver Assistance Systems (ADAS) to full self-driving vehicles. Driving automation is promising to reduce accidents, increase safety, and increase access to mobility for more people such as the elderly and the handicapped. However, one of the main challenges facing autonomous vehicles is robust perception which can enable safe interaction and decision making. With so many sensors to perceive the environment, each with its own capabilities and limitations, vision is by far one of the main sensing modalities. Cameras are cheap and can provide rich information of the observed scene. Therefore, this dissertation develops a set of visual perception algorithms with a focus on autonomous driving as the target application area. This dissertation starts by addressing the problem of real-time motion estimation of an agent using only the visual input from a camera attached to it, a problem known as visual odometry. The visual odometry algorithm can achieve low drift rates over long-traveled distances. This is made possible through the innovative local mapping approach used. This visual odometry algorithm was then combined with my multi-object detection and tracking system. The tracking system operates in a tracking-by-detection paradigm where an object detector based on convolution neural networks (CNNs) is used. Therefore, the combined system can detect and track other traffic participants both in image domain and in 3D world frame while simultaneously estimating vehicle motion. This is a necessary requirement for obstacle avoidance and safe navigation. Finally, the operational range of traditional monocular cameras was expanded with the capability to infer depth and thus replace stereo and RGB-D cameras. This is accomplished through a single-stream convolution neural network which can output both depth prediction and semantic segmentation. Semantic segmentation is the process of classifying each pixel in an image and is an important step toward scene understanding. Literature survey, algorithms descriptions, and comprehensive evaluations on real-world datasets are presented.Ph.D.College of Engineering & Computer ScienceUniversity of Michiganhttps://deepblue.lib.umich.edu/bitstream/2027.42/153989/1/Mohamed Aladem Final Dissertation.pdfDescription of Mohamed Aladem Final Dissertation.pdf : Dissertatio

    Evaluating the accuracy of vehicle tracking data obtained from Unmanned Aerial Vehicles

    Get PDF
    Abstract This paper presents a methodology for tracking moving vehicles that integrates Unmanned Aerial Vehicles with video processing techniques. The authors investigated the usefulness of Unmanned Aerial Vehicles to capture reliable individual vehicle data by using GPS technology as a benchmark. A video processing algorithm for vehicles trajectory acquisition is introduced. The algorithm is based on OpenCV libraries. In order to assess the accuracy of the proposed video processing algorithm an instrumented vehicle was equipped with a high precision GPS. The video capture experiments were performed in two case studies. From the field, about 24,000 positioning data were acquired for the analysis. The results of these experiments highlight the versatility of the Unmanned Aerial Vehicles technology combined with video processing technique in monitoring real traffic data

    The Need for Accurate Pre-processing and Data Integration for the Application of Hyperspectral Imaging in Mineral Exploration

    Get PDF
    Die hyperspektrale Bildgebung stellt eine SchlĂŒsseltechnologie in der nicht-invasiven Mineralanalyse dar, sei es im Labormaßstab oder als fernerkundliche Methode. Rasante Entwicklungen im Sensordesign und in der Computertechnik hinsichtlich Miniaturisierung, Bildauflösung und DatenqualitĂ€t ermöglichen neue Einsatzgebiete in der Erkundung mineralischer Rohstoffe, wie die drohnen-gestĂŒtzte Datenaufnahme oder digitale Aufschluss- und Bohrkernkartierung. AllgemeingĂŒltige Datenverarbeitungsroutinen fehlen jedoch meist und erschweren die Etablierung dieser vielversprechenden AnsĂ€tze. Besondere Herausforderungen bestehen hinsichtlich notwendiger radiometrischer und geometrischer Datenkorrekturen, der rĂ€umlichen Georeferenzierung sowie der Integration mit anderen Datenquellen. Die vorliegende Arbeit beschreibt innovative ArbeitsablĂ€ufe zur Lösung dieser Problemstellungen und demonstriert die Wichtigkeit der einzelnen Schritte. Sie zeigt das Potenzial entsprechend prozessierter spektraler Bilddaten fĂŒr komplexe Aufgaben in Mineralexploration und Geowissenschaften.Hyperspectral imaging (HSI) is one of the key technologies in current non-invasive material analysis. Recent developments in sensor design and computer technology allow the acquisition and processing of high spectral and spatial resolution datasets. In contrast to active spectroscopic approaches such as X-ray fluorescence or laser-induced breakdown spectroscopy, passive hyperspectral reflectance measurements in the visible and infrared parts of the electromagnetic spectrum are considered rapid, non-destructive, and safe. Compared to true color or multi-spectral imagery, a much larger range and even small compositional changes of substances can be differentiated and analyzed. Applications of hyperspectral reflectance imaging can be found in a wide range of scientific and industrial fields, especially when physically inaccessible or sensitive samples and processes need to be analyzed. In geosciences, this method offers a possibility to obtain spatially continuous compositional information of samples, outcrops, or regions that might be otherwise inaccessible or too large, dangerous, or environmentally valuable for a traditional exploration at reasonable expenditure. Depending on the spectral range and resolution of the deployed sensor, HSI can provide information about the distribution of rock-forming and alteration minerals, specific chemical compounds and ions. Traditional operational applications comprise space-, airborne, and lab-scale measurements with a usually (near-)nadir viewing angle. The diversity of available sensors, in particular the ongoing miniaturization, enables their usage from a wide range of distances and viewing angles on a large variety of platforms. Many recent approaches focus on the application of hyperspectral sensors in an intermediate to close sensor-target distance (one to several hundred meters) between airborne and lab-scale, usually implying exceptional acquisition parameters. These comprise unusual viewing angles as for the imaging of vertical targets, specific geometric and radiometric distortions associated with the deployment of small moving platforms such as unmanned aerial systems (UAS), or extreme size and complexity of data created by large imaging campaigns. Accurate geometric and radiometric data corrections using established methods is often not possible. Another important challenge results from the overall variety of spatial scales, sensors, and viewing angles, which often impedes a combined interpretation of datasets, such as in a 2D geographic information system (GIS). Recent studies mostly referred to work with at least partly uncorrected data that is not able to set the results in a meaningful spatial context. These major unsolved challenges of hyperspectral imaging in mineral exploration initiated the motivation for this work. The core aim is the development of tools that bridge data acquisition and interpretation, by providing full image processing workflows from the acquisition of raw data in the field or lab, to fully corrected, validated and spatially registered at-target reflectance datasets, which are valuable for subsequent spectral analysis, image classification, or fusion in different operational environments at multiple scales. I focus on promising emerging HSI approaches, i.e.: (1) the use of lightweight UAS platforms, (2) mapping of inaccessible vertical outcrops, sometimes at up to several kilometers distance, (3) multi-sensor integration for versatile sample analysis in the near-field or lab-scale, and (4) the combination of reflectance HSI with other spectroscopic methods such as photoluminescence (PL) spectroscopy for the characterization of valuable elements in low-grade ores. In each topic, the state of the art is analyzed, tailored workflows are developed to meet key challenges and the potential of the resulting dataset is showcased on prominent mineral exploration related examples. Combined in a Python toolbox, the developed workflows aim to be versatile in regard to utilized sensors and desired applications
    • 

    corecore