33 research outputs found

    Remote mining: from clustering to DTM

    Get PDF
    LIDAR data acquisition is becoming an indispensable task for terrain characterization in large surfaces. In Mediterranean woods this job results hard due to the great variety of heights and forms, as well as sparse vegetation that they present. A new data mining-based approach is proposed with the aim of classifying LIDAR data clouds as a first step in DTM generation. The developed methodology consists in a multi-step iterative process that splits the data into different classes (ground and low/med/high vegetation) by means of a clustering algorithm. This method has been tested on three different areas of the southern Spain with successful results, verging on 80% hitsMinisterio de Ciencia y Tecnología TIN2007-6808

    Effects of Aerial LiDAR Data Density on the Accuracy of Building Reconstruction

    Get PDF
    Previous work has identified a positive relationship between the density of aerial LiDAR input for building reconstruction and the accuracy of the resulting reconstructed models. We hypothesize a point of diminished returns at which higher data density no longer contributes meaningfully to higher accuracy in the end product. We investigate this relationship by subsampling a high-density dataset from the City of Surrey, BC to different densities and inputting each subsampled dataset to reconstruction using two different reconstruction methods. We then determine the accuracy of reconstruction based on manually created reference data, in terms of both 2D footprint accuracy and 3D model accuracy. We find that there is no quantitative evidence for meaningfully improved output accuracy from densities higher than 4 p/m2 for either method, although aesthetic improvements at higher point cloud densities are noted for one method

    Rapid Inspection of Pavement Markings Using Mobile Laser Scanning Point Clouds

    Get PDF
    Intelligent Transportation System (ITS) is the combination of information technology, sensors and communications for more efficient, safer, more secure and more eco-friendly surface transport. One of the most viable forms of ITS is the driverless car, which exist mainly as prototypes. Serval automobile manufacturers (e.g. Ford, GM, BMW, Toyota, Tesla, Honda) and non-automobile companies (e.g. Apple, Google, Nokia, Baidu, Huawei) have invested in this field, and wider commercialization of the driverless car is estimated in 2025 to 2030. Currently, the key elements of the driverless car are the sensors and a prior 3D map. The sensors mounted on the vehicle are the “eyes” of the driverless car to capture the 3D data of its environment. Comparing its environment and a pre-prepared prior known 3D map, the driverless car can distinguish moving targets (e.g. vehicles, pedestrians) and permanent surface features (e.g. buildings, trees, roads, traffic signs) and take relevant actions. With a centimetre-accuracy prior map, the intractable perception problem is transformed into a solvable localization task. The most important technology for generating the prior map is Mobile Laser Scanning (MLS). MLS technology can safely and rapidly acquire highly dense and accurate georeferenced 3D point clouds with the measurement of surface reflectivity. Therefore, the 3D point clouds with intensity data not only offer the detailed 3D surface of the road but also contains pavement marking information that are embedded in the prior map for automatic navigation. Relevant researches have been focused on the pavement marking extraction from MLS data to collect, update and maintain the 3D prior maps. However, the accuracy and efficiency of automatic extraction of pavement markings can be further improved by intensity correction and window-based enhancement. Thus, this study aims at building a robust method for semi-automated information extraction of pavement markings detected from MLS point clouds. The proposed workflow consists of three components: preprocessing, extraction, and classification. In preprocessing, the 3D MLS point clouds are converted into the radiometrically corrected and enhanced 2D intensity imagery of the road surface. Then the pavement markings are automatically extracted with the intensity using a set of algorithms, including Otsu’s thresholding, neighbour-counting filtering, and region growing. Finally, the extracted pavement markings are classified with the geometric parameters using a manually defined decision tree. Case studies are conducted using the MLS datasets acquired in Kingston (Ontario, Canada) and Xiamen (Fujian, China), respectively, with significantly different road environments by two RIEGL VMX-450 systems. The results demonstrated that the proposed workflow and method can achieve 93% in completeness, 95% in correctness, and 94% in F-score respectively when using Xiamen dataset, while 84%, 93%, 89% respectively when using Kingston dataset

    Estimation of Carbon Storage in Urban Trees Using Multispectral Airborne Laser Scanning Data

    Get PDF
    With the continued growth of global population, urbanization becomes an inevitable trend. As substantial urban expansion undergoes, ecosystem and global land cover have been altered consequently. Urban development becomes the biggest contributor to global carbon emissions while the process of urbanization results in urban heat islands, climate change, and losses of carbon sinks. Urban vegetation has drawn direct attention of city planners and policy makers by considering the importance of vegetation in urban climate modification and energy conservation in different ways. For instance, tree shading and wind shielding effects can attenuate the direct solar heat and air infiltration into individual houses. In city wide, vegetation contributes the largest proportion of carbon storage which reduces climate warming and urban heat island effects by sequestering CO2 and storing carbon in biomass. The carbon content stored in individual trees can be estimated by dendrometric parameters such as the diameter at breast height (DBH) using allometry-based models. With the development of airborne laser scanning (ALS) technology, ALS data and very high resolution multispectral imagery have proven to be promising tools for deriving dendrometric parameters in forest. With the emerging multispectral ALS technology, it became possible to obtain both the range and spectral information from a single source meanwhile the intensity of multispectral ALS showed its power in vegetation mapping. This study aims to develop a workflow that can quantify the carbon storage in urban trees using multispectral ALS data. The workflow consists of four steps: multispectral ALS data processing, vegetation isolation, dendrometric parameters estimation, and carbon storage modeling. First, the raw multispectral ALS data is intensity-rectified and filtered to generate a normalized Digital Surface Model (nDSM) and multispectral ALS intensity information at wavelengths: 532 nm (Green), 1064 nm (Near-infrared, NIR), and 1550 nm (Shortwave Infrared, SWIR), respectively. Vegetation covers are isolated by the support vector machine (SVM) classifier using multispectral ALS intensity information and nDSM in which total six classes including two vegetation classes (grass and tree) are classified. Individual tree crown is delineated by local maxima filtering and marker-controlled watershed segmentation. Tree height and crown width are derived from the crown segments and compared with field measurements. An ALS-DBH (diameter at breast height) multiple linear regression model is developed to predict field-measured DBH using ALS-derived tree height and crown width and assessed by cross validation. Then the carbon storage in individual trees is calculated by allometric equations using ALS-estimated DBH and height. A total of 40 trees are sampled in the field that four attributes: height, crown width, DBH, and biomass are recorded for each single tree. The results show that the land cover classification with multispectral ALS intensity images and nDSM achieves above 90% overall accuracy. The result of local maxima filtering is improved by using both multispectral ALS intensity and nDSM as input data. The ALS-derived tree height has a root mean square error (RMSE) of 1.21 m (relative RMSE = 6.8%) and the ALS-derived crown width has a RMSE of 1.47 m (relative RMSE = 16.4%). The prediction performance of the ALS-DBH model achieves R2 over 0.80 with a RMSE of 4.6 cm. The predicted carbon storage using ALS-modeled DBH corresponded to a RMSE of 142 kg (28.6%) and a bias of 14.4 kg. Results suggest that ALS-based dendrometric parameter estimation and allometric models can yield consistent performance and accurate estimation. Citywide carbon storage estimation is derived in this study by extrapolating the values within the study area to the entire city based on the specific proportion of each land cover type in the entire city. The proposed workflow also reveals the potential of multispectral ALS data in estimating carbon storage at individual-tree level and mapping vegetation in the urban environment

    Geometric data understanding : deriving case specific features

    Get PDF
    There exists a tradition using precise geometric modeling, where uncertainties in data can be considered noise. Another tradition relies on statistical nature of vast quantity of data, where geometric regularity is intrinsic to data and statistical models usually grasp this level only indirectly. This work focuses on point cloud data of natural resources and the silhouette recognition from video input as two real world examples of problems having geometric content which is intangible at the raw data presentation. This content could be discovered and modeled to some degree by such machine learning (ML) approaches like deep learning, but either a direct coverage of geometry in samples or addition of special geometry invariant layer is necessary. Geometric content is central when there is a need for direct observations of spatial variables, or one needs to gain a mapping to a geometrically consistent data representation, where e.g. outliers or noise can be easily discerned. In this thesis we consider transformation of original input data to a geometric feature space in two example problems. The first example is curvature of surfaces, which has met renewed interest since the introduction of ubiquitous point cloud data and the maturation of the discrete differential geometry. Curvature spectra can characterize a spatial sample rather well, and provide useful features for ML purposes. The second example involves projective methods used to video stereo-signal analysis in swimming analytics. The aim is to find meaningful local geometric representations for feature generation, which also facilitate additional analysis based on geometric understanding of the model. The features are associated directly to some geometric quantity, and this makes it easier to express the geometric constraints in a natural way, as shown in the thesis. Also, the visualization and further feature generation is much easier. Third, the approach provides sound baseline methods to more traditional ML approaches, e.g. neural network methods. Fourth, most of the ML methods can utilize the geometric features presented in this work as additional features.Geometriassa käytetään perinteisesti tarkkoja malleja, jolloin datassa esiintyvät epätarkkuudet edustavat melua. Toisessa perinteessä nojataan suuren datamäärän tilastolliseen luonteeseen, jolloin geometrinen säännönmukaisuus on datan sisäsyntyinen ominaisuus, joka hahmotetaan tilastollisilla malleilla ainoastaan epäsuorasti. Tämä työ keskittyy kahteen esimerkkiin: luonnonvaroja kuvaaviin pistepilviin ja videohahmontunnistukseen. Nämä ovat todellisia ongelmia, joissa geometrinen sisältö on tavoittamattomissa raakadatan tasolla. Tämä sisältö voitaisiin jossain määrin löytää ja mallintaa koneoppimisen keinoin, esim. syväoppimisen avulla, mutta joko geometria pitää kattaa suoraan näytteistämällä tai tarvitaan neuronien lisäkerros geometrisia invariansseja varten. Geometrinen sisältö on keskeinen, kun tarvitaan suoraa avaruudellisten suureiden havainnointia, tai kun tarvitaan kuvaus geometrisesti yhtenäiseen dataesitykseen, jossa poikkeavat näytteet tai melu voidaan helposti erottaa. Tässä työssä tarkastellaan datan muuntamista geometriseen piirreavaruuteen kahden esimerkkiohjelman suhteen. Ensimmäinen esimerkki on pintakaarevuus, joka on uudelleen virinneen kiinnostuksen kohde kaikkialle saatavissa olevan datan ja diskreetin geometrian kypsymisen takia. Kaarevuusspektrit voivat luonnehtia avaruudellista kohdetta melko hyvin ja tarjota koneoppimisessa hyödyllisiä piirteitä. Toinen esimerkki koskee projektiivisia menetelmiä käytettäessä stereovideosignaalia uinnin analytiikkaan. Tavoite on löytää merkityksellisiä paikallisen geometrian esityksiä, jotka samalla mahdollistavat muun geometrian ymmärrykseen perustuvan analyysin. Piirteet liittyvät suoraan johonkin geometriseen suureeseen, ja tämä helpottaa luonnollisella tavalla geometristen rajoitteiden käsittelyä, kuten väitöstyössä osoitetaan. Myös visualisointi ja lisäpiirteiden luonti muuttuu helpommaksi. Kolmanneksi, lähestymistapa suo selkeän vertailumenetelmän perinteisemmille koneoppimisen lähestymistavoille, esim. hermoverkkomenetelmille. Neljänneksi, useimmat koneoppimismenetelmät voivat hyödyntää tässä työssä esitettyjä geometrisia piirteitä lisäämällä ne muiden piirteiden joukkoon

    Extraction of Digital Terrain Models from Airborne Laser Scanning Data based on Transfer-Learning

    Get PDF
    With the rapid urbanization, timely and comprehensive urban thematic and topographic information is highly needed. Digital Terrain Models (DTMs), as one of unique urban topographic information, directly affect subsequent urban applications such as smart cities, urban microclimate studies, emergency and disaster management. Therefore, both the accuracy and resolution of DTMs define the quality of consequent tasks. Current workflows for DTM extraction vary in accuracy and resolution due to the complexity of terrain and off-terrain objects. Traditional filters, which rely on certain assumptions of surface morphology, insufficiently generalize complex terrain. Recent development in semantic labeling of point clouds has shed light on this problem. Under the semantic labeling context, DTM extraction can be viewed as a binary classification task. This study aims at developing a workflow for automated point-wise DTM extraction from Airborne Laser Scanning (ALS) point clouds using a transfer-learning approach on ResNet. The workflow consists of three parts: feature image generation, transfer learning using ResNet, and accuracy assessment. First, each point is transformed into a feature image based on its elevation differences with neighbouring points. Then, the feature images are classified into ground and non-ground using ResNet models. The ground points are extracted by remapping each feature image to its corresponding points. Lastly, the proposed workflow is compared with two traditional filters, namely the Progressive Morphological Filter (PMF) and the Progress TIN Densification (PTD). Results show that the proposed workflow establishes an advantageous accuracy of DTM extraction, which yields only 0.522% Type I error, 4.84% Type II error and 2.43% total error. In comparison, Type I, Type II and total error for PMF are 7.82%, 11.6%, and 9.48%, for PTD are 1.55%, 5.37%, and 3.22%, respectively. The root mean squared error of interpolated DTM of 1 m resolution is only 7.3 cm. Moreover, the use of pre-trained weights largely accelerated the training process and enabled the network to reach unprecedented accuracy even on a small amount of training set. Qualitative analysis is further conducted to investigate the reliability and limitations of the proposed workflow

    Multiple Instance Choquet Integral for multiresolution sensor fusion

    Get PDF
    Imagine you are traveling to Columbia, MO for the first time. On your flight to Columbia, the woman sitting next to you recommended a bakery by a large park with a big yellow umbrella outside. After you land, you need directions to the hotel from the airport. Suppose you are driving a rental car, you will need to park your car at a parking lot or a parking structure. After a good night's sleep in the hotel, you may decide to go for a run in the morning on the closest trail and stop by that recommended bakery under a big yellow umbrella. It would be helpful in the course of completing all these tasks to accurately distinguish the proper car route and walking trail, find a parking lot, and pinpoint the yellow umbrella. Satellite imagery and other geo-tagged data such as Open Street Maps provide effective information for this goal. Open Street Maps can provide road information and suggest bakery within a five-mile radius. The yellow umbrella is a distinctive color and, perhaps, is made of a distinctive material that can be identified from a hyperspectral camera. Open Street Maps polygons are tagged with information such as "parking lot" and "sidewalk." All these information can and should be fused to help identify and offer better guidance on the tasks you are completing. Supervised learning methods generally require precise labels for each training data point. It is hard (and probably at an extra cost) to manually go through and label each pixel in the training imagery. GPS coordinates cannot always be fully trusted as a GPS device may only be accurate to the level of several pixels. In many cases, it is practically infeasible to obtain accurate pixel-level training labels to perform fusion for all the imagery and maps available. Besides, the training data may come in a variety of data types, such as imagery or as a 3D point cloud. The imagery may have different resolutions, scales and, even, coordinate systems. Previous fusion methods are generally only limited to data mapped to the same pixel grid, with accurate labels. Furthermore, most fusion methods are restricted to only two sources, even if certain methods, such as pan-sharpening, can deal with different geo-spatial types or data of different resolution. It is, therefore, necessary and important, to come up with a way to perform fusion on multiple sources of imagery and map data, possibly with different resolutions and of different geo-spatial types with consideration of uncertain labels. I propose a Multiple Instance Choquet Integral framework for multi-resolution multisensor fusion with uncertain training labels. The Multiple Instance Choquet Integral (MICI) framework addresses uncertain training labels and performs both classification and regression. Three classifier fusion models, i.e. the noisy-or, min-max, and generalized-mean models, are derived under MICI. The Multi-Resolution Multiple Instance Choquet Integral (MR-MICI) framework is built upon the MICI framework and further addresses multiresolution in the fusion sources in addition to the uncertainty in training labels. For both MICI and MR-MICI, a monotonic normalized fuzzy measure is learned to be used with the Choquet integral to perform two-class classifier fusion given bag-level training labels. An optimization scheme based on the evolutionary algorithm is used to optimize the models proposed. For regression problems where the desired prediction is real-valued, the primary instance assumption is adopted. The algorithms are applied to target detection, regression and scene understanding applications. Experiments are conducted on the fusion of remote sensing data (hyperspectral and LiDAR) over the campus of University of Southern Mississippi - Gulfpark. Clothpanel sub-pixel and super-pixel targets were placed on campus with varying levels of occlusion and the proposed algorithms can successfully detect the targets in the scene. A semi-supervised approach is developed to automatically generate training labels based on data from Google Maps, Google Earth and Open Street Map. Based on such training labels with uncertainty, the proposed algorithms can also identify materials on campus for scene understanding, such as road, buildings, sidewalks, etc. In addition, the algorithms are used for weed detection and real-valued crop yield prediction experiments based on remote sensing data that can provide information for agricultural applications.Includes biblographical reference

    International benchmarking of terrestrial laser scanning approaches for forest inventories

    Get PDF
    The last two decades have witnessed increasing awareness of the potential of terrestrial laser scanning (TLS) in forest applications in both public and commercial sectors, along with tremendous research efforts and progress. It is time to inspect the achievements of and the remaining barriers to TLS-based forest investigations, so further research and application are clearly orientated in operational uses of TLS. In such context, the international TLS benchmarking project was launched in 2014 by the European Spatial Data Research Organization and coordinated by the Finnish Geospatial Research Institute. The main objectives of this benchmarking study are to evaluate the potential of applying TLS in characterizing forests, to clarify the strengths and the weaknesses of TLS as a measure of forest digitization, and to reveal the capability of recent algorithms for tree-attribute extraction. The project is designed to benchmark the TLS algorithms by processing identical TLS datasets for a standardized set of forest attribute criteria and by evaluating the results through a common procedure respecting reliable references. Benchmarking results reflect large variances in estimating accuracies, which were unveiled through the 18 compared algorithms and through the evaluation framework, i.e., forest complexity categories, TLS data acquisition approaches, tree attributes and evaluation procedures. The evaluation framework includes three new criteria proposed in this benchmarking and the algorithm performances are investigated through combining two or more criteria (e.g., the accuracy of the individual tree attributes are inspected in conjunction with plot-level completeness) in order to reveal algorithms’ overall performance. The results also reveal some best available forest attribute estimates at this time, which clarify the status quo of TLS-based forest investigations. Some results are well expected, while some are new, e.g., the variances of estimating accuracies between single-/multi-scan, the principle of the algorithm designs and the possibility of a computer outperforming human operation. With single-scan data, i.e., one hemispherical scan per plot, most of the recent algorithms are capable of achieving stem detection with approximately 75% completeness and 90% correctness in the easy forest stands (easy plots: 600 stems/ha, 20 cm mean DBH). The detection rate decreases when the stem density increases and the average DBH decreases, i.e., 60% completeness with 90% correctness (medium plots: 1000 stem/ha, 15 cm mean DBH) and 30% completeness with 90% correctness (difficult plots: 2000 stems/ha, 10 cm mean DBH). The application of the multi-scan approach, i.e., five scans per plot at the center and four quadrant angles, is more effective in complex stands, increasing the completeness to approximately 90% for medium plots and to approximately 70% for difficult plots, with almost 100% correctness. The results of this benchmarking also show that the TLS-based approaches can provide the estimates of the DBH and the stem curve at a 1–2 cm accuracy that are close to what is required in practical applications, e.g., national forest inventories (NFIs). In terms of algorithm development, a high level of automation is a commonly shared standard, but a bottleneck occurs at stem detection and tree height estimation, especially in multilayer and dense forest stands. The greatest challenge is that even with the multi-scan approach, it is still hard to completely and accurately record stems of all trees in a plot due to the occlusion effects of the trees and bushes in forests. Future development must address the redundant yet incomplete point clouds of forest sample plots and recognize trees more accurately and efficiently. It is worth noting that TLS currently provides the best quality terrestrial point clouds in comparison with all other technologies, meaning that all the benchmarks labeled in this paper can also serve as a reference for other terrestrial point clouds sources.</p

    Autonomous 3D Urban and Complex Terrain Geometry Generation and Micro-Climate Modelling Using CFD and Deep Learning

    Get PDF
    Sustainable building design requires a clear understanding and realistic modelling of the complex interaction between climate and built environment to create safe and comfortable outdoor and indoor spaces. This necessitates unprecedented urban climate modelling at high temporal and spatial resolution. The interaction between complex urban geometries and the microclimate is characterized by complex transport mechanisms. The challenge to generate geometric and physics boundary conditions in an automated manner is hindering the progress of computational methods in urban design. Thus, the challenge of modelling realistic and pragmatic numerical urban micro-climate for wind engineering, environmental, and building energy simulation applications should address the complexity of the geometry and the variability of surface types involved in urban exposures. The original contribution to knowledge in this research is the proposed an end-to-end workflow that employs a cutting-edge deep learning model for image segmentation to generate building footprint polygons autonomously and combining those polygons with LiDAR data to generate level of detail three (LOD3) 3D building models to tackle the geometry modelling issue in climate modelling and solar power potential assessment. Urban and topography geometric modelling is a challenging task when undertaking climate model assessment. This paper describes a deep learning technique that is based on U-Net architecture to automate 3D building model generation by combining satellite imagery with LiDAR data. The deep learning model used registered a mean squared error of 0.02. The extracted building polygons were extruded using height information from corresponding LiDAR data. The building roof structures were also modelled from the same point cloud data. The method used has the potential to automate the task of generating urban scale 3D building models and can be used for city-wide applications. The advantage of applying a deep learning model in an image processing task is that it can be applied to a new set of input image data to extract building footprint polygons for autonomous application once it has been trained. In addition, the model can be improved over time with minimum adjustments when an improved quality dataset is available, and the trained parameters can be improved further building on previously learned features. Application examples for pedestrian level wind and solar energy availability assessment as well as modeling wind flow over complex terrain are presented
    corecore