10,870 research outputs found

    Hierarchical Object Parsing from Structured Noisy Point Clouds

    Full text link
    Object parsing and segmentation from point clouds are challenging tasks because the relevant data is available only as thin structures along object boundaries or other features, and is corrupted by large amounts of noise. To handle this kind of data, flexible shape models are desired that can accurately follow the object boundaries. Popular models such as Active Shape and Active Appearance models lack the necessary flexibility for this task, while recent approaches such as the Recursive Compositional Models make model simplifications in order to obtain computational guarantees. This paper investigates a hierarchical Bayesian model of shape and appearance in a generative setting. The input data is explained by an object parsing layer, which is a deformation of a hidden PCA shape model with Gaussian prior. The paper also introduces a novel efficient inference algorithm that uses informed data-driven proposals to initialize local searches for the hidden variables. Applied to the problem of object parsing from structured point clouds such as edge detection images, the proposed approach obtains state of the art parsing errors on two standard datasets without using any intensity information.Comment: 13 pages, 16 figure

    Discriminating small wooded elements in rural landscape from aerial photography: a hybrid pixel/object-based analysis approach

    Get PDF
    While small, fragmented wooded elements do not represent a large surface area in agricultural landscape, their role in the sustainability of ecological processes is recognized widely. Unfortunately, landscape ecology studies suffer from the lack of methods for automatic detection of these elements. We propose a hybrid approach using both aerial photographs and ancillary data of coarser resolution to automatically discriminate small wooded elements. First, a spectral and textural analysis is performed to identify all the planted-tree areas in the digital photograph. Secondly, an object-orientated spatial analysis using the two data sources and including a multi-resolution segmentation is applied to distinguish between large and small woods, copses, hedgerows and scattered trees. The results show the usefulness of the hybrid approach and the prospects for future ecological applications

    ROAM: a Rich Object Appearance Model with Application to Rotoscoping

    Get PDF
    Rotoscoping, the detailed delineation of scene elements through a video shot, is a painstaking task of tremendous importance in professional post-production pipelines. While pixel-wise segmentation techniques can help for this task, professional rotoscoping tools rely on parametric curves that offer the artists a much better interactive control on the definition, editing and manipulation of the segments of interest. Sticking to this prevalent rotoscoping paradigm, we propose a novel framework to capture and track the visual aspect of an arbitrary object in a scene, given a first closed outline of this object. This model combines a collection of local foreground/background appearance models spread along the outline, a global appearance model of the enclosed object and a set of distinctive foreground landmarks. The structure of this rich appearance model allows simple initialization, efficient iterative optimization with exact minimization at each step, and on-line adaptation in videos. We demonstrate qualitatively and quantitatively the merit of this framework through comparisons with tools based on either dynamic segmentation with a closed curve or pixel-wise binary labelling

    MILD-Net: Minimal Information Loss Dilated Network for Gland Instance Segmentation in Colon Histology Images

    Get PDF
    The analysis of glandular morphology within colon histopathology images is an important step in determining the grade of colon cancer. Despite the importance of this task, manual segmentation is laborious, time-consuming and can suffer from subjectivity among pathologists. The rise of computational pathology has led to the development of automated methods for gland segmentation that aim to overcome the challenges of manual segmentation. However, this task is non-trivial due to the large variability in glandular appearance and the difficulty in differentiating between certain glandular and non-glandular histological structures. Furthermore, a measure of uncertainty is essential for diagnostic decision making. To address these challenges, we propose a fully convolutional neural network that counters the loss of information caused by max-pooling by re-introducing the original image at multiple points within the network. We also use atrous spatial pyramid pooling with varying dilation rates for preserving the resolution and multi-level aggregation. To incorporate uncertainty, we introduce random transformations during test time for an enhanced segmentation result that simultaneously generates an uncertainty map, highlighting areas of ambiguity. We show that this map can be used to define a metric for disregarding predictions with high uncertainty. The proposed network achieves state-of-the-art performance on the GlaS challenge dataset and on a second independent colorectal adenocarcinoma dataset. In addition, we perform gland instance segmentation on whole-slide images from two further datasets to highlight the generalisability of our method. As an extension, we introduce MILD-Net+ for simultaneous gland and lumen segmentation, to increase the diagnostic power of the network.Comment: Initial version published at Medical Imaging with Deep Learning (MIDL) 201

    Estimation of forest variables using airborne laser scanning

    Get PDF
    Airborne laser scanning can provide three-dimensional measurements of the forest canopy with high efficiency and precision. There are presently a large number of airborne laser scanning instruments in operation. The aims of the studies reported in this thesis were, to develop and validate methods for estimation of forest variables using laser data, and to investigate the influence of laser system parameters on the estimates. All studies were carried out in hemi-boreal forest at a test area in southwestern Sweden (lat. 58°30’N, long. 13°40’ E). Forest variables were estimated using regression models. On plot level, the Root Mean Square Error (RMSE) for mean tree height estimations ranged between 6% and 11% of the average value for different datasets and methods. The RMSE for stem volume estimations ranged between 19% and 26% of the average value for different datasets and methods. On stand level (area 0.64 ha), the RMSE was 3% and 11% of the average value for mean tree height and stem volume estimations, respectively. A simulation model was used to investigate the effect of different scanning angles on laser measurement of tree height and canopy closure. The effect of different scanning angles was different within different simulated forest types, e.g., different tree species. High resolution laser data were used for detection of individual trees. In total, 71% of the field measurements were detected representing 91% of the total stem volume. Height and crown diameter of the detected trees could be estimated with a RMSE of 0.63 m and 0.61 m, respectively. The magnitude of the height estimation errors was similar to what is usually achieved using field inventory. Using different laser footprint diameters (0.26 to 3.68 m) gave similar estimation accuracies. The tree species Norway spruce (Picea abies L. Karst.) and Scots pine (Pinus sylvestris L.) were discriminated at individual tree level with an accuracy of 95%. The results in this thesis show that airborne laser scanners are useful as forest inventory tools. Forest variables can be estimated on tree level, plot level and stand level with similar accuracies as traditional field inventories

    Quantifying the urban forest environment using dense discrete return LiDAR and aerial color imagery for segmentation and object-level biomass assessment

    Get PDF
    The urban forest is becoming increasingly important in the contexts of urban green space and recreation, carbon sequestration and emission offsets, and socio-economic impacts. In addition to aesthetic value, these green spaces remove airborne pollutants, preserve natural resources, and mitigate adverse climate changes, among other benefits. A great deal of attention recently has been paid to urban forest management. However, the comprehensive monitoring of urban vegetation for carbon sequestration and storage is an under-explored research area. Such an assessment of carbon stores often requires information at the individual tree level, necessitating the proper masking of vegetation from the built environment, as well as delineation of individual tree crowns. As an alternative to expensive and time-consuming manual surveys, remote sensing can be used effectively in characterizing the urban vegetation and man-made objects. Many studies in this field have made use of aerial and multispectral/hyperspectral imagery over cities. The emergence of light detection and ranging (LiDAR) technology, however, has provided new impetus to the effort of extracting objects and characterizing their 3D attributes - LiDAR has been used successfully to model buildings and urban trees. However, challenges remain when using such structural information only, and researchers have investigated the use of fusion-based approaches that combine LiDAR and aerial imagery to extract objects, thereby allowing the complementary characteristics of the two modalities to be utilized. In this study, a fusion-based classification method was implemented between high spatial resolution aerial color (RGB) imagery and co-registered LiDAR point clouds to classify urban vegetation and buildings from other urban classes/cover types. Structural, as well as spectral features, were used in the classification method. These features included height, flatness, and the distribution of normal surface vectors from LiDAR data, along with a non-calibrated LiDAR-based vegetation index, derived from combining LiDAR intensity at 1064 nm with the red channel of the RGB imagery. This novel index was dubbed the LiDAR-infused difference vegetation index (LDVI). Classification results indicated good separation between buildings and vegetation, with an overall accuracy of 92% and a kappa statistic of 0.85. A multi-tiered delineation algorithm subsequently was developed to extract individual tree crowns from the identified tree clusters, followed by the application of species-independent biomass models based on LiDAR-derived tree attributes in regression analysis. These LiDAR-based biomass assessments were conducted for individual trees, as well as for clusters of trees, in cases where proper delineation of individual trees was impossible. The detection accuracy of the tree delineation algorithm was 70%. The LiDAR-derived biomass estimates were validated against allometry-based biomass estimates that were computed from field-measured tree data. It was found out that LiDAR-derived tree volume, area, and different distribution parameters of height (e.g., maximum height, mean of height) are important to model biomass. The best biomass model for the tree clusters and the individual trees showed an adjusted R-Squared value of 0.93 and 0.58, respectively. The results of this study showed that the developed fusion-based classification approach using LiDAR and aerial color (RGB) imagery is capable of producing good object detection accuracy. It was concluded that the LDVI can be used in vegetation detection and can act as a substitute for the normalized difference vegetation index (NDVI), when near-infrared multiband imagery is not available. Furthermore, the utility of LiDAR for characterizing the urban forest and associated biomass was proven. This work could have significant impact on the rapid and accurate assessment of urban green spaces and associated carbon monitoring and management

    ATD: a multiplatform for semiautomatic 3-D detection of kidneys and their pathology in real time

    Get PDF
    This research presents a novel multi-functional system for medical Imaging-enabled Assistive Diagnosis (IAD). Although the IAD demonstrator has focused on abdominal images and supports the clinical diagnosis of kidneys using CT/MRI imaging, it can be adapted to work on image delineation, annotation and 3D real-size volumetric modelling of other organ structures such as the brain, spine, etc. The IAD provides advanced real-time 3D visualisation and measurements with fully automated functionalities as developed in two stages. In the first stage, via the clinically driven user interface, specialist clinicians use CT/MRI imaging datasets to accurately delineate and annotate the kidneys and their possible abnormalities, thus creating “3D Golden Standard Models”. Based on these models, in the second stage, clinical support staff i.e. medical technicians interactively define model-based rules and parameters for the integrated “Automatic Recognition Framework” to achieve results which are closest to that of the clinicians. These specific rules and parameters are stored in “Templates” and can later be used by any clinician to automatically identify organ structures i.e. kidneys and their possible abnormalities. The system also supports the transmission of these “Templates” to another expert for a second opinion. A 3D model of the body, the organs and their possible pathology with real metrics is also integrated. The automatic functionality was tested on eleven MRI datasets (comprising of 286 images) and the 3D models were validated by comparing them with the metrics from the corresponding “3D Golden Standard Models”. The system provides metrics for the evaluation of the results, in terms of Accuracy, Precision, Sensitivity, Specificity and Dice Similarity Coefficient (DSC) so as to enable benchmarking of its performance. The first IAD prototype has produced promising results as its performance accuracy based on the most widely deployed evaluation metric, DSC, yields 97% for the recognition of kidneys and 96% for their abnormalities; whilst across all the above evaluation metrics its performance ranges between 96% and 100%. Further development of the IAD system is in progress to extend and evaluate its clinical diagnostic support capability through development and integration of additional algorithms to offer fully computer-aided identification of other organs and their abnormalities based on CT/MRI/Ultra-sound Imaging
    • …
    corecore