5 research outputs found

    Uncertainties of Human Perception in Visual Image Interpretation in Complex Urban Environments

    Get PDF
    Today satellite images are mostly exploited automatically due to advances in image classification methods. Manual visual image interpretation (MVII), however, still plays a significant role e.g., to generate training data for machine-learning algorithms or for validation purposes. In certain urban environments, however, of e.g., highest densities and structural complexity, textural and spectral complications in overlapping roof-structures still demand the human interpreter if one aims to capture individual building structures. The cognitive perception and real-world experience are still inevitable. Against these backgrounds, this article aims at quantifying and interpreting the uncertainties of mapping rooftop footprints of such areas. We focus on the agreement among interpreters and which aspects of perception and elements of image interpretation affect mapping. Ten test persons digitized six complex built-up areas. Hereby, we receive quantitative information about spatial variables of buildings to systematically check the consistency and congruence of results. An additional questionnaire reveals qualitative information about obstacles. Generally, we find large differences among interpreters’ mapping results and a high consistency of results for the same interpreter. We measure rising deviations correlate with a rising morphologic complexity. High degrees of individuality are expressed e.g., in time consumption, insitu-or geographic information system (GIS)-precognition whereas data source mostly influences the mapping procedure. By this study, we aim to fill a gap as prior research using MVII often does not implement an uncertainty analysis or quantify mapping aberrations. We conclude that remote sensing studies should not only rely unquestioned on MVII for validation; furthermore, data and methods are needed to suspend uncertainty

    Improved fully convolutional network with conditional random field for building extraction

    Get PDF
    Dissertation submitted in partial fulfilment of the requirements for the degree of Master of Science in Geospatial TechnologiesBuilding extraction from remotely sensed imagery plays an important role in urban planning, disaster management, navigation, updating geographic databases and several other geospatial applications. Several published contributions are dedicated to the applications of Deep Convolutional Neural Network (DCNN) for building extraction using aerial/satellite imagery exists; however, in all these contributions a good accuracy is always paid at the price of extremely complex and large network architectures. In this paper, we present an enhanced Fully Convolutional Network (FCN) framework especially molded for building extraction of remotely sensed images by applying Conditional Random Field (CRF). The main purpose here is to propose a framework which balances maximum accuracy with less network complexity. The modern activation function called Exponential Linear Unit (ELU) is applied to improve the performance of the Fully Convolutional Network (FCN), resulting in more, yet accurate building prediction. To further reduce the noise (false classified buildings) and to sharpen the boundary of the buildings, a post processing CRF is added at the end of the adopted Convolutional Neural Network (CNN) framework. The experiments were conducted on Massachusetts building aerial imagery. The results show that our proposed framework outperformed FCN baseline, which is the existing baseline framework for semantic segmentation, in term of performance measure, the F1-score and Intersection Over Union (IoU) measure. Additionally, the proposed method stood superior to the pre-existing classifier for building extraction using the same dataset in terms of performance measure and network complexity at once

    Learning to count buildings in diverse aerial scenes

    No full text
    corecore