318 research outputs found

    3D Remote Sensing Applications in Forest Ecology: Composition, Structure and Function

    Get PDF
    Dear Colleagues, The composition, structure and function of forest ecosystems are the key features characterizing their ecological properties, and can thus be crucially shaped and changed by various biotic and abiotic factors on multiple spatial scales. The magnitude and extent of these changes in recent decades calls for enhanced mitigation and adaption measures. Remote sensing data and methods are the main complementary sources of up-to-date synoptic and objective information of forest ecology. Due to the inherent 3D nature of forest ecosystems, the analysis of 3D sources of remote sensing data is considered to be most appropriate for recreating the forest’s compositional, structural and functional dynamics. In this Special Issue of Forests, we published a set of state-of-the-art scientific works including experimental studies, methodological developments and model validations, all dealing with the general topic of 3D remote sensing-assisted applications in forest ecology. We showed applications in forest ecology from a broad collection of method and sensor combinations, including fusion schemes. All in all, the studies and their focuses are as broad as a forest’s ecology or the field of remote sensing and, thus, reflect the very diverse usages and directions toward which future research and practice will be directed

    Urban tree classification using discrete-return LiDAR and an object-level local binary pattern algorithm

    Full text link
    © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group. Urban trees have the potential to mitigate some of the harm brought about by rapid urbanization and population growth, as well as serious environmental degradation (e.g. soil erosion, carbon pollution and species extirpation), in cities. This paper presents a novel urban tree extraction modelling approach that uses discrete laser scanning point clouds and object-based textural analysis to (1) develop a model characterised by four sub-models, including (a) height-based split segmentation, (b) feature extraction, (c) texture analysis and (d) classification, and (2) apply this model to classify urban trees. The canopy height model is integrated with the object-level local binary pattern algorithm (LBP) to achieve high classification accuracy. The results of each sub-model reveal that the classification of urban trees based on the height at 47.14 (high) and 2.12 m (low), respectively, while based on crown widths were highest and lowest at 22.5 and 2.55 m, respectively. Results also indicate that the proposed algorithm of urban tree modelling is effective for practical use

    Anomalous change detection in multi-temporal hyperspectral images

    Get PDF
    In latest years, the possibility to exploit the high amount of spectral information has made hyperspectral remote sensing a very promising approach to detect changes occurred in multi-temporal images. Detection of changes in images of the same area collected at different times is of crucial interest in military and civilian applications, spanning from wide area surveillance and damage assessment to geology and land cover. In military operations, the interest is in rapid location and tracking of objects of interest, people, vehicles or equipment that pose a potential threat. In civilian contexts, changes of interest may include different types of natural or manmade threats, such as the path of an impending storm or the source of a hazardous material spill. In this PhD thesis, the focus is on Anomalous Change Detection (ACD) in airborne hyperspectral images. The goal is the detection of small changes occurred in two images of the same scene, i.e. changes having size comparable with the sensor ground resolution. The objects of interest typically occupy few pixels of the image and change detection must be accomplished in a pixel-wise fashion. Moreover, since the images are in general not radiometrically comparable, because illumination, atmospheric and environmental conditions change from one acquisition to the other, pervasive and uninteresting changes must be accounted for in developing ACD strategies. ACD process can be distinguished into two main phases: a pre-processing step, which includes radiometric correction, image co-registration and noise filtering, and a detection step, where the pre-processed images are compared according to a defined criterion in order to derive a statistical ACD map highlighting the anomalous changes occurred in the scene. In the literature, ACD has been widely investigated providing valuable methods in order to cope with these problems. In this work, a general overview of ACD methods is given reviewing the most known pre-processing and detection methods proposed in the literature. The analysis has been conducted unifying different techniques in a common framework based on binary decision theory, where one has to test the two competing hypotheses H0 (change absent) and H1 (change present) on the basis of an observation vector derived from the radiance measured on each pixel of the two images. Particular emphasis has been posed on statistical approaches, where ACD is derived in the framework of Neymann Pearson theory and the decision rule is carried out on the basis of the statistical properties assumed for the two hypotheses distribution, the observation vector space and the secondary data exploited for the estimation of the unknown parameters. Typically, ACD techniques assume that the observation represents the realization of jointly Gaussian spatially stationary random process. Though such assumption is adopted because of its mathematical tractability, it may be quite simplistic to model the multimodality usually met in real data. A more appropriate model is that adopted to derive the well known RX anomaly detector which assumes the local Gaussianity of the hyperspectral data. In this framework, a new statistical ACD method has been proposed considering the local Gaussianity of the hyperspectral data. The assumption of local stationarity for the observations in the two hypotheses is taken into account by considering two different models, leading to two different detectors. In addition, when data are collected by airborne platforms, perfect co-registration between images is very difficult to achieve. As a consequence, a residual misregistration (RMR) error should be taken into account in developing ACD techniques. Different techniques have been proposed to cope with the performance degradation problem due to the RMR, embedding the a priori knowledge on the statistical properties of the RMR in the change detection scheme. In this context, a new method has been proposed for the estimation of the first and second order statistics of the RMR. The technique is based on a sequential strategy that exploits the Scale Invariant Feature Transform (SIFT) algorithm cascaded with the Minimum Covariance Determinant algorithm. The proposed method adapts the SIFT procedure to hyperspectral images and improves the robustness of the outliers filtering by means of a highly robust estimator of multivariate location. Then, the attention has been focused on noise filtering techniques aimed at enforcing the consistency of the ACD process. To this purpose, a new method has been proposed to mitigate the negative effects due to random noise. In particular, this is achieved by means of a band selection technique aimed at discarding spectral channels whose useful signal content is low compared with the noise contribution. Band selection is performed on a per-pixel basis by exploiting the estimates of the noise variance accounting also for the presence of the signal dependent noise component. Finally, the effectiveness of the proposed techniques has been extensively evaluated by employing different real hyperspectral datasets containing anomalous changes collected in different acquisition conditions and on different scenarios, highlighting advantages and drawbacks of each method. In summary, the main issues related to ACD in multi-temporal hyperspectral images have been examined in this PhD thesis. With reference to the pre-processing step, two original contributions have been offered: i) an unsupervised technique for the estimation of the RMR noise affecting hyperspectral images, and ii) an adaptive approach for ACD which mitigates the negative effects due to random noise. As to the detection step, a survey of the existing techniques has been carried out, highlighting the major drawbacks and disadvantages, and a novel contribution has been offered by presenting a new statistical ACD method which considers the local Gaussianity of the hyperspectral data

    A Review of Landcover Classification with Very-High Resolution Remotely Sensed Optical Images—Analysis Unit, Model Scalability and Transferability

    Get PDF
    As an important application in remote sensing, landcover classification remains one of the most challenging tasks in very-high-resolution (VHR) image analysis. As the rapidly increasing number of Deep Learning (DL) based landcover methods and training strategies are claimed to be the state-of-the-art, the already fragmented technical landscape of landcover mapping methods has been further complicated. Although there exists a plethora of literature review work attempting to guide researchers in making an informed choice of landcover mapping methods, the articles either focus on the review of applications in a specific area or revolve around general deep learning models, which lack a systematic view of the ever advancing landcover mapping methods. In addition, issues related to training samples and model transferability have become more critical than ever in an era dominated by data-driven approaches, but these issues were addressed to a lesser extent in previous review articles regarding remote sensing classification. Therefore, in this paper, we present a systematic overview of existing methods by starting from learning methods and varying basic analysis units for landcover mapping tasks, to challenges and solutions on three aspects of scalability and transferability with a remote sensing classification focus including (1) sparsity and imbalance of data; (2) domain gaps across different geographical regions; and (3) multi-source and multi-view fusion. We discuss in detail each of these categorical methods and draw concluding remarks in these developments and recommend potential directions for the continued endeavor

    A Review of Landcover Classification with Very-High Resolution Remotely Sensed Optical Images—Analysis Unit, Model Scalability and Transferability

    Get PDF
    As an important application in remote sensing, landcover classification remains one of the most challenging tasks in very-high-resolution (VHR) image analysis. As the rapidly increasing number of Deep Learning (DL) based landcover methods and training strategies are claimed to be the state-of-the-art, the already fragmented technical landscape of landcover mapping methods has been further complicated. Although there exists a plethora of literature review work attempting to guide researchers in making an informed choice of landcover mapping methods, the articles either focus on the review of applications in a specific area or revolve around general deep learning models, which lack a systematic view of the ever advancing landcover mapping methods. In addition, issues related to training samples and model transferability have become more critical than ever in an era dominated by data-driven approaches, but these issues were addressed to a lesser extent in previous review articles regarding remote sensing classification. Therefore, in this paper, we present a systematic overview of existing methods by starting from learning methods and varying basic analysis units for landcover mapping tasks, to challenges and solutions on three aspects of scalability and transferability with a remote sensing classification focus including (1) sparsity and imbalance of data; (2) domain gaps across different geographical regions; and (3) multi-source and multi-view fusion. We discuss in detail each of these categorical methods and draw concluding remarks in these developments and recommend potential directions for the continued endeavor

    Relation Network for Multi-label Aerial Image Classification

    Get PDF
    Multi-label classification plays a momentous role in perceiving intricate contents of an aerial image and triggers several related studies over the last years. However, most of them deploy few efforts in exploiting label relations, while such dependencies are crucial for making accurate predictions. Although an LSTM layer can be introduced to modeling such label dependencies in a chain propagation manner, the efficiency might be questioned when certain labels are improperly inferred. To address this, we propose a novel aerial image multi-label classification network, attention-aware label relational reasoning network. Particularly, our network consists of three elemental modules: 1) a label-wise feature parcel learning module, 2) an attentional region extraction module, and 3) a label relational inference module. To be more specific, the label-wise feature parcel learning module is designed for extracting high-level label-specific features. The attentional region extraction module aims at localizing discriminative regions in these features and yielding attentional label-specific features. The label relational inference module finally predicts label existences using label relations reasoned from outputs of the previous module. The proposed network is characterized by its capacities of extracting discriminative label-wise features in a proposal-free way and reasoning about label relations naturally and interpretably. In our experiments, we evaluate the proposed model on the UCM multi-label dataset and a newly produced dataset, AID multi-label dataset. Quantitative and qualitative results on these two datasets demonstrate the effectiveness of our model. To facilitate progress in the multi-label aerial image classification, the AID multi-label dataset will be made publicly available

    Recurrently Exploring Class-wise Attention in A Hybrid Convolutional and Bidirectional LSTM Network for Multi-label Aerial Image Classification

    Get PDF
    Aerial image classification is of great significance in remote sensing community, and many researches have been conducted over the past few years. Among these studies, most of them focus on categorizing an image into one semantic label, while in the real world, an aerial image is often associated with multiple labels, e.g., multiple object-level labels in our case. Besides, a comprehensive picture of present objects in a given high resolution aerial image can provide more in-depth understanding of the studied region. For these reasons, aerial image multi-label classification has been attracting increasing attention. However, one common limitation shared by existing methods in the community is that the co-occurrence relationship of various classes, so called class dependency, is underexplored and leads to an inconsiderate decision. In this paper, we propose a novel end-to-end network, namely class-wise attention-based convolutional and bidirectional LSTM network (CA-Conv-BiLSTM), for this task. The proposed network consists of three indispensable components: 1) a feature extraction module, 2) a class attention learning layer, and 3) a bidirectional LSTM-based sub-network. Particularly, the feature extraction module is designed for extracting fine-grained semantic feature maps, while the class attention learning layer aims at capturing discriminative class-specific features. As the most important part, the bidirectional LSTM-based sub-network models the underlying class dependency in both directions and produce structured multiple object labels. Experimental results on UCM multi-label dataset and DFC15 multi-label dataset validate the effectiveness of our model quantitatively and qualitatively

    Spatial Pyramid Context-Aware Moving Object Detection and Tracking for Full Motion Video and Wide Aerial Motion Imagery

    Get PDF
    A robust and fast automatic moving object detection and tracking system is essential to characterize target object and extract spatial and temporal information for different functionalities including video surveillance systems, urban traffic monitoring and navigation, robotic. In this dissertation, I present a collaborative Spatial Pyramid Context-aware moving object detection and Tracking system. The proposed visual tracker is composed of one master tracker that usually relies on visual object features and two auxiliary trackers based on object temporal motion information that will be called dynamically to assist master tracker. SPCT utilizes image spatial context at different level to make the video tracking system resistant to occlusion, background noise and improve target localization accuracy and robustness. We chose a pre-selected seven-channel complementary features including RGB color, intensity and spatial pyramid of HoG to encode object color, shape and spatial layout information. We exploit integral histogram as building block to meet the demands of real-time performance. A novel fast algorithm is presented to accurately evaluate spatially weighted local histograms in constant time complexity using an extension of the integral histogram method. Different techniques are explored to efficiently compute integral histogram on GPU architecture and applied for fast spatio-temporal median computations and 3D face reconstruction texturing. We proposed a multi-component framework based on semantic fusion of motion information with projected building footprint map to significantly reduce the false alarm rate in urban scenes with many tall structures. The experiments on extensive VOTC2016 benchmark dataset and aerial video confirm that combining complementary tracking cues in an intelligent fusion framework enables persistent tracking for Full Motion Video and Wide Aerial Motion Imagery.Comment: PhD Dissertation (162 pages

    Detecting, Tracking, And Recognizing Activities In Aerial Video

    Get PDF
    In this dissertation, we address the problem of detecting humans and vehicles, tracking them in crowded scenes, and finally determining their activities in aerial video. Even though this is a well explored problem in the field of computer vision, many challenges still remain when one is presented with realistic data. These challenges include large camera motion, strong scene parallax, fast object motion, large object density, strong shadows, and insufficiently large action datasets. Therefore, we propose a number of novel methods based on exploiting scene constraints from the imagery itself to aid in the detection and tracking of objects. We show, via experiments on several datasets, that superior performance is achieved with the use of proposed constraints. First, we tackle the problem of detecting moving, as well as stationary, objects in scenes that contain parallax and shadows. We do this on both regular aerial video, as well as the new and challenging domain of wide area surveillance. This problem poses several challenges: large camera motion, strong parallax, large number of moving objects, small number of pixels on target, single channel data, and low frame-rate of video. We propose a method for detecting moving and stationary objects that overcomes these challenges, and evaluate it on CLIF and VIVID datasets. In order to find moving objects, we use median background modelling which requires few frames to obtain a workable model, and is very robust when there is a large number of moving objects in the scene while the model is being constructed. We then iii remove false detections from parallax and registration errors using gradient information from the background image. Relying merely on motion to detect objects in aerial video may not be sufficient to provide complete information about the observed scene. First of all, objects that are permanently stationary may be of interest as well, for example to determine how long a particular vehicle has been parked at a certain location. Secondly, moving vehicles that are being tracked through the scene may sometimes stop and remain stationary at traffic lights and railroad crossings. These prolonged periods of non-motion make it very difficult for the tracker to maintain the identities of the vehicles. Therefore, there is a clear need for a method that can detect stationary pedestrians and vehicles in UAV imagery. This is a challenging problem due to small number of pixels on the target, which makes it difficult to distinguish objects from background clutter, and results in a much larger search space. We propose a method for constraining the search based on a number of geometric constraints obtained from the metadata. Specifically, we obtain the orientation of the ground plane normal, the orientation of the shadows cast by out of plane objects in the scene, and the relationship between object heights and the size of their corresponding shadows. We utilize the above information in a geometry-based shadow and ground plane normal blob detector, which provides an initial estimation for the locations of shadow casting out of plane (SCOOP) objects in the scene. These SCOOP candidate locations are then classified as either human or clutter using a combination of wavelet features, and a Support Vector Machine. Additionally, we combine regular SCOOP and inverted SCOOP candidates to obtain vehicle candidates. We show impressive results on sequences from VIVID and CLIF datasets, and provide comparative quantitative and qualitative analysis. We also show that we can extend the SCOOP detection method to automatically estimate the iv orientation of the shadow in the image without relying on metadata. This is useful in cases where metadata is either unavailable or erroneous. Simply detecting objects in every frame does not provide sufficient understanding of the nature of their existence in the scene. It may be necessary to know how the objects have travelled through the scene over time and which areas they have visited. Hence, there is a need to maintain the identities of the objects across different time instances. The task of object tracking can be very challenging in videos that have low frame rate, high density, and a very large number of objects, as is the case in the WAAS data. Therefore, we propose a novel method for tracking a large number of densely moving objects in an aerial video. In order to keep the complexity of the tracking problem manageable when dealing with a large number of objects, we divide the scene into grid cells, solve the tracking problem optimally within each cell using bipartite graph matching and then link the tracks across the cells. Besides tractability, grid cells also allow us to define a set of local scene constraints, such as road orientation and object context. We use these constraints as part of cost function to solve the tracking problem; This allows us to track fast-moving objects in low frame rate videos. In addition to moving through the scene, the humans that are present may be performing individual actions that should be detected and recognized by the system. A number of different approaches exist for action recognition in both aerial and ground level video. One of the requirements for the majority of these approaches is the existence of a sizeable dataset of examples of a particular action from which a model of the action can be constructed. Such a luxury is not always possible in aerial scenarios since it may be difficult to fly a large number of missions to observe a particular event multiple times. Therefore, we propose a method for v recognizing human actions in aerial video from as few examples as possible (a single example in the extreme case). We use the bag of words action representation and a 1vsAll multi-class classification framework. We assume that most of the classes have many examples, and construct Support Vector Machine models for each class. Then, we use Support Vector Machines that were trained for classes with many examples to improve the decision function of the Support Vector Machine that was trained using few examples, via late weighted fusion of decision values
    • …
    corecore