13 research outputs found
Recommended from our members
Difference Enhancement and Spatial-Spectral Non-Local Network for Change Detection in VHR Remote Sensing Images
Copyright © The Author(s) 2021. The popular Siamese convolutional neural networks (CNNs) for remote sensing (RS) image change detection (CD) often suffer from two problems. First, they either ignore the original information of bitemporal images or insufficiently utilize the difference information between bitemporal images, which leads to the low tightness of the changed objects. Second, Siamese CNNs always employ dual-branch encoders for CD, which increases computational cost. To address the above issues, this article proposes a network based on difference enhancement and spatial–spectral nonlocal (DESSN) for CD in very-high-resolution (VHR) images. This article makes threefold contributions. First, we design a difference enhancement (DE) module that can effectively learn the difference representation between foreground and background to reduce the impact of irrelevant changes on the detection results. Second, we present a spatial–spectral nonlocal (SSN) module that is different from vanilla nonlocal because multiscale spatial global features are incorporated to model the large-scale variation of objects during CD. The module can be used to strengthen the edge integrity and internal tightness of changed objects. Third, the asymmetric double convolution with Ghost (ADCG) module is exploited instead of standard convolution. The ADCG can not only refine the edge information of the changed objects, since horizontal and vertical convolutional kernels have good contour preservation advantages, but also greatly reduce the computational complexity of the proposed model. The experiments on two public VHR CD datasets demonstrate that the proposed network can provide higher detection accuracy and requires smaller memory usage than state-of-the-art networks.10.13039/501100017596-Natural Science Basic Research Program of Shaanxi (Grant Number: 2021JC-47);
10.13039/501100001809-National Natural Science Foundation of China (Grant Number: 61871259 and 61861024);
10.13039/501100000288-National Natural Science Foundation of China-Royal Society, U.K. (Grant Number: 61811530325 (IECnNSFCn170396));
Key Research and Development Program of Shaanxi (Grant Number: 2021ZDLGY08-07);
Shaanxi Joint Laboratory of Artificial Intelligence (Grant Number: 2020SS-03);
Special Construction Fund for Key Disciplines of Shaanxi Provincial Higher Education
Land-cover change detection using paired OpenStreetMap data and optical high-resolution imagery via object-guided Transformer
Optical high-resolution imagery and OpenStreetMap (OSM) data are two
important data sources for land-cover change detection. Previous studies in
these two data sources focus on utilizing the information in OSM data to aid
the change detection on multi-temporal optical high-resolution images. This
paper pioneers the direct detection of land-cover changes utilizing paired OSM
data and optical imagery, thereby broadening the horizons of change detection
tasks to encompass more dynamic earth observations. To this end, we propose an
object-guided Transformer (ObjFormer) architecture by naturally combining the
prevalent object-based image analysis (OBIA) technique with the advanced vision
Transformer architecture. The introduction of OBIA can significantly reduce the
computational overhead and memory burden in the self-attention module.
Specifically, the proposed ObjFormer has a hierarchical pseudo-siamese encoder
consisting of object-guided self-attention modules that extract representative
features of different levels from OSM data and optical images; a decoder
consisting of object-guided cross-attention modules can progressively recover
the land-cover changes from the extracted heterogeneous features. In addition
to the basic supervised binary change detection task, this paper raises a new
semi-supervised semantic change detection task that does not require any
manually annotated land-cover labels of optical images to train semantic change
detectors. Two lightweight semantic decoders are added to ObjFormer to
accomplish this task efficiently. A converse cross-entropy loss is designed to
fully utilize the negative samples, thereby contributing to the great
performance improvement in this task. The first large-scale benchmark dataset
containing 1,287 map-image pairs (1024 1024 pixels for each sample)
covering 40 regions on six continents ...(see the manuscript for the full
abstract
A Review of Landcover Classification with Very-High Resolution Remotely Sensed Optical Images—Analysis Unit, Model Scalability and Transferability
As an important application in remote sensing, landcover classification remains one of the most challenging tasks in very-high-resolution (VHR) image analysis. As the rapidly increasing number of Deep Learning (DL) based landcover methods and training strategies are claimed to be the state-of-the-art, the already fragmented technical landscape of landcover mapping methods has been further complicated. Although there exists a plethora of literature review work attempting to guide researchers in making an informed choice of landcover mapping methods, the articles either focus on the review of applications in a specific area or revolve around general deep learning models, which lack a systematic view of the ever advancing landcover mapping methods. In addition, issues related to training samples and model transferability have become more critical than ever in an era dominated by data-driven approaches, but these issues were addressed to a lesser extent in previous review articles regarding remote sensing classification. Therefore, in this paper, we present a systematic overview of existing methods by starting from learning methods and varying basic analysis units for landcover mapping tasks, to challenges and solutions on three aspects of scalability and transferability with a remote sensing classification focus including (1) sparsity and imbalance of data; (2) domain gaps across different geographical regions; and (3) multi-source and multi-view fusion. We discuss in detail each of these categorical methods and draw concluding remarks in these developments and recommend potential directions for the continued endeavor