459 research outputs found
Pansharpening via Frequency-Aware Fusion Network with Explicit Similarity Constraints
The process of fusing a high spatial resolution (HR) panchromatic (PAN) image
and a low spatial resolution (LR) multispectral (MS) image to obtain an HRMS
image is known as pansharpening. With the development of convolutional neural
networks, the performance of pansharpening methods has been improved, however,
the blurry effects and the spectral distortion still exist in their fusion
results due to the insufficiency in details learning and the frequency mismatch
between MSand PAN. Therefore, the improvement of spatial details at the premise
of reducing spectral distortion is still a challenge. In this paper, we propose
a frequency-aware fusion network (FAFNet) together with a novel high-frequency
feature similarity loss to address above mentioned problems. FAFNet is mainly
composed of two kinds of blocks, where the frequency aware blocks aim to
extract features in the frequency domain with the help of discrete wavelet
transform (DWT) layers, and the frequency fusion blocks reconstruct and
transform the features from frequency domain to spatial domain with the
assistance of inverse DWT (IDWT) layers. Finally, the fusion results are
obtained through a convolutional block. In order to learn the correspondence,
we also propose a high-frequency feature similarity loss to constrain the HF
features derived from PAN and MS branches, so that HF features of PAN can
reasonably be used to supplement that of MS. Experimental results on three
datasets at both reduced- and full-resolution demonstrate the superiority of
the proposed method compared with several state-of-the-art pansharpening
models.Comment: 14 page
CrossDiff: Exploring Self-Supervised Representation of Pansharpening via Cross-Predictive Diffusion Model
Fusion of a panchromatic (PAN) image and corresponding multispectral (MS)
image is also known as pansharpening, which aims to combine abundant spatial
details of PAN and spectral information of MS. Due to the absence of
high-resolution MS images, available deep-learning-based methods usually follow
the paradigm of training at reduced resolution and testing at both reduced and
full resolution. When taking original MS and PAN images as inputs, they always
obtain sub-optimal results due to the scale variation. In this paper, we
propose to explore the self-supervised representation of pansharpening by
designing a cross-predictive diffusion model, named CrossDiff. It has two-stage
training. In the first stage, we introduce a cross-predictive pretext task to
pre-train the UNet structure based on conditional DDPM, while in the second
stage, the encoders of the UNets are frozen to directly extract spatial and
spectral features from PAN and MS, and only the fusion head is trained to adapt
for pansharpening task. Extensive experiments show the effectiveness and
superiority of the proposed model compared with state-of-the-art supervised and
unsupervised methods. Besides, the cross-sensor experiments also verify the
generalization ability of proposed self-supervised representation learners for
other satellite's datasets. We will release our code for reproducibility
- …