4,105 research outputs found
Recent Advances in Image Restoration with Applications to Real World Problems
In the past few decades, imaging hardware has improved tremendously in terms of resolution, making widespread usage of images in many diverse applications on Earth and planetary missions. However, practical issues associated with image acquisition are still affecting image quality. Some of these issues such as blurring, measurement noise, mosaicing artifacts, low spatial or spectral resolution, etc. can seriously affect the accuracy of the aforementioned applications. This book intends to provide the reader with a glimpse of the latest developments and recent advances in image restoration, which includes image super-resolution, image fusion to enhance spatial, spectral resolution, and temporal resolutions, and the generation of synthetic images using deep learning techniques. Some practical applications are also included
Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors
Publiher's version (ΓΊtgefin grein)We propose an unmixing framework for enhancing endmember fraction maps using a combination of spectral and visible images. The new method, data fusion through spatial information-aided learning (DFuSIAL), is based on a learning process for the fusion of a multispectral image of low spatial resolution and a visible RGB image of high spatial resolution. Unlike commonly used methods, DFuSIAL allows for fusing data from different sensors. To achieve this objective, we apply a learning process using automatically extracted invariant points, which are assumed to have the same land cover type in both images. First, we estimate the fraction maps of a set of endmembers for the spectral image. Then, we train a spatial-features aided neural network (SFFAN) to learn the relationship between the fractions, the visible bands, and rotation-invariant spatial features for learning (RISFLs) that we extract from the RGB image. Our experiments show that the proposed DFuSIAL method obtains fraction maps with significantly enhanced spatial resolution and an average mean absolute error between 2% and 4% compared to the reference ground truth. Furthermore, it is shown that the proposed method is preferable to other examined state-of-the-art methods, especially when data is obtained from different instruments and in cases with missing-data pixels.This research was partially funded by the Icelandic Research Fund through the EMMIRS project, and bythe Israel Science Ministry and Space Agency through the Venus project.Peer Reviewe
Deep learning in remote sensing: a review
Standing at the paradigm shift towards data-intensive science, machine
learning techniques are becoming increasingly important. In particular, as a
major breakthrough in the field, deep learning has proven as an extremely
powerful tool in many fields. Shall we embrace deep learning as the key to all?
Or, should we resist a 'black-box' solution? There are controversial opinions
in the remote sensing community. In this article, we analyze the challenges of
using deep learning for remote sensing data analysis, review the recent
advances, and provide resources to make deep learning in remote sensing
ridiculously simple to start with. More importantly, we advocate remote sensing
scientists to bring their expertise into deep learning, and use it as an
implicit general model to tackle unprecedented large-scale influential
challenges, such as climate change and urbanization.Comment: Accepted for publication IEEE Geoscience and Remote Sensing Magazin
Development of an unsupervised remote sensing methodology of detect surface leakage from terrestrial CO2 storage sites
Imperial Users onl
Multisource and Multitemporal Data Fusion in Remote Sensing
The sharp and recent increase in the availability of data captured by
different sensors combined with their considerably heterogeneous natures poses
a serious challenge for the effective and efficient processing of remotely
sensed data. Such an increase in remote sensing and ancillary datasets,
however, opens up the possibility of utilizing multimodal datasets in a joint
manner to further improve the performance of the processing approaches with
respect to the application at hand. Multisource data fusion has, therefore,
received enormous attention from researchers worldwide for a wide variety of
applications. Moreover, thanks to the revisit capability of several spaceborne
sensors, the integration of the temporal information with the spatial and/or
spectral/backscattering information of the remotely sensed data is possible and
helps to move from a representation of 2D/3D data to 4D data structures, where
the time variable adds new information as well as challenges for the
information extraction algorithms. There are a huge number of research works
dedicated to multisource and multitemporal data fusion, but the methods for the
fusion of different modalities have expanded in different paths according to
each research community. This paper brings together the advances of multisource
and multitemporal data fusion approaches with respect to different research
communities and provides a thorough and discipline-specific starting point for
researchers at different levels (i.e., students, researchers, and senior
researchers) willing to conduct novel investigations on this challenging topic
by supplying sufficient detail and references
A Deep Learning Framework in Selected Remote Sensing Applications
The main research topic is designing and implementing a deep learning framework applied to remote sensing. Remote sensing techniques and applications play a crucial role in observing the Earth evolution, especially nowadays, where the effects of climate change on our life is more and more evident.
A considerable amount of data are daily acquired all over the Earth. Effective exploitation of this information requires the robustness, velocity and accuracy of deep learning. This emerging need inspired the choice of this topic.
The conducted studies mainly focus on two European Space Agency (ESA) missions: Sentinel 1 and Sentinel 2. Images provided by the ESA Sentinel-2 mission are rapidly becoming the main source of information for the entire remote sensing community, thanks to their unprecedented combination of spatial, spectral and temporal resolution, as well as their open access policy. The increasing interest gained by these satellites in the research laboratory and applicative scenarios pushed us to utilize them in the considered framework. The combined use of Sentinel 1 and Sentinel 2 is crucial and very prominent in different contexts and different kinds of monitoring when the growing (or changing) dynamics are very rapid.
Starting from this general framework, two specific research activities were identified and investigated, leading to the results presented in this dissertation. Both these studies can be placed in the context of data fusion.
The first activity deals with a super-resolution framework to improve Sentinel 2 bands supplied at 20 meters up to 10 meters. Increasing the spatial resolution of these bands is of great interest in many remote sensing applications, particularly in monitoring vegetation, rivers, forests, and so on.
The second topic of the deep learning framework has been applied to the multispectral Normalized Difference Vegetation Index (NDVI) extraction, and the semantic segmentation obtained fusing Sentinel 1 and S2 data. The S1 SAR data is of great importance for the quantity of information extracted in the context of monitoring wetlands, rivers and forests, and many other contexts.
In both cases, the problem was addressed with deep learning techniques, and in both cases, very lean architectures were used, demonstrating that even without the availability of computing power, it is possible to obtain high-level results.
The core of this framework is a Convolutional Neural Network (CNN).
{CNNs have been successfully applied
to many image processing problems,
like super-resolution,
pansharpening,
classification,
and others,
because of several advantages such as
(i) the capability to approximate complex non-linear
functions,
(ii) the ease of training that allows
to avoid time-consuming handcraft filter design,
(iii) the parallel computational architecture.
Even if a large amount of "labelled" data is required for training, the CNN performances pushed me to this architectural choice.} In our S1 and S2 integration task, we have faced and overcome the problem of manually labelled data with an approach based on integrating these two different sensors.
Therefore, apart from the investigation in Sentinel-1 and Sentinel-2 integration, the main contribution in both cases of these works is, in particular, the possibility of designing a CNN-based solution that can be distinguished by its lightness from a computational point of view and consequent substantial saving of time compared to more complex deep learning state-of-the-art solutions
μκ³΅κ° ν΄μλ ν₯μμ ν΅ν μμ λ³ν λͺ¨λν°λ§
νμλ
Όλ¬Έ(λ°μ¬) -- μμΈλνκ΅λνμ : νκ²½λνμ νλκ³Όμ μ‘°κ²½ν, 2023. 2. λ₯μλ ¬.μ‘μ μνκ³μμ λκΈ°κΆκ³Ό μλ¬ΌκΆμ μνΈ μμ©μ μ΄ν΄νκΈ° μν΄μλ μμ λ³νμ λͺ¨λν°λ§μ΄ νμνλ€. μ΄ λ, μμ±μμμ μ§νλ©΄μ κ΄μΈ‘νμ¬ μμμ§λλ₯Ό μ 곡ν μ μμ§λ§, μ§νλ³νμ μμΈν μ 보λ ꡬλ¦μ΄λ μμ± μ΄λ―Έμ§μ κ³΅κ° ν΄μλμ μν΄ μ νλμλ€. λν μμ±μμμ μκ³΅κ° ν΄μλκ° μμμ§λλ₯Ό ν΅ν κ΄ν©μ± λͺ¨λν°λ§μ λ―ΈμΉλ μν₯μ μμ ν λ°νμ§μ§ μμλ€.
λ³Έ λ
Όλ¬Έμμλ κ³ ν΄μλ μμ μ§λλ₯Ό μΌλ¨μλ‘ μμ±νκΈ° μμ± μμμ μκ³΅κ° ν΄μλλ₯Ό ν₯μμν€λ κ²μ λͺ©νλ‘ νμλ€. κ³ ν΄μλ μμ±μμμ νμ©ν μμ λ³ν λͺ¨λν°λ§μ μ곡κ°μ μΌλ‘ νμ₯νκΈ° μν΄ 1) μ μ§κΆ€λ μμ±μ νμ©ν μμμ΅ν©μ ν΅ν΄ μκ°ν΄μλ ν₯μ, 2) μ λμ μμ±λ€νΈμν¬λ₯Ό νμ©ν 곡κ°ν΄μλ ν₯μ, 3) μ곡κ°ν΄μλκ° λμ μμ±μμμ ν μ§νΌλ³΅μ΄ κ· μ§νμ§ μμ 곡κ°μμ μλ¬Ό κ΄ν©μ± λͺ¨λν°λ§μ μννμλ€. μ΄μ²λΌ, μμ±κΈ°λ° μ격νμ§μμ μλ‘μ΄ κΈ°μ μ΄ λ±μ₯ν¨μ λ°λΌ νμ¬ λ° κ³Όκ±°μ μμ±μμμ μκ³΅κ° ν΄μλ μΈ‘λ©΄μμ ν₯μλμ΄ μμ λ³νμ λͺ¨λν°λ§ ν μ μλ€.
μ 2μ₯μμλ μ μ§κΆ€λμμ±μμμ νμ©νλ μκ³΅κ° μμμ΅ν©μΌλ‘ μλ¬Όμ κ΄ν©μ±μ λͺ¨λν°λ§ νμ λ, μκ°ν΄μλκ° ν₯μλ¨μ 보μλ€. μκ³΅κ° μμμ΅ν© μ, ꡬλ¦νμ§, μλ°©ν₯ λ°μ¬ ν¨μ μ‘°μ , κ³΅κ° λ±λ‘, μκ³΅κ° μ΅ν©, μκ³΅κ° κ²°μΈ‘μΉ λ³΄μ λ±μ κ³Όμ μ κ±°μΉλ€. μ΄ μμμ΅ν© μ°μΆλ¬Όμ κ²½μκ΄λ¦¬ λ±μΌλ‘ μμ μ§μμ μ°κ° λ³λμ΄ ν° λ μ₯μ(λκ²½μ§μ λμ½μλ¦Ό)μμ νκ°νμλ€. κ·Έ κ²°κ³Ό, μκ³΅κ° μμμ΅ν© μ°μΆλ¬Όμ κ²°μΈ‘μΉ μμ΄ νμ₯κ΄μΈ‘μ μμΈ‘νμλ€ (R2 = 0.71, μλ νΈν₯ = 5.64% λκ²½μ§; R2 = 0.79, μλ νΈν₯ = -13.8%, νμ½μλ¦Ό). μκ³΅κ° μμμ΅ν©μ μμ μ§λμ μκ³΅κ° ν΄μλλ₯Ό μ μ§μ μΌλ‘ κ°μ νμ¬, μλ¬Ό μμ₯κΈ°λμ μμ±μμμ΄ νμ₯ κ΄μΈ‘μ κ³Όμ νκ°λ₯Ό μ€μλ€. μμμ΅ν©μ λμ μκ³΅κ° ν΄μλλ‘ κ΄ν©μ± μ§λλ₯Ό μΌκ°κ²©μΌλ‘ μμ±νκΈ°μ μ΄λ₯Ό νμ©νμ¬ μμ± μμμ μ νλ μκ³΅κ° ν΄μλλ‘ λ°νμ§μ§ μμ μλ¬Όλ³νμ κ³Όμ μ λ°κ²¬νκΈΈ κΈ°λνλ€.
μμμ 곡κ°λΆν¬μ μ λ°λμ
κ³Ό ν μ§ νΌλ³΅ λ³ν λͺ¨λν°λ§μ μν΄ νμμ μ΄λ€. κ³ ν΄μλ μμ±μμμΌλ‘ μ§κ΅¬ νλ©΄μ κ΄μΈ‘νλ κ²μ μ©μ΄νκ² ν΄μ‘λ€. νΉν Planet Fusionμ μ΄μνμμ±κ΅° λ°μ΄ν°λ₯Ό μ΅λν νμ©ν΄ λ°μ΄ν° κ²°μΈ‘μ΄ μλ 3m κ³΅κ° ν΄μλμ μ§ν νλ©΄ λ°μ¬λμ΄λ€. κ·Έλ¬λ κ³Όκ±° μμ± μΌμ(Landsatμ κ²½μ° 30~60m)μ κ³΅κ° ν΄μλλ μμμ 곡κ°μ λ³νλ₯Ό μμΈ λΆμνλ κ²μ μ ννλ€. μ 3μ₯μμλ Landsat λ°μ΄ν°μ κ³΅κ° ν΄μλλ₯Ό ν₯μνκΈ° μν΄ Planet Fusion λ° Landsat 8 λ°μ΄ν°λ₯Ό μ¬μ©νμ¬ μ΄μ€ μ λμ μμ± λ€νΈμν¬(the dual RSS-GAN)λ₯Ό νμ΅μμΌ, κ³ ν΄μλ μ κ·ν μμ μ§μ(NDVI)μ μλ¬Ό κ·Όμ μΈμ λ°μ¬(NIRv)λλ₯Ό μμ±νλ νλ€. νμκΈ°λ° νμ₯ μμμ§μ(μ΅λ 8λ
)μ λλ‘ κΈ°λ° μ΄λΆκ΄μ§λλ‘ the dual RSS-GANμ μ±λ₯μ λνλ―Όκ΅ λ΄ λ λμμ§(λκ²½μ§μ νμ½μλ¦Ό)μμ νκ°νλ€. The dual RSS-GANμ Landsat 8 μμμ 곡κ°ν΄μλλ₯Ό ν₯μμμΌ κ³΅κ° ννμ 보μνκ³ μμ μ§μμ κ³μ μ λ³νλ₯Ό ν¬μ°©νλ€(R2> 0.96). κ·Έλ¦¬κ³ the dual RSS-GANμ Landsat 8 μμ μ§μκ° νμ₯μ λΉν΄ κ³Όμ νκ°λλ κ²μ μννλ€. νμ₯ κ΄μΈ‘μ λΉν΄ μ΄μ€ RSS-GANκ³Ό Landsat 8μ μλ νΈν₯ κ° κ°κ° -0.8% μμ -1.5%, -10.3% μμ -4.6% μλ€. μ΄λ¬ν κ°μ μ Planet Fusionμ 곡κ°μ 보λ₯Ό μ΄μ€ RSS-GANλ‘ νμ΅νμκΈ°μ κ°λ₯νλ€. ν€λΉ μ°κ΅¬ κ²°κ³Όλ Landsat μμμ κ³΅κ° ν΄μλλ₯Ό ν₯μμμΌ μ¨κ²¨μ§ κ³΅κ° μ 보λ₯Ό μ 곡νλ μλ‘μ΄ μ κ·Ό λ°©μμ΄λ€.
κ³ ν΄μλμμ μλ¬Ό κ΄ν©μ± μ§λλ ν μ§νΌλ³΅μ΄ 볡μ‘ν 곡κ°μμ νμ μν λͺ¨λν°λ§μ νμμ μ΄λ€. κ·Έλ¬λ Sentinel-2, Landsat λ° MODISμ κ°μ΄ νμ λμ‘° κΆ€λμ μλ μμ±μ κ³΅κ° ν΄μλκ° λκ±°λ μκ° ν΄μλ λμ μμ±μμλ§ μ 곡ν μ μλ€. μ΅κ·Ό λ°μ¬λ μ΄μνμμ±κ΅°μ μ΄λ¬ν ν΄μλ νκ³μ 극볡ν μ μλ€. νΉν Planet Fusionμ μ΄μνμμ± μλ£μ μκ³΅κ° ν΄μλλ‘ μ§νλ©΄μ κ΄μΈ‘ν μ μλ€. 4μ₯μμ, Planet Fusion μ§νλ°μ¬λλ₯Ό μ΄μ©νμ¬ μμμμ λ°μ¬λ κ·Όμ μΈμ 볡μ¬(NIRvP)λ₯Ό 3m ν΄μλ μ§λλ₯Ό μΌκ°κ²©μΌλ‘ μμ±νλ€. κ·Έλ° λ€μ λ―Έκ΅ μΊλ¦¬ν¬λμμ£Ό μν¬λΌλ©ν -μ νΈμν¨ λΈνμ νλμ€ νμ λ€νΈμν¬ λ°μ΄ν°μ λΉκ΅νμ¬ μλ¬Ό κ΄ν©μ±μ μΆμ νκΈ° μν NIRvP μ§λμ μ±λ₯μ νκ°νμλ€. μ 체μ μΌλ‘ NIRvP μ§λλ μ΅μ§μ μ¦μ μμ λ³νμλ λΆκ΅¬νκ³ κ°λ³ λμμ§μ μλ¬Ό κ΄ν©μ±μ μκ°μ λ³νλ₯Ό ν¬μ°©νμλ€. κ·Έλ¬λ λμμ§ μ 체μ λν NIRvP μ§λμ μλ¬Ό κ΄ν©μ± μ¬μ΄μ κ΄κ³λ NIRvP μ§λλ₯Ό νλμ€ νμ κ΄μΈ‘λ²μμ μΌμΉμν¬ λλ§ λμ μκ΄κ΄κ³λ₯Ό 보μλ€. κ΄μΈ‘λ²μλ₯Ό μΌμΉμν¬ κ²½μ°, NIRvP μ§λλ μλ¬Ό κ΄ν©μ±μ μΆμ νλ λ° μμ΄ νμ₯ NIRvPλ³΄λ€ μ°μν μ±λ₯μ 보μλ€. μ΄λ¬ν μ±λ₯ μ°¨μ΄λ νλμ€ νμ κ΄μΈ‘λ²μλ₯Ό μΌμΉμν¬ λ, μ°κ΅¬ λμμ§ κ°μ NIRvP-μλ¬Ό κ΄ν©μ± κ΄κ³μ κΈ°μΈκΈ°κ° μΌκ΄μ±μ 보μκΈ° λλ¬Έμ΄λ€. λ³Έ μ°κ΅¬ κ²°κ³Όλ μμ± κ΄μΈ‘μ νλμ€ νμ κ΄μΈ‘λ²μμ μΌμΉμν€λ κ²μ μ€μμ±μ 보μ¬μ£Όκ³ λμ μκ³΅κ° ν΄μλλ‘ μλ¬Ό κ΄ν©μ±μ μ격μΌλ‘ λͺ¨λν°λ§νλ μ΄μνμμ±κ΅° μλ£μ μ μ¬λ ₯μ 보μ¬μ€λ€.Monitoring changes in terrestrial vegetation is essential to understanding interactions between atmosphere and biosphere, especially terrestrial ecosystem. To this end, satellite remote sensing offer maps for examining land surface in different scales. However, the detailed information was hindered under the clouds or limited by the spatial resolution of satellite imagery. Moreover, the impacts of spatial and temporal resolution in photosynthesis monitoring were not fully revealed.
In this dissertation, I aimed to enhance the spatial and temporal resolution of satellite imagery towards daily gap-free vegetation maps with high spatial resolution. In order to expand vegetation change monitoring in time and space using high-resolution satellite images, I 1) improved temporal resolution of satellite dataset through image fusion using geostationary satellites, 2) improved spatial resolution of satellite dataset using generative adversarial networks, and 3) showed the use of high spatiotemporal resolution maps for monitoring plant photosynthesis especially over heterogeneous landscapes. With the advent of new techniques in satellite remote sensing, current and past datasets can be fully utilized for monitoring vegetation changes in the respect of spatial and temporal resolution.
In Chapter 2, I developed the integrated system that implemented geostationary satellite products in the spatiotemporal image fusion method for monitoring canopy photosynthesis. The integrated system contains the series of process (i.e., cloud masking, nadir bidirectional reflectance function adjustment, spatial registration, spatiotemporal image fusion, spatial gap-filling, temporal-gap-filling). I conducted the evaluation of the integrated system over heterogeneous rice paddy landscape where the drastic land cover changes were caused by cultivation management and deciduous forest where consecutive changes occurred in time. The results showed that the integrated system well predict in situ measurements without data gaps (R2 = 0.71, relative bias = 5.64% at rice paddy site; R2 = 0.79, relative bias = -13.8% at deciduous forest site). The integrated system gradually improved the spatiotemporal resolution of vegetation maps, reducing the underestimation of in situ measurements, especially during peak growing season. Since the integrated system generates daily canopy photosynthesis maps for monitoring dynamics among regions of interest worldwide with high spatial resolution. I anticipate future efforts to reveal the hindered information by the limited spatial and temporal resolution of satellite imagery.
Detailed spatial representations of terrestrial vegetation are essential for precision agricultural applications and the monitoring of land cover changes in heterogeneous landscapes. The advent of satellite-based remote sensing has facilitated daily observations of the Earths surface with high spatial resolution. In particular, a data fusion product such as Planet Fusion has realized the delivery of daily, gap-free surface reflectance data with 3-m pixel resolution through full utilization of relatively recent (i.e., 2018-) CubeSat constellation data. However, the spatial resolution of past satellite sensors (i.e., 30β60 m for Landsat) has restricted the detailed spatial analysis of past changes in vegetation. In Chapter 3, to overcome the spatial resolution constraint of Landsat data for long-term vegetation monitoring, we propose a dual remote-sensing super-resolution generative adversarial network (dual RSS-GAN) combining Planet Fusion and Landsat 8 data to simulate spatially enhanced long-term time-series of the normalized difference vegetation index (NDVI) and near-infrared reflectance from vegetation (NIRv). We evaluated the performance of the dual RSS-GAN against in situ tower-based continuous measurements (up to 8 years) and remotely piloted aerial system-based maps of cropland and deciduous forest in the Republic of Korea. The dual RSS-GAN enhanced spatial representations in Landsat 8 images and captured seasonal variation in vegetation indices (R2 > 0.95, for the dual RSS-GAN maps vs. in situ data from all sites). Overall, the dual RSS-GAN reduced Landsat 8 vegetation index underestimations compared with in situ measurements; relative bias values of NDVI ranged from β3.2% to 1.2% and β12.4% to β3.7% for the dual RSS-GAN and Landsat 8, respectively. This improvement was caused by spatial enhancement through the dual RSS-GAN, which captured fine-scale information from Planet Fusion. This study presents a new approach for the restoration of hidden sub-pixel spatial information in Landsat images.
Mapping canopy photosynthesis in both high spatial and temporal resolution is essential for carbon cycle monitoring in heterogeneous areas. However, well established satellites in sun-synchronous orbits such as Sentinel-2, Landsat and MODIS can only provide either high spatial or high temporal resolution but not both. Recently established CubeSat satellite constellations have created an opportunity to overcome this resolution trade-off. In particular, Planet Fusion allows full utilization of the CubeSat data resolution and coverage while maintaining high radiometric quality. In Chapter 4, I used the Planet Fusion surface reflectance product to calculate daily, 3-m resolution, gap-free maps of the near-infrared radiation reflected from vegetation (NIRvP). I then evaluated the performance of these NIRvP maps for estimating canopy photosynthesis by comparing with data from a flux tower network in Sacramento-San Joaquin Delta, California, USA. Overall, NIRvP maps captured temporal variations in canopy photosynthesis of individual sites, despite changes in water extent in the wetlands and frequent mowing in the crop fields. When combining data from all sites, however, I found that robust agreement between NIRvP maps and canopy photosynthesis could only be achieved when matching NIRvP maps to the flux tower footprints. In this case of matched footprints, NIRvP maps showed considerably better performance than in situ NIRvP in estimating canopy photosynthesis both for daily sum and data around the time of satellite overpass (R2 = 0.78 vs. 0.60, for maps vs. in situ for the satellite overpass time case). This difference in performance was mostly due to the higher degree of consistency in slopes of NIRvP-canopy photosynthesis relationships across the study sites for flux tower footprint-matched maps. Our results show the importance of matching satellite observations to the flux tower footprint and demonstrate the potential of CubeSat constellation imagery to monitor canopy photosynthesis remotely at high spatio-temporal resolution.Chapter 1. Introduction 2
1. Background 2
1.1 Daily gap-free surface reflectance using geostationary satellite products 2
1.2 Monitoring past vegetation changes with high-spatial-resolution 3
1.3 High spatiotemporal resolution vegetation photosynthesis maps 4
2. Purpose of Research 4
Chapter 2. Generating daily gap-filled BRDF adjusted surface reflectance product at 10 m resolution using geostationary satellite product for monitoring daily canopy photosynthesis 6
1. Introduction 6
2. Methods 11
2.1 Study sites 11
2.2 In situ measurements 13
2.3 Satellite products 14
2.4 Integrated system 17
2.5 Canopy photosynthesis 21
2.6 Evaluation 23
3. Results and discussion 24
3.1 Comparison of STIF NDVI and NIRv with in situ NDVI and NIRv 24
3.2 Comparison of STIF NIRvP with in situ NIRvP 28
4. Conclusion 31
Chapter 3. Super-resolution of historic Landsat imagery using a dual Generative Adversarial Network (GAN) model with CubeSat constellation imagery for monitoring vegetation changes 32
1. Introduction 32
2. Methods 38
2.1 Real-ESRGAN model 38
2.2 Study sites 40
2.3 In situ measurements 42
2.4 Vegetation index 44
2.5 Satellite data 45
2.6 Planet Fusion 48
2.7 Dual RSS-GAN via fine-tuned Real-ESRGAN 49
2.8 Evaluation 54
3. Results 57
3.1 Comparison of NDVI and NIRv maps from Planet Fusion, Sentinel 2 NBAR, and Landsat 8 NBAR data with in situ NDVI and NIRv 57
3.2 Comparison of dual RSS-SRGAN model results with Landsat 8 NDVI and NIRv 60
3.3 Comparison of dual RSS-GAN model results with respect to in situ time-series NDVI and NIRv 63
3.4 Comparison of the dual RSS-GAN model with NDVI and NIRv maps derived from RPAS 66
4. Discussion 70
4.1 Monitoring changes in terrestrial vegetation using the dual RSS-GAN model 70
4.2 CubeSat data in the dual RSS-GAN model 72
4.3 Perspectives and limitations 73
5. Conclusion 78
Appendices 79
Supplementary material 82
Chapter 4. Matching high resolution satellite data and flux tower footprints improves their agreement in photosynthesis estimates 85
1. Introduction 85
2. Methods 89
2.1 Study sites 89
2.2 In situ measurements 92
2.3 Planet Fusion NIRvP 94
2.4 Flux footprint model 98
2.5 Evaluation 98
3. Results 105
3.1 Comparison of Planet Fusion NIRv and NIRvP with in situ NIRv and NIRvP 105
3.2 Comparison of instantaneous Planet Fusion NIRv and NIRvP with against tower GPP estimates 108
3.3 Daily GPP estimation from Planet Fusion -derived NIRvP 114
4. Discussion 118
4.1 Flux tower footprint matching and effects of spatial and temporal resolution on GPP estimation 118
4.2 Roles of radiation component in GPP mapping 123
4.3 Limitations and perspectives 126
5. Conclusion 133
Appendix 135
Supplementary Materials 144
Chapter 5. Conclusion 153
Bibliography 155
Abstract in Korea 199
Acknowledgements 202λ°
Deep Learning based data-fusion methods for remote sensing applications
In the last years, an increasing number of remote sensing sensors have been launched to orbit around the Earth, with a continuously growing production of massive data, that are useful for a large number of monitoring applications, especially for the monitoring task. Despite modern optical sensors provide rich spectral information about Earth's surface, at very high resolution, they are weather-sensitive. On the other hand, SAR images are always available also in presence of clouds and are almost weather-insensitive, as well as daynight available, but they do not provide a rich spectral information and are severely affected by speckle "noise" that make difficult the information extraction. For the above reasons it is worth and challenging to fuse data provided by different sources and/or acquired at different times, in order to leverage on their diversity and complementarity to retrieve the target information. Motivated by the success of the employment of Deep Learning methods in many image processing tasks, in this thesis it has been faced different typical remote sensing data-fusion problems by means of suitably designed Convolutional Neural Networks
- β¦