133 research outputs found
Radiometrically-Accurate Hyperspectral Data Sharpening
Improving the spatial resolution of hyperpsectral image (HSI) has traditionally been an important topic in the field of remote sensing. Many approaches have been proposed based on various theories including component substitution, multiresolution analysis, spectral unmixing, Bayesian probability, and tensor representation. However, these methods have some common disadvantages, such as that they are not robust to different up-scale ratios and they have little concern for the per-pixel radiometric accuracy of the sharpened image. Moreover, many learning-based methods have been proposed through decades of innovations, but most of them require a large set of training pairs, which is unpractical for many real problems. To solve these problems, we firstly proposed an unsupervised Laplacian Pyramid Fusion Network (LPFNet) to generate a radiometrically-accurate high-resolution HSI. First, with the low-resolution hyperspectral image (LR-HSI) and the high-resolution multispectral image (HR-MSI), the preliminary high-resolution hyperspectral image (HR-HSI) is calculated via linear regression. Next, the high-frequency details of the preliminary HR-HSI are estimated via the subtraction between it and the CNN-generated-blurry version. By injecting the details to the output of the generative CNN with the low-resolution hyperspectral image (LR-HSI) as input, the final HR-HSI is obtained. LPFNet is designed for fusing the LR-HSI and HR-MSI covers the same Visible-Near-Infrared (VNIR) bands, while the short-wave infrared (SWIR) bands of HSI are ignored. SWIR bands are equally important to VNIR bands, but their spatial details are more challenging to be enhanced because the HR-MSI, used to provide the spatial details in the fusion process, usually has no SWIR coverage or lower-spatial-resolution SWIR. To this end, we designed an unsupervised cascade fusion network (UCFNet) to sharpen the Vis-NIR-SWIR LR-HSI. First, the preliminary high-resolution VNIR hyperspectral image (HR-VNIR-HSI) is obtained with a conventional hyperspectral algorithm. Then, the HR-MSI, the preliminary HR-VNIR-HSI, and the LR-SWIR-HSI are passed to the generative convolutional neural network to produce an HR-HSI. In the training process, the cascade sharpening method is employed to improve stability. Furthermore, the self-supervising loss is introduced based on the cascade strategy to further improve the spectral accuracy. Experiments are conducted on both LPFNet and UCFNet with different datasets and up-scale ratios. Also, state-of-the-art baseline methods are implemented and compared with the proposed methods with different quantitative metrics. Results demonstrate that proposed methods outperform the competitors in all cases in terms of spectral and spatial accuracy
Fast Hierarchical Deep Unfolding Network for Image Compressed Sensing
By integrating certain optimization solvers with deep neural network, deep
unfolding network (DUN) has attracted much attention in recent years for image
compressed sensing (CS). However, there still exist several issues in existing
DUNs: 1) For each iteration, a simple stacked convolutional network is usually
adopted, which apparently limits the expressiveness of these models. 2) Once
the training is completed, most hyperparameters of existing DUNs are fixed for
any input content, which significantly weakens their adaptability. In this
paper, by unfolding the Fast Iterative Shrinkage-Thresholding Algorithm
(FISTA), a novel fast hierarchical DUN, dubbed FHDUN, is proposed for image
compressed sensing, in which a well-designed hierarchical unfolding
architecture is developed to cooperatively explore richer contextual prior
information in multi-scale spaces. To further enhance the adaptability, series
of hyperparametric generation networks are developed in our framework to
dynamically produce the corresponding optimal hyperparameters according to the
input content. Furthermore, due to the accelerated policy in FISTA, the newly
embedded acceleration module makes the proposed FHDUN save more than 50% of the
iterative loops against recent DUNs. Extensive CS experiments manifest that the
proposed FHDUN outperforms existing state-of-the-art CS methods, while
maintaining fewer iterations.Comment: Accepted by ACM MM 202
Dual-Octave Convolution for Accelerated Parallel MR Image Reconstruction
Magnetic resonance (MR) image acquisition is an inherently prolonged process,
whose acceleration by obtaining multiple undersampled images simultaneously
through parallel imaging has always been the subject of research. In this
paper, we propose the Dual-Octave Convolution (Dual-OctConv), which is capable
of learning multi-scale spatial-frequency features from both real and imaginary
components, for fast parallel MR image reconstruction. By reformulating the
complex operations using octave convolutions, our model shows a strong ability
to capture richer representations of MR images, while at the same time greatly
reducing the spatial redundancy. More specifically, the input feature maps and
convolutional kernels are first split into two components (i.e., real and
imaginary), which are then divided into four groups according to their spatial
frequencies. Then, our Dual-OctConv conducts intra-group information updating
and inter-group information exchange to aggregate the contextual information
across different groups. Our framework provides two appealing benefits: (i) it
encourages interactions between real and imaginary components at various
spatial frequencies to achieve richer representational capacity, and (ii) it
enlarges the receptive field by learning multiple spatial-frequency features of
both the real and imaginary components. We evaluate the performance of the
proposed model on the acceleration of multi-coil MR image reconstruction.
Extensive experiments are conducted on an {in vivo} knee dataset under
different undersampling patterns and acceleration factors. The experimental
results demonstrate the superiority of our model in accelerated parallel MR
image reconstruction. Our code is available at:
github.com/chunmeifeng/Dual-OctConv.Comment: Proceedings of the 35th AAAI Conference on Artificial Intelligence
(AAAI) 202
Unsupervised Learning from Shollow to Deep
Machine learning plays a pivotal role in most state-of-the-art systems in many application research domains. With the rising of deep learning, massive labeled data become the solution of feature learning, which enables the model to learn automatically. Unfortunately, the trained deep learning model is hard to adapt to other datasets without fine-tuning, and the applicability of machine learning methods is limited by the amount of available labeled data. Therefore, the aim of this thesis is to alleviate the limitations of supervised learning by exploring algorithms to learn good internal representations, and invariant feature hierarchies from unlabelled data.
Firstly, we extend the traditional dictionary learning and sparse coding algorithms onto hierarchical image representations in a principled way. To achieve dictionary atoms capture additional information from extended receptive fields and attain improved descriptive capacity, we present a two-pass multi-resolution cascade framework for dictionary learning and sparse coding. This cascade method allows collaborative reconstructions at different resolutions using only the same dimensional dictionary atoms. The jointly learned dictionary comprises atoms that adapt to the information available at the coarsest layer, where the support of atoms reaches a maximum range, and the residual images, where the supplementary details refine progressively a reconstruction objective. Our method generates flexible and accurate representations using only a small number of coefficients, and is efficient in computation.
In the following work, we propose to incorporate the traditional self-expressiveness property into deep learning to explore better representation for subspace clustering. This architecture is built upon deep auto-encoders, which non-linearly map the input data into a latent space. Our key idea is to introduce a novel self-expressive layer between the encoder and the decoder to mimic the ``self-expressiveness'' property that has proven effective in traditional subspace clustering. Being differentiable, our new self-expressive layer provides a simple but effective way to learn pairwise affinities between all data points through a standard back-propagation procedure. Being nonlinear, our neural-network based method is able to cluster data points having complex (often nonlinear) structures.
However, Subspace clustering algorithms are notorious for their scalability issues because building and processing large affinity matrices are demanding. We propose two methods to tackle this problem. One method is based on -Subspace Clustering, where we introduce a method that simultaneously learns an embedding space along subspaces within it to minimize a notion of reconstruction error, thus addressing the problem of subspace clustering in an end-to-end learning paradigm. This in turn frees us from the need of having an affinity matrix to perform clustering. The other way starts from using a feed forward network to replace the spectral clustering and learn the affinities of each data from "self-expressive" layer. We introduce the Neural Collaborative Subspace Clustering, where it benefits from a classifier which determines whether a pair of points lies on the same subspace under supervision of "self-expressive" layer. Essential to our model is the construction of two affinity matrices, one from the classifier and the other from a notion of subspace self-expressiveness, to supervise training in a collaborative scheme.
In summary, we make constributions on how to perform the unsupervised learning in several tasks in this thesis. It starts from traditional sparse coding and dictionary learning perspective in low-level vision. Then, we exploit how to incorporate unsupervised learning in convolutional neural networks without label information and make subspace clustering to large scale dataset. Furthermore, we also extend the clustering on dense prediction task (saliency detection)
Recent Advances in Image Restoration with Applications to Real World Problems
In the past few decades, imaging hardware has improved tremendously in terms of resolution, making widespread usage of images in many diverse applications on Earth and planetary missions. However, practical issues associated with image acquisition are still affecting image quality. Some of these issues such as blurring, measurement noise, mosaicing artifacts, low spatial or spectral resolution, etc. can seriously affect the accuracy of the aforementioned applications. This book intends to provide the reader with a glimpse of the latest developments and recent advances in image restoration, which includes image super-resolution, image fusion to enhance spatial, spectral resolution, and temporal resolutions, and the generation of synthetic images using deep learning techniques. Some practical applications are also included
Impact of GAN-based Lesion-Focused Medical Image Super-Resolution on Radiomic Feature Robustness
Dissertation presented as the partial requirement for obtaining a Master's degree in Data Science and Advanced Analytics, specialization in Data ScienceRobust machine learning models based on radiomic features might allow for accurate diagnosis, prognosis, and medical decision-making. Unfortunately, the lack of
standardized radiomic feature extraction has hampered their clinical use. Since the
radiomic features tend to be affected by low voxel statistics in regions of interest, increasing the sample size would improve their robustness in clinical studies. Therefore,
we propose a Generative Adversarial Network (GAN)-based lesion-focused framework
for Computed Tomography (CT) image Super-Resolution (SR); for the lesion (i.e., cancer) patch-focused training, we incorporate Spatial Pyramid Pooling (SPP) into GANConstrained by the Identical, Residual, and Cycle Learning Ensemble (GAN-CIRCLE).
At 2× SR, the proposed model achieved better perceptual quality with less blurring
than the other considered state-of-the-art SR methods, while producing comparable
results at 4× SR. We also evaluated the robustness of our model’s radiomic feature in
terms of quantization on a different lung cancer CT dataset using Principal Component Analysis (PCA). Intriguingly, the most important radiomic features in our PCAbased analysis were the most robust features extracted on the GAN-super-resolved
images. These achievements pave the way for the application of GAN-based image
Super-Resolution techniques for studies of radiomics for robust biomarker discoveryModelos de machine learning robustos baseados em atributos radiômicos possibilitam
diagnósticos e decisões médicas mais precisas. Infelizmente, por causa da falta de padronização na extração de atributos radiômicos, sua utilização em contextos clÃnicos
tem sido restrita. Considerando que atributos radiômics tendem a ser afetados pelas estatÃtiscas de voxels de baixo volume nas regiões de interesse, o aumento to tamanho da
amostra tem o potencial de melhorar a robustez desses atributos em estudos clÃnicos.
Portanto, esse trabalho propões um framework baseado numa rede neural generativa
(GAN) focada na região de interesse para a super-resolução de imagens de Tomografia
Computadorizada (CT). Para treinar a rede de forma concentrada na lesão (i.e. cancer),
incorporamos a tecnica de Spatial Pyramid Pooling no framework da GAN-CIRCLE.
Nos experimentos de super-resolução 2×, o modelo proposto alcançou melhor qualidade perceptual com menos embaçamento do que outros métodos estado-da-arte
considerados. A robustez dos atributos radiômics das imagens super-resolvidas geradas pelo modelo também foram analizadas em termos de quantização em um banco
de imagens diferente, contendo imagens de tomografia computadorizada de câncer
de pulmão, usando anaálise de componentes principaiss (PCA). Intrigantemente, os
atributos radiômicos mais importantes nessa análise foram também os atributos mais
robustos extraÃdos das imagens super-resolvidas pelo método proposto. Esses resultados abrem caminho para a aplicação de técnicas de super-resolução baseadas em redes
neurais generativas aplicadas a estudos de radômica para a descoberta de biomarcadores robustos.This work was partially supported by The Mark Foundation for Cancer Research and Cancer Research UK Cambridge Centre [C9685/A25177], the Wellcome Trust Innovator Award, UK [215733/Z/19/Z] and the CRUK National Cancer Imaging Translational Accelerator (NCITA) [C42780/A27066]. Additional support was also provided by the National Institute of Health Research (NIHR) Cambridge Biomedical Research Centre (BRC-1215-20014). This works was also finantially supported by national funds through the FCT (Fundação para a Ciência e a Tecnologia) by the projects GADgET (DSAIPA/DS/0022/2018) and the Slovenian Research Agency (research core funding no. P5-0410).This work was partially supported by The Mark Foundation for Cancer Research and Cancer Research UK Cambridge Centre [C9685/A25177], the Wellcome Trust Innovator Award, UK [215733/Z/19/Z] and the CRUK National Cancer Imaging Translational Accelerator (NCITA) [C42780/A27066]. Additional support was also provided by the National Institute of Health Research (NIHR) Cambridge Biomedical Research Centre (BRC-1215-20014). This works was also finantially supported by national funds through the FCT (Fundação para a Ciência e a Tecnologia) by the projects GADgET (DSAIPA/DS/0022/2018) and the Slovenian Research Agency (research core funding no. P5-0410)
A review on deep learning techniques for 3D sensed data classification
Over the past decade deep learning has driven progress in 2D image
understanding. Despite these advancements, techniques for automatic 3D sensed
data understanding, such as point clouds, is comparatively immature. However,
with a range of important applications from indoor robotics navigation to
national scale remote sensing there is a high demand for algorithms that can
learn to automatically understand and classify 3D sensed data. In this paper we
review the current state-of-the-art deep learning architectures for processing
unstructured Euclidean data. We begin by addressing the background concepts and
traditional methodologies. We review the current main approaches including;
RGB-D, multi-view, volumetric and fully end-to-end architecture designs.
Datasets for each category are documented and explained. Finally, we give a
detailed discussion about the future of deep learning for 3D sensed data, using
literature to justify the areas where future research would be most valuable.Comment: 25 pages, 9 figures. Review pape
- …