22 research outputs found
Rethinking PRL: A Multiscale Progressively Residual Learning Network for Inverse Halftoning
Image inverse halftoning is a classic image restoration task, aiming to
recover continuous-tone images from halftone images with only bilevel pixels.
Because the halftone images lose much of the original image content, inverse
halftoning is a classic ill-problem. Although existing inverse halftoning
algorithms achieve good performance, their results lose image details and
features. Therefore, it is still a challenge to recover high-quality
continuous-tone images. In this paper, we propose an end-to-end multiscale
progressively residual learning network (MSPRL), which has a UNet architecture
and takes multiscale input images. To make full use of different input image
information, we design a shallow feature extraction module to capture similar
features between images of different scales. We systematically study the
performance of different methods and compare them with our proposed method. In
addition, we employ different training strategies to optimize the model, which
is important for optimizing the training process and improving performance.
Extensive experiments demonstrate that our MSPRL model obtains considerable
performance gains in detail restoration
Layer Decomposition Learning Based on Gaussian Convolution Model and Residual Deblurring for Inverse Halftoning
Layer decomposition to separate an input image into base and detail layers
has been steadily used for image restoration. Existing residual networks based
on an additive model require residual layers with a small output range for fast
convergence and visual quality improvement. However, in inverse halftoning,
homogenous dot patterns hinder a small output range from the residual layers.
Therefore, a new layer decomposition network based on the Gaussian convolution
model (GCM) and structure-aware deblurring strategy is presented to achieve
residual learning for both the base and detail layers. For the base layer, a
new GCM-based residual subnetwork is presented. The GCM utilizes a statistical
distribution, in which the image difference between a blurred continuous-tone
image and a blurred halftoned image with a Gaussian filter can result in a
narrow output range. Subsequently, the GCM-based residual subnetwork uses a
Gaussian-filtered halftoned image as input and outputs the image difference as
residual, thereby generating the base layer, i.e., the Gaussian-blurred
continuous-tone image. For the detail layer, a new structure-aware residual
deblurring subnetwork (SARDS) is presented. To remove the Gaussian blurring of
the base layer, the SARDS uses the predicted base layer as input and outputs
the deblurred version. To more effectively restore image structures such as
lines and texts, a new image structure map predictor is incorporated into the
deblurring network to induce structure-adaptive learning. This paper provides a
method to realize the residual learning of both the base and detail layers
based on the GCM and SARDS. In addition, it is verified that the proposed
method surpasses state-of-the-art methods based on U-Net, direct deblurring
networks, and progressively residual networks
Robust Computer Vision Against Adversarial Examples and Domain Shifts
Recent advances in deep learning have achieved remarkable success in various computer vision problems. Driven by progressive computing resources and a vast amount of data, deep learning technology is reshaping human life. However, Deep Neural Networks (DNNs) have been shown vulnerable to adversarial examples, in which carefully crafted perturbations can easily fool DNNs into making wrong predictions. On the other hand, DNNs have poor generalization to domain shifts, as they suffer from performance degradation when encountering data from new visual distributions. We view these issues from the perspective of robustness. More precisely, existing deep learning technology is not reliable enough for many scenarios, where adversarial examples and domain shifts are among the most critical. The lack of reliability inevitably limits DNNs from being deployed in more important computer vision applications, such as self-driving vehicles and medical instruments that have major safety concerns.
To overcome these challenges, we focus on investigating and addressing the robustness of deep learning-based computer vision approaches. The first part of this thesis attempts to robustify computer vision models against adversarial examples. We dive into such adversarial robustness from four aspects: novel attacks for strengthening benchmarks, empirical defenses validated by a third-party evaluator, generalizable defenses that can defend against multiple and unforeseen attacks, and defenses specifically designed for less explored tasks. The second part of this thesis improves the robustness against domain shifts via domain adaptation. We dive into two important domain adaptation settings: unsupervised domain adaptation, which is the most common, and source-free domain adaptation, which is more practical in real-world scenarios. The last part explores the intersection of adversarial robustness and domain adaptation fields to provide new insights for robust DNNs. We study two directions: adversarial defense for domain adaptation and adversarial defense via domain adaptations. This dissertation aims at more robust, reliable, and trustworthy computer vision
Real-time Ultrasound Signals Processing: Denoising and Super-resolution
Ultrasound acquisition is widespread in the biomedical field, due to its properties of low cost, portability, and non-invasiveness for the patient. The processing and analysis of US signals, such as images, 2D videos, and volumetric images, allows the physician to monitor the evolution of the patient's disease, and support diagnosis, and treatments (e.g., surgery). US images are affected by speckle noise, generated by the overlap of US waves. Furthermore, low-resolution images are acquired when a high acquisition frequency is applied to accurately characterise the behaviour of anatomical features that quickly change over time. Denoising and super-resolution of US signals are relevant to improve the visual evaluation of the physician and the performance and accuracy of processing methods, such as segmentation and classification. The main requirements for the processing and analysis of US signals are real-time execution, preservation of anatomical features, and reduction of artefacts. In this context, we present a novel framework for the real-time denoising of US 2D images based on deep learning and high-performance computing, which reduces noise while preserving anatomical features in real-time execution. We extend our framework to the denoise of arbitrary US signals, such as 2D videos and 3D images, and we apply denoising algorithms that account for spatio-temporal signal properties into an image-to-image deep learning model. As a building block of this framework, we propose a novel denoising method belonging to the class of low-rank approximations, which learns and predicts the optimal thresholds of the Singular Value Decomposition. While previous denoise work compromises the computational cost and effectiveness of the method, the proposed framework achieves the results of the best denoising algorithms in terms of noise removal, anatomical feature preservation, and geometric and texture properties conservation, in a real-time execution that respects industrial constraints. The framework reduces the artefacts (e.g., blurring) and preserves the spatio-temporal consistency among frames/slices; also, it is general to the denoising algorithm, anatomical district, and noise intensity. Then, we introduce a novel framework for the real-time reconstruction of the non-acquired scan lines through an interpolating method; a deep learning model improves the results of the interpolation to match the target image (i.e., the high-resolution image). We improve the accuracy of the prediction of the reconstructed lines through the design of the network architecture and the loss function. %The design of the deep learning architecture and the loss function allow the network to improve the accuracy of the prediction of the reconstructed lines. In the context of signal approximation, we introduce our kernel-based sampling method for the reconstruction of 2D and 3D signals defined on regular and irregular grids, with an application to US 2D and 3D images. Our method improves previous work in terms of sampling quality, approximation accuracy, and geometry reconstruction with a slightly higher computational cost. For both denoising and super-resolution, we evaluate the compliance with the real-time requirement of US applications in the medical domain and provide a quantitative evaluation of denoising and super-resolution methods on US and synthetic images. Finally, we discuss the role of denoising and super-resolution as pre-processing steps for segmentation and predictive analysis of breast pathologies
Application and Theory of Multimedia Signal Processing Using Machine Learning or Advanced Methods
This Special Issue is a book composed by collecting documents published through peer review on the research of various advanced technologies related to applications and theories of signal processing for multimedia systems using ML or advanced methods. Multimedia signals include image, video, audio, character recognition and optimization of communication channels for networks. The specific contents included in this book are data hiding, encryption, object detection, image classification, and character recognition. Academics and colleagues who are interested in these topics will find it interesting to read
Project Tech Top study of lunar, planetary and solar topography Final report
Data acquisition techniques for information on lunar, planetary, and solar topograph
Fine spatial scale modelling of Trentino past forest landscape and future change scenarios to study ecosystem services through the years
Ciolli, MarcoCantiani, Maria Giulia1openLandscape in Europe has dramatically changed in the last decades. This has been especially
true for Alpine regions, where the progressive urbanization of the valleys has been accom-
panied by the abandonment of smaller villages and areas at higher elevation. This trend
has been clearly observable in the Provincia Autonoma di Trento (PAT) region in the Italian
Alps. The impact has been substantial for many rural areas, with the progressive shrinking
of meadows and pastures due to the forest natural recolonization. These modifications of the
landscape affect biodiversity, social and cultural dynamics, including landscape perception
and some ecosystem services. Literature review showed that this topic has been addressed
by several authors across the Alps, but their researches are limited in space coverage, spatial
resolution and time span. This thesis aims to create a comprehensive dataset of historical
maps and multitemporal orthophotos in the area of PAT to perform data analysis to identify
the changes in forest and open areas, being an evaluation of how these changes affected land-
scape structure and ecosystems, create a future change scenario for a test area and highlight
some major changes in ecosystem services through time.
In this study a high resolution dataset of maps covering the whole PAT area for over
a century was developed. The earlier representation of the PAT territory which contained
reliable data about forest coverage was considered is the Historic Cadastral maps of the 1859.
These maps in fact systematically and accurately represented the land use of each parcel in
the Habsburg Empire, included the PAT. Then, the Italian Kingdom Forest Maps, was the
next important source of information about the forest coverage after World War I, before
coming to the most recent datasets of the greyscale images of 1954, 1994 and the multiband
images of 2006 and 2015.
The purpose of the dataset development is twofold: to create a series of maps describing
the forest and open areas coverage in the last 160 years for the whole PAT on one hand and
to setup and test procedures to extract the relevant information from imagery and historical
maps on the other. The datasets were archived, processed and analysed using the Free and
Open Source Software (FOSS) GIS GRASS, QGIS and R.
The goal set by this work was achieved by a remote sensed analysis of said maps and
aerial imagery. A series of procedures were applied to extract a land use map, with the forest
categories reaching a level of detail rarely achieved for a study area of such an extension
(6200 km2
). The resolution of the original maps is in fact at a meter level, whereas the coarser
resampling adopted is 10mx10m pixels.
The great variety and size of the input data required the development, along the main part
of the research, of a series of new tools for automatizing the analysis of the aerial imagery,
to reduce the user intervention. New tools for historic map classification were as well developed, for eliminating from the resulting maps of land use from symbols (e.g.: signs), thus
enhancing the results.
Once the multitemporal forest maps were obtained, the second phase of the current work
was a qualitative and quantitative assessment of the forest coverage and how it changed.
This was performed by the evaluation of a number of landscape metrics, indexes used to
quantify the compaction or the rarefaction of the forest areas.
A recurring issue in the current Literature on the topic of landscape metrics was identified
along their analysis in the current work, that was extensively studied. This highlighted the
importance of specifying some parameters in the most used landscape fragmentation analy-
sis software to make the results of different studies properly comparable.
Within this analysis a set of data coming from other maps were used to characterize the process of afforestation in PAT, such as the potential forest maps, which were used to quantify
the area of potential forest which were actually afforested through the years, the Digital Ele-
vation Model, which was used to quantify the changes in forest area at a different ranges of
altitude, and finally the forest class map, which was used to estimate how afforestation has
affected each single forest type.
The output forest maps were used to analyse and estimate some ecosystem services, in par-
ticular the protection from soil erosion, the changes in biodiversity and the landscape of the
forests.
Finally, a procedure for the analysis of future changes scenarios was set up to study how
afforestation will proceed in absence of external factors in a protected area of PAT. The pro-
cedure was developed using Agent Based Models, which considers trees as thinking agents,
able to choose where to expand the forest area.
The first part of the results achieved consists in a temporal series of maps representing the
situation of the forest in each year of the considered dataset. The analysis of these maps
suggests a trend of afforestation across the PAT territory. The forest maps were then reclassi-
fied by altitude ranges and forest types to show how the afforestation proceeded at different
altitudes and forest types. The results showed that forest expansion acted homogeneously
through different altitude and forest types. The analysis of a selected set of landscape met-
rics showed a progressive compaction of the forests at the expenses of the open areas, in each
altitude range and for each forest type. This generated on one hand a benefit for all those
ecosystem services linked to a high forest cover, while reduced ecotonal habitats and affected
biodiversity distribution and quality. Finally the ABM procedure resulted in a set of maps
representing a possible evolution of the forest in an area of PAT, which represented a similar
situation respect to other simulations developed using different models in the same area. A
second part of the result achieved in the current work consisted in new open source tools
for image analysis developed for achieving the results showed, but with a potentially wider
field of application, along with new procedure for the evaluation of the image classification.
The current work fulfilled its aims, while providing in the meantime new tools and enhance-
ment of existing tools for remote sensing and leaving as heritage a large dataset that will be
used to deepen he knowledge of the territory of PAT, and, more widely to study emerging
pattern in afforestation in an alpine environment.openGobbi, S