76 research outputs found
Deep Learning in Cardiology
The medical field is creating large amount of data that physicians are unable
to decipher and use efficiently. Moreover, rule-based expert systems are
inefficient in solving complicated medical tasks or for creating insights using
big data. Deep learning has emerged as a more accurate and effective technology
in a wide range of medical problems such as diagnosis, prediction and
intervention. Deep learning is a representation learning method that consists
of layers that transform the data non-linearly, thus, revealing hierarchical
relationships and structures. In this review we survey deep learning
application papers that use structured data, signal and imaging modalities from
cardiology. We discuss the advantages and limitations of applying deep learning
in cardiology that also apply in medicine in general, while proposing certain
directions as the most viable for clinical use.Comment: 27 pages, 2 figures, 10 table
Deep Learning for Face Anti-Spoofing: A Survey
Face anti-spoofing (FAS) has lately attracted increasing attention due to its
vital role in securing face recognition systems from presentation attacks
(PAs). As more and more realistic PAs with novel types spring up, traditional
FAS methods based on handcrafted features become unreliable due to their
limited representation capacity. With the emergence of large-scale academic
datasets in the recent decade, deep learning based FAS achieves remarkable
performance and dominates this area. However, existing reviews in this field
mainly focus on the handcrafted features, which are outdated and uninspiring
for the progress of FAS community. In this paper, to stimulate future research,
we present the first comprehensive review of recent advances in deep learning
based FAS. It covers several novel and insightful components: 1) besides
supervision with binary label (e.g., '0' for bonafide vs. '1' for PAs), we also
investigate recent methods with pixel-wise supervision (e.g., pseudo depth
map); 2) in addition to traditional intra-dataset evaluation, we collect and
analyze the latest methods specially designed for domain generalization and
open-set FAS; and 3) besides commercial RGB camera, we summarize the deep
learning applications under multi-modal (e.g., depth and infrared) or
specialized (e.g., light field and flash) sensors. We conclude this survey by
emphasizing current open issues and highlighting potential prospects.Comment: IEEE Transactions on Pattern Analysis and Machine Intelligence
(TPAMI
Explainable, Domain-Adaptive, and Federated Artificial Intelligence in Medicine
Artificial intelligence (AI) continues to transform data analysis in many
domains. Progress in each domain is driven by a growing body of annotated data,
increased computational resources, and technological innovations. In medicine,
the sensitivity of the data, the complexity of the tasks, the potentially high
stakes, and a requirement of accountability give rise to a particular set of
challenges. In this review, we focus on three key methodological approaches
that address some of the particular challenges in AI-driven medical decision
making. (1) Explainable AI aims to produce a human-interpretable justification
for each output. Such models increase confidence if the results appear
plausible and match the clinicians expectations. However, the absence of a
plausible explanation does not imply an inaccurate model. Especially in highly
non-linear, complex models that are tuned to maximize accuracy, such
interpretable representations only reflect a small portion of the
justification. (2) Domain adaptation and transfer learning enable AI models to
be trained and applied across multiple domains. For example, a classification
task based on images acquired on different acquisition hardware. (3) Federated
learning enables learning large-scale models without exposing sensitive
personal health information. Unlike centralized AI learning, where the
centralized learning machine has access to the entire training data, the
federated learning process iteratively updates models across multiple sites by
exchanging only parameter updates, not personal health data. This narrative
review covers the basic concepts, highlights relevant corner-stone and
state-of-the-art research in the field, and discusses perspectives.Comment: This paper is accepted in IEEE CAA Journal of Automatica Sinica, Nov.
10 202
Optimising complexity of CNN models for resource constrained devices: QRS detection case study
Traditional DL models are complex and resource hungry and thus, care needs to
be taken in designing Internet of (medical) things (IoT, or IoMT) applications
balancing efficiency-complexity trade-off. Recent IoT solutions tend to avoid
using deep-learning methods due to such complexities, and rather classical
filter-based methods are commonly used. We hypothesize that a shallow CNN model
can offer satisfactory level of performance in combination by leveraging other
essential solution-components, such as post-processing that is suitable for
resource constrained environment. In an IoMT application context, QRS-detection
and R-peak localisation from ECG signal as a case study, the complexities of
CNN models and post-processing were varied to identify a set of combinations
suitable for a range of target resource-limited environments. To the best of
our knowledge, finding a deploy-able configuration, by incrementally increasing
the CNN model complexity, as required to match the target's resource capacity,
and leveraging the strength of post-processing, is the first of its kind. The
results show that a shallow 2-layer CNN with a suitable post-processing can
achieve 90\% F1-score, and the scores continue to improving for 8-32 layer
CNNs, which can be used to profile target constraint environment. The outcome
shows that it is possible to design an optimal DL solution with known target
performance characteristics and resource (computing capacity, and memory)
constraints
Combining Synthesis of Cardiorespiratory Signals and Artifacts with Deep Learning for Robust Vital Sign Estimation
Healthcare has been remarkably morphing on the account of Big Data. As Machine Learning
(ML) consolidates its place in simpler clinical chores, more complex Deep Learning (DL) algorithms
have struggled to keep up, despite their superior capabilities. This is mainly attributed
to the need for large amounts of data for training, which the scientific community is unable to
satisfy.
The number of promising DL algorithms is considerable, although solutions directly targeting
the shortage of data lack. Currently, dynamical generative models are the best bet, but focus on
single, classical modalities and tend to complicate significantly with the amount of physiological
effects they can simulate.
This thesis aims at providing and validating a framework, specifically addressing the data
deficit in the scope of cardiorespiratory signals. Firstly, a multimodal statistical synthesizer was
designed to generate large, annotated artificial signals. By expressing data through coefficients of
pre-defined, fitted functions and describing their dependence with Gaussian copulas, inter- and
intra-modality associations were learned. Thereafter, new coefficients are sampled to generate
artificial, multimodal signals with the original physiological dynamics. Moreover, normal and
pathological beats along with artifacts were included by employing Markov models. Secondly,
a convolutional neural network (CNN) was conceived with a novel sensor-fusion architecture
and trained with synthesized data under real-world experimental conditions to evaluate how its
performance is affected.
Both the synthesizer and the CNN not only performed at state of the art level but also innovated
with multiple types of generated data and detection error improvements, respectively.
Cardiorespiratory data augmentation corrected performance drops when not enough data is available,
enhanced the CNN’s ability to perform on noisy signals and to carry out new tasks when
introduced to, otherwise unavailable, types of data. Ultimately, the framework was successfully
validated showing potential to leverage future DL research on Cardiology into clinical standards
Seamless Multimodal Biometrics for Continuous Personalised Wellbeing Monitoring
Artificially intelligent perception is increasingly present in the lives of
every one of us. Vehicles are no exception, (...) In the near future, pattern
recognition will have an even stronger role in vehicles, as self-driving cars
will require automated ways to understand what is happening around (and within)
them and act accordingly. (...) This doctoral work focused on advancing
in-vehicle sensing through the research of novel computer vision and pattern
recognition methodologies for both biometrics and wellbeing monitoring. The
main focus has been on electrocardiogram (ECG) biometrics, a trait well-known
for its potential for seamless driver monitoring. Major efforts were devoted to
achieving improved performance in identification and identity verification in
off-the-person scenarios, well-known for increased noise and variability. Here,
end-to-end deep learning ECG biometric solutions were proposed and important
topics were addressed such as cross-database and long-term performance,
waveform relevance through explainability, and interlead conversion. Face
biometrics, a natural complement to the ECG in seamless unconstrained
scenarios, was also studied in this work. The open challenges of masked face
recognition and interpretability in biometrics were tackled in an effort to
evolve towards algorithms that are more transparent, trustworthy, and robust to
significant occlusions. Within the topic of wellbeing monitoring, improved
solutions to multimodal emotion recognition in groups of people and
activity/violence recognition in in-vehicle scenarios were proposed. At last,
we also proposed a novel way to learn template security within end-to-end
models, dismissing additional separate encryption processes, and a
self-supervised learning approach tailored to sequential data, in order to
ensure data security and optimal performance. (...)Comment: Doctoral thesis presented and approved on the 21st of December 2022
to the University of Port
DepthFake: a depth-based strategy for detecting Deepfake videos
Fake content has grown at an incredible rate over the past few years. The
spread of social media and online platforms makes their dissemination on a
large scale increasingly accessible by malicious actors. In parallel, due to
the growing diffusion of fake image generation methods, many Deep
Learning-based detection techniques have been proposed. Most of those methods
rely on extracting salient features from RGB images to detect through a binary
classifier if the image is fake or real. In this paper, we proposed DepthFake,
a study on how to improve classical RGB-based approaches with depth-maps. The
depth information is extracted from RGB images with recent monocular depth
estimation techniques. Here, we demonstrate the effective contribution of
depth-maps to the deepfake detection task on robust pre-trained architectures.
The proposed RGBD approach is in fact able to achieve an average improvement of
3.20% and up to 11.7% for some deepfake attacks with respect to standard RGB
architectures over the FaceForensic++ dataset.Comment: 2022 ICPR Workshop on Artificial Intelligence for Multimedia
Forensics and Disinformation Detectio
- …