130 research outputs found
Advancing Biomedicine with Graph Representation Learning: Recent Progress, Challenges, and Future Directions
Graph representation learning (GRL) has emerged as a pivotal field that has
contributed significantly to breakthroughs in various fields, including
biomedicine. The objective of this survey is to review the latest advancements
in GRL methods and their applications in the biomedical field. We also
highlight key challenges currently faced by GRL and outline potential
directions for future research.Comment: Accepted by 2023 IMIA Yearbook of Medical Informatic
Uncertainty-Aware Decision Transformer for Stochastic Driving Environments
Offline Reinforcement Learning (RL) has emerged as a promising framework for
learning policies without active interactions, making it especially appealing
for autonomous driving tasks. Recent successes of Transformers inspire casting
offline RL as sequence modeling, which performs well in long-horizon tasks.
However, they are overly optimistic in stochastic environments with incorrect
assumptions that the same goal can be consistently achieved by identical
actions. In this paper, we introduce an UNcertainty-awaRE deciSion Transformer
(UNREST) for planning in stochastic driving environments without introducing
additional transition or complex generative models. Specifically, UNREST
estimates state uncertainties by the conditional mutual information between
transitions and returns, and segments sequences accordingly. Discovering the
`uncertainty accumulation' and `temporal locality' properties of driving
environments, UNREST replaces the global returns in decision transformers with
less uncertain truncated returns, to learn from true outcomes of agent actions
rather than environment transitions. We also dynamically evaluate environmental
uncertainty during inference for cautious planning. Extensive experimental
results demonstrate UNREST's superior performance in various driving scenarios
and the power of our uncertainty estimation strategy
Boosting Offline Reinforcement Learning for Autonomous Driving with Hierarchical Latent Skills
Learning-based vehicle planning is receiving increasing attention with the
emergence of diverse driving simulators and large-scale driving datasets. While
offline reinforcement learning (RL) is well suited for these safety-critical
tasks, it still struggles to plan over extended periods. In this work, we
present a skill-based framework that enhances offline RL to overcome the
long-horizon vehicle planning challenge. Specifically, we design a variational
autoencoder (VAE) to learn skills from offline demonstrations. To mitigate
posterior collapse of common VAEs, we introduce a two-branch sequence encoder
to capture both discrete options and continuous variations of the complex
driving skills. The final policy treats learned skills as actions and can be
trained by any off-the-shelf offline RL algorithms. This facilitates a shift in
focus from per-step actions to temporally extended skills, thereby enabling
long-term reasoning into the future. Extensive results on CARLA prove that our
model consistently outperforms strong baselines at both training and new
scenarios. Additional visualizations and experiments demonstrate the
interpretability and transferability of extracted skills
Boosting Operational DNN Testing Efficiency through Conditioning
With the increasing adoption of Deep Neural Network (DNN) models as integral
parts of software systems, efficient operational testing of DNNs is much in
demand to ensure these models' actual performance in field conditions. A
challenge is that the testing often needs to produce precise results with a
very limited budget for labeling data collected in field.
Viewing software testing as a practice of reliability estimation through
statistical sampling, we re-interpret the idea behind conventional structural
coverages as conditioning for variance reduction. With this insight we propose
an efficient DNN testing method based on the conditioning on the representation
learned by the DNN model under testing. The representation is defined by the
probability distribution of the output of neurons in the last hidden layer of
the model. To sample from this high dimensional distribution in which the
operational data are sparsely distributed, we design an algorithm leveraging
cross entropy minimization.
Experiments with various DNN models and datasets were conducted to evaluate
the general efficiency of the approach. The results show that, compared with
simple random sampling, this approach requires only about a half of labeled
inputs to achieve the same level of precision.Comment: Published in the Proceedings of the 27th ACM Joint European Software
Engineering Conference and Symposium on the Foundations of Software
Engineering (ESEC/FSE 2019
Learning with Logical Constraints but without Shortcut Satisfaction
Recent studies in neuro-symbolic learning have explored the integration of
logical knowledge into deep learning via encoding logical constraints as an
additional loss function. However, existing approaches tend to vacuously
satisfy logical constraints through shortcuts, failing to fully exploit the
knowledge. In this paper, we present a new framework for learning with logical
constraints. Specifically, we address the shortcut satisfaction issue by
introducing dual variables for logical connectives, encoding how the constraint
is satisfied. We further propose a variational framework where the encoded
logical constraint is expressed as a distributional loss that is compatible
with the model's original training loss. The theoretical analysis shows that
the proposed approach bears salient properties, and the experimental
evaluations demonstrate its superior performance in both model generalizability
and constraint satisfaction.Comment: Published as a conference paper at ICLR 2023, and code is available
at https://github.com/SoftWiser-group/NeSy-without-Shortcut
From moments of the distribution function to hydrodynamics the non-conformal case
We study the one-dimensional boost-invariant Boltzmann equation in the
relaxation-time approximation using special moments of the distribution
function for a system with a finite particle mass. The infinite hierarchy of
moments can be truncated by keeping only the three lowest moments that
correspond to the three independent components of the energy-momentum tensor.
We show that such a three-moment truncation reproduces accurately the exact
solution of the kinetic equation after a simple renormalization that takes into
account the effects of the neglected higher moments. We derive second-order
Israel-Stewart hydrodynamic equations from the three-moment equations, and show
that, for most physically relevant initial conditions, these equations yield
results comparable to those of the three-moment truncation, albeit less
accurate. We attribute this feature to the fact that the structure of
Israel-Stewart equations is similar to that of the three-moment truncation. In
particular, the presence of the relaxation term in the Israel-Stewart
equations, yields an early-time regime that mimics approximately the
collisionless regime. A detailed comparison of the three-moment truncation with
second-order non-conformal hydrodynamics reveals ambiguities in the definition
of second-order transport coefficients. These ambiguities affect the ability of
Israel-Stewart hydrodynamics to reproduce results of kinetic theory.Comment: 44 pages, 11 figure
Softened Symbol Grounding for Neuro-symbolic Systems
Neuro-symbolic learning generally consists of two separated worlds, i.e.,
neural network training and symbolic constraint solving, whose success hinges
on symbol grounding, a fundamental problem in AI. This paper presents a novel,
softened symbol grounding process, bridging the gap between the two worlds, and
resulting in an effective and efficient neuro-symbolic learning framework.
Technically, the framework features (1) modeling of symbol solution states as a
Boltzmann distribution, which avoids expensive state searching and facilitates
mutually beneficial interactions between network training and symbolic
reasoning;(2) a new MCMC technique leveraging projection and SMT solvers, which
efficiently samples from disconnected symbol solution spaces; (3) an annealing
mechanism that can escape from %being trapped into sub-optimal symbol
groundings. Experiments with three representative neuro symbolic learning tasks
demonstrate that, owining to its superior symbol grounding capability, our
framework successfully solves problems well beyond the frontier of the existing
proposals.Comment: Published as a conference paper at ICLR 2023. Code is available at
https://github.com/SoftWiser-group/Soften-NeSy-learnin
Is Underwater Image Enhancement All Object Detectors Need?
Underwater object detection is a crucial and challenging problem in marine
engineering and aquatic robot. The difficulty is partly because of the
degradation of underwater images caused by light selective absorption and
scattering. Intuitively, enhancing underwater images can benefit high-level
applications like underwater object detection. However, it is still unclear
whether all object detectors need underwater image enhancement as
pre-processing. We therefore pose the questions "Does underwater image
enhancement really improve underwater object detection?" and "How does
underwater image enhancement contribute to underwater object detection?". With
these two questions, we conduct extensive studies. Specifically, we use 18
state-of-the-art underwater image enhancement algorithms, covering traditional,
CNN-based, and GAN-based algorithms, to pre-process underwater object detection
data. Then, we retrain 7 popular deep learning-based object detectors using the
corresponding results enhanced by different algorithms, obtaining 126
underwater object detection models. Coupled with 7 object detection models
retrained using raw underwater images, we employ these 133 models to
comprehensively analyze the effect of underwater image enhancement on
underwater object detection. We expect this study can provide sufficient
exploration to answer the aforementioned questions and draw more attention of
the community to the joint problem of underwater image enhancement and
underwater object detection. The pre-trained models and results are publicly
available and will be regularly updated. Project page:
https://github.com/BIGWangYuDong/lqit/tree/main/configs/detection/uw_enhancement_affect_detection.Comment: 17 pages, 9 figure
Pretreatment plasma fibrinogen level as a prognostic biomarker for patients with lung cancer
Many researchers have shown that pretreatment plasma fibrinogen levels are closely correlated with the prognosis of patients with lung cancer (LC). In this study, we thus performed a meta-analysis to systematically assess the prognostic value of pretreatment plasma fibrinogen levels in LC patients. A computerized systematic search in PubMed, EMBASE, Web of Science and China National Knowledge Infrastructure (CNKI) was performed up to March 15, 2018. Studies with available data on the prognostic value of plasma fibrinogen in LC patients were eligible for inclusion. The pooled hazard ratios (HRs) and odd ratios (ORs) with 95% confidence intervals (CIs) were used to evaluate the correlation between pretreatment plasma fibrinogen levels and prognosis as well as clinicopathological characteristics. A total of 17 studies with 6,460 LC patients were included in this meta-analysis. A higher pretreatment plasma fibrinogen level was significantly associated with worse overall survival (OS) (HR: 1.57; 95% CI: 1.39-1.77; p=0.001), disease-free survival (DFS) (HR: 1.53; 95% CI: 1.33-1.76; p=0.003), and progression-free survival (PFS) (HR: 3.14; 95% CI: 2.15-4.59; po0.001). Furthermore, our subgroup and sensitivity analyses demonstrated that the pooled HR for OS was robust and reliable. In addition, we also found that a higher fibrinogen level predicted advanced TNM stage (III-IV) (OR=2.18, 95% CI: 1.79-2.66; po0.001) and a higher incidence of lymph node metastasis (OR=1.74, 95% CI: 1.44-2.10; p=0.02). Our study suggested that higher pretreatment plasma fibrinogen levels predict worse prognoses in LC patients
Explainable Graph Neural Network for Alzheimer's Disease And Related Dementias Risk Prediction
Alzheimer's disease and related dementias (ADRD) ranks as the sixth leading
cause of death in the US, underlining the importance of accurate ADRD risk
prediction. While recent advancement in ADRD risk prediction have primarily
relied on imaging analysis, yet not all patients undergo medical imaging before
an ADRD diagnosis. Merging machine learning with claims data can reveal
additional risk factors and uncover interconnections among diverse medical
codes. Our goal is to utilize Graph Neural Networks (GNNs) with claims data for
ADRD risk prediction. Addressing the lack of human-interpretable reasons behind
these predictions, we introduce an innovative method to evaluate relationship
importance and its influence on ADRD risk prediction, ensuring comprehensive
interpretation.
We employed Variationally Regularized Encoder-decoder Graph Neural Network
(VGNN) for estimating ADRD likelihood. We created three scenarios to assess the
model's efficiency, using Random Forest and Light Gradient Boost Machine as
baselines. We further used our relation importance method to clarify the key
relationships for ADRD risk prediction. VGNN surpassed other baseline models by
10% in the area under the receiver operating characteristic. The integration of
the GNN model and relation importance interpretation could potentially play an
essential role in providing valuable insight into factors that may contribute
to or delay ADRD progression.
Employing a GNN approach with claims data enhances ADRD risk prediction and
provides insights into the impact of interconnected medical code relationships.
This methodology not only enables ADRD risk modeling but also shows potential
for other image analysis predictions using claims data
- …