130 research outputs found

    Advancing Biomedicine with Graph Representation Learning: Recent Progress, Challenges, and Future Directions

    Full text link
    Graph representation learning (GRL) has emerged as a pivotal field that has contributed significantly to breakthroughs in various fields, including biomedicine. The objective of this survey is to review the latest advancements in GRL methods and their applications in the biomedical field. We also highlight key challenges currently faced by GRL and outline potential directions for future research.Comment: Accepted by 2023 IMIA Yearbook of Medical Informatic

    Uncertainty-Aware Decision Transformer for Stochastic Driving Environments

    Full text link
    Offline Reinforcement Learning (RL) has emerged as a promising framework for learning policies without active interactions, making it especially appealing for autonomous driving tasks. Recent successes of Transformers inspire casting offline RL as sequence modeling, which performs well in long-horizon tasks. However, they are overly optimistic in stochastic environments with incorrect assumptions that the same goal can be consistently achieved by identical actions. In this paper, we introduce an UNcertainty-awaRE deciSion Transformer (UNREST) for planning in stochastic driving environments without introducing additional transition or complex generative models. Specifically, UNREST estimates state uncertainties by the conditional mutual information between transitions and returns, and segments sequences accordingly. Discovering the `uncertainty accumulation' and `temporal locality' properties of driving environments, UNREST replaces the global returns in decision transformers with less uncertain truncated returns, to learn from true outcomes of agent actions rather than environment transitions. We also dynamically evaluate environmental uncertainty during inference for cautious planning. Extensive experimental results demonstrate UNREST's superior performance in various driving scenarios and the power of our uncertainty estimation strategy

    Boosting Offline Reinforcement Learning for Autonomous Driving with Hierarchical Latent Skills

    Full text link
    Learning-based vehicle planning is receiving increasing attention with the emergence of diverse driving simulators and large-scale driving datasets. While offline reinforcement learning (RL) is well suited for these safety-critical tasks, it still struggles to plan over extended periods. In this work, we present a skill-based framework that enhances offline RL to overcome the long-horizon vehicle planning challenge. Specifically, we design a variational autoencoder (VAE) to learn skills from offline demonstrations. To mitigate posterior collapse of common VAEs, we introduce a two-branch sequence encoder to capture both discrete options and continuous variations of the complex driving skills. The final policy treats learned skills as actions and can be trained by any off-the-shelf offline RL algorithms. This facilitates a shift in focus from per-step actions to temporally extended skills, thereby enabling long-term reasoning into the future. Extensive results on CARLA prove that our model consistently outperforms strong baselines at both training and new scenarios. Additional visualizations and experiments demonstrate the interpretability and transferability of extracted skills

    Boosting Operational DNN Testing Efficiency through Conditioning

    Full text link
    With the increasing adoption of Deep Neural Network (DNN) models as integral parts of software systems, efficient operational testing of DNNs is much in demand to ensure these models' actual performance in field conditions. A challenge is that the testing often needs to produce precise results with a very limited budget for labeling data collected in field. Viewing software testing as a practice of reliability estimation through statistical sampling, we re-interpret the idea behind conventional structural coverages as conditioning for variance reduction. With this insight we propose an efficient DNN testing method based on the conditioning on the representation learned by the DNN model under testing. The representation is defined by the probability distribution of the output of neurons in the last hidden layer of the model. To sample from this high dimensional distribution in which the operational data are sparsely distributed, we design an algorithm leveraging cross entropy minimization. Experiments with various DNN models and datasets were conducted to evaluate the general efficiency of the approach. The results show that, compared with simple random sampling, this approach requires only about a half of labeled inputs to achieve the same level of precision.Comment: Published in the Proceedings of the 27th ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering (ESEC/FSE 2019

    Learning with Logical Constraints but without Shortcut Satisfaction

    Full text link
    Recent studies in neuro-symbolic learning have explored the integration of logical knowledge into deep learning via encoding logical constraints as an additional loss function. However, existing approaches tend to vacuously satisfy logical constraints through shortcuts, failing to fully exploit the knowledge. In this paper, we present a new framework for learning with logical constraints. Specifically, we address the shortcut satisfaction issue by introducing dual variables for logical connectives, encoding how the constraint is satisfied. We further propose a variational framework where the encoded logical constraint is expressed as a distributional loss that is compatible with the model's original training loss. The theoretical analysis shows that the proposed approach bears salient properties, and the experimental evaluations demonstrate its superior performance in both model generalizability and constraint satisfaction.Comment: Published as a conference paper at ICLR 2023, and code is available at https://github.com/SoftWiser-group/NeSy-without-Shortcut

    From moments of the distribution function to hydrodynamics - the non-conformal case

    Full text link
    We study the one-dimensional boost-invariant Boltzmann equation in the relaxation-time approximation using special moments of the distribution function for a system with a finite particle mass. The infinite hierarchy of moments can be truncated by keeping only the three lowest moments that correspond to the three independent components of the energy-momentum tensor. We show that such a three-moment truncation reproduces accurately the exact solution of the kinetic equation after a simple renormalization that takes into account the effects of the neglected higher moments. We derive second-order Israel-Stewart hydrodynamic equations from the three-moment equations, and show that, for most physically relevant initial conditions, these equations yield results comparable to those of the three-moment truncation, albeit less accurate. We attribute this feature to the fact that the structure of Israel-Stewart equations is similar to that of the three-moment truncation. In particular, the presence of the relaxation term in the Israel-Stewart equations, yields an early-time regime that mimics approximately the collisionless regime. A detailed comparison of the three-moment truncation with second-order non-conformal hydrodynamics reveals ambiguities in the definition of second-order transport coefficients. These ambiguities affect the ability of Israel-Stewart hydrodynamics to reproduce results of kinetic theory.Comment: 44 pages, 11 figure

    Softened Symbol Grounding for Neuro-symbolic Systems

    Full text link
    Neuro-symbolic learning generally consists of two separated worlds, i.e., neural network training and symbolic constraint solving, whose success hinges on symbol grounding, a fundamental problem in AI. This paper presents a novel, softened symbol grounding process, bridging the gap between the two worlds, and resulting in an effective and efficient neuro-symbolic learning framework. Technically, the framework features (1) modeling of symbol solution states as a Boltzmann distribution, which avoids expensive state searching and facilitates mutually beneficial interactions between network training and symbolic reasoning;(2) a new MCMC technique leveraging projection and SMT solvers, which efficiently samples from disconnected symbol solution spaces; (3) an annealing mechanism that can escape from %being trapped into sub-optimal symbol groundings. Experiments with three representative neuro symbolic learning tasks demonstrate that, owining to its superior symbol grounding capability, our framework successfully solves problems well beyond the frontier of the existing proposals.Comment: Published as a conference paper at ICLR 2023. Code is available at https://github.com/SoftWiser-group/Soften-NeSy-learnin

    Is Underwater Image Enhancement All Object Detectors Need?

    Full text link
    Underwater object detection is a crucial and challenging problem in marine engineering and aquatic robot. The difficulty is partly because of the degradation of underwater images caused by light selective absorption and scattering. Intuitively, enhancing underwater images can benefit high-level applications like underwater object detection. However, it is still unclear whether all object detectors need underwater image enhancement as pre-processing. We therefore pose the questions "Does underwater image enhancement really improve underwater object detection?" and "How does underwater image enhancement contribute to underwater object detection?". With these two questions, we conduct extensive studies. Specifically, we use 18 state-of-the-art underwater image enhancement algorithms, covering traditional, CNN-based, and GAN-based algorithms, to pre-process underwater object detection data. Then, we retrain 7 popular deep learning-based object detectors using the corresponding results enhanced by different algorithms, obtaining 126 underwater object detection models. Coupled with 7 object detection models retrained using raw underwater images, we employ these 133 models to comprehensively analyze the effect of underwater image enhancement on underwater object detection. We expect this study can provide sufficient exploration to answer the aforementioned questions and draw more attention of the community to the joint problem of underwater image enhancement and underwater object detection. The pre-trained models and results are publicly available and will be regularly updated. Project page: https://github.com/BIGWangYuDong/lqit/tree/main/configs/detection/uw_enhancement_affect_detection.Comment: 17 pages, 9 figure

    Pretreatment plasma fibrinogen level as a prognostic biomarker for patients with lung cancer

    Get PDF
    Many researchers have shown that pretreatment plasma fibrinogen levels are closely correlated with the prognosis of patients with lung cancer (LC). In this study, we thus performed a meta-analysis to systematically assess the prognostic value of pretreatment plasma fibrinogen levels in LC patients. A computerized systematic search in PubMed, EMBASE, Web of Science and China National Knowledge Infrastructure (CNKI) was performed up to March 15, 2018. Studies with available data on the prognostic value of plasma fibrinogen in LC patients were eligible for inclusion. The pooled hazard ratios (HRs) and odd ratios (ORs) with 95% confidence intervals (CIs) were used to evaluate the correlation between pretreatment plasma fibrinogen levels and prognosis as well as clinicopathological characteristics. A total of 17 studies with 6,460 LC patients were included in this meta-analysis. A higher pretreatment plasma fibrinogen level was significantly associated with worse overall survival (OS) (HR: 1.57; 95% CI: 1.39-1.77; p=0.001), disease-free survival (DFS) (HR: 1.53; 95% CI: 1.33-1.76; p=0.003), and progression-free survival (PFS) (HR: 3.14; 95% CI: 2.15-4.59; po0.001). Furthermore, our subgroup and sensitivity analyses demonstrated that the pooled HR for OS was robust and reliable. In addition, we also found that a higher fibrinogen level predicted advanced TNM stage (III-IV) (OR=2.18, 95% CI: 1.79-2.66; po0.001) and a higher incidence of lymph node metastasis (OR=1.74, 95% CI: 1.44-2.10; p=0.02). Our study suggested that higher pretreatment plasma fibrinogen levels predict worse prognoses in LC patients

    Explainable Graph Neural Network for Alzheimer's Disease And Related Dementias Risk Prediction

    Full text link
    Alzheimer's disease and related dementias (ADRD) ranks as the sixth leading cause of death in the US, underlining the importance of accurate ADRD risk prediction. While recent advancement in ADRD risk prediction have primarily relied on imaging analysis, yet not all patients undergo medical imaging before an ADRD diagnosis. Merging machine learning with claims data can reveal additional risk factors and uncover interconnections among diverse medical codes. Our goal is to utilize Graph Neural Networks (GNNs) with claims data for ADRD risk prediction. Addressing the lack of human-interpretable reasons behind these predictions, we introduce an innovative method to evaluate relationship importance and its influence on ADRD risk prediction, ensuring comprehensive interpretation. We employed Variationally Regularized Encoder-decoder Graph Neural Network (VGNN) for estimating ADRD likelihood. We created three scenarios to assess the model's efficiency, using Random Forest and Light Gradient Boost Machine as baselines. We further used our relation importance method to clarify the key relationships for ADRD risk prediction. VGNN surpassed other baseline models by 10% in the area under the receiver operating characteristic. The integration of the GNN model and relation importance interpretation could potentially play an essential role in providing valuable insight into factors that may contribute to or delay ADRD progression. Employing a GNN approach with claims data enhances ADRD risk prediction and provides insights into the impact of interconnected medical code relationships. This methodology not only enables ADRD risk modeling but also shows potential for other image analysis predictions using claims data
    corecore