239 research outputs found
Faithful and Consistent Graph Neural Network Explanations with Rationale Alignment
Uncovering rationales behind predictions of graph neural networks (GNNs) has
received increasing attention over recent years. Instance-level GNN explanation
aims to discover critical input elements, like nodes or edges, that the target
GNN relies upon for making predictions. %These identified sub-structures can
provide interpretations of GNN's behavior. Though various algorithms are
proposed, most of them formalize this task by searching the minimal subgraph
which can preserve original predictions. However, an inductive bias is
deep-rooted in this framework: several subgraphs can result in the same or
similar outputs as the original graphs. Consequently, they have the danger of
providing spurious explanations and failing to provide consistent explanations.
Applying them to explain weakly-performed GNNs would further amplify these
issues. To address this problem, we theoretically examine the predictions of
GNNs from the causality perspective. Two typical reasons for spurious
explanations are identified: confounding effect of latent variables like
distribution shift, and causal factors distinct from the original input.
Observing that both confounding effects and diverse causal rationales are
encoded in internal representations, \tianxiang{we propose a new explanation
framework with an auxiliary alignment loss, which is theoretically proven to be
optimizing a more faithful explanation objective intrinsically. Concretely for
this alignment loss, a set of different perspectives are explored: anchor-based
alignment, distributional alignment based on Gaussian mixture models,
mutual-information-based alignment, etc. A comprehensive study is conducted
both on the effectiveness of this new framework in terms of explanation
faithfulness/consistency and on the advantages of these variants.Comment: TIST2023. arXiv admin note: substantial text overlap with
arXiv:2205.1373
3D FE modelling of die stress and wear in non-isothermal forging of Ti–6Al–4V turbine blade
RegExplainer: Generating Explanations for Graph Neural Networks in Regression Task
Graph regression is a fundamental task and has received increasing attention
in a wide range of graph learning tasks. However, the inference process is
often not interpretable. Most existing explanation techniques are limited to
understanding GNN behaviors in classification tasks. In this work, we seek an
explanation to interpret the graph regression models (XAIG-R). We show that
existing methods overlook the distribution shifting and continuously ordered
decision boundary, which hinders them away from being applied in the regression
tasks. To address these challenges, we propose a novel objective based on the
information bottleneck theory and introduce a new mix-up framework, which could
support various GNNs in a model-agnostic manner. We further present a
contrastive learning strategy to tackle the continuously ordered labels in
regression task. To empirically verify the effectiveness of the proposed
method, we introduce three benchmark datasets and a real-life dataset for
evaluation. Extensive experiments show the effectiveness of the proposed method
in interpreting GNN models in regression tasks
DyExplainer: Explainable Dynamic Graph Neural Networks
Graph Neural Networks (GNNs) resurge as a trending research subject owing to
their impressive ability to capture representations from graph-structured data.
However, the black-box nature of GNNs presents a significant challenge in terms
of comprehending and trusting these models, thereby limiting their practical
applications in mission-critical scenarios. Although there has been substantial
progress in the field of explaining GNNs in recent years, the majority of these
studies are centered on static graphs, leaving the explanation of dynamic GNNs
largely unexplored. Dynamic GNNs, with their ever-evolving graph structures,
pose a unique challenge and require additional efforts to effectively capture
temporal dependencies and structural relationships. To address this challenge,
we present DyExplainer, a novel approach to explaining dynamic GNNs on the fly.
DyExplainer trains a dynamic GNN backbone to extract representations of the
graph at each snapshot, while simultaneously exploring structural relationships
and temporal dependencies through a sparse attention technique. To preserve the
desired properties of the explanation, such as structural consistency and
temporal continuity, we augment our approach with contrastive learning
techniques to provide priori-guided regularization. To model longer-term
temporal dependencies, we develop a buffer-based live-updating scheme for
training. The results of our extensive experiments on various datasets
demonstrate the superiority of DyExplainer, not only providing faithful
explainability of the model predictions but also significantly improving the
model prediction accuracy, as evidenced in the link prediction task.Comment: 9 page
Adaptive Policy Learning for Offline-to-Online Reinforcement Learning
Conventional reinforcement learning (RL) needs an environment to collect
fresh data, which is impractical when online interactions are costly. Offline
RL provides an alternative solution by directly learning from the previously
collected dataset. However, it will yield unsatisfactory performance if the
quality of the offline datasets is poor. In this paper, we consider an
offline-to-online setting where the agent is first learned from the offline
dataset and then trained online, and propose a framework called Adaptive Policy
Learning for effectively taking advantage of offline and online data.
Specifically, we explicitly consider the difference between the online and
offline data and apply an adaptive update scheme accordingly, that is, a
pessimistic update strategy for the offline dataset and an optimistic/greedy
update scheme for the online dataset. Such a simple and effective method
provides a way to mix the offline and online RL and achieve the best of both
worlds. We further provide two detailed algorithms for implementing the
framework through embedding value or policy-based RL algorithms into it.
Finally, we conduct extensive experiments on popular continuous control tasks,
and results show that our algorithm can learn the expert policy with high
sample efficiency even when the quality of offline dataset is poor, e.g.,
random dataset.Comment: AAAI202
Label Propagation for Graph Label Noise
Label noise is a common challenge in large datasets, as it can significantly
degrade the generalization ability of deep neural networks. Most existing
studies focus on noisy labels in computer vision; however, graph models
encompass both node features and graph topology as input, and become more
susceptible to label noise through message-passing mechanisms. Recently, only a
few works have been proposed to tackle the label noise on graphs. One major
limitation is that they assume the graph is homophilous and the labels are
smoothly distributed. Nevertheless, real-world graphs may contain varying
degrees of heterophily or even be heterophily-dominated, leading to the
inadequacy of current methods. In this paper, we study graph label noise in the
context of arbitrary heterophily, with the aim of rectifying noisy labels and
assigning labels to previously unlabeled nodes. We begin by conducting two
empirical analyses to explore the impact of graph homophily on graph label
noise. Following observations, we propose a simple yet efficient algorithm,
denoted as LP4GLN. Specifically, LP4GLN is an iterative algorithm with three
steps: (1) reconstruct the graph to recover the homophily property, (2) utilize
label propagation to rectify the noisy labels, (3) select high-confidence
labels to retain for the next iteration. By iterating these steps, we obtain a
set of correct labels, ultimately achieving high accuracy in the node
classification task. The theoretical analysis is also provided to demonstrate
its remarkable denoising "effect". Finally, we conduct experiments on 10
benchmark datasets under varying graph heterophily levels and noise types,
comparing the performance of LP4GLN with 7 typical baselines. Our results
illustrate the superior performance of the proposed LP4GLN
Characterization of Diversity and Probiotic Efficiency of the Autochthonous Lactic Acid Bacteria in the Fermentation of Selected Raw Fruit and Vegetable Juices
The diversity of indigenous lactic acid bacteria (LAB) in fermented broccoli, cherry, ginger, white radish, and white-fleshed pitaya juices was analyzed using culture-independent and -dependent approaches. The major properties of selected probiotic strains, including dynamic variations in pH, viable cell counts, antibiotic resistance, bacterial adhesion to hydrophobic compounds, and survivability during simulated gastrointestinal transit, were investigated using broccoli as the fermentation substrate. In broccoli and ginger juices, the genus Lactobacillus occupied the dominant position (abundances of 79.0 and 30.3%, respectively); in cherry and radish juices, Weissella occupied the dominant position (abundances of 78.3 and 83.2%, respectively); and in pitaya juice, Streptococcus and Lactococcus occupied the dominant positions (52.2 and 37.0%, respectively). Leuconostoc mesenteroides, Weissella cibaria/soli/confusa, Enterococcus gallinarum/durans/hirae, Pediococcus pentosaceus, Bacillus coagulans, and Lactococcus garvieae/lactis subspecies were identified by partial 16S rRNA gene sequencing. In general, the selected autochthonous LAB isolates displayed no significant differences in comparison with commercial strains with regard to growth rates or acidification in fermented broccoli juice. Among all the isolates, L. mesenteroides B4-25 exhibited the highest antibiotic resistance profile (equal to that of L. plantarum CICC20265), and suitable adhesion properties (adhesion of 13.4 ± 5.2% ∼ 36.4 ± 3.2% and 21.6 ± 1.4% ∼ 69.6 ± 2.3% to ethyl acetate and xylene, respectively). Furthermore, P. pentosaceus Ca-4 and L. mesenteroides B-25 featured the highest survival rates (22.4 ± 2.6 and 21.2 ± 1.4%, respectively), after simulated gastrointestinal transit. These results indicated a high level of diversity among the autochthonous bacterial community in fermented fruit and vegetable juices, and demonstrated the potential of these candidate probiotics for applications in fermentation
TNC-UTM: A Holistic Solution to Secure Enterprise Networks
This paper presents TNC-UTM, a holistic solution to secure enterprise networks from gateway to endpoints. Just as its name suggested, the TNC-UTM solution combines two popular techniques TNC and UTM together by defining an interface between them that integrates their security capacity to provide efficiently network access control and security protection for enterprise network. Not only TNC-UTM provides the features of TNC and UTM, but also it achieves stronger security and higher performance by introducing intelligent configuration decisions and RBAC mechanism. Experiment demonstrated the superior advantages of the TNC-UTM solution
Random Walk on Multiple Networks
Random Walk is a basic algorithm to explore the structure of networks, which
can be used in many tasks, such as local community detection and network
embedding. Existing random walk methods are based on single networks that
contain limited information. In contrast, real data often contain entities with
different types or/and from different sources, which are comprehensive and can
be better modeled by multiple networks. To take advantage of rich information
in multiple networks and make better inferences on entities, in this study, we
propose random walk on multiple networks, RWM. RWM is flexible and supports
both multiplex networks and general multiple networks, which may form
many-to-many node mappings between networks. RWM sends a random walker on each
network to obtain the local proximity (i.e., node visiting probabilities)
w.r.t. the starting nodes. Walkers with similar visiting probabilities
reinforce each other. We theoretically analyze the convergence properties of
RWM. Two approximation methods with theoretical performance guarantees are
proposed for efficient computation. We apply RWM in link prediction, network
embedding, and local community detection. Comprehensive experiments conducted
on both synthetic and real-world datasets demonstrate the effectiveness and
efficiency of RWM.Comment: Accepted to IEEE TKD
- …