333 research outputs found
Graph Contrastive Invariant Learning from the Causal Perspective
Graph contrastive learning (GCL), learning the node representation by
contrasting two augmented graphs in a self-supervised way, has attracted
considerable attention. GCL is usually believed to learn the invariant
representation. However, does this understanding always hold in practice? In
this paper, we first study GCL from the perspective of causality. By analyzing
GCL with the structural causal model (SCM), we discover that traditional GCL
may not well learn the invariant representations due to the non-causal
information contained in the graph. How can we fix it and encourage the current
GCL to learn better invariant representations? The SCM offers two requirements
and motives us to propose a novel GCL method. Particularly, we introduce the
spectral graph augmentation to simulate the intervention upon non-causal
factors. Then we design the invariance objective and independence objective to
better capture the causal factors. Specifically, (i) the invariance objective
encourages the encoder to capture the invariant information contained in causal
variables, and (ii) the independence objective aims to reduce the influence of
confounders on the causal variables. Experimental results demonstrate the
effectiveness of our approach on node classification tasks
Effects of Arbuscular Mycorrhizal Fungi on Root Growth and Architecture of Tulip Gesneriana
Arbuscular mycorrhizal fungi(AMF) can promote the absorption of soil water and mineral nutrients, improve photosynthesis, and make host attain higher quality finally by establishing symbiotic relationship between AMF and host root. To improve Tulip gesneriana quality have practical meaning under no bad affect to cultivation soil, in the light of its economical and ecological values. However, some AMF may be diverse from others, the concrete function of AMF on commercial tulip varieties need to explore. Therefore, three different sets of arbuscular mycorrhizal fungi were inoculated into tulip rhizosphere soil, which were set as 4(Diversispora versiformis), 7(Diversispora spurca) and 1 + 3 + 4 (Rhizophagus intraradias + Funneliformis mosseae + Diversispora versiformis), respectively. The results showed that the activity of most of the measured indices increased, the average root diameter and sucrose content decreased in those three mycorrhizal treatments. Our research provide some theoretical basis for the application of AMF on T.gesneriana ecological cultivation in future
Generalizing Graph Neural Networks on Out-Of-Distribution Graphs
Graph Neural Networks (GNNs) are proposed without considering the agnostic
distribution shifts between training and testing graphs, inducing the
degeneration of the generalization ability of GNNs on Out-Of-Distribution (OOD)
settings. The fundamental reason for such degeneration is that most GNNs are
developed based on the I.I.D hypothesis. In such a setting, GNNs tend to
exploit subtle statistical correlations existing in the training set for
predictions, even though it is a spurious correlation. However, such spurious
correlations may change in testing environments, leading to the failure of
GNNs. Therefore, eliminating the impact of spurious correlations is crucial for
stable GNNs. To this end, we propose a general causal representation framework,
called StableGNN. The main idea is to extract high-level representations from
graph data first and resort to the distinguishing ability of causal inference
to help the model get rid of spurious correlations. Particularly, we exploit a
graph pooling layer to extract subgraph-based representations as high-level
representations. Furthermore, we propose a causal variable distinguishing
regularizer to correct the biased training distribution. Hence, GNNs would
concentrate more on the stable correlations. Extensive experiments on both
synthetic and real-world OOD graph datasets well verify the effectiveness,
flexibility and interpretability of the proposed framework.Comment: IEEE TPAMI 202
Efficient Multimodal Semantic Segmentation via Dual-Prompt Learning
Multimodal (e.g., RGB-Depth/RGB-Thermal) fusion has shown great potential for
improving semantic segmentation in complex scenes (e.g., indoor/low-light
conditions). Existing approaches often fully fine-tune a dual-branch
encoder-decoder framework with a complicated feature fusion strategy for
achieving multimodal semantic segmentation, which is training-costly due to the
massive parameter updates in feature extraction and fusion. To address this
issue, we propose a surprisingly simple yet effective dual-prompt learning
network (dubbed DPLNet) for training-efficient multimodal (e.g., RGB-D/T)
semantic segmentation. The core of DPLNet is to directly adapt a frozen
pre-trained RGB model to multimodal semantic segmentation, reducing parameter
updates. For this purpose, we present two prompt learning modules, comprising
multimodal prompt generator (MPG) and multimodal feature adapter (MFA). MPG
works to fuse the features from different modalities in a compact manner and is
inserted from shadow to deep stages to generate the multi-level multimodal
prompts that are injected into the frozen backbone, while MPG adapts prompted
multimodal features in the frozen backbone for better multimodal semantic
segmentation. Since both the MPG and MFA are lightweight, only a few trainable
parameters (3.88M, 4.4% of the pre-trained backbone parameters) are introduced
for multimodal feature fusion and learning. Using a simple decoder (3.27M
parameters), DPLNet achieves new state-of-the-art performance or is on a par
with other complex approaches on four RGB-D/T semantic segmentation datasets
while satisfying parameter efficiency. Moreover, we show that DPLNet is general
and applicable to other multimodal tasks such as salient object detection and
video semantic segmentation. Without special design, DPLNet outperforms many
complicated models. Our code will be available at
github.com/ShaohuaDong2021/DPLNet.Comment: 11 pages, 4 figures, 9 table
- …