147 research outputs found
Numerical Methods for Distributed Stochastic Compositional Optimization Problems with Aggregative Structure
The paper studies the distributed stochastic compositional optimization
problems over networks, where all the agents' inner-level function is the sum
of each agent's private expectation function. Focusing on the aggregative
structure of the inner-level function, we employ the hybrid variance reduction
method to obtain the information on each agent's private expectation function,
and apply the dynamic consensus mechanism to track the information on each
agent's inner-level function. Then by combining with the standard distributed
stochastic gradient descent method, we propose a distributed aggregative
stochastic compositional gradient descent method. When the objective function
is smooth, the proposed method achieves the optimal convergence rate
. We further combine the proposed method with
the communication compression and propose the communication compressed variant
distributed aggregative stochastic compositional gradient descent method. The
compressed variant of the proposed method maintains the optimal convergence
rate . Simulated experiments on decentralized
reinforcement learning verify the effectiveness of the proposed methods
A quatum inspired neural network for geometric modeling
By conceiving physical systems as 3D many-body point clouds, geometric graph
neural networks (GNNs), such as SE(3)/E(3) equivalent GNNs, have showcased
promising performance. In particular, their effective message-passing mechanics
make them adept at modeling molecules and crystalline materials. However,
current geometric GNNs only offer a mean-field approximation of the many-body
system, encapsulated within two-body message passing, thus falling short in
capturing intricate relationships within these geometric graphs. To address
this limitation, tensor networks, widely employed by computational physics to
handle manybody systems using high-order tensors, have been introduced.
Nevertheless, integrating these tensorized networks into the message-passing
framework of GNNs faces scalability and symmetry conservation (e.g.,
permutation and rotation) challenges. In response, we introduce an innovative
equivariant Matrix Product State (MPS)-based message-passing strategy, through
achieving an efficient implementation of the tensor contraction operation. Our
method effectively models complex many-body relationships, suppressing
mean-field approximations, and captures symmetries within geometric graphs.
Importantly, it seamlessly replaces the standard message-passing and
layer-aggregation modules intrinsic to geometric GNNs. We empirically validate
the superior accuracy of our approach on benchmark tasks, including predicting
classical Newton systems and quantum tensor Hamiltonian matrices. To our
knowledge, our approach represents the inaugural utilization of parameterized
geometric tensor networks
Evaluating Self-Supervised Learning for Molecular Graph Embeddings
Graph Self-Supervised Learning (GSSL) provides a robust pathway for acquiring
embeddings without expert labelling, a capability that carries profound
implications for molecular graphs due to the staggering number of potential
molecules and the high cost of obtaining labels. However, GSSL methods are
designed not for optimisation within a specific domain but rather for
transferability across a variety of downstream tasks. This broad applicability
complicates their evaluation. Addressing this challenge, we present "Molecular
Graph Representation Evaluation" (MOLGRAPHEVAL), generating detailed profiles
of molecular graph embeddings with interpretable and diversified attributes.
MOLGRAPHEVAL offers a suite of probing tasks grouped into three categories: (i)
generic graph, (ii) molecular substructure, and (iii) embedding space
properties. By leveraging MOLGRAPHEVAL to benchmark existing GSSL methods
against both current downstream datasets and our suite of tasks, we uncover
significant inconsistencies between inferences drawn solely from existing
datasets and those derived from more nuanced probing. These findings suggest
that current evaluation methodologies fail to capture the entirety of the
landscape.Comment: update result
A Group Symmetric Stochastic Differential Equation Model for Molecule Multi-modal Pretraining
Molecule pretraining has quickly become the go-to schema to boost the
performance of AI-based drug discovery. Naturally, molecules can be represented
as 2D topological graphs or 3D geometric point clouds. Although most existing
pertaining methods focus on merely the single modality, recent research has
shown that maximizing the mutual information (MI) between such two modalities
enhances the molecule representation ability. Meanwhile, existing molecule
multi-modal pretraining approaches approximate MI based on the representation
space encoded from the topology and geometry, thus resulting in the loss of
critical structural information of molecules. To address this issue, we propose
MoleculeSDE. MoleculeSDE leverages group symmetric (e.g., SE(3)-equivariant and
reflection-antisymmetric) stochastic differential equation models to generate
the 3D geometries from 2D topologies, and vice versa, directly in the input
space. It not only obtains tighter MI bound but also enables prosperous
downstream tasks than the previous work. By comparing with 17 pretraining
baselines, we empirically verify that MoleculeSDE can learn an expressive
representation with state-of-the-art performance on 26 out of 32 downstream
tasks
UniDistill: A Universal Cross-Modality Knowledge Distillation Framework for 3D Object Detection in Bird's-Eye View
In the field of 3D object detection for autonomous driving, the sensor
portfolio including multi-modality and single-modality is diverse and complex.
Since the multi-modal methods have system complexity while the accuracy of
single-modal ones is relatively low, how to make a tradeoff between them is
difficult. In this work, we propose a universal cross-modality knowledge
distillation framework (UniDistill) to improve the performance of
single-modality detectors. Specifically, during training, UniDistill projects
the features of both the teacher and the student detector into Bird's-Eye-View
(BEV), which is a friendly representation for different modalities. Then, three
distillation losses are calculated to sparsely align the foreground features,
helping the student learn from the teacher without introducing additional cost
during inference. Taking advantage of the similar detection paradigm of
different detectors in BEV, UniDistill easily supports LiDAR-to-camera,
camera-to-LiDAR, fusion-to-LiDAR and fusion-to-camera distillation paths.
Furthermore, the three distillation losses can filter the effect of misaligned
background information and balance between objects of different sizes,
improving the distillation effectiveness. Extensive experiments on nuScenes
demonstrate that UniDistill effectively improves the mAP and NDS of student
detectors by 2.0%~3.2%
- …