16 research outputs found
Distributed Training of Graph Convolutional Networks
The aim of this work is to develop a fully-distributed algorithmic framework
for training graph convolutional networks (GCNs). The proposed method is able
to exploit the meaningful relational structure of the input data, which are
collected by a set of agents that communicate over a sparse network topology.
After formulating the centralized GCN training problem, we first show how to
make inference in a distributed scenario where the underlying data graph is
split among different agents. Then, we propose a distributed gradient descent
procedure to solve the GCN training problem. The resulting model distributes
computation along three lines: during inference, during back-propagation, and
during optimization. Convergence to stationary solutions of the GCN training
problem is also established under mild conditions. Finally, we propose an
optimization criterion to design the communication topology between agents in
order to match with the graph describing data relationships. A wide set of
numerical results validate our proposal. To the best of our knowledge, this is
the first work combining graph convolutional neural networks with distributed
optimization.Comment: Published on IEEE Transactions on Signal and Information Processing
over Network
Adaptive Point Transformer
The recent surge in 3D data acquisition has spurred the development of
geometric deep learning models for point cloud processing, boosted by the
remarkable success of transformers in natural language processing. While point
cloud transformers (PTs) have achieved impressive results recently, their
quadratic scaling with respect to the point cloud size poses a significant
scalability challenge for real-world applications. To address this issue, we
propose the Adaptive Point Cloud Transformer (AdaPT), a standard PT model
augmented by an adaptive token selection mechanism. AdaPT dynamically reduces
the number of tokens during inference, enabling efficient processing of large
point clouds. Furthermore, we introduce a budget mechanism to flexibly adjust
the computational cost of the model at inference time without the need for
retraining or fine-tuning separate models. Our extensive experimental
evaluation on point cloud classification tasks demonstrates that AdaPT
significantly reduces computational complexity while maintaining competitive
accuracy compared to standard PTs. The code for AdaPT is made publicly
available.Comment: 26 pages, 8 figures, submitted to Neural Networ
Explainability in subgraphs-enhanced Graph Neural Networks
Paper submitted to https://septentrio.uit.no/index.php/nldl/indexRecently, subgraphs-enhanced Graph Neural Networks (SGNNs) have been introduced to enhance
the expressive power of Graph Neural Networks
(GNNs), which was proved to be not higher than
the 1-dimensional Weisfeiler-Leman isomorphism
test. The new paradigm suggests using subgraphs
extracted from the input graph to improve the
model’s expressiveness, but the additional complexity exacerbates an already challenging problem in
GNNs: explaining their predictions. In this work,
we adapt PGExplainer, one of the most recent explainers for GNNs, to SGNNs. The proposed explainer accounts for the contribution of all the different subgraphs and can produce a meaningful explanation that humans can interpret. The experiments that we performed both on real and synthetic datasets show that our framework is successful in explaining the decision process of an SGNN
on graph classification tasks
Explainability in subgraphs-enhanced Graph Neural Networks
Recently, subgraphs-enhanced Graph Neural Networks (SGNNs) have been
introduced to enhance the expressive power of Graph Neural Networks (GNNs),
which was proved to be not higher than the 1-dimensional Weisfeiler-Leman
isomorphism test. The new paradigm suggests using subgraphs extracted from the
input graph to improve the model's expressiveness, but the additional
complexity exacerbates an already challenging problem in GNNs: explaining their
predictions. In this work, we adapt PGExplainer, one of the most recent
explainers for GNNs, to SGNNs. The proposed explainer accounts for the
contribution of all the different subgraphs and can produce a meaningful
explanation that humans can interpret. The experiments that we performed both
on real and synthetic datasets show that our framework is successful in
explaining the decision process of an SGNN on graph classification tasks
MoDiPO: text-to-motion alignment via AI-feedback-driven Direct Preference Optimization
Diffusion Models have revolutionized the field of human motion generation by
offering exceptional generation quality and fine-grained controllability
through natural language conditioning. Their inherent stochasticity, that is
the ability to generate various outputs from a single input, is key to their
success. However, this diversity should not be unrestricted, as it may lead to
unlikely generations. Instead, it should be confined within the boundaries of
text-aligned and realistic generations. To address this issue, we propose
MoDiPO (Motion Diffusion DPO), a novel methodology that leverages Direct
Preference Optimization (DPO) to align text-to-motion models. We streamline the
laborious and expensive process of gathering human preferences needed in DPO by
leveraging AI feedback instead. This enables us to experiment with novel DPO
strategies, using both online and offline generated motion-preference pairs. To
foster future research we contribute with a motion-preference dataset which we
dub Pick-a-Move. We demonstrate, both qualitatively and quantitatively, that
our proposed method yields significantly more realistic motions. In particular,
MoDiPO substantially improves Frechet Inception Distance (FID) while retaining
the same RPrecision and Multi-Modality performances
Interpreting Black-Box Models: A Review on Explainable Artificial Intelligence
Recent years have seen a tremendous growth in Artificial Intelligence (AI)-based methodological development in a broad range of domains. In this rapidly evolving field, large number of methods are being reported using machine learning (ML) and Deep Learning (DL) models. Majority of these models are inherently complex and lacks explanations of the decision making process causing these models to be termed as 'Black-Box'. One of the major bottlenecks to adopt such models in mission-critical application domains, such as banking, e-commerce, healthcare, and public services and safety, is the difficulty in interpreting them. Due to the rapid proleferation of these AI models, explaining their learning and decision making process are getting harder which require transparency and easy predictability. Aiming to collate the current state-of-the-art in interpreting the black-box models, this study provides a comprehensive analysis of the explainable AI (XAI) models. To reduce false negative and false positive outcomes of these back-box models, finding flaws in them is still difficult and inefficient. In this paper, the development of XAI is reviewed meticulously through careful selection and analysis of the current state-of-the-art of XAI research. It also provides a comprehensive and in-depth evaluation of the XAI frameworks and their efficacy to serve as a starting point of XAI for applied and theoretical researchers. Towards the end, it highlights emerging and critical issues pertaining to XAI research to showcase major, model-specific trends for better explanation, enhanced transparency, and improved prediction accuracy
ICML 2023 Topological Deep Learning Challenge : Design and Results
This paper presents the computational challenge on topological deep learning
that was hosted within the ICML 2023 Workshop on Topology and Geometry in
Machine Learning. The competition asked participants to provide open-source
implementations of topological neural networks from the literature by
contributing to the python packages TopoNetX (data processing) and TopoModelX
(deep learning). The challenge attracted twenty-eight qualifying submissions in
its two-month duration. This paper describes the design of the challenge and
summarizes its main findings