853,045 research outputs found
Deep Learning at Scale with Nearest Neighbours Communications
As deep learning techniques become more and more popular, there is the need to move these applications from the data scientist’s Jupyter notebook to efficient and reliable enterprise solutions. Moreover, distributed training of deep learning models will happen more and more outside the well-known borders of cloud and HPC infrastructure and will move to edge and mobile platforms. Current techniques for distributed deep learning have drawbacks in both these scenarios, limiting their long-term applicability.
After a critical review of the established techniques for Data Parallel training from both a distributed computing and deep learning perspective, a novel approach based on nearest-neighbour communications is presented in order to overcome some of the issues related to mainstream approaches, such as global communication patterns. Moreover, in order to validate the proposed strategy, the Flexible Asynchronous Scalable Training (FAST) framework is introduced, which allows to apply the nearest-neighbours communications approach to a deep learning framework of choice.
Finally, a relevant use-case is deployed on a medium-scale infrastructure to demonstrate both the framework and the methodology presented. Training convergence and scalability results are presented and discussed in comparison to a baseline defined by using state-of-the-art distributed training tools provided by a well-known deep learning framework
Neural Distributed Autoassociative Memories: A Survey
Introduction. Neural network models of autoassociative, distributed memory
allow storage and retrieval of many items (vectors) where the number of stored
items can exceed the vector dimension (the number of neurons in the network).
This opens the possibility of a sublinear time search (in the number of stored
items) for approximate nearest neighbors among vectors of high dimension. The
purpose of this paper is to review models of autoassociative, distributed
memory that can be naturally implemented by neural networks (mainly with local
learning rules and iterative dynamics based on information locally available to
neurons). Scope. The survey is focused mainly on the networks of Hopfield,
Willshaw and Potts, that have connections between pairs of neurons and operate
on sparse binary vectors. We discuss not only autoassociative memory, but also
the generalization properties of these networks. We also consider neural
networks with higher-order connections and networks with a bipartite graph
structure for non-binary data with linear constraints. Conclusions. In
conclusion we discuss the relations to similarity search, advantages and
drawbacks of these techniques, and topics for further research. An interesting
and still not completely resolved question is whether neural autoassociative
memories can search for approximate nearest neighbors faster than other index
structures for similarity search, in particular for the case of very high
dimensional vectors.Comment: 31 page
FLock: Defending Malicious Behaviors in Federated Learning with Blockchain
Federated learning (FL) is a promising way to allow multiple data owners
(clients) to collaboratively train machine learning models without compromising
data privacy. Yet, existing FL solutions usually rely on a centralized
aggregator for model weight aggregation, while assuming clients are honest.
Even if data privacy can still be preserved, the problem of single-point
failure and data poisoning attack from malicious clients remains unresolved. To
tackle this challenge, we propose to use distributed ledger technology (DLT) to
achieve FLock, a secure and reliable decentralized Federated Learning system
built on blockchain. To guarantee model quality, we design a novel peer-to-peer
(P2P) review and reward/slash mechanism to detect and deter malicious clients,
powered by on-chain smart contracts. The reward/slash mechanism, in addition,
serves as incentives for participants to honestly upload and review model
parameters in the FLock system. FLock thus improves the performance and the
robustness of FL systems in a fully P2P manner.Comment: Accepted by NeurIPS 2022 Worksho
- …