43,945 research outputs found
Logic Negation with Spiking Neural P Systems
Nowadays, the success of neural networks as reasoning systems is doubtless.
Nonetheless, one of the drawbacks of such reasoning systems is that they work
as black-boxes and the acquired knowledge is not human readable. In this paper,
we present a new step in order to close the gap between connectionist and logic
based reasoning systems. We show that two of the most used inference rules for
obtaining negative information in rule based reasoning systems, the so-called
Closed World Assumption and Negation as Finite Failure can be characterized by
means of spiking neural P systems, a formal model of the third generation of
neural networks born in the framework of membrane computing.Comment: 25 pages, 1 figur
Small-variance asymptotics for Bayesian neural networks
Bayesian neural networks (BNNs) are a rich and flexible class of models that have several advantages over standard feedforward networks, but are typically expensive to train on large-scale data. In this thesis, we explore the use of small-variance asymptotics-an approach to yielding fast algorithms from probabilistic models-on various Bayesian neural network models. We first demonstrate how small-variance asymptotics shows precise connections between standard neural networks and BNNs; for example, particular sampling algorithms for BNNs reduce to standard backpropagation in the small-variance limit. We then explore a more complex BNN where the number of hidden units is additionally treated as a random variable in the model. While standard sampling schemes would be too slow to be practical, our asymptotic approach yields a simple method for extending standard backpropagation to the case where the number of hidden units is not fixed. We show on several data sets that the resulting algorithm has benefits over backpropagation on networks with a fixed architecture.2019-01-02T00:00:00
Deciding How to Decide: Dynamic Routing in Artificial Neural Networks
We propose and systematically evaluate three strategies for training
dynamically-routed artificial neural networks: graphs of learned
transformations through which different input signals may take different paths.
Though some approaches have advantages over others, the resulting networks are
often qualitatively similar. We find that, in dynamically-routed networks
trained to classify images, layers and branches become specialized to process
distinct categories of images. Additionally, given a fixed computational
budget, dynamically-routed networks tend to perform better than comparable
statically-routed networks.Comment: ICML 2017. Code at https://github.com/MasonMcGill/multipath-nn Video
abstract at https://youtu.be/NHQsDaycwy
IEEE Access special section editorial: Artificial intelligence enabled networking
With today’s computer networks becoming increasingly dynamic, heterogeneous, and complex, there is great interest in deploying artificial intelligence (AI) based techniques for optimization and management of computer networks. AI techniques—that subsume multidisciplinary techniques from machine learning, optimization theory, game theory, control theory, and meta-heuristics—have long been applied to optimize computer networks in many diverse settings. Such an approach is gaining increased traction with the emergence of novel networking paradigms that promise to simplify network management (e.g., cloud computing, network functions virtualization, and software-defined networking) and provide intelligent services (e.g., future 5G mobile networks). Looking ahead, greater integration of AI into networking architectures can help develop a future vision of cognitive networks that will show network-wide intelligent behavior to solve problems of network heterogeneity, performance, and quality of service (QoS)
- …