324,028 research outputs found
Truncated Variational EM for Semi-Supervised Neural Simpletrons
Inference and learning for probabilistic generative networks is often very
challenging and typically prevents scalability to as large networks as used for
deep discriminative approaches. To obtain efficiently trainable, large-scale
and well performing generative networks for semi-supervised learning, we here
combine two recent developments: a neural network reformulation of hierarchical
Poisson mixtures (Neural Simpletrons), and a novel truncated variational EM
approach (TV-EM). TV-EM provides theoretical guarantees for learning in
generative networks, and its application to Neural Simpletrons results in
particularly compact, yet approximately optimal, modifications of learning
equations. If applied to standard benchmarks, we empirically find, that
learning converges in fewer EM iterations, that the complexity per EM iteration
is reduced, and that final likelihood values are higher on average. For the
task of classification on data sets with few labels, learning improvements
result in consistently lower error rates if compared to applications without
truncation. Experiments on the MNIST data set herein allow for comparison to
standard and state-of-the-art models in the semi-supervised setting. Further
experiments on the NIST SD19 data set show the scalability of the approach when
a manifold of additional unlabeled data is available
Neurogenesis Deep Learning
Neural machine learning methods, such as deep neural networks (DNN), have
achieved remarkable success in a number of complex data processing tasks. These
methods have arguably had their strongest impact on tasks such as image and
audio processing - data processing domains in which humans have long held clear
advantages over conventional algorithms. In contrast to biological neural
systems, which are capable of learning continuously, deep artificial networks
have a limited ability for incorporating new information in an already trained
network. As a result, methods for continuous learning are potentially highly
impactful in enabling the application of deep networks to dynamic data sets.
Here, inspired by the process of adult neurogenesis in the hippocampus, we
explore the potential for adding new neurons to deep layers of artificial
neural networks in order to facilitate their acquisition of novel information
while preserving previously trained data representations. Our results on the
MNIST handwritten digit dataset and the NIST SD 19 dataset, which includes
lower and upper case letters and digits, demonstrate that neurogenesis is well
suited for addressing the stability-plasticity dilemma that has long challenged
adaptive machine learning algorithms.Comment: 8 pages, 8 figures, Accepted to 2017 International Joint Conference
on Neural Networks (IJCNN 2017
Forming of Learning Set for Neural Networks in Problems of Losless Data Compression
questions of forming of learning sets for artificial neural networks in problems of lossless data
compression are considered. Methods of construction and use of learning sets are studied. The way of forming of
learning set during training an artificial neural network on the data stream is offered
Graph Neural Networks in Particle Physics
Particle physics is a branch of science aiming at discovering the fundamental
laws of matter and forces. Graph neural networks are trainable functions which
operate on graphs -- sets of elements and their pairwise relations -- and are a
central method within the broader field of geometric deep learning. They are
very expressive and have demonstrated superior performance to other classical
deep learning approaches in a variety of domains. The data in particle physics
are often represented by sets and graphs and as such, graph neural networks
offer key advantages. Here we review various applications of graph neural
networks in particle physics, including different graph constructions, model
architectures and learning objectives, as well as key open problems in particle
physics for which graph neural networks are promising.Comment: 29 pages, 11 figures, submitted to Machine Learning: Science and
Technology, Focus on Machine Learning for Fundamental Physics collectio
Graph Neural Networks in Particle Physics
Particle physics is a branch of science aiming at discovering the fundamental laws of matter and forces. Graph neural networks are trainable functions which operate on graphs—sets of elements and their pairwise relations—and are a central method within the broader field of geometric deep learning. They are very expressive and have demonstrated superior performance to other classical deep learning approaches in a variety of domains. The data in particle physics are often represented by sets and graphs and as such, graph neural networks offer key advantages. Here we review various applications of graph neural networks in particle physics, including different graph constructions, model architectures and learning objectives, as well as key open problems in particle physics for which graph neural networks are promising
Exact ICL maximization in a non-stationary time extension of the latent block model for dynamic networks
The latent block model (LBM) is a flexible probabilistic tool to describe
interactions between node sets in bipartite networks, but it does not account
for interactions of time varying intensity between nodes in unknown classes. In
this paper we propose a non stationary temporal extension of the LBM that
clusters simultaneously the two node sets of a bipartite network and constructs
classes of time intervals on which interactions are stationary. The number of
clusters as well as the membership to classes are obtained by maximizing the
exact complete-data integrated likelihood relying on a greedy search approach.
Experiments on simulated and real data are carried out in order to assess the
proposed methodology.Comment: European Symposium on Artificial Neural Networks, Computational
Intelligence and Machine Learning (ESANN), Apr 2015, Bruges, Belgium.
pp.225-230, 2015, Proceedings of the 23-th European Symposium on Artificial
Neural Networks, Computational Intelligence and Machine Learning (ESANN 2015
An Empirical Evaluation of Deep Learning on Highway Driving
Numerous groups have applied a variety of deep learning techniques to
computer vision problems in highway perception scenarios. In this paper, we
presented a number of empirical evaluations of recent deep learning advances.
Computer vision, combined with deep learning, has the potential to bring about
a relatively inexpensive, robust solution to autonomous driving. To prepare
deep learning for industry uptake and practical applications, neural networks
will require large data sets that represent all possible driving environments
and scenarios. We collect a large data set of highway data and apply deep
learning and computer vision algorithms to problems such as car and lane
detection. We show how existing convolutional neural networks (CNNs) can be
used to perform lane and vehicle detection while running at frame rates
required for a real-time system. Our results lend credence to the hypothesis
that deep learning holds promise for autonomous driving.Comment: Added a video for lane detectio
- …