17,332 research outputs found
Parameters estimation for spatio-temporal maximum entropy distributions: application to neural spike trains
We propose a numerical method to learn Maximum Entropy (MaxEnt) distributions
with spatio-temporal constraints from experimental spike trains. This is an
extension of two papers [10] and [4] who proposed the estimation of parameters
where only spatial constraints were taken into account. The extension we
propose allows to properly handle memory effects in spike statistics, for large
sized neural networks.Comment: 34 pages, 33 figure
One-class classifiers based on entropic spanning graphs
One-class classifiers offer valuable tools to assess the presence of outliers
in data. In this paper, we propose a design methodology for one-class
classifiers based on entropic spanning graphs. Our approach takes into account
the possibility to process also non-numeric data by means of an embedding
procedure. The spanning graph is learned on the embedded input data and the
outcoming partition of vertices defines the classifier. The final partition is
derived by exploiting a criterion based on mutual information minimization.
Here, we compute the mutual information by using a convenient formulation
provided in terms of the -Jensen difference. Once training is
completed, in order to associate a confidence level with the classifier
decision, a graph-based fuzzy model is constructed. The fuzzification process
is based only on topological information of the vertices of the entropic
spanning graph. As such, the proposed one-class classifier is suitable also for
data characterized by complex geometric structures. We provide experiments on
well-known benchmarks containing both feature vectors and labeled graphs. In
addition, we apply the method to the protein solubility recognition problem by
considering several representations for the input samples. Experimental results
demonstrate the effectiveness and versatility of the proposed method with
respect to other state-of-the-art approaches.Comment: Extended and revised version of the paper "One-Class Classification
Through Mutual Information Minimization" presented at the 2016 IEEE IJCNN,
Vancouver, Canad
Deep Learning as a Parton Shower
We make the connection between certain deep learning architectures and the
renormalisation group explicit in the context of QCD by using a deep learning
network to construct a toy parton shower model. The model aims to describe
proton-proton collisions at the Large Hadron Collider. A convolutional
autoencoder learns a set of kernels that efficiently encode the behaviour of
fully showered QCD collision events. The network is structured recursively so
as to ensure self-similarity, and the number of trained network parameters is
low. Randomness is introduced via a novel custom masking layer, which also
preserves existing parton splittings by using layer-skipping connections. By
applying a shower merging procedure, the network can be evaluated on unshowered
events produced by a matrix element calculation. The trained network behaves as
a parton shower that qualitatively reproduces jet-based observables.Comment: 26 pages, 13 figure
- …