630 research outputs found
The Little-Hopfield model on a Random Graph
We study the Hopfield model on a random graph in scaling regimes where the
average number of connections per neuron is a finite number and where the spin
dynamics is governed by a synchronous execution of the microscopic update rule
(Little-Hopfield model).We solve this model within replica symmetry and by
using bifurcation analysis we prove that the spin-glass/paramagnetic and the
retrieval/paramagnetictransition lines of our phase diagram are identical to
those of sequential dynamics.The first-order retrieval/spin-glass transition
line follows by direct evaluation of our observables using population dynamics.
Within the accuracy of numerical precision and for sufficiently small values of
the connectivity parameter we find that this line coincides with the
corresponding sequential one. Comparison with simulation experiments shows
excellent agreement.Comment: 14 pages, 4 figure
On-Line AdaTron Learning of Unlearnable Rules
We study the on-line AdaTron learning of linearly non-separable rules by a
simple perceptron. Training examples are provided by a perceptron with a
non-monotonic transfer function which reduces to the usual monotonic relation
in a certain limit. We find that, although the on-line AdaTron learning is a
powerful algorithm for the learnable rule, it does not give the best possible
generalization error for unlearnable problems. Optimization of the learning
rate is shown to greatly improve the performance of the AdaTron algorithm,
leading to the best possible generalization error for a wide range of the
parameter which controls the shape of the transfer function.)Comment: RevTeX 17 pages, 8 figures, to appear in Phys.Rev.
Thermodynamic properties of extremely diluted symmetric Q-Ising neural networks
Using the replica-symmetric mean-field theory approach the thermodynamic and
retrieval properties of extremely diluted {\it symmetric} -Ising neural
networks are studied. In particular, capacity-gain parameter and
capacity-temperature phase diagrams are derived for and .
The zero-temperature results are compared with those obtained from a study of
the dynamics of the model. Furthermore, the de Almeida-Thouless line is
determined. Where appropriate, the difference with other -Ising
architectures is outlined.Comment: 16 pages Latex including 6 eps-figures. Corrections, also in most of
the figures have been mad
Online Learning with Ensembles
Supervised online learning with an ensemble of students randomized by the
choice of initial conditions is analyzed. For the case of the perceptron
learning rule, asymptotically the same improvement in the generalization error
of the ensemble compared to the performance of a single student is found as in
Gibbs learning. For more optimized learning rules, however, using an ensemble
yields no improvement. This is explained by showing that for any learning rule
a transform exists, such that a single student using
has the same generalization behaviour as an ensemble of
-students.Comment: 8 pages, 1 figure. Submitted to J.Phys.
On the conditions for the existence of Perfect Learning and power law in learning from stochastic examples by Ising perceptrons
In a previous letter, we studied learning from stochastic examples by
perceptrons with Ising weights in the framework of statistical mechanics. Under
the one-step replica symmetry breaking ansatz, the behaviours of learning
curves were classified according to some local property of the rules by which
examples were drawn. Further, the conditions for the existence of the Perfect
Learning together with other behaviors of the learning curves were given. In
this paper, we give the detailed derivation about these results and further
argument about the Perfect Learning together with extensive numerical
calculations.Comment: 28 pages, 43 figures. Submitted to J. Phys.
Statistical Mechanics of Learning in the Presence of Outliers
Using methods of statistical mechanics, we analyse the effect of outliers on
the supervised learning of a classification problem. The learning strategy aims
at selecting informative examples and discarding outliers. We compare two
algorithms which perform the selection either in a soft or a hard way. When the
fraction of outliers grows large, the estimation errors undergo a first order
phase transition.Comment: 24 pages, 7 figures (minor extensions added
Correlated patterns in non-monotonic graded-response perceptrons
The optimal capacity of graded-response perceptrons storing biased and
spatially correlated patterns with non-monotonic input-output relations is
studied. It is shown that only the structure of the output patterns is
important for the overall performance of the perceptrons.Comment: 4 pages, 4 figure
Statistical Mechanics of Support Vector Networks
Using methods of Statistical Physics, we investigate the generalization
performance of support vector machines (SVMs), which have been recently
introduced as a general alternative to neural networks. For nonlinear
classification rules, the generalization error saturates on a plateau, when the
number of examples is too small to properly estimate the coefficients of the
nonlinear part. When trained on simple rules, we find that SVMs overfit only
weakly. The performance of SVMs is strongly enhanced, when the distribution of
the inputs has a gap in feature space.Comment: REVTeX, 4 pages, 2 figures, accepted by Phys. Rev. Lett (typos
corrected
Phase transitions in optimal unsupervised learning
We determine the optimal performance of learning the orientation of the
symmetry axis of a set of P = alpha N points that are uniformly distributed in
all the directions but one on the N-dimensional sphere. The components along
the symmetry breaking direction, of unitary vector B, are sampled from a
mixture of two gaussians of variable separation and width. The typical optimal
performance is measured through the overlap Ropt=B.J* where J* is the optimal
guess of the symmetry breaking direction. Within this general scenario, the
learning curves Ropt(alpha) may present first order transitions if the clusters
are narrow enough. Close to these transitions, high performance states can be
obtained through the minimization of the corresponding optimal potential,
although these solutions are metastable, and therefore not learnable, within
the usual bayesian scenario.Comment: 9 pages, 8 figures, submitted to PRE, This new version of the paper
contains one new section, Bayesian versus optimal solutions, where we explain
in detail the results supporting our claim that bayesian learning may not be
optimal. Figures 4 of the first submission was difficult to understand. We
replaced it by two new figures (Figs. 4 and 5 in this new version) containing
more detail
On the center of mass of Ising vectors
We show that the center of mass of Ising vectors that obey some simple
constraints, is again an Ising vector.Comment: 8 pages, 3 figures, LaTeX; Claims in connection with disordered
systems have been withdrawn; More detailed description of the simulations;
Inset added to figure
- …