5 research outputs found

    Dynamics of Supervised Learning with Restricted Training Sets

    Full text link
    We study the dynamics of supervised learning in layered neural networks, in the regime where the size pp of the training set is proportional to the number NN of inputs. Here the local fields are no longer described by Gaussian probability distributions. We show how dynamical replica theory can be used to predict the evolution of macroscopic observables, including the relevant performance measures, incorporating the old formalism in the limit Ξ±=p/Nβ†’βˆž\alpha=p/N\to\infty as a special case. For simplicity we restrict ourselves to single-layer networks and realizable tasks.Comment: 36 pages, latex2e, 12 eps figures (to be publ in: Proc Newton Inst Workshop on On-Line Learning '97

    Dynamics of Supervised Learning with Restricted Training Sets and Noisy Teachers

    No full text
    We generalize a recent formalism to describe the dynamics of supervised learning in layered neural networks, in the regime where data recycling is inevitable, to the case of noisy teachers. Our theory generates predictions for the evolution in time of training- and generalization errors, and extends the class of mathematically solvable learning processes in large neural networks to those complicated situations where overfitting occurs. 1 Introduction Tools from statistical mechanics have been used successfully over the last decade to study the dynamics of learning in large layered neural networks (for a review see e.g. [1] or [2]). The simplest mathematical theories result upon assuming the data set to be much larger than the number of weight updates made, which rules out recycling and ensures that any distribution of relevance will be Gaussian. Unfortunately, this regime is also not the most relevant one, both in terms of applications of neural networks and in terms of mathematical i..
    corecore