972 research outputs found
Why and When Can Deep -- but Not Shallow -- Networks Avoid the Curse of Dimensionality: a Review
The paper characterizes classes of functions for which deep learning can be
exponentially better than shallow learning. Deep convolutional networks are a
special case of these conditions, though weight sharing is not the main reason
for their exponential advantage
Decision region approximation by polynomials or neural networks
We give degree of approximation results for decision regions which are defined by polynomial and neural network
parametrizations. The volume of the misclassified region is used to measure the approximation error, and results for the degree
of L1 approximation of functions are used. For polynomial parametrizations, we show that the degree of approximation is at
least 1, whereas for neural network parametrizations we prove the slightly weaker result that the degree of approximation is at least r, where r can be any number in the open interval (0, 1)
Neural network representation and learning of mappings and their derivatives
Discussed here are recent theorems proving that artificial neural networks are capable of approximating an arbitrary mapping and its derivatives as accurately as desired. This fact forms the basis for further results establishing the learnability of the desired approximations, using results from non-parametric statistics. These results have potential applications in robotics, chaotic dynamics, control, and sensitivity analysis. An example involving learning the transfer function and its derivatives for a chaotic map is discussed
Artificial Neural Networks
Artificial neural networks (ANNs) constitute a class of flexible nonlinear models designed to mimic biological neural systems. In this entry, we introduce ANN using familiar econometric terminology and provide an overview of ANN modeling approach and its implementation methods.
A Theory of Networks for Appxoimation and Learning
Learning an input-output mapping from a set of examples, of the type that many neural networks have been constructed to perform, can be regarded as synthesizing an approximation of a multi-dimensional function, that is solving the problem of hypersurface reconstruction. From this point of view, this form of learning is closely related to classical approximation techniques, such as generalized splines and regularization theory. This paper considers the problems of an exact representation and, in more detail, of the approximation of linear and nolinear mappings in terms of simpler functions of fewer variables. Kolmogorov's theorem concerning the representation of functions of several variables in terms of functions of one variable turns out to be almost irrelevant in the context of networks for learning. We develop a theoretical framework for approximation based on regularization techniques that leads to a class of three-layer networks that we call Generalized Radial Basis Functions (GRBF), since they are mathematically related to the well-known Radial Basis Functions, mainly used for strict interpolation tasks. GRBF networks are not only equivalent to generalized splines, but are also closely related to pattern recognition methods such as Parzen windows and potential functions and to several neural network algorithms, such as Kanerva's associative memory, backpropagation and Kohonen's topology preserving map. They also have an interesting interpretation in terms of prototypes that are synthesized and optimally combined during the learning stage. The paper introduces several extensions and applications of the technique and discusses intriguing analogies with neurobiological data
Elementary Derivative Tasks and Neural Net Multiscale Analysis of Tasks
Neural nets are known to be universal approximators. In particular, formal
neurons implementing wavelets have been shown to build nets able to approximate
any multidimensional task. Such very specialized formal neurons may be,
however, difficult to obtain biologically and/or industrially. In this paper we
relax the constraint of a strict ``Fourier analysis'' of tasks. Rather, we use
a finite number of more realistic formal neurons implementing elementary tasks
such as ``window'' or ``Mexican hat'' responses, with adjustable widths. This
is shown to provide a reasonably efficient, practical and robust,
multifrequency analysis. A training algorithm, optimizing the task with respect
to the widths of the responses, reveals two distinct training modes. The first
mode induces some of the formal neurons to become identical, hence promotes
``derivative tasks''. The other mode keeps the formal neurons distinct.Comment: latex neurondlt.tex, 7 files, 6 figures, 9 pages [SPhT-T01/064],
submitted to Phys. Rev.
- …