226 research outputs found
A wavelet-based CMAC for enhanced multidimensional learning
The CMAC (Cerebellar Model Articulation Controller) neural network has been successfully used in control systems and other applications for many years. The network structure is modular and associative, allowing for rapid learning convergence with an ease of implementation in either hardware or software. The rate of convergence of the network is determined largely by the choice of the receptive field shape and the generalization parameter. This research contains a rigorous analysis of the rate of convergence with the standard CMAC, as well as the rate of convergence of networks using other receptive field shape. The effects of decimation from state-space to weight space are examined in detail. This analysis shows CMAC to be an adaptive lowpass filter, where the filter dynamics are governed by the generalization parameter. A more general CMAC is derived using wavelet-based receptive fields and a controllable decimation scheme, that is capable of convergence at any frequency within the Nyquist limits. The flexible decimation structure facilitates the optimization of computation for complex multidimensional problems. The stability of the wavelet-based CMAC is also examined
A Theory of Networks for Appxoimation and Learning
Learning an input-output mapping from a set of examples, of the type that many neural networks have been constructed to perform, can be regarded as synthesizing an approximation of a multi-dimensional function, that is solving the problem of hypersurface reconstruction. From this point of view, this form of learning is closely related to classical approximation techniques, such as generalized splines and regularization theory. This paper considers the problems of an exact representation and, in more detail, of the approximation of linear and nolinear mappings in terms of simpler functions of fewer variables. Kolmogorov's theorem concerning the representation of functions of several variables in terms of functions of one variable turns out to be almost irrelevant in the context of networks for learning. We develop a theoretical framework for approximation based on regularization techniques that leads to a class of three-layer networks that we call Generalized Radial Basis Functions (GRBF), since they are mathematically related to the well-known Radial Basis Functions, mainly used for strict interpolation tasks. GRBF networks are not only equivalent to generalized splines, but are also closely related to pattern recognition methods such as Parzen windows and potential functions and to several neural network algorithms, such as Kanerva's associative memory, backpropagation and Kohonen's topology preserving map. They also have an interesting interpretation in terms of prototypes that are synthesized and optimally combined during the learning stage. The paper introduces several extensions and applications of the technique and discusses intriguing analogies with neurobiological data
Integrating Statistical and Machine Learning Approaches to Identify Receptive Field Structure in Neural Populations
Neurons can code for multiple variables simultaneously and neuroscientists
are often interested in classifying neurons based on their receptive field
properties. Statistical models provide powerful tools for determining the
factors influencing neural spiking activity and classifying individual neurons.
However, as neural recording technologies have advanced to produce simultaneous
spiking data from massive populations, classical statistical methods often lack
the computational efficiency required to handle such data. Machine learning
(ML) approaches are known for enabling efficient large scale data analyses;
however, they typically require massive training sets with balanced data, along
with accurate labels to fit well. Additionally, model assessment and
interpretation are often more challenging for ML than for classical statistical
methods. To address these challenges, we develop an integrated framework,
combining statistical modeling and machine learning approaches to identify the
coding properties of neurons from large populations. In order to demonstrate
this framework, we apply these methods to data from a population of neurons
recorded from rat hippocampus to characterize the distribution of spatial
receptive fields in this region
Combining case based reasoning with neural networks
This paper presents a neural network based technique for mapping problem situations to problem solutions for Case-Based Reasoning (CBR) applications. Both neural networks and
CBR are instance-based learning techniques, although neural nets work with numerical data and CBR systems work with symbolic data. This paper discusses how the application scope of both paradigms could be enhanced by the use of hybrid concepts. To make the use of neural networks possible, the problem's situation and solution features are transformed into continuous features, using techniques similar to CBR's definition of similarity metrics. Radial Basis Function (RBF) neural nets are used to create a multivariable, continuous input-output mapping. As the mapping is continuous, this technique also provides generalisation between cases, replacing the domain specific
solution adaptation techniques required by conventional CBR. This continuous representation also allows, as in
fuzzy logic, an associated membership measure to be output with each symbolic feature, aiding the prioritisation of various possible solutions. A further advantage is that, as the RBF neurons are only active in a limited area of the input space, the solution can be accompanied by local estimates of accuracy, based on the sufficiency of the cases present in that area as well as the results measured during testing. We describe how the application of this technique could be of benefit to the real world problem of sales advisory systems, among others
A survey of visual preprocessing and shape representation techniques
Many recent theories and methods proposed for visual preprocessing and shape representation are summarized. The survey brings together research from the fields of biology, psychology, computer science, electrical engineering, and most recently, neural networks. It was motivated by the need to preprocess images for a sparse distributed memory (SDM), but the techniques presented may also prove useful for applying other associative memories to visual pattern recognition. The material of this survey is divided into three sections: an overview of biological visual processing; methods of preprocessing (extracting parts of shape, texture, motion, and depth); and shape representation and recognition (form invariance, primitives and structural descriptions, and theories of attention)
I-theory on depth vs width: hierarchical function composition
Deep learning networks with convolution, pooling and subsampling are a special case of hierar- chical architectures, which can be represented by trees (such as binary trees). Hierarchical as well as shallow networks can approximate functions of several variables, in particular those that are com- positions of low dimensional functions. We show that the power of a deep network architecture with respect to a shallow network is rather independent of the specific nonlinear operations in the network and depends instead on the the behavior of the VC-dimension. A shallow network can approximate compositional functions with the same error of a deep network but at the cost of a VC-dimension that is exponential instead than quadratic in the dimensionality of the function. To complete the argument we argue that there exist visual computations that are intrinsically compositional. In particular, we prove that recognition invariant to translation cannot be computed by shallow networks in the presence of clutter. Finally, a general framework that includes the compositional case is sketched. The key con- dition that allows tall, thin networks to be nicer that short, fat networks is that the target input-output function must be sparse in a certain technical sense.This work was supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF - 1231216
Diagnosis of Malignant Melanoma using a Neural Network
Malignant melanoma is the deadliest form of all skin cancers. Approximately 32,000 new cases of malignant melanoma were diagnosed in 1991, with approximately 80 percent of patients expected to survive five years [1], Fortunately, if detected early, even malignant melanoma may be treated successfully. Thus, in recent years, there has been a rising interest in the automated detection and diagnosis of skin cancer, particularly malignant melanoma [2]. In this thesis, a novel neural network approach for the automated distinction of melanoma from three benign categories of tumors which exhibit melanoma-like characteristics is presented. The approach is based on devising new and discriminant features which are used as inputs to an artificial neural network for classification of tumor images as malignant or benign. Promising results have been obtained using this method on real skin cancer images
ANN Models and Bayesian Spline Models for Analysis of Exchange Rates and Gold Price
ANN (Artificial Neural Network) models and Spline techniques have been applied to
economic analysis, to handle economic problems, evaluate portfolio risk and stock
performance, and to forecast stock exchange rates and gold prices. These techniques are
improving nowadays and continue to serve as powerful predictive tools.
In this study, we compare the performance of ANN models and Bayesian Spline models
in forecasting economic datasets. We consider the most commonly used ANN models,
which are Generalized Regression Neural Networks (GRNN), Multilayer Perceptron
(MLP), and Radial Basis Function Neural Networks (RBFNN). We compare these
models using BayesX and Statistica software with three important economic datasets: on
the exchange rate of Turkish Liras (TL) to Euro, exchange rate of Turkish Liras (TL) to
United States Dollars (USD), and Gold Price for Turkey. With these three economic
datasets, we made a comparative study of these models, using the criterions MSE and
MAPE to evaluate their forecasting performance. The results demonstrate that the
penalized spline model performed best amongst the spline techniques and their Bayesian
versions. Amongst the ANN models, the MLP model obtained the best performance
criterion results
- …