436 research outputs found

    Effect of Input Noise and Output Node Stochastic on Wang's k WTA

    Get PDF

    Analysis and design of a distributed k-winners-take-all model

    Get PDF
    The -winners-take-all (WTA) problem is to find the largest inputs from inputs. In this paper, we design and propose a novel distributed WTA model, for which no central unit is needed to realize the computation of the winners. As a result, the proposed model has the general advantages of distributed models over centralized ones, such as better robustness to faults of agents. The global asymptotic convergence of the proposed distributed model is proven. Besides, two numerical examples on networks of agents with static inputs and time-varying inputs are presented to validate the performance of the proposed model

    Random Projection in the Brain and Computation with Assemblies of Neurons

    Get PDF
    It has been recently shown via simulations [Dasgupta et al., 2017] that random projection followed by a cap operation (setting to one the k largest elements of a vector and everything else to zero), a map believed to be an important part of the insect olfactory system, has strong locality sensitivity properties. We calculate the asymptotic law whereby the overlap in the input vectors is conserved, verifying mathematically this empirical finding. We then focus on the far more complex homologous operation in the mammalian brain, the creation through successive projections and caps of an assembly (roughly, a set of excitatory neurons representing a memory or concept) in the presence of recurrent synapses and plasticity. After providing a careful definition of assemblies, we prove that the operation of assembly projection converges with high probability, over the randomness of synaptic connectivity, even if plasticity is relatively small (previous proofs relied on high plasticity). We also show that assembly projection has itself some locality preservation properties. Finally, we propose a large repertoire of assembly operations, including associate, merge, reciprocal project, and append, each of them both biologically plausible and consistent with what we know from experiments, and show that this computational system is capable of simulating, again with high probability, arbitrary computation in a quite natural way. We hope that this novel way of looking at brain computation, open-ended and based on reasonably mainstream ideas in neuroscience, may prove an attractive entry point for computer scientists to work on understanding the brain

    Computation in Dynamically Bounded Asymmetric Systems

    Get PDF
    Previous explanations of computations performed by recurrent networks have focused on symmetrically connected saturating neurons and their convergence toward attractors. Here we analyze the behavior of asymmetrical connected networks of linear threshold neurons, whose positive response is unbounded. We show that, for a wide range of parameters, this asymmetry brings interesting and computationally useful dynamical properties. When driven by input, the network explores potential solutions through highly unstable ‘expansion’ dynamics. This expansion is steered and constrained by negative divergence of the dynamics, which ensures that the dimensionality of the solution space continues to reduce until an acceptable solution manifold is reached. Then the system contracts stably on this manifold towards its final solution trajectory. The unstable positive feedback and cross inhibition that underlie expansion and divergence are common motifs in molecular and neuronal networks. Therefore we propose that very simple organizational constraints that combine these motifs can lead to spontaneous computation and so to the spontaneous modification of entropy that is characteristic of living systems

    Active Sensing of Robot Arms Based on Zeroing Neural Networks: A Biological-Heuristic Optimization Model

    Get PDF
    Conventional biological-heuristic solutions via zeroing neural network (ZNN) models have achieved preliminary efficiency on time-dependent nonlinear optimization problems handling. However, the investigation on finding a feasible ZNN model to solve the time-dependent nonlinear optimization problems with both inequality and equality constraints still remains stagnant because of the nonlinearity and complexity. To make new progresses on the ZNN for time-dependent nonlinear optimization problems solving, this paper proposes a biological-heuristic optimization model, i.e., inequality and equality constrained optimization ZNN (IECO-ZNN). Such a proposed IECO-ZNN breaks the conditionality that the solutions via ZNN for solving nonlinear optimization problems can not consider the inequality and equality constraints at the same time. The time-dependent nonlinear optimization problem subject to inequality and equality constraints is skillfully converted to a time-dependent equality system by exploiting the Lagrange multiplier rule. The design process for the IECO-ZNN model is presented together with its new architecture illustrated in details. In addition, the conversion equivalence, global stability as well as exponential convergence property are theoretically proven. Moreover, numerical studies, real-world applications to robot arm active sensing, and comparisons sufficiently verify the effectiveness and superiority of the proposed IECO-ZNN model for the time-dependent nonlinear optimization with inequality and equality constraints

    Continuous-time recurrent neural networks for quadratic programming: theory and engineering applications.

    Get PDF
    Liu Shubao.Thesis (M.Phil.)--Chinese University of Hong Kong, 2005.Includes bibliographical references (leaves 90-98).Abstracts in English and Chinese.Abstract --- p.i摘要 --- p.iiiAcknowledgement --- p.ivChapter 1 --- Introduction --- p.1Chapter 1.1 --- Time-Varying Quadratic Optimization --- p.1Chapter 1.2 --- Recurrent Neural Networks --- p.3Chapter 1.2.1 --- From Feedforward to Recurrent Networks --- p.3Chapter 1.2.2 --- Computational Power and Complexity --- p.6Chapter 1.2.3 --- Implementation Issues --- p.7Chapter 1.3 --- Thesis Organization --- p.9Chapter I --- Theory and Models --- p.11Chapter 2 --- Linearly Constrained QP --- p.13Chapter 2.1 --- Model Description --- p.14Chapter 2.2 --- Convergence Analysis --- p.17Chapter 3 --- Quadratically Constrained QP --- p.26Chapter 3.1 --- Problem Formulation --- p.26Chapter 3.2 --- Model Description --- p.27Chapter 3.2.1 --- Model 1 (Dual Model) --- p.28Chapter 3.2.2 --- Model 2 (Improved Dual Model) --- p.28Chapter II --- Engineering Applications --- p.29Chapter 4 --- KWTA Network Circuit Design --- p.31Chapter 4.1 --- Introduction --- p.31Chapter 4.2 --- Equivalent Reformulation --- p.32Chapter 4.3 --- KWTA Network Model --- p.36Chapter 4.4 --- Simulation Results --- p.40Chapter 4.5 --- Conclusions --- p.40Chapter 5 --- Dynamic Control of Manipulators --- p.43Chapter 5.1 --- Introduction --- p.43Chapter 5.2 --- Problem Formulation --- p.44Chapter 5.3 --- Simplified Dual Neural Network --- p.47Chapter 5.4 --- Simulation Results --- p.51Chapter 5.5 --- Concluding Remarks --- p.55Chapter 6 --- Robot Arm Obstacle Avoidance --- p.56Chapter 6.1 --- Introduction --- p.56Chapter 6.2 --- Obstacle Avoidance Scheme --- p.58Chapter 6.2.1 --- Equality Constrained Formulation --- p.58Chapter 6.2.2 --- Inequality Constrained Formulation --- p.60Chapter 6.3 --- Simplified Dual Neural Network Model --- p.64Chapter 6.3.1 --- Existing Approaches --- p.64Chapter 6.3.2 --- Model Derivation --- p.65Chapter 6.3.3 --- Convergence Analysis --- p.67Chapter 6.3.4 --- Model Comparision --- p.69Chapter 6.4 --- Simulation Results --- p.70Chapter 6.5 --- Concluding Remarks --- p.71Chapter 7 --- Multiuser Detection --- p.77Chapter 7.1 --- Introduction --- p.77Chapter 7.2 --- Problem Formulation --- p.78Chapter 7.3 --- Neural Network Architecture --- p.82Chapter 7.4 --- Simulation Results --- p.84Chapter 8 --- Conclusions and Future Works --- p.88Chapter 8.1 --- Concluding Remarks --- p.88Chapter 8.2 --- Future Prospects --- p.88Bibliography --- p.8

    Neuron Clustering for Mitigating Catastrophic Forgetting in Supervised and Reinforcement Learning

    Get PDF
    Neural networks have had many great successes in recent years, particularly with the advent of deep learning and many novel training techniques. One issue that has affected neural networks and prevented them from performing well in more realistic online environments is that of catastrophic forgetting. Catastrophic forgetting affects supervised learning systems when input samples are temporally correlated or are non-stationary. However, most real-world problems are non-stationary in nature, resulting in prolonged periods of time separating inputs drawn from different regions of the input space. Reinforcement learning represents a worst-case scenario when it comes to precipitating catastrophic forgetting in neural networks. Meaningful training examples are acquired as the agent explores different regions of its state/action space. When the agent is in one such region, only highly correlated samples from that region are typically acquired. Moreover, the regions that the agent is likely to visit will depend on its current policy, suggesting that an agent that has a good policy may avoid exploring particular regions. The confluence of these factors means that without some mitigation techniques, supervised neural networks as function approximation in temporal-difference learning will be restricted to the simplest test cases. This work explores catastrophic forgetting in neural networks in terms of supervised and reinforcement learning. A simple mathematical model is introduced to argue that catastrophic forgetting is a result of overlapping representations in the hidden layers in which updates to the weights can affect multiple unrelated regions of the input space. A novel neural network architecture, dubbed cluster-select, is introduced which utilizes online clustering for the selection of a subset of hidden neurons to be activated in the feedforward and backpropagation stages. Clusterselect is demonstrated to outperform leading techniques in both classification nd regression. In the context of reinforcement learning, cluster-select is studied for both fully and partially observable Markov decision processes and is demonstrated to converge faster and behave in a more stable manner when compared to other state-of-the-art algorithms

    Distributed learning in sensor networks

    Get PDF
    corecore