5,598 research outputs found
An empirical methodology for developing stockmarket trading systems using artificial neural networks
STNet: Selective Tuning of Convolutional Networks for Object Localization
Visual attention modeling has recently gained momentum in developing visual
hierarchies provided by Convolutional Neural Networks. Despite recent successes
of feedforward processing on the abstraction of concepts form raw images, the
inherent nature of feedback processing has remained computationally
controversial. Inspired by the computational models of covert visual attention,
we propose the Selective Tuning of Convolutional Networks (STNet). It is
composed of both streams of Bottom-Up and Top-Down information processing to
selectively tune the visual representation of Convolutional networks. We
experimentally evaluate the performance of STNet for the weakly-supervised
localization task on the ImageNet benchmark dataset. We demonstrate that STNet
not only successfully surpasses the state-of-the-art results but also generates
attention-driven class hypothesis maps
Solving constraint-satisfaction problems with distributed neocortical-like neuronal networks
Finding actions that satisfy the constraints imposed by both external inputs
and internal representations is central to decision making. We demonstrate that
some important classes of constraint satisfaction problems (CSPs) can be solved
by networks composed of homogeneous cooperative-competitive modules that have
connectivity similar to motifs observed in the superficial layers of neocortex.
The winner-take-all modules are sparsely coupled by programming neurons that
embed the constraints onto the otherwise homogeneous modular computational
substrate. We show rules that embed any instance of the CSPs planar four-color
graph coloring, maximum independent set, and Sudoku on this substrate, and
provide mathematical proofs that guarantee these graph coloring problems will
convergence to a solution. The network is composed of non-saturating linear
threshold neurons. Their lack of right saturation allows the overall network to
explore the problem space driven through the unstable dynamics generated by
recurrent excitation. The direction of exploration is steered by the constraint
neurons. While many problems can be solved using only linear inhibitory
constraints, network performance on hard problems benefits significantly when
these negative constraints are implemented by non-linear multiplicative
inhibition. Overall, our results demonstrate the importance of instability
rather than stability in network computation, and also offer insight into the
computational role of dual inhibitory mechanisms in neural circuits.Comment: Accepted manuscript, in press, Neural Computation (2018
Letter to the Editor: Physiological Interpretation of the Self-Organizing Map Algorithm
Air Force Office of Scientific Research (F49620-92-J-0499); Office of Naval Research (N00014-92-J-4015, N00014-91-J-4100
Information theory, complexity and neural networks
Some of the main results in the mathematical evaluation of neural networks as information processing systems are discussed. The basic operation of feedback and feed-forward neural networks is described. Their memory capacity and computing power are considered. The concept of learning by example as it applies to neural networks is examined
- …