376 research outputs found
Semantic learning in autonomously active recurrent neural networks
The human brain is autonomously active, being characterized by a
self-sustained neural activity which would be present even in the absence of
external sensory stimuli. Here we study the interrelation between the
self-sustained activity in autonomously active recurrent neural nets and
external sensory stimuli.
There is no a priori semantical relation between the influx of external
stimuli and the patterns generated internally by the autonomous and ongoing
brain dynamics. The question then arises when and how are semantic correlations
between internal and external dynamical processes learned and built up?
We study this problem within the paradigm of transient state dynamics for the
neural activity in recurrent neural nets, i.e. for an autonomous neural
activity characterized by an infinite time-series of transiently stable
attractor states. We propose that external stimuli will be relevant during the
sensitive periods, {\it viz} the transition period between one transient state
and the subsequent semi-stable attractor. A diffusive learning signal is
generated unsupervised whenever the stimulus influences the internal dynamics
qualitatively.
For testing we have presented to the model system stimuli corresponding to
the bars and stripes problem. We found that the system performs a non-linear
independent component analysis on its own, being continuously and autonomously
active. This emergent cognitive capability results here from a general
principle for the neural dynamics, the competition between neural ensembles.Comment: Journal of Algorithms in Cognition, Informatics and Logic, special
issue on `Perspectives and Challenges for Recurrent Neural Networks', in
pres
Reinforcement learning or active inference?
This paper questions the need for reinforcement learning or control theory when optimising behaviour. We show that it is fairly simple to teach an agent complicated and adaptive behaviours using a free-energy formulation of perception. In this formulation, agents adjust their internal states and sampling of the environment to minimize their free-energy. Such agents learn causal structure in the environment and sample it in an adaptive and self-supervised fashion. This results in behavioural policies that reproduce those optimised by reinforcement learning and dynamic programming. Critically, we do not need to invoke the notion of reward, value or utility. We illustrate these points by solving a benchmark problem in dynamic programming; namely the mountain-car problem, using active perception or inference under the free-energy principle. The ensuing proof-of-concept may be important because the free-energy formulation furnishes a unified account of both action and perception and may speak to a reappraisal of the role of dopamine in the brain
Fully Trainable and Interpretable Non-Local Sparse Models for Image Restoration
Non-local self-similarity and sparsity principles have proven to be powerful
priors for natural image modeling. We propose a novel differentiable relaxation
of joint sparsity that exploits both principles and leads to a general
framework for image restoration which is (1) trainable end to end, (2) fully
interpretable, and (3) much more compact than competing deep learning
architectures. We apply this approach to denoising, jpeg deblocking, and
demosaicking, and show that, with as few as 100K parameters, its performance on
several standard benchmarks is on par or better than state-of-the-art methods
that may have an order of magnitude or more parameters.Comment: ECCV 202
Human identification via unsupervised feature learning from UWB radar data
This paper presents an automated approach to automatically distinguish the identity of multiple residents in smart homes. Without using any intrusive video surveillance devices or wearable tags, we achieve the goal of human identification through properly processing and analyzing the received signals from the ultra-wideband (UWB) radar installed in indoor environments. Because the UWB signals are very noisy and unstable, we employ unsupervised feature learning techniques to automatically learn local, discriminative features that can incorporate intra-class variations of the same identity, and yet reflect differences in distinguishing different human identities. The learned features are then used to train an SVM classifier and recognize the identity of residents. We validate our proposed solution via extensive experiments using real data collected in real-life situations. Our findings show that feature learning based on K-means clustering, coupled with whitening and pooling, achieves the highest accuracy, when only limited training data is available. This shows that the proposed feature learning and classification framework combined with the UWB radar technology provides an effective solution to human identification in multi-residential smart homes
Parametric study of EEG sensitivity to phase noise during face processing
<b>Background: </b>
The present paper examines the visual processing speed of complex objects, here faces, by mapping the relationship between object physical properties and single-trial brain responses. Measuring visual processing speed is challenging because uncontrolled physical differences that co-vary with object categories might affect brain measurements, thus biasing our speed estimates. Recently, we demonstrated that early event-related potential (ERP) differences between faces and objects are preserved even when images differ only in phase information, and amplitude spectra are equated across image categories. Here, we use a parametric design to study how early ERP to faces are shaped by phase information. Subjects performed a two-alternative force choice discrimination between two faces (Experiment 1) or textures (two control experiments). All stimuli had the same amplitude spectrum and were presented at 11 phase noise levels, varying from 0% to 100% in 10% increments, using a linear phase interpolation technique. Single-trial ERP data from each subject were analysed using a multiple linear regression model.
<b>Results: </b>
Our results show that sensitivity to phase noise in faces emerges progressively in a short time window between the P1 and the N170 ERP visual components. The sensitivity to phase noise starts at about 120–130 ms after stimulus onset and continues for another 25–40 ms. This result was robust both within and across subjects. A control experiment using pink noise textures, which had the same second-order statistics as the faces used in Experiment 1, demonstrated that the sensitivity to phase noise observed for faces cannot be explained by the presence of global image structure alone. A second control experiment used wavelet textures that were matched to the face stimuli in terms of second- and higher-order image statistics. Results from this experiment suggest that higher-order statistics of faces are necessary but not sufficient to obtain the sensitivity to phase noise function observed in response to faces.
<b>Conclusion: </b>
Our results constitute the first quantitative assessment of the time course of phase information processing by the human visual brain. We interpret our results in a framework that focuses on image statistics and single-trial analyses
Catalyzing next-generation Artificial Intelligence through NeuroAI
Neuroscience has long been an essential driver of progress in artificial intelligence (AI). We propose that to accelerate progress in AI, we must invest in fundamental research in NeuroAI. A core component of this is the embodied Turing test, which challenges AI animal models to interact with the sensorimotor world at skill levels akin to their living counterparts. The embodied Turing test shifts the focus from those capabilities like game playing and language that are especially well-developed or uniquely human to those capabilities - inherited from over 500 million years of evolution - that are shared with all animals. Building models that can pass the embodied Turing test will provide a roadmap for the next generation of AI
Consequences of converting graded to action potentials upon neural information coding and energy efficiency
Information is encoded in neural circuits using both graded and action potentials, converting between them within single neurons and successive processing layers. This conversion is accompanied by information loss and a drop in energy efficiency. We investigate the biophysical causes of this loss of information and efficiency by comparing spiking neuron models, containing stochastic voltage-gated Na+ and K+ channels, with generator potential and graded potential models lacking voltage-gated Na+ channels. We identify three causes of information loss in the generator potential that are the by-product of action potential generation: (1) the voltage-gated Na+ channels necessary for action potential generation increase intrinsic noise and (2) introduce non-linearities, and (3) the finite duration of the action potential creates a ‘footprint’ in the generator potential that obscures incoming signals. These three processes reduce information rates by ~50% in generator potentials, to ~3 times that of spike trains. Both generator potentials and graded potentials consume almost an order of magnitude less energy per second than spike trains. Because of the lower information rates of generator potentials they are substantially less energy efficient than graded potentials. However, both are an order of magnitude more efficient than spike trains due to the higher energy costs and low information content of spikes, emphasizing that there is a two-fold cost of converting analogue to digital; information loss and cost inflation
1/f2 Characteristics and Isotropy in the Fourier Power Spectra of Visual Art, Cartoons, Comics, Mangas, and Different Categories of Photographs
Art images and natural scenes have in common that their radially averaged (1D) Fourier spectral power falls according to a power-law with increasing spatial frequency (1/f2 characteristics), which implies that the power spectra have scale-invariant properties. In the present study, we show that other categories of man-made images, cartoons and graphic novels (comics and mangas), have similar properties. Further on, we extend our investigations to 2D power spectra. In order to determine whether the Fourier power spectra of man-made images differed from those of other categories of images (photographs of natural scenes, objects, faces and plants and scientific illustrations), we analyzed their 2D power spectra by principal component analysis. Results indicated that the first fifteen principal components allowed a partial separation of the different image categories. The differences between the image categories were studied in more detail by analyzing whether the mean power and the slope of the power gradients from low to high spatial frequencies varied across orientations in the power spectra. Mean power was generally higher in cardinal orientations both in real-world photographs and artworks, with no systematic difference between the two types of images. However, the slope of the power gradients showed a lower degree of mean variability across spectral orientations (i.e., more isotropy) in art images, cartoons and graphic novels than in photographs of comparable subject matters. Taken together, these results indicate that art images, cartoons and graphic novels possess relatively uniform 1/f2 characteristics across all orientations. In conclusion, the man-made stimuli studied, which were presumably produced to evoke pleasant and/or enjoyable visual perception in human observers, form a subset of all images and share statistical properties in their Fourier power spectra. Whether these properties are necessary or sufficient to induce aesthetic perception remains to be investigated
Age-related delay in information accrual for faces: Evidence from a parametric, single-trial EEG approach
Background: In this study, we quantified age-related changes in the time-course of face processing
by means of an innovative single-trial ERP approach. Unlike analyses used in previous studies, our
approach does not rely on peak measurements and can provide a more sensitive measure of
processing delays. Young and old adults (mean ages 22 and 70 years) performed a non-speeded
discrimination task between two faces. The phase spectrum of these faces was manipulated
parametrically to create pictures that ranged between pure noise (0% phase information) and the
undistorted signal (100% phase information), with five intermediate steps.
Results: Behavioural 75% correct thresholds were on average lower, and maximum accuracy was
higher, in younger than older observers. ERPs from each subject were entered into a single-trial
general linear regression model to identify variations in neural activity statistically associated with
changes in image structure. The earliest age-related ERP differences occurred in the time window
of the N170. Older observers had a significantly stronger N170 in response to noise, but this age
difference decreased with increasing phase information. Overall, manipulating image phase
information had a greater effect on ERPs from younger observers, which was quantified using a
hierarchical modelling approach. Importantly, visual activity was modulated by the same stimulus
parameters in younger and older subjects. The fit of the model, indexed by R2, was computed at
multiple post-stimulus time points. The time-course of the R2 function showed a significantly slower
processing in older observers starting around 120 ms after stimulus onset. This age-related delay
increased over time to reach a maximum around 190 ms, at which latency younger observers had
around 50 ms time lead over older observers.
Conclusion: Using a component-free ERP analysis that provides a precise timing of the visual
system sensitivity to image structure, the current study demonstrates that older observers
accumulate face information more slowly than younger subjects. Additionally, the N170 appears to
be less face-sensitive in older observers
- …