374,479 research outputs found
Modelling Pattern Recognition in Cricket Phonotaxis
A spiking neuron implementation of pattern recognition of the calling songs in
Gryllus bimaculatus is proposed. A simplified model of the auditory
interneuron AN1 has been fitted to extracellular physiological data. The model
captures the aspects of AN1’s rate-response to acoustic stimulation which are
believed to be sufficient for pattern recognition. Stimulation patterns can be
induced into the model via current injecton of the signals envelope-shapes.
The model was used as the input stage to the pattern recognition mechanisms. A
biologically plausible filter mechanism for pulse-pause patterns is proposed
which is based on short term synaptic plasticity. Three simple filter
mechanism are described, based on either isolated synaptic depression or
synaptic facilitation. These filters are able to reproduce physiological
findings from the cricket’s auditory brain neurons. Further, it is argued that
more complex filters can be produced by using combinations of depression and
facilitation, and that a complete model of the cricket’s pattern recognition
apparatus may be implemented in this way. This however is left as a subject of
further studies
A Reverse Hierarchy Model for Predicting Eye Fixations
A number of psychological and physiological evidences suggest that early
visual attention works in a coarse-to-fine way, which lays a basis for the
reverse hierarchy theory (RHT). This theory states that attention propagates
from the top level of the visual hierarchy that processes gist and abstract
information of input, to the bottom level that processes local details.
Inspired by the theory, we develop a computational model for saliency detection
in images. First, the original image is downsampled to different scales to
constitute a pyramid. Then, saliency on each layer is obtained by image
super-resolution reconstruction from the layer above, which is defined as
unpredictability from this coarse-to-fine reconstruction. Finally, saliency on
each layer of the pyramid is fused into stochastic fixations through a
probabilistic model, where attention initiates from the top layer and
propagates downward through the pyramid. Extensive experiments on two standard
eye-tracking datasets show that the proposed method can achieve competitive
results with state-of-the-art models.Comment: CVPR 2014, 27th IEEE Conference on Computer Vision and Pattern
Recognition (CVPR). CVPR 201
A Leaf Recognition Algorithm for Plant Classification Using Probabilistic Neural Network
In this paper, we employ Probabilistic Neural Network (PNN) with image and
data processing techniques to implement a general purpose automated leaf
recognition algorithm. 12 leaf features are extracted and orthogonalized into 5
principal variables which consist the input vector of the PNN. The PNN is
trained by 1800 leaves to classify 32 kinds of plants with an accuracy greater
than 90%. Compared with other approaches, our algorithm is an accurate
artificial intelligence approach which is fast in execution and easy in
implementation.Comment: 6 pages, 3 figures, 2 table
A frequency-selective feedback model of auditory efferent suppression and its implications for the recognition of speech in noise
The potential contribution of the peripheral auditory efferent system to our understanding of speech in a background of competing noise was studied using a computer model of the auditory periphery and assessed using an automatic speech recognition system. A previous study had shown that a fixed efferent attenuation applied to all channels of a multi-channel model could improve the recognition of connected digit triplets in noise [G. J. Brown, R. T. Ferry, and R. Meddis, J. Acoust. Soc. Am. 127, 943?954 (2010)]. In the current study an anatomically justified feedback loop was used to automatically regulate separate attenuation values for each auditory channel. This arrangement resulted in a further enhancement of speech recognition over fixed-attenuation conditions. Comparisons between multi-talker babble and pink noise interference conditions suggest that the benefit originates from the model?s ability to modify the amount of suppression in each channel separately according to the spectral shape of the interfering sounds
Neurophysiological Assessment of Affective Experience
In the field of Affective Computing the affective experience (AX) of the user during the interaction with computers is of great interest. The automatic recognition of the affective state, or emotion, of the user is one of the big challenges. In this proposal I focus on the affect recognition via physiological and neurophysiological signals. Long‐standing evidence from psychophysiological research and more recently from research in affective neuroscience suggests that both, body and brain physiology, are able to indicate the current affective state of a subject. However, regarding the classification of AX several questions are still unanswered. The principal possibility of AX classification was repeatedly shown, but its generalisation over different task contexts, elicitating stimuli modalities, subjects or time is seldom addressed. In this proposal I will discuss a possible agenda for the further exploration of physiological and neurophysiological correlates of AX over different elicitation modalities and task contexts
Prerequisites for Affective Signal Processing (ASP)
Although emotions are embraced by science, their recognition has not reached a satisfying level. Through a concise overview of affect, its signals, features, and classification methods, we provide understanding for the problems encountered. Next, we identify the prerequisites for successful Affective Signal Processing: validation (e.g., mapping of constructs on signals), triangulation, a physiology-driven approach, and contributions of the signal processing community. Using these directives, a critical analysis of a real-world case is provided. This illustrates that the prerequisites can become a valuable guide for Affective Signal Processing (ASP)
Biometrics for Emotion Detection (BED): Exploring the combination of Speech and ECG
The paradigm Biometrics for Emotion Detection (BED) is introduced, which enables unobtrusive emotion recognition, taking into account varying environments. It uses the electrocardiogram (ECG) and speech, as a powerful but rarely used combination to unravel people’s emotions. BED was applied in two environments (i.e., office and home-like) in which 40 people watched 6 film scenes. It is shown that both heart rate variability (derived from the ECG) and, when people’s gender is taken into account, the standard deviation of the fundamental frequency of speech indicate people’s experienced emotions. As such, these measures validate each other. Moreover, it is found that people’s environment can indeed of influence experienced emotions. These results indicate that BED might become an important paradigm for unobtrusive emotion detection
- …
