2,750 research outputs found
Recognition of Facial Expressions by Cortical Multi-scale Line and Edge Coding
Face-to-face communications between humans involve emotions, which often are unconsciously conveyed by facial expressions and body gestures. Intelligent human-machine interfaces, for example in cognitive robotics, need to recognize emotions. This paper addresses facial expressions and their neural correlates on the basis of a model of the visual cortex: the multi-scale line and edge coding. The recognition model links the cortical representation with Paul Ekman's Action Units which are related to the different facial muscles. The model applies a top-down categorization with trends and magnitudes of displacements of the mouth and eyebrows based on expected displacements relative to a neutral expression. The happy vs. not-happy categorization yielded a. correct recognition rate of 91%, whereas final recognition of the six expressions happy, anger, disgust, fear, sadness and surprise resulted in a. rate of 78%
Recognition of facial expressions by cortical multi-scale line and edge coding
Empirical studies concerning face recognition suggest that faces may be stored in memory by a few canonical representations. Models of visual perception are based on image representations in cortical area V1 and beyond, which contain many cell layers for feature extraction.
Simple, complex and end-stopped cells provide input for line, edge and keypoint detection. Detected events provide a rich, multi-scale object representation, and this representation can be stored in memory in
order to identify objects. In this paper, the above context is applied to face recognition. The multi-scale line/edge representation is explored in conjunction with keypoint-based saliency maps for Focus-of-Attention.
Recognition rates of up to 96% were achieved by combining frontal and 3/4 views, and recognition was quite robust against partial occlusions
Face segregation and recognition by cortical multi-scale line and edge coding
Models of visual perception are based on image representations in
cortical area V1 and higher areas which contain many cell layers for feature
extraction. Basic simple, complex and end-stopped cells provide input for line,
edge and keypoint detection. In this paper we present an improved method for
multi-scale line/edge detection based on simple and complex cells. We illustrate
the line/edge representation for object reconstruction, and we present models for
multi-scale face (object) segregation and recognition that can be embedded into
feedforward dorsal and ventral data streams (the “what” and “where” subsystems)
with feedback streams from higher areas for obtaining translation, rotation
and scale invariance
Face recognition by cortical multi-scale line and edge representations
Empirical studies concerning face recognition suggest that
faces may be stored in memory by a few canonical representations. Models
of visual perception are based on image representations in cortical
area V1 and beyond, which contain many cell layers for feature extraction.
Simple, complex and end-stopped cells provide input for line, edge
and keypoint detection. Detected events provide a rich, multi-scale object
representation, and this representation can be stored in memory in
order to identify objects. In this paper, the above context is applied to
face recognition. The multi-scale line/edge representation is explored in
conjunction with keypoint-based saliency maps for Focus-of-Attention.
Recognition rates of up to 96% were achieved by combining frontal and
3/4 views, and recognition was quite robust against partial occlusions
Recommended from our members
Background suppressing Gabor energy filtering
In the field of facial emotion recognition, early research advanced with the use of Gabor filters. However, these filters lack generalization and result in undesirably large feature vector size. In recent work, more attention has been given to other local appearance features. Two desired characteristics in a facial appearance feature are generalization capability, and the compactness of representation. In this paper, we propose a novel texture feature inspired by Gabor energy filters, called background suppressing Gabor energy filtering. The feature has a generalization component that removes background texture. It has a reduced feature vector size due to maximal representation and soft orientation histograms, and it is awhite box representation. We demonstrate improved performance on the non-trivial Audio/Visual Emotion Challenge 2012 grand-challenge dataset by a factor of 7.17 over the Gabor filter on the development set. We also demonstrate applicability of our approach beyond facial emotion recognition which yields improved classification rate over the Gabor filter for four bioimaging datasets by an average of 8.22%
A biological and real-time framework for hand gestures and head poses
Human-robot interaction is an interdisciplinary research area that aims at the development of social robots. Since social robots are expected to interact with humans and understand their behavior through gestures and body movements, cognitive psychology and robot technology must be integrated. In this paper we present a biological and real-time framework for detecting and tracking hands and heads. This framework is based on keypoints extracted by means of cortical V1 end-stopped cells. Detected keypoints and the cells’ responses are used to classify the junction type. Through the combination of annotated keypoints in a hierarchical, multi-scale tree structure, moving and deformable hands can be segregated and tracked over time. By using hand templates with lines and edges at only a few scales, a hand’s gestures can be recognized. Head tracking and pose detection are also implemented, which can be integrated with detection of facial expressions in the future. Through the combinations of head poses and hand gestures a large number of commands can be given to a robot
Multi-scale lines and edges in V1 and beyond: brightness, object categorization and recognition, and consciousness
In this paper we present an improved model for line and edge detection in cortical area V1. This model is based on responses of simple and complex cells, and it is multi-scale with no free parameters. We illustrate the use of the multi-scale line/edge representation in different processes: visual reconstruction or brightness perception, automatic scale selection and object segregation. A two-level object categorization scenario is tested in which pre-categorization is based on coarse scales only and final categorization on coarse plus fine scales. We also present a multi-scale object and face recognition model. Processing schemes are discussed in the framework of a complete cortical architecture. The fact that brightness perception and object recognition may be based on the same symbolic image representation is an indication that the entire (visual) cortex is involved in consciousness
Cortical 3D Face Recognition Framework
Empirical studies concerning face recognition suggest that faces may be stored in memory by a few canonical representations. In cortical area V1 exist double-opponent colour blobs, also simple, complex and end-stopped cells which provide input for a multiscale line/edge representation, keypoints for dynamic routing and saliency maps for Focus-of-Attention. All these combined allow us to segregate faces. Events of different facial views are stored in memory and combined in order to identify the view and recognise the face including facial expression. In this paper we show that with five 2D views and their cortical representations it is possible to determine the left-right and frontal-lateral-profile views and to achieve view-invariant recognition of 3D faces
Change blindness: eradication of gestalt strategies
Arrays of eight, texture-defined rectangles were used as stimuli in a one-shot change blindness (CB) task where there was a 50% chance that one rectangle would change orientation between two successive presentations separated by an interval. CB was eliminated by cueing the target rectangle in the first stimulus, reduced by cueing in the interval and unaffected by cueing in the second presentation. This supports the idea that a representation was formed that persisted through the interval before being 'overwritten' by the second presentation (Landman et al, 2003 Vision Research 43149–164]. Another possibility is that participants used some kind of grouping or Gestalt strategy. To test this we changed the spatial position of the rectangles in the second presentation by shifting them along imaginary spokes (by ±1 degree) emanating from the central fixation point. There was no significant difference seen in performance between this and the standard task [F(1,4)=2.565, p=0.185]. This may suggest two things: (i) Gestalt grouping is not used as a strategy in these tasks, and (ii) it gives further weight to the argument that objects may be stored and retrieved from a pre-attentional store during this task
Improved line/edge detection and visual reconstruction
Lines and edges provide important information for object categorization and recognition. In addition, one
brightness model is based on a symbolic interpretation of the cortical multi-scale line/edge representation. In
this paper we present an improved scheme for line/edge extraction from simple and complex cells and we illustrate
the multi-scale representation. This representation can be used for visual reconstruction, but also for nonphotorealistic
rendering. Together with keypoints and a new model of disparity estimation, a 3D wireframe representation
of e.g. faces can be obtained in the future
- …