132,317 research outputs found
3D Face Morphing Attacks: Generation, Vulnerability and Detection
Face Recognition systems (FRS) have been found to be vulnerable to morphing
attacks, where the morphed face image is generated by blending the face images
from contributory data subjects. This work presents a novel direction for
generating face-morphing attacks in 3D. To this extent, we introduced a novel
approach based on blending 3D face point clouds corresponding to contributory
data subjects. The proposed method generates 3D face morphing by projecting the
input 3D face point clouds onto depth maps and 2D color images, followed by
image blending and wrapping operations performed independently on the color
images and depth maps. We then back-projected the 2D morphing color map and the
depth map to the point cloud using the canonical (fixed) view. Given that the
generated 3D face morphing models will result in holes owing to a single
canonical view, we have proposed a new algorithm for hole filling that will
result in a high-quality 3D face morphing model. Extensive experiments were
conducted on the newly generated 3D face dataset comprising 675 3D scans
corresponding to 41 unique data subjects and a publicly available database
(Facescape) with 100 data subjects. Experiments were performed to benchmark the
vulnerability of the {proposed 3D morph-generation scheme against} automatic
2D, 3D FRS, and human observer analysis. We also presented a quantitative
assessment of the quality of the generated 3D face-morphing models using eight
different quality metrics. Finally, we propose three different 3D face Morphing
Attack Detection (3D-MAD) algorithms to benchmark the performance of 3D face
morphing attack detection techniques.Comment: The paper is accepted at IEEE Transactions on Biometrics, Behavior
and Identity Scienc
A Neural Model of Visually Guided Steering, Obstacle Avoidance, and Route Selection
A neural model is developed to explain how humans can approach a goal object on foot while steering around obstacles to avoid collisions in a cluttered environment. The model uses optic flow from a 3D virtual reality environment to determine the position of objects based on motion discontinuities, and computes heading direction, or the direction of self-motion, from global optic flow. The cortical representation of heading interacts with the representations of a goal and obstacles such that the goal acts as an attractor of heading, while obstacles act as repellers. In addition the model maintains fixation on the goal object by generating smooth pursuit eye movements. Eye rotations can distort the optic flow field, complicating heading perception, and the model uses extraretinal signals to correct for this distortion and accurately represent heading. The model explains how motion processing mechanisms in cortical areas MT, MST, and posterior parietal cortex can be used to guide steering. The model quantitatively simulates human psychophysical data about visually-guided steering, obstacle avoidance, and route selection.Air Force Office of Scientific Research (F4960-01-1-0397); National Geospatial-Intelligence Agency (NMA201-01-1-2016); National Science Foundation (SBE-0354378); Office of Naval Research (N00014-01-1-0624
- …