62,388 research outputs found
Fast and exact bi-directional fitting of active appearance models
Finding landmarks on objects like faces is a challenging computer vision problem, especially in real life conditions (or in-the-wild) and Active Appearance Models have been widely used to solve it. State-of-the-art algorithms for fitting an AAM to a new image are based on Gauss-Newton (GN) optimization. Recently fast GN algorithms have been proposed for both forward additive and inverse compositional fitting frameworks. In this paper, we propose a fast and exact bi-directional (Fast-Bd) approach to AAM fitting by combining both approaches. Although such a method might appear to increase computational burden, we show that by capitalizing on results from optimization theory, an exact solution, as computationally efficient as the original forward or inverse formulation, can be derived. Our proposed bi-directional approach achieves state-of-the-art performance and superior convergence properties. These findings are validated on two challenging, in-the-wild data sets, LFPW and Helen, and comparison is provided to the state-of-the art methods for Active Appearance Models fitting
A Unified Framework for Compositional Fitting of Active Appearance Models
Active Appearance Models (AAMs) are one of the most popular and
well-established techniques for modeling deformable objects in computer vision.
In this paper, we study the problem of fitting AAMs using Compositional
Gradient Descent (CGD) algorithms. We present a unified and complete view of
these algorithms and classify them with respect to three main characteristics:
i) cost function; ii) type of composition; and iii) optimization method.
Furthermore, we extend the previous view by: a) proposing a novel Bayesian cost
function that can be interpreted as a general probabilistic formulation of the
well-known project-out loss; b) introducing two new types of composition,
asymmetric and bidirectional, that combine the gradients of both image and
appearance model to derive better conver- gent and more robust CGD algorithms;
and c) providing new valuable insights into existent CGD algorithms by
reinterpreting them as direct applications of the Schur complement and the
Wiberg method. Finally, in order to encourage open research and facilitate
future comparisons with our work, we make the implementa- tion of the
algorithms studied in this paper publicly available as part of the Menpo
Project.Comment: 39 page
Optic nerve head segmentation
Reliable and efficient optic disk localization and segmentation are important tasks in automated retinal screening. General-purpose edge detection algorithms often fail to segment the optic disk due to fuzzy boundaries, inconsistent image contrast or missing edge features. This paper presents an algorithm for the localization and segmentation of the optic nerve head boundary in low-resolution images (about 20 /spl mu//pixel). Optic disk localization is achieved using specialized template matching, and segmentation by a deformable contour model. The latter uses a global elliptical model and a local deformable model with variable edge-strength dependent stiffness. The algorithm is evaluated against a randomly selected database of 100 images from a diabetic screening programme. Ten images were classified as unusable; the others were of variable quality. The localization algorithm succeeded on all bar one usable image; the contour estimation algorithm was qualitatively assessed by an ophthalmologist as having Excellent-Fair performance in 83% of cases, and performs well even on blurred image
GAGAN: Geometry-Aware Generative Adversarial Networks
Deep generative models learned through adversarial training have become
increasingly popular for their ability to generate naturalistic image textures.
However, aside from their texture, the visual appearance of objects is
significantly influenced by their shape geometry; information which is not
taken into account by existing generative models. This paper introduces the
Geometry-Aware Generative Adversarial Networks (GAGAN) for incorporating
geometric information into the image generation process. Specifically, in GAGAN
the generator samples latent variables from the probability space of a
statistical shape model. By mapping the output of the generator to a canonical
coordinate frame through a differentiable geometric transformation, we enforce
the geometry of the objects and add an implicit connection from the prior to
the generated object. Experimental results on face generation indicate that the
GAGAN can generate realistic images of faces with arbitrary facial attributes
such as facial expression, pose, and morphology, that are of better quality
than current GAN-based methods. Our method can be used to augment any existing
GAN architecture and improve the quality of the images generated
Efficient illumination independent appearance-based face tracking
One of the major challenges that visual tracking algorithms face nowadays is being
able to cope with changes in the appearance of the target during tracking. Linear
subspace models have been extensively studied and are possibly the most popular
way of modelling target appearance. We introduce a linear subspace representation
in which the appearance of a face is represented by the addition of two approxi-
mately independent linear subspaces modelling facial expressions and illumination
respectively. This model is more compact than previous bilinear or multilinear ap-
proaches. The independence assumption notably simplifies system training. We only
require two image sequences. One facial expression is subject to all possible illumina-
tions in one sequence and the face adopts all facial expressions under one particular
illumination in the other. This simple model enables us to train the system with
no manual intervention. We also revisit the problem of efficiently fitting a linear
subspace-based model to a target image and introduce an additive procedure for
solving this problem. We prove that Matthews and Baker’s Inverse Compositional
Approach makes a smoothness assumption on the subspace basis that is equiva-
lent to Hager and Belhumeur’s, which worsens convergence. Our approach differs
from Hager and Belhumeur’s additive and Matthews and Baker’s compositional ap-
proaches in that we make no smoothness assumptions on the subspace basis. In the
experiments conducted we show that the model introduced accurately represents
the appearance variations caused by illumination changes and facial expressions.
We also verify experimentally that our fitting procedure is more accurate and has
better convergence rate than the other related approaches, albeit at the expense of
a slight increase in computational cost. Our approach can be used for tracking a
human face at standard video frame rates on an average personal computer
A Comprehensive Performance Evaluation of Deformable Face Tracking "In-the-Wild"
Recently, technologies such as face detection, facial landmark localisation
and face recognition and verification have matured enough to provide effective
and efficient solutions for imagery captured under arbitrary conditions
(referred to as "in-the-wild"). This is partially attributed to the fact that
comprehensive "in-the-wild" benchmarks have been developed for face detection,
landmark localisation and recognition/verification. A very important technology
that has not been thoroughly evaluated yet is deformable face tracking
"in-the-wild". Until now, the performance has mainly been assessed
qualitatively by visually assessing the result of a deformable face tracking
technology on short videos. In this paper, we perform the first, to the best of
our knowledge, thorough evaluation of state-of-the-art deformable face tracking
pipelines using the recently introduced 300VW benchmark. We evaluate many
different architectures focusing mainly on the task of on-line deformable face
tracking. In particular, we compare the following general strategies: (a)
generic face detection plus generic facial landmark localisation, (b) generic
model free tracking plus generic facial landmark localisation, as well as (c)
hybrid approaches using state-of-the-art face detection, model free tracking
and facial landmark localisation technologies. Our evaluation reveals future
avenues for further research on the topic.Comment: E. Antonakos and P. Snape contributed equally and have joint second
authorshi
The early B-type star Rho Oph A is an X-ray lighthouse
We present the results of a 140 ks XMM-Newton observation of the B2 star
Ophiuchi A. The star has exhibited strong X-ray variability: a
cusp-shaped increase of rate, similar to that which we partially observed in
2013, and a bright flare. These events are separated in time by about 104 ks,
which likely corresponds to the rotational period of the star (1.2 days). Time
resolved spectroscopy of the X-ray spectra shows that the first event is caused
by an increase of the plasma emission measure, while the second increase of
rate is a major flare with temperatures in excess of 60 MK ( keV).
From the analysis of its rise, we infer a magnetic field of G and a
size of the flaring region of cm, which corresponds
to of the stellar radius. We speculate that either an intrinsic
magnetism that produces a hot spot on its surface or an unknown low mass
companion are the source of such X-rays and variability. A hot spot of magnetic
origin should be a stable structure over a time span of 2.5 years, and
suggests an overall large scale dipolar magnetic field that produces an
extended feature on the stellar surface. In the second scenario, a low mass
unknown companion is the emitter of X-rays and it should orbit extremely close
to the surface of the primary in a locked spin-orbit configuration, almost on
the verge of collapsing onto the primary. As such, the X-ray activity of the
secondary star would be enhanced by its young age, and the tight orbit as in RS
Cvn systems and Ophiuchi would constitute an extreme system that is
worthy of further investigation.Comment: 10 pages, 7 figures, 2 tables, A&A accepted, this is the version
after the language editor correction
- …