2,274 research outputs found
Real-Time Hand Tracking Using a Sum of Anisotropic Gaussians Model
Real-time marker-less hand tracking is of increasing importance in
human-computer interaction. Robust and accurate tracking of arbitrary hand
motion is a challenging problem due to the many degrees of freedom, frequent
self-occlusions, fast motions, and uniform skin color. In this paper, we
propose a new approach that tracks the full skeleton motion of the hand from
multiple RGB cameras in real-time. The main contributions include a new
generative tracking method which employs an implicit hand shape representation
based on Sum of Anisotropic Gaussians (SAG), and a pose fitting energy that is
smooth and analytically differentiable making fast gradient based pose
optimization possible. This shape representation, together with a full
perspective projection model, enables more accurate hand modeling than a
related baseline method from literature. Our method achieves better accuracy
than previous methods and runs at 25 fps. We show these improvements both
qualitatively and quantitatively on publicly available datasets.Comment: 8 pages, Accepted version of paper published at 3DV 201
Content based image pose manipulation
This thesis proposes the application of space-frequency transformations to the domain of pose estimation in images. This idea is explored using the Wavelet Transform with illustrative applications in pose estimation for face images, and images of planar scenes. The approach is based
on examining the spatial frequency components in an image, to allow the inherent scene symmetry balance to be recovered. For face images with restricted pose variation (looking left or right), an algorithm is proposed to maximise this symmetry in order to transform the image
into a fronto-parallel pose. This scheme is further employed to identify the optimal frontal facial pose from a video sequence to automate facial capture processes. These features are an important pre-requisite in facial recognition and expression classification systems. The under
lying principles of this spatial-frequency approach are examined with respect to images with planar scenes. Using the Continuous Wavelet Transform, full perspective planar transformations are estimated within a featureless framework. Restoring central symmetry to the wavelet
transformed images in an iterative optimisation scheme removes this perspective pose. This advances upon existing spatial approaches that require segmentation and feature matching, and frequency only techniques that are limited to affine transformation recovery. To evaluate the proposed techniques, the pose of a database of subjects portraying varying yaw orientations is estimated and the accuracy is measured against the captured ground truth information. Additionally, full perspective homographies for synthesised and imaged textured planes are estimated. Experimental results are presented for both situations that compare favourably with existing techniques in the literature
Accelerated Parameter Estimation with DALE
We consider methods for improving the estimation of constraints on a
high-dimensional parameter space with a computationally expensive likelihood
function. In such cases Markov chain Monte Carlo (MCMC) can take a long time to
converge and concentrates on finding the maxima rather than the often-desired
confidence contours for accurate error estimation. We employ DALE (Direct
Analysis of Limits via the Exterior of ) for determining confidence
contours by minimizing a cost function parametrized to incentivize points in
parameter space which are both on the confidence limit and far from previously
sampled points. We compare DALE to the nested sampling algorithm
implemented in MultiNest on a toy likelihood function that is highly
non-Gaussian and non-linear in the mapping between parameter values and
. We find that in high-dimensional cases DALE finds the same
confidence limit as MultiNest using roughly an order of magnitude fewer
evaluations of the likelihood function. DALE is open-source and available
at https://github.com/danielsf/Dalex.git
Biologically informed ecological niche models for an example pelagic, highly mobile species
Background: Although pelagic seabirds are broadly recognised as indicators of the health of marine systems, numerous gaps exist in knowledge of their at-sea distributions at the species level. These gaps have profound negative impacts on the robustness of marine conservation policies. Correlative modelling techniques have provided some information, but few studies have explored model development for non-breeding pelagic seabirds. Here, I present a first phase in developing robust niche models for highly mobile species as a baseline for further development.Methodology: Using observational data from a 12-year time period, 217 unique model parameterisations across three correlative modelling algorithms (boosted regression trees, Maxent and minimum volume ellipsoids) were tested in a time-averaged approach for their ability to recreate the at-sea distribution of non-breeding Wandering Albatrosses (Diomedea exulans) to provide a baseline for further development.Principle Findings/Results: Overall, minimum volume ellipsoids outperformed both boosted regression trees and Maxent. However, whilst the latter two algorithms generally overfit the data, minimum volume ellipsoids tended to underfit the data. Conclusions: The results of this exercise suggest a necessary evolution in how correlative modelling for highly mobile species such as pelagic seabirds should be approached. These insights are crucial for understanding seabird–environment interactions at macroscales, which can facilitate the ability to address population declines and inform effective marine conservation policy in the wake of rapid global change
High-speed tensor tomography: iterative reconstruction tensor tomography (IRTT) algorithm
The recent advent of tensor tomography techniques has enabled tomographic investigations of the 3D nanostructure organization of biological and material science samples. These techniques extended the concept of conventional X-ray tomography by reconstructing not only a scalar value such as the attenuation coefficient per voxel, but also a set of parameters that capture the local anisotropy of nanostructures within every voxel of the sample. Tensor tomography data sets are intrinsically large as each pixel of a conventional X-ray projection is substituted by a scattering pattern, and projections have to be recorded at different sample angular orientations with several tilts of the rotation axis with respect to the X-ray propagation direction. Currently available reconstruction approaches for such large data sets are computationally expensive. Here, a novel, fast reconstruction algorithm, named iterative reconstruction tensor tomography (IRTT), is presented to simplify and accelerate tensor tomography reconstructions. IRTT is based on a second-rank tensor model to describe the anisotropy of the nanostructure in every voxel and on an iterative error backpropagation reconstruction algorithm to achieve high convergence speed. The feasibility and accuracy of IRTT are demonstrated by reconstructing the nanostructure anisotropy of three samples: a carbon fiber knot, a human bone trabecula specimen and a fixed mouse brain. Results and reconstruction speed were compared with those obtained by the small-angle scattering tensor tomography (SASTT) reconstruction method introduced by Liebi et al. [Nature (2015), 527, 349–352]. The principal orientation of the nanostructure within each voxel revealed a high level of agreement between the two methods. Yet, for identical data sets and computer hardware used, IRTT was shown to be more than an order of magnitude faster. IRTT was found to yield robust results, it does not require prior knowledge of the sample for initializing parameters, and can be used in cases where simple anisotropy metrics are sufficient, i.e. the tensor approximation adequately captures the level of anisotropy and the dominant orientation within a voxel. In addition, by greatly accelerating the reconstruction, IRTT is particularly suitable for handling large tomographic data sets of samples with internal structure or as a real-time analysis tool during the experiment for online feedback during data acquisition. Alternatively, the IRTT results might be used as an initial guess for models capturing a higher complexity of structural anisotropy such as spherical harmonics based SASTT in Liebi et al. (2015), improving both overall convergence speed and robustness of the reconstruction
Registration techniques for computer assisted orthopaedic surgery
The registration of 3D preoperative medical data to patients is a key task in developing computer assisted surgery systems. In computer assisted surgery, the patient in the operation theatre must be aligned with the coordinate system in which the preoperative data has been acquired, so that the planned surgery based on the preoperative data can be carried out under the guidance of the computer assisted surgery system.The aim of this research is to investigate registration algorithms for developing computer assisted bone surgery systems. We start with reference mark registration. New interpretations are given to the development of well knowm algorithms based on singular value decomposition, polar decomposition techniques and the unit quaternion representation of the rotation matrix. In addition, a new algorithm is developed based on the estimate of the rotation axis. For non-land mark registration, we first develop iterative closest line segment and iterative closest triangle patch registrations, similar to the well known iterative closest point registration, when the preoperative data are dense enough. We then move to the situation where the preoperative data are not dense enough. Implicit fitting is considered to interpolate the gaps between the data . A new ellipsoid fitting algorithm and a new constructive implicit fitting strategy are developed. Finally, a region to region matching procedure is proposed based on our novel constructive implicit fitting technique. Experiments demonstrate that the new algorithm is very stable and very efficient
Robust computational intelligence techniques for visual information processing
The third part is exclusively dedicated to the super-resolution of Magnetic Resonance Images. In one of these works, an algorithm based on the random shifting technique is developed. Besides, we studied noise removal and resolution enhancement simultaneously. To end, the cost function of deep networks has been modified by different combinations of norms in order to improve their training.
Finally, the general conclusions of the research are presented and discussed, as well as the possible future research lines that are able to make use of the results obtained in this Ph.D. thesis.This Ph.D. thesis is about image processing by computational intelligence techniques. Firstly, a general overview of this book is carried out, where the motivation, the hypothesis, the objectives, and the methodology employed are described. The use and analysis of different mathematical norms will be our goal. After that, state of the art focused on the applications of the image processing proposals is presented. In addition, the fundamentals of the image modalities, with particular attention to magnetic resonance, and the learning techniques used in this research, mainly based on neural networks, are summarized. To end up, the mathematical framework on which this work is based on, ₚ-norms, is defined.
Three different parts associated with image processing techniques follow. The first non-introductory part of this book collects the developments which are about image segmentation. Two of them are applications for video surveillance tasks and try to model the background of a scenario using a specific camera. The other work is centered on the medical field, where the goal of segmenting diabetic wounds of a very heterogeneous dataset is addressed.
The second part is focused on the optimization and implementation of new models for curve and surface fitting in two and three dimensions, respectively. The first work presents a parabola fitting algorithm based on the measurement of the distances of the interior and exterior points to the focus and the directrix. The second work changes to an ellipse shape, and it ensembles the information of multiple fitting methods. Last, the ellipsoid problem is addressed in a similar way to the parabola
Structural properties of disk galaxies I. The intrinsic ellipticity of bulges
(Abridged) A variety of formation scenarios was proposed to explain the
diversity of properties observed in bulges. Studying their intrinsic shape can
help in constraining the dominant mechanism at the epochs of their assembly.
The structural parameters of a magnitude-limited sample of 148 unbarred S0--Sb
galaxies were derived in order to study the correlations between bulges and
disks as well as the probability distribution function (PDF) of the intrinsic
equatorial ellipticity of bulges. It is presented a new fitting algorithm
(GASP2D) to perform the two-dimensional photometric decomposition of galaxy
surface-brightness distribution. This was assumed to be the sum of the
contribution of a bulge and disk component characterized by elliptical and
concentric isophotes with constant (but possibly different) ellipticity and
position angles. Bulge and disk parameters of the sample galaxies were derived
from the J-band images which were available in the Two Micron All Sky Survey.
The PDF of the equatorial ellipticity of the bulges was derived from the
distribution of the observed ellipticities of bulges and misalignments between
bulges and disks. Strong correlations between the bulge and disk parameters
were found. About 80% of bulges in unbarred lenticular and
early-to-intermediate spiral galaxies are not oblate but triaxial ellipsoids.
Their mean axial ratio in the equatorial plane is = 0.85. There is not
significant dependence of their PDF on morphology, light concentration, and
luminosity. The interplay between bulge and disk parameters favors scenarios in
which bulges assembled from mergers and/or grew over long times through disk
secular evolution. But all these mechanisms have to be tested against the
derived distribution of bulge intrinsic ellipticities.Comment: 24 pages, 13 figures, accepted for publication in A&A, corrected
proof
HfS, Hyperfine Structure Fitting Tool
Hyperfine Structure Fitting (HfS) is a tool to fit the hyperfine structure of spectral lines with multiple velocity components. The HfS_nh3 procedures included in HfS simultaneously fit the hyperfine structure of the NH3 (J, K) = (1, 1) and (2, 2) transitions, and perform a standard analysis to derive , NH3 column density, , and . HfS uses a Monte Carlo approach for fitting the line parameters. Special attention is paid to the derivation of the parameter uncertainties. HfS includes procedures that make use of parallel computing for fitting spectra from a data cube
- …