25,958 research outputs found
Substructure and Boundary Modeling for Continuous Action Recognition
This paper introduces a probabilistic graphical model for continuous action
recognition with two novel components: substructure transition model and
discriminative boundary model. The first component encodes the sparse and
global temporal transition prior between action primitives in state-space model
to handle the large spatial-temporal variations within an action class. The
second component enforces the action duration constraint in a discriminative
way to locate the transition boundaries between actions more accurately. The
two components are integrated into a unified graphical structure to enable
effective training and inference. Our comprehensive experimental results on
both public and in-house datasets show that, with the capability to incorporate
additional information that had not been explicitly or efficiently modeled by
previous methods, our proposed algorithm achieved significantly improved
performance for continuous action recognition.Comment: Detailed version of the CVPR 2012 paper. 15 pages, 6 figure
Fast and easy blind deblurring using an inverse filter and PROBE
PROBE (Progressive Removal of Blur Residual) is a recursive framework for
blind deblurring. Using the elementary modified inverse filter at its core,
PROBE's experimental performance meets or exceeds the state of the art, both
visually and quantitatively. Remarkably, PROBE lends itself to analysis that
reveals its convergence properties. PROBE is motivated by recent ideas on
progressive blind deblurring, but breaks away from previous research by its
simplicity, speed, performance and potential for analysis. PROBE is neither a
functional minimization approach, nor an open-loop sequential method (blur
kernel estimation followed by non-blind deblurring). PROBE is a feedback
scheme, deriving its unique strength from the closed-loop architecture rather
than from the accuracy of its algorithmic components
Single camera pose estimation using Bayesian filtering and Kinect motion priors
Traditional approaches to upper body pose estimation using monocular vision
rely on complex body models and a large variety of geometric constraints. We
argue that this is not ideal and somewhat inelegant as it results in large
processing burdens, and instead attempt to incorporate these constraints
through priors obtained directly from training data. A prior distribution
covering the probability of a human pose occurring is used to incorporate
likely human poses. This distribution is obtained offline, by fitting a
Gaussian mixture model to a large dataset of recorded human body poses, tracked
using a Kinect sensor. We combine this prior information with a random walk
transition model to obtain an upper body model, suitable for use within a
recursive Bayesian filtering framework. Our model can be viewed as a mixture of
discrete Ornstein-Uhlenbeck processes, in that states behave as random walks,
but drift towards a set of typically observed poses. This model is combined
with measurements of the human head and hand positions, using recursive
Bayesian estimation to incorporate temporal information. Measurements are
obtained using face detection and a simple skin colour hand detector, trained
using the detected face. The suggested model is designed with analytical
tractability in mind and we show that the pose tracking can be
Rao-Blackwellised using the mixture Kalman filter, allowing for computational
efficiency while still incorporating bio-mechanical properties of the upper
body. In addition, the use of the proposed upper body model allows reliable
three-dimensional pose estimates to be obtained indirectly for a number of
joints that are often difficult to detect using traditional object recognition
strategies. Comparisons with Kinect sensor results and the state of the art in
2D pose estimation highlight the efficacy of the proposed approach.Comment: 25 pages, Technical report, related to Burke and Lasenby, AMDO 2014
conference paper. Code sample: https://github.com/mgb45/SignerBodyPose Video:
https://www.youtube.com/watch?v=dJMTSo7-uF
Material Recognition CNNs and Hierarchical Planning for Biped Robot Locomotion on Slippery Terrain
In this paper we tackle the problem of visually predicting surface friction
for environments with diverse surfaces, and integrating this knowledge into
biped robot locomotion planning. The problem is essential for autonomous robot
locomotion since diverse surfaces with varying friction abound in the real
world, from wood to ceramic tiles, grass or ice, which may cause difficulties
or huge energy costs for robot locomotion if not considered. We propose to
estimate friction and its uncertainty from visual estimation of material
classes using convolutional neural networks, together with probability
distribution functions of friction associated with each material. We then
robustly integrate the friction predictions into a hierarchical (footstep and
full-body) planning method using chance constraints, and optimize the same
trajectory costs at both levels of the planning method for consistency. Our
solution achieves fully autonomous perception and locomotion on slippery
terrain, which considers not only friction and its uncertainty, but also
collision, stability and trajectory cost. We show promising friction prediction
results in real pictures of outdoor scenarios, and planning experiments on a
real robot facing surfaces with different friction
- …