60,373 research outputs found
Event-based Vision: A Survey
Event cameras are bio-inspired sensors that differ from conventional frame
cameras: Instead of capturing images at a fixed rate, they asynchronously
measure per-pixel brightness changes, and output a stream of events that encode
the time, location and sign of the brightness changes. Event cameras offer
attractive properties compared to traditional cameras: high temporal resolution
(in the order of microseconds), very high dynamic range (140 dB vs. 60 dB), low
power consumption, and high pixel bandwidth (on the order of kHz) resulting in
reduced motion blur. Hence, event cameras have a large potential for robotics
and computer vision in challenging scenarios for traditional cameras, such as
low-latency, high speed, and high dynamic range. However, novel methods are
required to process the unconventional output of these sensors in order to
unlock their potential. This paper provides a comprehensive overview of the
emerging field of event-based vision, with a focus on the applications and the
algorithms developed to unlock the outstanding properties of event cameras. We
present event cameras from their working principle, the actual sensors that are
available and the tasks that they have been used for, from low-level vision
(feature detection and tracking, optic flow, etc.) to high-level vision
(reconstruction, segmentation, recognition). We also discuss the techniques
developed to process events, including learning-based techniques, as well as
specialized processors for these novel sensors, such as spiking neural
networks. Additionally, we highlight the challenges that remain to be tackled
and the opportunities that lie ahead in the search for a more efficient,
bio-inspired way for machines to perceive and interact with the world
Low computational complexity variable block size (VBS) partitioning for motion estimation using the Walsh Hadamard transform (WHT)
Variable Block Size (VBS) based motion estimation has
been adapted in state of the art video coding, such as
H.264/AVC, VC-1. However, a low complexity H.264/AVC
encoder cannot take advantage of VBS due to its power consumption
requirements. In this paper, we present a VBS partition
algorithm based on a binary motion edge map without
either initial motion estimation or Rate-Distortion (R-D)
optimization for selecting modes. The proposed algorithm
uses the Walsh Hadamard Transform (WHT) to create a binary
edge map, which provides a computational complexity
cost effectiveness compared to other light segmentation
methods typically used to detect the required region
Single camera pose estimation using Bayesian filtering and Kinect motion priors
Traditional approaches to upper body pose estimation using monocular vision
rely on complex body models and a large variety of geometric constraints. We
argue that this is not ideal and somewhat inelegant as it results in large
processing burdens, and instead attempt to incorporate these constraints
through priors obtained directly from training data. A prior distribution
covering the probability of a human pose occurring is used to incorporate
likely human poses. This distribution is obtained offline, by fitting a
Gaussian mixture model to a large dataset of recorded human body poses, tracked
using a Kinect sensor. We combine this prior information with a random walk
transition model to obtain an upper body model, suitable for use within a
recursive Bayesian filtering framework. Our model can be viewed as a mixture of
discrete Ornstein-Uhlenbeck processes, in that states behave as random walks,
but drift towards a set of typically observed poses. This model is combined
with measurements of the human head and hand positions, using recursive
Bayesian estimation to incorporate temporal information. Measurements are
obtained using face detection and a simple skin colour hand detector, trained
using the detected face. The suggested model is designed with analytical
tractability in mind and we show that the pose tracking can be
Rao-Blackwellised using the mixture Kalman filter, allowing for computational
efficiency while still incorporating bio-mechanical properties of the upper
body. In addition, the use of the proposed upper body model allows reliable
three-dimensional pose estimates to be obtained indirectly for a number of
joints that are often difficult to detect using traditional object recognition
strategies. Comparisons with Kinect sensor results and the state of the art in
2D pose estimation highlight the efficacy of the proposed approach.Comment: 25 pages, Technical report, related to Burke and Lasenby, AMDO 2014
conference paper. Code sample: https://github.com/mgb45/SignerBodyPose Video:
https://www.youtube.com/watch?v=dJMTSo7-uF
Event-Based Motion Segmentation by Motion Compensation
In contrast to traditional cameras, whose pixels have a common exposure time,
event-based cameras are novel bio-inspired sensors whose pixels work
independently and asynchronously output intensity changes (called "events"),
with microsecond resolution. Since events are caused by the apparent motion of
objects, event-based cameras sample visual information based on the scene
dynamics and are, therefore, a more natural fit than traditional cameras to
acquire motion, especially at high speeds, where traditional cameras suffer
from motion blur. However, distinguishing between events caused by different
moving objects and by the camera's ego-motion is a challenging task. We present
the first per-event segmentation method for splitting a scene into
independently moving objects. Our method jointly estimates the event-object
associations (i.e., segmentation) and the motion parameters of the objects (or
the background) by maximization of an objective function, which builds upon
recent results on event-based motion-compensation. We provide a thorough
evaluation of our method on a public dataset, outperforming the
state-of-the-art by as much as 10%. We also show the first quantitative
evaluation of a segmentation algorithm for event cameras, yielding around 90%
accuracy at 4 pixels relative displacement.Comment: When viewed in Acrobat Reader, several of the figures animate. Video:
https://youtu.be/0q6ap_OSBA
Topology-Guided Path Integral Approach for Stochastic Optimal Control in Cluttered Environment
This paper addresses planning and control of robot motion under uncertainty
that is formulated as a continuous-time, continuous-space stochastic optimal
control problem, by developing a topology-guided path integral control method.
The path integral control framework, which forms the backbone of the proposed
method, re-writes the Hamilton-Jacobi-Bellman equation as a statistical
inference problem; the resulting inference problem is solved by a sampling
procedure that computes the distribution of controlled trajectories around the
trajectory by the passive dynamics. For motion control of robots in a highly
cluttered environment, however, this sampling can easily be trapped in a local
minimum unless the sample size is very large, since the global optimality of
local minima depends on the degree of uncertainty. Thus, a homology-embedded
sampling-based planner that identifies many (potentially) local-minimum
trajectories in different homology classes is developed to aid the sampling
process. In combination with a receding-horizon fashion of the optimal control
the proposed method produces a dynamically feasible and collision-free motion
plans without being trapped in a local minimum. Numerical examples on a
synthetic toy problem and on quadrotor control in a complex obstacle field
demonstrate the validity of the proposed method.Comment: arXiv admin note: text overlap with arXiv:1510.0534
- âŠ