18,995 research outputs found
Multi-Target Tracking in Distributed Sensor Networks using Particle PHD Filters
Multi-target tracking is an important problem in civilian and military
applications. This paper investigates multi-target tracking in distributed
sensor networks. Data association, which arises particularly in multi-object
scenarios, can be tackled by various solutions. We consider sequential Monte
Carlo implementations of the Probability Hypothesis Density (PHD) filter based
on random finite sets. This approach circumvents the data association issue by
jointly estimating all targets in the region of interest. To this end, we
develop the Diffusion Particle PHD Filter (D-PPHDF) as well as a centralized
version, called the Multi-Sensor Particle PHD Filter (MS-PPHDF). Their
performance is evaluated in terms of the Optimal Subpattern Assignment (OSPA)
metric, benchmarked against a distributed extension of the Posterior
Cram\'er-Rao Lower Bound (PCRLB), and compared to the performance of an
existing distributed PHD Particle Filter. Furthermore, the robustness of the
proposed tracking algorithms against outliers and their performance with
respect to different amounts of clutter is investigated.Comment: 27 pages, 6 figure
A survey of localization in wireless sensor network
Localization is one of the key techniques in wireless sensor network. The location estimation methods can be classified into target/source localization and node self-localization. In target localization, we mainly introduce the energy-based method. Then we investigate the node self-localization methods. Since the widespread adoption of the wireless sensor network, the localization methods are different in various applications. And there are several challenges in some special scenarios. In this paper, we present a comprehensive survey of these challenges: localization in non-line-of-sight, node selection criteria for localization in energy-constrained network, scheduling the sensor node to optimize the tradeoff between localization performance and energy consumption, cooperative node localization, and localization algorithm in heterogeneous network. Finally, we introduce the evaluation criteria for localization in wireless sensor network
A Differential Approach for Gaze Estimation
Non-invasive gaze estimation methods usually regress gaze directions directly
from a single face or eye image. However, due to important variabilities in eye
shapes and inner eye structures amongst individuals, universal models obtain
limited accuracies and their output usually exhibit high variance as well as
biases which are subject dependent. Therefore, increasing accuracy is usually
done through calibration, allowing gaze predictions for a subject to be mapped
to his/her actual gaze. In this paper, we introduce a novel image differential
method for gaze estimation. We propose to directly train a differential
convolutional neural network to predict the gaze differences between two eye
input images of the same subject. Then, given a set of subject specific
calibration images, we can use the inferred differences to predict the gaze
direction of a novel eye sample. The assumption is that by allowing the
comparison between two eye images, annoyance factors (alignment, eyelid
closing, illumination perturbations) which usually plague single image
prediction methods can be much reduced, allowing better prediction altogether.
Experiments on 3 public datasets validate our approach which constantly
outperforms state-of-the-art methods even when using only one calibration
sample or when the latter methods are followed by subject specific gaze
adaptation.Comment: Extension to our paper A differential approach for gaze estimation
with calibration (BMVC 2018) Submitted to PAMI on Aug. 7th, 2018 Accepted by
PAMI short on Dec. 2019, in IEEE Transactions on Pattern Analysis and Machine
Intelligenc
Flow-Guided Feature Aggregation for Video Object Detection
Extending state-of-the-art object detectors from image to video is
challenging. The accuracy of detection suffers from degenerated object
appearances in videos, e.g., motion blur, video defocus, rare poses, etc.
Existing work attempts to exploit temporal information on box level, but such
methods are not trained end-to-end. We present flow-guided feature aggregation,
an accurate and end-to-end learning framework for video object detection. It
leverages temporal coherence on feature level instead. It improves the
per-frame features by aggregation of nearby features along the motion paths,
and thus improves the video recognition accuracy. Our method significantly
improves upon strong single-frame baselines in ImageNet VID, especially for
more challenging fast moving objects. Our framework is principled, and on par
with the best engineered systems winning the ImageNet VID challenges 2016,
without additional bells-and-whistles. The proposed method, together with Deep
Feature Flow, powered the winning entry of ImageNet VID challenges 2017. The
code is available at
https://github.com/msracver/Flow-Guided-Feature-Aggregation
Lookahead Strategies for Sequential Monte Carlo
Based on the principles of importance sampling and resampling, sequential
Monte Carlo (SMC) encompasses a large set of powerful techniques dealing with
complex stochastic dynamic systems. Many of these systems possess strong
memory, with which future information can help sharpen the inference about the
current state. By providing theoretical justification of several existing
algorithms and introducing several new ones, we study systematically how to
construct efficient SMC algorithms to take advantage of the "future"
information without creating a substantially high computational burden. The
main idea is to allow for lookahead in the Monte Carlo process so that future
information can be utilized in weighting and generating Monte Carlo samples, or
resampling from samples of the current state.Comment: Published in at http://dx.doi.org/10.1214/12-STS401 the Statistical
Science (http://www.imstat.org/sts/) by the Institute of Mathematical
Statistics (http://www.imstat.org
- …