81,670 research outputs found
Twins:Device-free Object Tracking using Passive Tags
Without requiring objects to carry any transceiver, device-free based object
tracking provides a promising solution for many localization and tracking
systems to monitor non-cooperative objects such as intruders. However, existing
device-free solutions mainly use sensors and active RFID tags, which are much
more expensive compared to passive tags. In this paper, we propose a novel
motion detection and tracking method using passive RFID tags, named Twins. The
method leverages a newly observed phenomenon called critical state caused by
interference among passive tags. We contribute to both theory and practice of
such phenomenon by presenting a new interference model that perfectly explains
this phenomenon and using extensive experiments to validate it. We design a
practical Twins based intrusion detection scheme and implement a real prototype
with commercial off-the-shelf reader and tags. The results show that Twins is
effective in detecting the moving object, with low location error of 0.75m in
average
Features-based moving objects tracking for smart video surveillances: A review
Video surveillance is one of the most active research topics in the computer vision due to the increasing need for security. Although surveillance systems are getting cheaper, the cost of having human operators to monitor the video feed can be very expensive and inefficient. To overcome this problem, the automated visual surveillance system can be used to detect any suspicious activities that require immediate action. The framework of a video surveillance system encompasses a large scope in machine vision, they are background modelling, object detection, moving objects classification, tracking, motion analysis, and require fusion of information from the camera networks. This paper reviews recent techniques used by researchers for detection of moving object detection and tracking in order to solve many surveillance problems. The features and algorithms used for modelling the object appearance and tracking multiple objects in outdoor and indoor environment are also reviewed in this paper. This paper summarizes the recent works done by previous researchers in moving objects tracking for single camera view and multiple cameras views. Nevertheless, despite of the recent progress in surveillance technologies, there still are challenges that need to be solved before the system can come out with a reliable automated video surveillance
3D hand tracking.
The hand is often considered as one of the most natural and intuitive interaction modalities for human-to-human interaction. In human-computer interaction (HCI), proper 3D hand tracking is the first step in developing a more intuitive HCI system which can be used in applications such as gesture recognition, virtual object manipulation and gaming. However, accurate 3D hand tracking, remains a challenging problem due to the hand’s deformation, appearance similarity, high inter-finger occlusion and complex articulated motion. Further, 3D hand tracking is also interesting from a theoretical point of view as it deals with three major areas of computer vision- segmentation (of hand), detection (of hand parts), and tracking (of hand). This thesis proposes a region-based skin color detection technique, a model-based and an appearance-based 3D hand tracking techniques to bring the human-computer interaction applications one step closer. All techniques are briefly described below. Skin color provides a powerful cue for complex computer vision applications. Although skin color detection has been an active research area for decades, the mainstream technology is based on individual pixels. This thesis presents a new region-based technique for skin color detection which outperforms the current state-of-the-art pixel-based skin color detection technique on the popular Compaq dataset (Jones & Rehg 2002). The proposed technique achieves 91.17% true positive rate with 13.12% false negative rate on the Compaq dataset tested over approximately 14,000 web images. Hand tracking is not a trivial task as it requires tracking of 27 degreesof- freedom of hand. Hand deformation, self occlusion, appearance similarity and irregular motion are major problems that make 3D hand tracking a very challenging task. This thesis proposes a model-based 3D hand tracking technique, which is improved by using proposed depth-foreground-background ii feature, palm deformation module and context cue. However, the major problem of model-based techniques is, they are computationally expensive. This can be overcome by discriminative techniques as described below. Discriminative techniques (for example random forest) are good for hand part detection, however they fail due to sensor noise and high interfinger occlusion. Additionally, these techniques have difficulties in modelling kinematic or temporal constraints. Although model-based descriptive (for example Markov Random Field) or generative (for example Hidden Markov Model) techniques utilize kinematic and temporal constraints well, they are computationally expensive and hardly recover from tracking failure. This thesis presents a unified framework for 3D hand tracking, using the best of both methodologies, which out performs the current state-of-the-art 3D hand tracking techniques. The proposed 3D hand tracking techniques in this thesis can be used to extract accurate hand movement features and enable complex human machine interaction such as gaming and virtual object manipulation
Real-time detection and tracking of multiple objects with partial decoding in H.264/AVC bitstream domain
In this paper, we show that we can apply probabilistic spatiotemporal
macroblock filtering (PSMF) and partial decoding processes to effectively
detect and track multiple objects in real time in H.264|AVC bitstreams with
stationary background. Our contribution is that our method cannot only show
fast processing time but also handle multiple moving objects that are
articulated, changing in size or internally have monotonous color, even though
they contain a chaotic set of non-homogeneous motion vectors inside. In
addition, our partial decoding process for H.264|AVC bitstreams enables to
improve the accuracy of object trajectories and overcome long occlusion by
using extracted color information.Comment: SPIE Real-Time Image and Video Processing Conference 200
Note: An object detection method for active camera
To solve the problems caused by a changing background during object detection in active camera, this paper proposes a new method based on SURF (speeded up robust features) and data clustering. The SURF feature points of each image are extracted, and each cluster center is calculated by processing the data clustering of k adjacent frames. Templates for each class are obtained by calculating the histograms within the regions around the center points of the clustering classes. The window of the moving object can be located by finding the region that satisfies the histogram matching result between adjacent frames. Experimental results demonstrate that the proposed method can improve the effectiveness of object detection.Yong Chen, Ronghua Zhang, Lei Shang, and Eric H
An extension of min/max flow framework
In this paper, the min/max flow scheme for image restoration is revised. The novelty consists of the fol-
24 lowing three parts. The first is to analyze the reason of the speckle generation and then to modify the
25 original scheme. The second is to point out that the continued application of this scheme cannot result
26 in an adaptive stopping of the curvature flow. This is followed by modifications of the original scheme
27 through the introduction of the Gradient Vector Flow (GVF) field and the zero-crossing detector, so as
28 to control the smoothing effect. Our experimental results with image restoration show that the proposed
29 schemes can reach a steady state solution while preserving the essential structures of objects. The third is
30 to extend the min/max flow scheme to deal with the boundary leaking problem, which is indeed an
31 intrinsic shortcoming of the familiar geodesic active contour model. The min/max flow framework pro-
32 vides us with an effective way to approximate the optimal solution. From an implementation point of
33 view, this extended scheme makes the speed function simpler and more flexible. The experimental
34 results of segmentation and region tracking show that the boundary leaking problem can be effectively
35 suppressed
MobiFace: A Novel Dataset for Mobile Face Tracking in the Wild
Face tracking serves as the crucial initial step in mobile applications
trying to analyse target faces over time in mobile settings. However, this
problem has received little attention, mainly due to the scarcity of dedicated
face tracking benchmarks. In this work, we introduce MobiFace, the first
dataset for single face tracking in mobile situations. It consists of 80
unedited live-streaming mobile videos captured by 70 different smartphone users
in fully unconstrained environments. Over bounding boxes are manually
labelled. The videos are carefully selected to cover typical smartphone usage.
The videos are also annotated with 14 attributes, including 6 newly proposed
attributes and 8 commonly seen in object tracking. 36 state-of-the-art
trackers, including facial landmark trackers, generic object trackers and
trackers that we have fine-tuned or improved, are evaluated. The results
suggest that mobile face tracking cannot be solved through existing approaches.
In addition, we show that fine-tuning on the MobiFace training data
significantly boosts the performance of deep learning-based trackers,
suggesting that MobiFace captures the unique characteristics of mobile face
tracking. Our goal is to offer the community a diverse dataset to enable the
design and evaluation of mobile face trackers. The dataset, annotations and the
evaluation server will be on \url{https://mobiface.github.io/}.Comment: To appear on The 14th IEEE International Conference on Automatic Face
and Gesture Recognition (FG 2019
- …