234,696 research outputs found
Visual Learning in Multiple-Object Tracking
Tracking moving objects in space is important for the maintenance of spatiotemporal continuity in everyday visual tasks. In the laboratory, this ability is tested using the Multiple Object Tracking (MOT) task, where participants track a subset of moving objects with attention over an extended period of time. The ability to track multiple objects with attention is severely limited. Recent research has shown that this ability may improve with extensive practice (e.g., from action videogame playing). However, whether tracking also improves in a short training session with repeated trajectories has rarely been investigated. In this study we examine the role of visual learning in multiple-object tracking and characterize how varieties of attention interact with visual learning.Participants first conducted attentive tracking on trials with repeated motion trajectories for a short session. In a transfer phase we used the same motion trajectories but changed the role of tracking targets and nontargets. We found that compared with novel trials, tracking was enhanced only when the target subset was the same as that used during training. Learning did not transfer when the previously trained targets and nontargets switched roles or mixed up. However, learning was not specific to the trained temporal order as it transferred to trials where the motion was played backwards.These findings suggest that a demanding task of tracking multiple objects can benefit from learning of repeated motion trajectories. Such learning potentially facilitates tracking in natural vision, although learning is largely confined to the trajectories of attended objects. Furthermore, we showed that learning in attentive tracking relies on relational coding of all target trajectories. Surprisingly, learning was not specific to the trained temporal context, probably because observers have learned motion paths of each trajectory independently of the exact temporal order
Variational Tracking and Redetection for Closely-spaced Objects in Heavy Clutter
The non-homogeneous Poisson process (NHPP) is a widely used measurement model
that allows for an object to generate multiple measurements over time. However,
it can be difficult to efficiently and reliably track multiple objects under
this NHPP model in scenarios with a high density of closely-spaced objects and
heavy clutter. Therefore, based on the general coordinate ascent variational
filtering framework, this paper presents a variational Bayes association-based
NHPP tracker (VB-AbNHPP) that can efficiently perform tracking, data
association, and learning of target and clutter rates with a parallelisable
implementation. In addition, a variational localisation strategy is proposed,
which enables rapid rediscovery of missed targets from a large surveillance
area under extremely heavy clutter. This strategy is integrated into the
VB-AbNHPP tracker, resulting in a robust methodology that can automatically
detect and recover from track loss. This tracker demonstrates improved tracking
performance compared with existing trackers in challenging scenarios, in terms
of both accuracy and efficiency
YOLORe-IDNet: An Efficient Multi-Camera System for Person-Tracking
The growing need for video surveillance in public spaces has created a demand
for systems that can track individuals across multiple cameras feeds in
real-time. While existing tracking systems have achieved impressive performance
using deep learning models, they often rely on pre-existing images of suspects
or historical data. However, this is not always feasible in cases where
suspicious individuals are identified in real-time and without prior knowledge.
We propose a person-tracking system that combines correlation filters and
Intersection Over Union (IOU) constraints for robust tracking, along with a
deep learning model for cross-camera person re-identification (Re-ID) on top of
YOLOv5. The proposed system quickly identifies and tracks suspect in real-time
across multiple cameras and recovers well after full or partial occlusion,
making it suitable for security and surveillance applications. It is
computationally efficient and achieves a high F1-Score of 79% and an IOU of 59%
comparable to existing state-of-the-art algorithms, as demonstrated in our
evaluation on a publicly available OTB-100 dataset. The proposed system offers
a robust and efficient solution for the real-time tracking of individuals
across multiple camera feeds. Its ability to track targets without prior
knowledge or historical data is a significant improvement over existing
systems, making it well-suited for public safety and surveillance applications
Mode Selection and Target Classification in Cognitive Radar Networks
Cognitive Radar Networks were proposed by Simon Haykin in 2006 to address
problems with large legacy radar implementations - primarily, single-point
vulnerabilities and lack of adaptability. This work proposes to leverage the
adaptability of cognitive radar networks to trade between active radar
observation, which uses high power and risks interception, and passive signal
parameter estimation, which uses target emissions to gain side information and
lower the power necessary to accurately track multiple targets. The goal of the
network is to learn over many target tracks both the characteristics of the
targets as well as the optimal action choices for each type of target. In order
to select between the available actions, we utilize a multi-armed bandit model,
using current class information as prior information. When the active radar
action is selected, the node estimates the physical behavior of targets through
the radar emissions. When the passive action is selected, the node estimates
the radio behavior of targets through passive sensing. Over many target tracks,
the network collects the observed behavior of targets and forms clusters of
similarly-behaved targets. In this way, the network meta-learns the target
class distributions while learning the optimal mode selections for each target
class.Comment: 6 pages, 5 figure
A Gaussian Process Regression based Dynamical Models Learning Algorithm for Target Tracking
Maneuvering target tracking is a challenging problem for sensor systems
because of the unpredictability of the targets' motions. This paper proposes a
novel data-driven method for learning the dynamical motion model of a target.
Non-parametric Gaussian process regression (GPR) is used to learn a target's
naturally shift invariant motion (NSIM) behavior, which is translationally
invariant and does not need to be constantly updated as the target moves. The
learned Gaussian processes (GPs) can be applied to track targets within
different surveillance regions from the surveillance region of the training
data by being incorporated into the particle filter (PF) implementation. The
performance of our proposed approach is evaluated over different maneuvering
scenarios by being compared with commonly used interacting multiple model
(IMM)-PF methods and provides around performance improvement for a
multi-target tracking (MTT) highly maneuvering scenario.Comment: 11 pages, 10 figure
Recommended from our members
Multitarget Tracking in Nonoverlapping Cameras Using a Reference Set
Tracking multiple targets in nonoverlapping cameras are challenging since the observations of the same targets are often separated by time and space. There might be significant appearance change of a target across camera views caused by variations in illumination conditions, poses, and camera imaging characteristics. Consequently, the same target may appear very different in two cameras. Therefore, associating tracks in different camera views directly based on their appearance similarity is difficult and prone to error. In most previous methods, the appearance similarity is computed either using color histograms or based on pretrained brightness transfer function that maps color between cameras. In this paper, a novel reference set based appearance model is proposed to improve multitarget tracking in a network of nonoverlapping cameras. Contrary to previous work, a reference set is constructed for a pair of cameras, containing subjects appearing in both camera views. For track association, instead of directly comparing the appearance of two targets in different camera views, they are compared indirectly via the reference set. Besides global color histograms, texture and shape features are extracted at different locations of a target, and AdaBoost is used to learn the discriminative power of each feature. The effectiveness of the proposed method over the state of the art on two challenging real-world multicamera video data sets is demonstrated by thorough experiments
Recommended from our members
Integrating Social Grouping for Multitarget Tracking Across Cameras in a CRF Model
- …