96,227 research outputs found
Enhancing Multi-Camera People Tracking with Anchor-Guided Clustering and Spatio-Temporal Consistency ID Re-Assignment
Multi-camera multiple people tracking has become an increasingly important
area of research due to the growing demand for accurate and efficient indoor
people tracking systems, particularly in settings such as retail, healthcare
centers, and transit hubs. We proposed a novel multi-camera multiple people
tracking method that uses anchor-guided clustering for cross-camera
re-identification and spatio-temporal consistency for geometry-based
cross-camera ID reassigning. Our approach aims to improve the accuracy of
tracking by identifying key features that are unique to every individual and
utilizing the overlap of views between cameras to predict accurate trajectories
without needing the actual camera parameters. The method has demonstrated
robustness and effectiveness in handling both synthetic and real-world data.
The proposed method is evaluated on CVPR AI City Challenge 2023 dataset,
achieving IDF1 of 95.36% with the first-place ranking in the challenge. The
code is available at: https://github.com/ipl-uw/AIC23_Track1_UWIPL_ETRI
Integration of multiple vision systems and toolbox development
Depending on the required coverage, multiple cameras with different fields of view, positions and orientations can be employed to form a motion tracking system. Correctly and efficiently designing and setting up a multi-camera vision system presents a technical challenge. This thesis describes the development and application of a toolbox that can help the user to design a multi-camera vision system. Using the parameters of cameras, including their positions and orientations, the toolbox can calculate the volume covered by the system and generate its visualization for a given tracking area. The cameras can be repositioned and reoriented using toolbox to generate the visualization of the volume covered. Finally, this thesis describes how to practically implement and achieve a proper multi-camera setup.
This thesis describes the integration of multiple cameras for vision system development based on Svoboda\u27s and Horn\u27s algorithms. Also, Dijkstra\u27s algorithm is implemented to estimate the tracking error between the master vision system and any of the slave vision systems. The toolbox is evaluated by comparing the calculated and actual covered volumes of a multi-camera system. The toolbox also is evaluated for its error estimation. The multi-camera vision system design is implemented using the developed toolbox for a virtual fastening operation of an aircraft fuselage in a computer-automated virtual environment (CAVE) --Abstract, page iii
MAVIS: Multi-Camera Augmented Visual-Inertial SLAM using SE2(3) Based Exact IMU Pre-integration
We present a novel optimization-based Visual-Inertial SLAM system designed
for multiple partially overlapped camera systems, named MAVIS. Our framework
fully exploits the benefits of wide field-of-view from multi-camera systems,
and the metric scale measurements provided by an inertial measurement unit
(IMU). We introduce an improved IMU pre-integration formulation based on the
exponential function of an automorphism of SE_2(3), which can effectively
enhance tracking performance under fast rotational motion and extended
integration time. Furthermore, we extend conventional front-end tracking and
back-end optimization module designed for monocular or stereo setup towards
multi-camera systems, and introduce implementation details that contribute to
the performance of our system in challenging scenarios. The practical validity
of our approach is supported by our experiments on public datasets. Our MAVIS
won the first place in all the vision-IMU tracks (single and multi-session
SLAM) on Hilti SLAM Challenge 2023 with 1.7 times the score compared to the
second place.Comment: video link: https://youtu.be/Q_jZSjhNFf
Exploratory Experimentation of Three Techniques for Rotating a 3D Scene by Primary School Students
International audienceMulti-touch mobile devices are now commonly used in any area, including education. In this context we focus on applications for 3D geometry learning in primary schools. Manipulating a 3D scene based on a 2D input touch space is one key challenge of such applications for pupils. In this paper we present the results of an exploratory experimentation with pupils. We compare three different interaction techniques for rotating a 3D scene based on different interaction metaphors by using: multi-touch input, movements of the tablet captured with a gyroscope sensor and movements of the head captured by a camera-based head tracking. We ran the exploratory experiment with 28 pupils in a primary school to compare these three techniques by considering the performance and the subjective preferences. Results indicate worst performance for head-tracking and similar performance for multi-touch input and gyroscope-based movement. Qualitative results indicate participant preference for multi-touch interaction
MUlti-Store Tracker (MUSTer): a Cognitive Psychology Inspired Approach to Object Tracking
International audienceVariations in the appearance of a tracked object, such as changes in geometry/photometry, camera viewpoint, illumination , or partial occlusion, pose a major challenge to object tracking. Here, we adopt cognitive psychology principles to design a flexible representation that can adapt to changes in object appearance during tracking. Inspired by the well-known Atkinson-Shiffrin Memory Model, we propose MUlti-Store Tracker (MUSTer), a dual-component approach consisting of short-and long-term memory stores to process target appearance memories. A powerful and efficient Integrated Correlation Filter (ICF) is employed in the short-term store for short-term tracking. The integrated long-term component, which is based on keypoint matching-tracking and RANSAC estimation, can interact with the long-term memory and provide additional information for output control. MUSTer was extensively evaluated on the CVPR2013 Online Object Tracking Benchmark (OOTB) and ALOV++ datasets. The experimental results demonstrated the superior performance of MUSTer in comparison with other state-of-art trackers
mvHOTA: A multi-view higher order tracking accuracy metric to measure spatial and temporal associations in multi-point detection
Multi-point tracking is a challenging task that involves detecting points in
the scene and tracking them across a sequence of frames. Computing
detection-based measures like the F-measure on a frame-by-frame basis is not
sufficient to assess the overall performance, as it does not interpret
performance in the temporal domain. The main evaluation metric available comes
from Multi-object tracking (MOT) methods to benchmark performance on datasets
such as KITTI with the recently proposed higher order tracking accuracy (HOTA)
metric, which is capable of providing a better description of the performance
over metrics such as MOTA, DetA, and IDF1. While the HOTA metric takes into
account temporal associations, it does not provide a tailored means to analyse
the spatial associations of a dataset in a multi-camera setup. Moreover, there
are differences in evaluating the detection task for points when compared to
objects (point distances vs. bounding box overlap). Therefore in this work, we
propose a multi-view higher order tracking metric (mvHOTA) to determine the
accuracy of multi-point (multi-instance and multi-class) tracking methods,
while taking into account temporal and spatial associations.mvHOTA can be
interpreted as the geometric mean of detection, temporal, and spatial
associations, thereby providing equal weighting to each of the factors. We
demonstrate the use of this metric to evaluate the tracking performance on an
endoscopic point detection dataset from a previously organised surgical data
science challenge. Furthermore, we compare with other adjusted MOT metrics for
this use-case, discuss the properties of mvHOTA, and show how the proposed
multi-view Association and the Occlusion index (OI) facilitate analysis of
methods with respect to handling of occlusions. The code is available at
https://github.com/Cardio-AI/mvhota.Comment: 16 pages, 9 figure
- …