6 research outputs found
An Accelerated Correlation Filter Tracker
Recent visual object tracking methods have witnessed a continuous improvement
in the state-of-the-art with the development of efficient discriminative
correlation filters (DCF) and robust deep neural network features. Despite the
outstanding performance achieved by the above combination, existing advanced
trackers suffer from the burden of high computational complexity of the deep
feature extraction and online model learning. We propose an accelerated ADMM
optimisation method obtained by adding a momentum to the optimisation sequence
iterates, and by relaxing the impact of the error between DCF parameters and
their norm. The proposed optimisation method is applied to an innovative
formulation of the DCF design, which seeks the most discriminative spatially
regularised feature channels. A further speed up is achieved by an adaptive
initialisation of the filter optimisation process. The significantly increased
convergence of the DCF filter is demonstrated by establishing the optimisation
process equivalence with a continuous dynamical system for which the
convergence properties can readily be derived. The experimental results
obtained on several well-known benchmarking datasets demonstrate the efficiency
and robustness of the proposed ACFT method, with a tracking accuracy comparable
to the start-of-the-art trackers
Tracking more than 100 arbitrary objects at 25 FPS through deep learning
Most video analytics applications rely on object detectors to localize objects in frames. However, when real-time is a requirement, running the detector at all the frames is usually not possible. This is somewhat circumvented by instantiating visual object trackers between detector calls, but this does not scale with the number of objects. To tackle this problem, we present SiamMT, a new deep learning multiple visual object tracking solution that applies single-object tracking principles to multiple arbitrary objects in real-time. To achieve this, SiamMT reuses feature computations, implements a novel crop-and-resize operator, and defines a new and efficient pairwise similarity operator. SiamMT naturally scales up to several dozens of targets, reaching 25 fps with 122 simultaneous objects for VGA videos, or up to 100 simultaneous objects in HD720 video. SiamMT has been validated on five large real-time benchmarks, achieving leading performance against current state-of-the-art trackersThis research was partially funded by the Spanish Ministerio de Ciencia e Innovación [grant numbers PID2020-112623GB-I00, RTI2018-097088-B-C32], and the Galician ConsellerÃa de Cultura, Educación e Universidade [grant numbers ED431C 2018/29, ED431C 2017/69, accreditation 2016–2019, ED431G/08].
These grants are co-funded by the European Regional Development Fund (ERDF). Lorenzo Vaquero is supported by the Spanish Ministerio de Universidades under the FPU national plan (FPU18/03174)S
Real-time siamese multiple object tracker with enhanced proposals
Maintaining the identity of multiple objects in real-time video is a challenging task, as it is not always feasible to run a detector on every frame. Thus, motion estimation systems are often employed, which either do not scale well with the number of targets or produce features with limited semantic information. To solve the aforementioned problems and allow the tracking of dozens of arbitrary objects in real-time, we propose SiamMOTION. SiamMOTION includes a novel proposal engine that produces quality features through an attention mechanism and a region-of-interest extractor fed by an inertia module and powered by a feature pyramid network. Finally, the extracted tensors enter a comparison head that efficiently matches pairs of exemplars and search areas, generating quality predictions via a pairwise depthwise region proposal network and a multi-object penalization module. SiamMOTION has been validated on five public benchmarks, achieving leading performance against current state-of-the-art trackers. Code available at: https://www.github.com/lorenzovaquero/SiamMOTIONThis research was partially funded by the Spanish Ministerio de Ciencia e Innovación [grant numbers PID2020-112623GB-I00, RTI2018-097088-B-C32], and the Galician ConsellerÃa de Cultura, Educación e Universidade [grant numbers ED431C 2018/29, ED431C 2021/048, ED431G 2019/04]. These grants are co-funded by the European Regional Development Fund (ERDF). Lorenzo Vaquero is supported by the Spanish Ministerio de Universidades under the FPU national plan (FPU18/03174). We also gratefully acknowledge the support of NVIDIA Corporation for hardware donations used for this researchS