68,916 research outputs found

    BSUV-Net: a fully-convolutional neural network for background subtraction of unseen videos

    Full text link
    Background subtraction is a basic task in computer vision and video processing often applied as a pre-processing step for object tracking, people recognition, etc. Recently, a number of successful background-subtraction algorithms have been proposed, however nearly all of the top-performing ones are supervised. Crucially, their success relies upon the availability of some annotated frames of the test video during training. Consequently, their performance on completely “unseen” videos is undocumented in the literature. In this work, we propose a new, supervised, background subtraction algorithm for unseen videos (BSUV-Net) based on a fully-convolutional neural network. The input to our network consists of the current frame and two background frames captured at different time scales along with their semantic segmentation maps. In order to reduce the chance of overfitting, we also introduce a new data-augmentation technique which mitigates the impact of illumination difference between the background frames and the current frame. On the CDNet-2014 dataset, BSUV-Net outperforms stateof-the-art algorithms evaluated on unseen videos in terms of several metrics including F-measure, recall and precision.Accepted manuscrip

    Holographic Imaging of Crowded Fields: High Angular Resolution Imaging with Excellent Quality at Very Low Cost

    Full text link
    We present a method for speckle holography that is optimised for crowded fields. Its two key features are an iterativ improvement of the instantaneous Point Spread Functions (PSFs) extracted from each speckle frame and the (optional) simultaneous use of multiple reference stars. In this way, high signal-to-noise and accuracy can be achieved on the PSF for each short exposure, which results in sensitive, high-Strehl re- constructed images. We have tested our method with different instruments, on a range of targets, and from the N- to the I-band. In terms of PSF cosmetics, stability and Strehl ratio, holographic imaging can be equal, and even superior, to the capabilities of currently available Adaptive Optics (AO) systems, particularly at short near-infrared to optical wavelengths. It outperforms lucky imaging because it makes use of the entire PSF and reduces the need for frame selection, thus leading to higher Strehl and improved sensitivity. Image reconstruction a posteriori, the possibility to use multiple reference stars and the fact that these reference stars can be rather faint means that holographic imaging offers a simple way to image large, dense stellar fields near the diffraction limit of large telescopes, similar to, but much less technologically demanding than, the capabilities of a multi-conjugate adaptive optics system. The method can be used with a large range of already existing imaging instruments and can also be combined with AO imaging when the corrected PSF is unstable.Comment: Accepted for publication in MNRAS on 15 Nov 201

    A fully-convolutional neural network for background subtraction of unseen videos

    Full text link
    Background subtraction is a basic task in computer vision and video processing often applied as a pre-processing step for object tracking, people recognition, etc. Recently, a number of successful background-subtraction algorithms have been proposed, however nearly all of the top-performing ones are supervised. Crucially, their success relies upon the availability of some annotated frames of the test video during training. Consequently, their performance on completely “unseen” videos is undocumented in the literature. In this work, we propose a new, supervised, backgroundsubtraction algorithm for unseen videos (BSUV-Net) based on a fully-convolutional neural network. The input to our network consists of the current frame and two background frames captured at different time scales along with their semantic segmentation maps. In order to reduce the chance of overfitting, we also introduce a new data-augmentation technique which mitigates the impact of illumination difference between the background frames and the current frame. On the CDNet-2014 dataset, BSUV-Net outperforms stateof-the-art algorithms evaluated on unseen videos in terms of several metrics including F-measure, recall and precision.Accepted manuscrip

    Democratic Representations

    Full text link
    Minimization of the ℓ∞\ell_{\infty} (or maximum) norm subject to a constraint that imposes consistency to an underdetermined system of linear equations finds use in a large number of practical applications, including vector quantization, approximate nearest neighbor search, peak-to-average power ratio (or "crest factor") reduction in communication systems, and peak force minimization in robotics and control. This paper analyzes the fundamental properties of signal representations obtained by solving such a convex optimization problem. We develop bounds on the maximum magnitude of such representations using the uncertainty principle (UP) introduced by Lyubarskii and Vershynin, and study the efficacy of ℓ∞\ell_{\infty}-norm-based dynamic range reduction. Our analysis shows that matrices satisfying the UP, such as randomly subsampled Fourier or i.i.d. Gaussian matrices, enable the computation of what we call democratic representations, whose entries all have small and similar magnitude, as well as low dynamic range. To compute democratic representations at low computational complexity, we present two new, efficient convex optimization algorithms. We finally demonstrate the efficacy of democratic representations for dynamic range reduction in a DVB-T2-based broadcast system.Comment: Submitted to a Journa

    Tracking in Urban Traffic Scenes from Background Subtraction and Object Detection

    Full text link
    In this paper, we propose to combine detections from background subtraction and from a multiclass object detector for multiple object tracking (MOT) in urban traffic scenes. These objects are associated across frames using spatial, colour and class label information, and trajectory prediction is evaluated to yield the final MOT outputs. The proposed method was tested on the Urban tracker dataset and shows competitive performances compared to state-of-the-art approaches. Results show that the integration of different detection inputs remains a challenging task that greatly affects the MOT performance
    • 

    corecore