58,860 research outputs found
Dense Motion Estimation for Smoke
Motion estimation for highly dynamic phenomena such as smoke is an open
challenge for Computer Vision. Traditional dense motion estimation algorithms
have difficulties with non-rigid and large motions, both of which are
frequently observed in smoke motion. We propose an algorithm for dense motion
estimation of smoke. Our algorithm is robust, fast, and has better performance
over different types of smoke compared to other dense motion estimation
algorithms, including state of the art and neural network approaches. The key
to our contribution is to use skeletal flow, without explicit point matching,
to provide a sparse flow. This sparse flow is upgraded to a dense flow. In this
paper we describe our algorithm in greater detail, and provide experimental
evidence to support our claims.Comment: ACCV201
Dynamic texture recognition using time-causal and time-recursive spatio-temporal receptive fields
This work presents a first evaluation of using spatio-temporal receptive
fields from a recently proposed time-causal spatio-temporal scale-space
framework as primitives for video analysis. We propose a new family of video
descriptors based on regional statistics of spatio-temporal receptive field
responses and evaluate this approach on the problem of dynamic texture
recognition. Our approach generalises a previously used method, based on joint
histograms of receptive field responses, from the spatial to the
spatio-temporal domain and from object recognition to dynamic texture
recognition. The time-recursive formulation enables computationally efficient
time-causal recognition. The experimental evaluation demonstrates competitive
performance compared to state-of-the-art. Especially, it is shown that binary
versions of our dynamic texture descriptors achieve improved performance
compared to a large range of similar methods using different primitives either
handcrafted or learned from data. Further, our qualitative and quantitative
investigation into parameter choices and the use of different sets of receptive
fields highlights the robustness and flexibility of our approach. Together,
these results support the descriptive power of this family of time-causal
spatio-temporal receptive fields, validate our approach for dynamic texture
recognition and point towards the possibility of designing a range of video
analysis methods based on these new time-causal spatio-temporal primitives.Comment: 29 pages, 16 figure
Friction Variability in Planar Pushing Data: Anisotropic Friction and Data-collection Bias
Friction plays a key role in manipulating objects. Most of what we do with
our hands, and most of what robots do with their grippers, is based on the
ability to control frictional forces. This paper aims to better understand the
variability and predictability of planar friction. In particular, we focus on
the analysis of a recent dataset on planar pushing by Yu et al. [1] devised to
create a data-driven footprint of planar friction.
We show in this paper how we can explain a significant fraction of the
observed unconventional phenomena, e.g., stochasticity and multi-modality, by
combining the effects of material non-homogeneity, anisotropy of friction and
biases due to data collection dynamics, hinting that the variability is
explainable but inevitable in practice.
We introduce an anisotropic friction model and conduct simulation experiments
comparing with more standard isotropic friction models. The anisotropic
friction between object and supporting surface results in convergence of
initial condition during the automated data collection. Numerical results
confirm that the anisotropic friction model explains the bias in the dataset
and the apparent stochasticity in the outcome of a push. The fact that the data
collection process itself can originate biases in the collected datasets,
resulting in deterioration of trained models, calls attention to the data
collection dynamics.Comment: 8 pages, 13 figure
Learning Deep Representations of Appearance and Motion for Anomalous Event Detection
We present a novel unsupervised deep learning framework for anomalous event
detection in complex video scenes. While most existing works merely use
hand-crafted appearance and motion features, we propose Appearance and Motion
DeepNet (AMDN) which utilizes deep neural networks to automatically learn
feature representations. To exploit the complementary information of both
appearance and motion patterns, we introduce a novel double fusion framework,
combining both the benefits of traditional early fusion and late fusion
strategies. Specifically, stacked denoising autoencoders are proposed to
separately learn both appearance and motion features as well as a joint
representation (early fusion). Based on the learned representations, multiple
one-class SVM models are used to predict the anomaly scores of each input,
which are then integrated with a late fusion strategy for final anomaly
detection. We evaluate the proposed method on two publicly available video
surveillance datasets, showing competitive performance with respect to state of
the art approaches.Comment: Oral paper in BMVC 201
- …