16,389 research outputs found
Detection of Moving Object in Dynamic Background Using Gaussian Max-Pooling and Segmentation Constrained RPCA
Due to its efficiency and stability, Robust Principal Component Analysis
(RPCA) has been emerging as a promising tool for moving object detection.
Unfortunately, existing RPCA based methods assume static or quasi-static
background, and thereby they may have trouble in coping with the background
scenes that exhibit a persistent dynamic behavior. In this work, we shall
introduce two techniques to fill in the gap. First, instead of using the raw
pixel-value as features that are brittle in the presence of dynamic background,
we devise a so-called Gaussian max-pooling operator to estimate a
"stable-value" for each pixel. Those stable-values are robust to various
background changes and can therefore distinguish effectively the foreground
objects from the background. Then, to obtain more accurate results, we further
propose a Segmentation Constrained RPCA (SC-RPCA) model, which incorporates the
temporal and spatial continuity in images into RPCA. The inference process of
SC-RPCA is a group sparsity constrained nuclear norm minimization problem,
which is convex and easy to solve. Experimental results on seven videos from
the CDCNET 2014 database show the superior performance of the proposed method
Comparative study of motion detection methods for video surveillance systems
The objective of this study is to compare several change detection methods
for a mono static camera and identify the best method for different complex
environments and backgrounds in indoor and outdoor scenes. To this end, we used
the CDnet video dataset as a benchmark that consists of many challenging
problems, ranging from basic simple scenes to complex scenes affected by bad
weather and dynamic backgrounds. Twelve change detection methods, ranging from
simple temporal differencing to more sophisticated methods, were tested and
several performance metrics were used to precisely evaluate the results.
Because most of the considered methods have not previously been evaluated on
this recent large scale dataset, this work compares these methods to fill a
lack in the literature, and thus this evaluation joins as complementary
compared with the previous comparative evaluations. Our experimental results
show that there is no perfect method for all challenging cases, each method
performs well in certain cases and fails in others. However, this study enables
the user to identify the most suitable method for his or her needs.Comment: 69 pages, 18 figures, journal pape
Optimal Continuous State POMDP Planning with Semantic Observations: A Variational Approach
This work develops novel strategies for optimal planning with semantic
observations using continuous state partially observable markov decision
processes (CPOMDPs). Two major innovations are presented in relation to
Gaussian mixture (GM) CPOMDP policy approximation methods. While existing
methods have many desirable theoretical properties, they are unable to
efficiently represent and reason over hybrid continuous-discrete probabilistic
models. The first major innovation is the derivation of closed-form variational
Bayes GM approximations of Point-Based Value Iteration Bellman policy backups,
using softmax models of continuous-discrete semantic observation probabilities.
A key benefit of this approach is that dynamic decision-making tasks can be
performed with complex non-Gaussian uncertainties, while also exploiting
continuous dynamic state space models (thus avoiding cumbersome and costly
discretization). The second major innovation is a new clustering-based
technique for mixture condensation that scales well to very large GM policy
functions and belief functions. Simulation results for a target search and
interception task with semantic observations show that the GM policies
resulting from these innovations are more effective than those produced by
other state of the art policy approximations, but require significantly less
modeling overhead and online runtime cost. Additional results show the
robustness of this approach to model errors and scaling to higher dimensions.Comment: Final version accepted to IEEE Transactions on Robotics (in press as
of August 2019
Background Subtraction in Real Applications: Challenges, Current Models and Future Directions
Computer vision applications based on videos often require the detection of
moving objects in their first step. Background subtraction is then applied in
order to separate the background and the foreground. In literature, background
subtraction is surely among the most investigated field in computer vision
providing a big amount of publications. Most of them concern the application of
mathematical and machine learning models to be more robust to the challenges
met in videos. However, the ultimate goal is that the background subtraction
methods developed in research could be employed in real applications like
traffic surveillance. But looking at the literature, we can remark that there
is often a gap between the current methods used in real applications and the
current methods in fundamental research. In addition, the videos evaluated in
large-scale datasets are not exhaustive in the way that they only covered a
part of the complete spectrum of the challenges met in real applications. In
this context, we attempt to provide the most exhaustive survey as possible on
real applications that used background subtraction in order to identify the
real challenges met in practice, the current used background models and to
provide future directions. Thus, challenges are investigated in terms of
camera, foreground objects and environments. In addition, we identify the
background models that are effectively used in these applications in order to
find potential usable recent background models in terms of robustness, time and
memory requirements.Comment: Submitted to Computer Science Revie
A Survey Of Activity Recognition And Understanding The Behavior In Video Survelliance
This paper presents a review of human activity recognition and behaviour
understanding in video sequence. The key objective of this paper is to provide
a general review on the overall process of a surveillance system used in the
current trend. Visual surveillance system is directed on automatic
identification of events of interest, especially on tracking and classification
of moving objects. The processing step of the video surveillance system
includes the following stages: Surrounding model, object representation, object
tracking, activity recognition and behaviour understanding. It describes
techniques that use to define a general set of activities that are applicable
to a wide range of scenes and environments in video sequence.Comment: 14 pages, 5 figures, 5 table
Unsupervised Deep Context Prediction for Background Foreground Separation
In many advanced video based applications background modeling is a
pre-processing step to eliminate redundant data, for instance in tracking or
video surveillance applications. Over the past years background subtraction is
usually based on low level or hand-crafted features such as raw color
components, gradients, or local binary patterns. The background subtraction
algorithms performance suffer in the presence of various challenges such as
dynamic backgrounds, photometric variations, camera jitters, and shadows. To
handle these challenges for the purpose of accurate background modeling we
propose a unified framework based on the algorithm of image inpainting. It is
an unsupervised visual feature learning hybrid Generative Adversarial algorithm
based on context prediction. We have also presented the solution of random
region inpainting by the fusion of center region inpaiting and random region
inpainting with the help of poisson blending technique. Furthermore we also
evaluated foreground object detection with the fusion of our proposed method
and morphological operations. The comparison of our proposed method with 12
state-of-the-art methods shows its stability in the application of background
estimation and foreground detection.Comment: 17 page
Background subtraction using the factored 3-way restricted Boltzmann machines
In this paper, we proposed a method for reconstructing the 3D model based on
continuous sensory input. The robot can draw on extremely large data from the
real world using various sensors. However, the sensory inputs are usually too
noisy and high-dimensional data. It is very difficult and time consuming for
robot to process using such raw data when the robot tries to construct 3D
model. Hence, there needs to be a method that can extract useful information
from such sensory inputs. To address this problem our method utilizes the
concept of Object Semantic Hierarchy (OSH). Different from the previous work
that used this hierarchy framework, we extract the motion information using the
Deep Belief Network technique instead of applying classical computer vision
approaches. We have trained on two large sets of random dot images (10,000)
which are translated and rotated, respectively, and have successfully extracted
several bases that explain the translation and rotation motion. Based on this
translation and rotation bases, background subtraction have become possible
using Object Semantic Hierarchy.Comment: EECS545 (2011 Winter) class project report at the University of
Michigan. This is for archiving purpos
Vision-Guided Robot Hearing
Natural human-robot interaction in complex and unpredictable environments is
one of the main research lines in robotics. In typical real-world scenarios,
humans are at some distance from the robot and the acquired signals are
strongly impaired by noise, reverberations and other interfering sources. In
this context, the detection and localisation of speakers plays a key role since
it is the pillar on which several tasks (e.g.: speech recognition and speaker
tracking) rely. We address the problem of how to detect and localize people
that are both seen and heard by a humanoid robot. We introduce a hybrid
deterministic/probabilistic model. Indeed, the deterministic component allows
us to map the visual information into the auditory space. By means of the
probabilistic component, the visual features guide the grouping of the auditory
features in order to form AV objects. The proposed model and the associated
algorithm are implemented in real-time (17 FPS) using a stereoscopic camera
pair and two microphones embedded into the head of the humanoid robot NAO. We
performed experiments on (i) synthetic data, (ii) a publicly available data set
and (iii) data acquired using the robot. The results we obtained validate the
approach and encourage us to further investigate how vision can help robot
hearing.Comment: 26 pages, many figures and tables, journa
Audio Surveillance: a Systematic Review
Despite surveillance systems are becoming increasingly ubiquitous in our
living environment, automated surveillance, currently based on video sensory
modality and machine intelligence, lacks most of the time the robustness and
reliability required in several real applications. To tackle this issue, audio
sensory devices have been taken into account, both alone or in combination with
video, giving birth, in the last decade, to a considerable amount of research.
In this paper audio-based automated surveillance methods are organized into a
comprehensive survey: a general taxonomy, inspired by the more widespread video
surveillance field, is proposed in order to systematically describe the methods
covering background subtraction, event classification, object tracking and
situation analysis. For each of these tasks, all the significant works are
reviewed, detailing their pros and cons and the context for which they have
been proposed. Moreover, a specific section is devoted to audio features,
discussing their expressiveness and their employment in the above described
tasks. Differently, from other surveys on audio processing and analysis, the
present one is specifically targeted to automated surveillance, highlighting
the target applications of each described methods and providing the reader
tables and schemes useful to retrieve the most suited algorithms for a specific
requirement
Solar Potential Analysis of Rooftops Using Satellite Imagery
Solar energy is one of the most important sources of renewable energy and the
cleanest form of energy. In India, where solar energy could produce power
around trillion kilowatt-hours in a year, our country is only able to produce
power of around in gigawatts only. Many people are not aware of the solar
potential of their rooftop, and hence they always think that installing solar
panels is very much expensive. In this work, we introduce an approach through
which we can generate a report remotely that provides the amount of solar
potential of a building using only its latitude and longitude. We further
evaluated various types of rooftops to make our solution more robust. We also
provide an approximate area of rooftop that can be used for solar panels
placement and a visual analysis of how solar panels can be placed to maximize
the output of solar power at a location
- …