3 research outputs found
Deep invariant texture features for water image classification
Detecting potential issues in naturally captured images of water is a challenging task due to visual similarities between clean and polluted water, as well as causes posed by image acquisition with different camera angles and placements. This paper presents novel deep invariant texture features along with a deep network for detecting clean and polluted water images. The proposed method first divides an input image into H, S and V components to extract finer details. For each of the color spaces, the proposed approach generates two directional coherence images based on Eigen value analysis and gradient distribution, which results in enhanced images. Then the proposed method extracts scale invariant gradient orientations based on Gaussian first order derivative filters on different standard deviations to study texture of each smoothed image. To strengthen the above features, we explore the combination of Gabor-wavelet-binary pattern for extracting texture of the input water image. The proposed method integrates merits of aforementioned features and the features extracted by VGG16 deep learning model to obtain a single feature vector. Furthermore, the extracted feature is fed to a gradient boosting decision tree for water image detection. A variety of experimental results on a large dataset containing different types of clean and stagnant water images show that the proposed method outperforms the existing methods in terms of classification rate and accuracy
Vision based estimation, localization, and mapping for autonomous vehicles
In this dissertation, we focus on developing simultaneous localization and mapping (SLAM) algorithms with a robot-centric estimation framework primarily using monocular vision sensors. A primary contribution of this work is to use a robot-centric mapping framework concurrently with a world-centric localization method. We exploit the differential equation of motion of the normalized pixel coordinates of each point feature in the robot body frame. Another contribution of our work is to exploit a multiple-view geometry formulation with initial and current view projection of point features. We extract the features from objects surrounding the river and their reflections. The correspondences of the features are used along with the attitude and altitude information of the robot. We demonstrate that the observability of the estimation system is improved by applying our robot-centric mapping framework and multiple-view measurements. Using the robot-centric mapping framework and multiple-view measurements including reflection of features, we present a vision based localization and mapping algorithm that we developed for an unmanned aerial vehicle (UAV) flying in a riverine environment. Our algorithm estimates the 3D positions of point features along a river and the pose of the UAV. Our UAV is equipped with a lightweight monocular camera, an inertial measurement unit (IMU), a magnetometer, an altimeter, and an onboard computer. To our knowledge, we report the first result that exploits the reflections of features in a riverine environment for localization and mapping. We also present an omnidirectional vision based localization and mapping system for a lawn mowing robot. Our algorithm can detect whether the robotic mower is contained in a permitted area. Our robotic mower is modified with an omnidirectional camera, an IMU, a magnetometer, and a vehicle speed sensor. Here, we also exploit the robot-centric mapping framework. The estimator in our system generates a 3D point based map with landmarks. Concurrently, the estimator defines a boundary of the mowing area by using the estimated trajectory of the mower. The estimated boundary and the landmark map are provided for the estimation of the mowing location and for the containment detection. First, we derive a nonlinear observer with contraction analysis and pseudo-measurements of the depth of each landmark to prevent the map estimator from diverging. Of particular interest for this work is ensuring that the estimator for localization and mapping will not fail due to the nonlinearity of the system model. For batch estimation, we design a hybrid extended Kalman smoother for our localization and robot-centric mapping model. Finally, we present a single camera based SLAM algorithm using a convex optimization based nonlinear estimator. We validate the effectiveness of our algorithms through numerical simulations and outdoor experiments
Large Scale Pattern Detection in Videos and Images from the Wild
PhDPattern detection is a well-studied area of computer vision, but still current methods are
unstable in images of poor quality. This thesis describes improvements over contemporary
methods in the fast detection of unseen patterns in a large corpus of videos that vary
tremendously in colour and texture definition, captured “in the wild” by mobile devices
and surveillance cameras.
We focus on three key areas of this broad subject;
First, we identify consistency weaknesses in existing techniques of processing an image
and it’s horizontally reflected (mirror) image. This is important in police investigations
where subjects change their appearance to try to avoid recognition, and we propose that
invariance to horizontal reflection should be more widely considered in image description
and recognition tasks too. We observe online Deep Learning system behaviours in
this respect, and provide a comprehensive assessment of 10 popular low level feature
detectors.
Second, we develop simple and fast algorithms that combine to provide memory- and
processing-efficient feature matching. These involve static scene elimination in the presence
of noise and on-screen time indicators, a blur-sensitive feature detection that finds
a greater number of corresponding features in images of varying sharpness, and a combinatorial
texture and colour feature matching algorithm that matches features when
either attribute may be poorly defined. A comprehensive evaluation is given, showing
some improvements over existing feature correspondence methods.
Finally, we study random decision forests for pattern detection. A new method of
indexing patterns in video sequences is devised and evaluated. We automatically label
positive and negative image training data, reducing a task of unsupervised learning to
one of supervised learning, and devise a node split function that is invariant to mirror
reflection and rotation through 90 degree angles. A high dimensional vote accumulator
encodes the hypothesis support, yielding implicit back-projection for pattern detection.European Union’s Seventh Framework Programme, specific
topic “framework and tools for (semi-) automated exploitation of massive amounts of digital data
for forensic purposes”, under grant agreement number 607480 (LASIE IP project)