353 research outputs found
Bounding Box-Free Instance Segmentation Using Semi-Supervised Learning for Generating a City-Scale Vehicle Dataset
Vehicle classification is a hot computer vision topic, with studies ranging
from ground-view up to top-view imagery. In remote sensing, the usage of
top-view images allows for understanding city patterns, vehicle concentration,
traffic management, and others. However, there are some difficulties when
aiming for pixel-wise classification: (a) most vehicle classification studies
use object detection methods, and most publicly available datasets are designed
for this task, (b) creating instance segmentation datasets is laborious, and
(c) traditional instance segmentation methods underperform on this task since
the objects are small. Thus, the present research objectives are: (1) propose a
novel semi-supervised iterative learning approach using GIS software, (2)
propose a box-free instance segmentation approach, and (3) provide a city-scale
vehicle dataset. The iterative learning procedure considered: (1) label a small
number of vehicles, (2) train on those samples, (3) use the model to classify
the entire image, (4) convert the image prediction into a polygon shapefile,
(5) correct some areas with errors and include them in the training data, and
(6) repeat until results are satisfactory. To separate instances, we considered
vehicle interior and vehicle borders, and the DL model was the U-net with the
Efficient-net-B7 backbone. When removing the borders, the vehicle interior
becomes isolated, allowing for unique object identification. To recover the
deleted 1-pixel borders, we proposed a simple method to expand each prediction.
The results show better pixel-wise metrics when compared to the Mask-RCNN (82%
against 67% in IoU). On per-object analysis, the overall accuracy, precision,
and recall were greater than 90%. This pipeline applies to any remote sensing
target, being very efficient for segmentation and generating datasets.Comment: 38 pages, 10 figures, submitted to journa
R-Net: A Deep Network for Multi-oriented Vehicle Detection in Aerial Images and Videos
Vehicle detection is a significant and challenging task in aerial remote
sensing applications. Most existing methods detect vehicles with regular
rectangle boxes and fail to offer the orientation of vehicles. However, the
orientation information is crucial for several practical applications, such as
the trajectory and motion estimation of vehicles. In this paper, we propose a
novel deep network, called rotatable region-based residual network (R-Net),
to detect multi-oriented vehicles in aerial images and videos. More specially,
R-Net is utilized to generate rotatable rectangular target boxes in a half
coordinate system. First, we use a rotatable region proposal network (R-RPN) to
generate rotatable region of interests (R-RoIs) from feature maps produced by a
deep convolutional neural network. Here, a proposed batch averaging rotatable
anchor (BAR anchor) strategy is applied to initialize the shape of vehicle
candidates. Next, we propose a rotatable detection network (R-DN) for the final
classification and regression of the R-RoIs. In R-DN, a novel rotatable
position sensitive pooling (R-PS pooling) is designed to keep the position and
orientation information simultaneously while downsampling the feature maps of
R-RoIs. In our model, R-RPN and R-DN can be trained jointly. We test our
network on two open vehicle detection image datasets, namely DLR 3K Munich
Dataset and VEDAI Dataset, demonstrating the high precision and robustness of
our method. In addition, further experiments on aerial videos show the good
generalization capability of the proposed method and its potential for vehicle
tracking in aerial videos. The demo video is available at
https://youtu.be/xCYD-tYudN0
Advanced framework for microscopic and laneâlevel macroscopic traffic parameters estimation from UAV video
Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/166282/1/itr2bf00873.pd
Deep Learning-Based Object Detection in Maritime Unmanned Aerial Vehicle Imagery: Review and Experimental Comparisons
With the advancement of maritime unmanned aerial vehicles (UAVs) and deep
learning technologies, the application of UAV-based object detection has become
increasingly significant in the fields of maritime industry and ocean
engineering. Endowed with intelligent sensing capabilities, the maritime UAVs
enable effective and efficient maritime surveillance. To further promote the
development of maritime UAV-based object detection, this paper provides a
comprehensive review of challenges, relative methods, and UAV aerial datasets.
Specifically, in this work, we first briefly summarize four challenges for
object detection on maritime UAVs, i.e., object feature diversity, device
limitation, maritime environment variability, and dataset scarcity. We then
focus on computational methods to improve maritime UAV-based object detection
performance in terms of scale-aware, small object detection, view-aware,
rotated object detection, lightweight methods, and others. Next, we review the
UAV aerial image/video datasets and propose a maritime UAV aerial dataset named
MS2ship for ship detection. Furthermore, we conduct a series of experiments to
present the performance evaluation and robustness analysis of object detection
methods on maritime datasets. Eventually, we give the discussion and outlook on
future works for maritime UAV-based object detection. The MS2ship dataset is
available at
\href{https://github.com/zcj234/MS2ship}{https://github.com/zcj234/MS2ship}.Comment: 32 pages, 18 figure
Automatic Pipeline Surveillance Air-Vehicle
This thesis presents the developments of a vision-based system for
aerial pipeline Right-of-Way surveillance using optical/Infrared sensors mounted
on Unmanned Aerial Vehicles (UAV). The aim of research is to develop a highly
automated, on-board system for detecting and following the pipelines; while
simultaneously detecting any third-party interference. The proposed approach
of using a UAV platform could potentially reduce the cost of monitoring and
surveying pipelines when compared to manned aircraft. The main contributions
of this thesis are the development of the image-analysis algorithms, the overall
system architecture and validation of in hardware based on scaled down Test
environment.
To evaluate the performance of the system, the algorithms were coded using
Python programming language. A small-scale test-rig of the pipeline structure,
as well as expected third-party interference, was setup to simulate the
operational environment and capture/record data for the algorithm testing and
validation.
The pipeline endpoints are identified by transforming the 16-bits depth data of
the explored environment into 3D point clouds world coordinates. Then, using
the Random Sample Consensus (RANSAC) approach, the foreground and
background are separated based on the transformed 3D point cloud to extract
the plane that corresponds to the ground. Simultaneously, the boundaries of the
explored environment are detected based on the 16-bit depth data using a
canny detector. Following that, these boundaries were filtered out, after being
transformed into a 3D point cloud, based on the real height of the pipeline for fast and accurate measurements using a Euclidean distance of each boundary
point, relative to the plane of the ground extracted previously. The filtered
boundaries were used to detect the straight lines of the object boundary (Hough
lines), once transformed into 16-bit depth data, using a Hough transform
method. The pipeline is verified by estimating a centre line segment, using a 3D
point cloud of each pair of the Hough line segments, (transformed into 3D).
Then, the corresponding linearity of the pipeline points cloud is filtered within
the width of the pipeline using Euclidean distance in the foreground point cloud.
Then, the segment length of the detected centre line is enhanced to match the
exact pipeline segment by extending it along the filtered point cloud of the
pipeline.
The third-party interference is detected based on four parameters, namely:
foreground depth data; pipeline depth data; pipeline endpoints location in the
3D point cloud; and Right-of-Way distance. The techniques include detection,
classification, and localization algorithms.
Finally, a waypoints-based navigation system was implemented for the air-
vehicle to fly over the course waypoints that were generated online by a
heading angle demand to follow the pipeline structure in real-time based on the
online identification of the pipeline endpoints relative to a camera frame
Smartphone-based object recognition with embedded machine learning intelligence for unmanned aerial vehicles
Existing artificial intelligence solutions typically operate in powerful platforms with high computational resources availability. However, a growing number of emerging use cases such as those based on unmanned aerial systems (UAS) require new solutions with embedded artificial intelligence on a highly mobile platform. This paper proposes an innovative UAS that explores machine learning (ML) capabilities in a smartphoneâbased mobile platform for object detection and recognition applications. A new system framework tailored to this challenging use case is designed with a customized workflow specified. Furthermore, the design of the embedded ML leverages TensorFlow, a cuttingâedge openâsource ML framework. The prototype of the system integrates all the architectural components in a fully functional system, and it is suitable for realâworld operational environments such as seek and rescue use cases. Experimental results validate the design and prototyping of the system and demonstrate an overall improved performance compared with the state of the art in terms of a wide range of metrics
- âŠ