536 research outputs found
The Segment Anything Model (SAM) for Remote Sensing Applications: From Zero to One Shot
Segmentation is an essential step for remote sensing image processing. This
study aims to advance the application of the Segment Anything Model (SAM), an
innovative image segmentation model by Meta AI, in the field of remote sensing
image analysis. SAM is known for its exceptional generalization capabilities
and zero-shot learning, making it a promising approach to processing aerial and
orbital images from diverse geographical contexts. Our exploration involved
testing SAM across multi-scale datasets using various input prompts, such as
bounding boxes, individual points, and text descriptors. To enhance the model's
performance, we implemented a novel automated technique that combines a
text-prompt-derived general example with one-shot training. This adjustment
resulted in an improvement in accuracy, underscoring SAM's potential for
deployment in remote sensing imagery and reducing the need for manual
annotation. Despite the limitations encountered with lower spatial resolution
images, SAM exhibits promising adaptability to remote sensing data analysis. We
recommend future research to enhance the model's proficiency through
integration with supplementary fine-tuning techniques and other networks.
Furthermore, we provide the open-source code of our modifications on online
repositories, encouraging further and broader adaptations of SAM to the remote
sensing domain.Comment: 20 pages, 9 figure
Smart environment monitoring through micro unmanned aerial vehicles
In recent years, the improvements of small-scale Unmanned Aerial Vehicles (UAVs) in terms of flight time, automatic control, and remote transmission are promoting the development of a wide range of practical applications. In aerial video surveillance, the monitoring of broad areas still has many challenges due to the achievement of different tasks in real-time, including mosaicking, change detection, and object detection. In this thesis work, a small-scale UAV based vision system to maintain regular surveillance over target areas is proposed. The system works in two modes. The first mode allows to monitor an area of interest by performing several flights. During the first flight, it creates an incremental geo-referenced mosaic of an area of interest and classifies all the known elements (e.g., persons) found on the ground by an improved Faster R-CNN architecture previously trained. In subsequent reconnaissance flights, the system searches for any changes (e.g., disappearance of persons) that may occur in the mosaic by a histogram equalization and RGB-Local Binary Pattern (RGB-LBP) based algorithm. If present, the mosaic is updated. The second mode, allows to perform a real-time classification by using, again, our improved Faster R-CNN model, useful for time-critical operations. Thanks to different design features, the system works in real-time and performs mosaicking and change detection tasks at low-altitude, thus allowing the classification even of small objects. The proposed system was tested by using the whole set of challenging video sequences contained in the UAV Mosaicking and Change Detection (UMCD) dataset and other public datasets. The evaluation of the system by well-known performance metrics has shown remarkable results in terms of mosaic creation and updating, as well as in terms of change detection and object detection
Small-Object Detection in Remote Sensing Images with End-to-End Edge-Enhanced GAN and Object Detector Network
The detection performance of small objects in remote sensing images is not
satisfactory compared to large objects, especially in low-resolution and noisy
images. A generative adversarial network (GAN)-based model called enhanced
super-resolution GAN (ESRGAN) shows remarkable image enhancement performance,
but reconstructed images miss high-frequency edge information. Therefore,
object detection performance degrades for small objects on recovered noisy and
low-resolution remote sensing images. Inspired by the success of edge enhanced
GAN (EEGAN) and ESRGAN, we apply a new edge-enhanced super-resolution GAN
(EESRGAN) to improve the image quality of remote sensing images and use
different detector networks in an end-to-end manner where detector loss is
backpropagated into the EESRGAN to improve the detection performance. We
propose an architecture with three components: ESRGAN, Edge Enhancement Network
(EEN), and Detection network. We use residual-in-residual dense blocks (RRDB)
for both the ESRGAN and EEN, and for the detector network, we use the faster
region-based convolutional network (FRCNN) (two-stage detector) and single-shot
multi-box detector (SSD) (one stage detector). Extensive experiments on a
public (car overhead with context) and a self-assembled (oil and gas storage
tank) satellite dataset show superior performance of our method compared to the
standalone state-of-the-art object detectors.Comment: This paper contains 27 pages and accepted for publication in MDPI
remote sensing journal. GitHub Repository:
https://github.com/Jakaria08/EESRGAN (Implementation
EVALUATING SAMPLING-BASED TECHNIQUE FOR IMPROVED SMALL OBJECT DETECTION USING APPROPRIATE SIZES OF THE ANCHOR BOXES
Abstract
Small object detection is a challenging task in computer vision due to the limited spatial information and low resolution of these objects. Existing generic object detectors demonstrate satisfactory performance on medium and large-sized objects, but they often struggle to accurately recognize small objects. The challenges arise from the low resolution and simple shape characteristics typically associated with small objects. In this paper, we propose an up sampling-based method for end-to-end tiny object identification that outperforms the existing state-of-the-art methods. We, like other contemporary approaches, first produce suggestions before labeling them. In the event of somewhat little things, we recommend tweaks to both of these procedures
Deep Learning Methods for Remote Sensing
Remote sensing is a field where important physical characteristics of an area are exacted using emitted radiation generally captured by satellite cameras, sensors onboard aerial vehicles, etc. Captured data help researchers develop solutions to sense and detect various characteristics such as forest fires, flooding, changes in urban areas, crop diseases, soil moisture, etc. The recent impressive progress in artificial intelligence (AI) and deep learning has sparked innovations in technologies, algorithms, and approaches and led to results that were unachievable until recently in multiple areas, among them remote sensing. This book consists of sixteen peer-reviewed papers covering new advances in the use of AI for remote sensing
EXPEDITIONARY LOGISTICS: A LOW-COST, DEPLOYABLE, UNMANNED AERIAL SYSTEM FOR AIRFIELD DAMAGE ASSESSMENT
Airfield Damage Repair (ADR) is among the most important expeditionary activities for our military. The goal of ADR is to restore a damaged airfield to operational status as quickly as possible. Before the process of ADR can begin, however, the damage to the airfield needs to be assessed. As a result, Airfield Damage Assessment (ADA) has received considerable attention. Often in a damaged airfield, there is an expectation of unexploded ordnance, which makes ADA a slow, difficult, and dangerous process. For this reason, it is best to make ADA completely unmanned and automated. Additionally, ADA needs to be executed as quickly as possible so that ADR can begin and the airfield restored to a usable condition. Among other modalities, tower-based monitoring and remote sensor systems are often used for ADA. There is now an opportunity to investigate the use of commercial-off-the-shelf, low-cost, automated sensor systems for automatic damage detection. By developing a combination of ground-based and Unmanned Aerial Vehicle sensor systems, we demonstrate the completion of ADA in a safe, efficient, and cost-effective manner.http://archive.org/details/expeditionarylog1094561346Outstanding ThesisLieutenant, United States NavyApproved for public release; distribution is unlimited
Bounding Box-Free Instance Segmentation Using Semi-Supervised Learning for Generating a City-Scale Vehicle Dataset
Vehicle classification is a hot computer vision topic, with studies ranging
from ground-view up to top-view imagery. In remote sensing, the usage of
top-view images allows for understanding city patterns, vehicle concentration,
traffic management, and others. However, there are some difficulties when
aiming for pixel-wise classification: (a) most vehicle classification studies
use object detection methods, and most publicly available datasets are designed
for this task, (b) creating instance segmentation datasets is laborious, and
(c) traditional instance segmentation methods underperform on this task since
the objects are small. Thus, the present research objectives are: (1) propose a
novel semi-supervised iterative learning approach using GIS software, (2)
propose a box-free instance segmentation approach, and (3) provide a city-scale
vehicle dataset. The iterative learning procedure considered: (1) label a small
number of vehicles, (2) train on those samples, (3) use the model to classify
the entire image, (4) convert the image prediction into a polygon shapefile,
(5) correct some areas with errors and include them in the training data, and
(6) repeat until results are satisfactory. To separate instances, we considered
vehicle interior and vehicle borders, and the DL model was the U-net with the
Efficient-net-B7 backbone. When removing the borders, the vehicle interior
becomes isolated, allowing for unique object identification. To recover the
deleted 1-pixel borders, we proposed a simple method to expand each prediction.
The results show better pixel-wise metrics when compared to the Mask-RCNN (82%
against 67% in IoU). On per-object analysis, the overall accuracy, precision,
and recall were greater than 90%. This pipeline applies to any remote sensing
target, being very efficient for segmentation and generating datasets.Comment: 38 pages, 10 figures, submitted to journa
Deep Learning-Based Object Detection in Maritime Unmanned Aerial Vehicle Imagery: Review and Experimental Comparisons
With the advancement of maritime unmanned aerial vehicles (UAVs) and deep
learning technologies, the application of UAV-based object detection has become
increasingly significant in the fields of maritime industry and ocean
engineering. Endowed with intelligent sensing capabilities, the maritime UAVs
enable effective and efficient maritime surveillance. To further promote the
development of maritime UAV-based object detection, this paper provides a
comprehensive review of challenges, relative methods, and UAV aerial datasets.
Specifically, in this work, we first briefly summarize four challenges for
object detection on maritime UAVs, i.e., object feature diversity, device
limitation, maritime environment variability, and dataset scarcity. We then
focus on computational methods to improve maritime UAV-based object detection
performance in terms of scale-aware, small object detection, view-aware,
rotated object detection, lightweight methods, and others. Next, we review the
UAV aerial image/video datasets and propose a maritime UAV aerial dataset named
MS2ship for ship detection. Furthermore, we conduct a series of experiments to
present the performance evaluation and robustness analysis of object detection
methods on maritime datasets. Eventually, we give the discussion and outlook on
future works for maritime UAV-based object detection. The MS2ship dataset is
available at
\href{https://github.com/zcj234/MS2ship}{https://github.com/zcj234/MS2ship}.Comment: 32 pages, 18 figure
Computer Vision in Wind Turbine Blade Inspections: An Analysis of Resolution Impact on Detection and Classification of Leading-Edge Erosion
Wind turbines, as critical components of the renewable energy industry, present unique maintenance challenges, particularly in remote or challenging locations such as offshore wind farms. These are amplified in the inspection of leading-edge erosion on wind turbine blades, a task still largely reliant on traditional methods. Emerging technologies like computer vision and object detection offer promising avenues for enhancing inspections, potentially reducing operational costs and human-associated risks. However, variability in image resolution, a critical factor for these technologies, remains a largely underexplored aspect in the wind energy context. This study explores the application of machine learning in detecting and categorizing leading edge erosion damage on wind turbine blades. YOLOv7, a state-of-the-art object detection model, is trained with a custom dataset consisting of images displaying various forms of leading edge erosion, representing multiple categories of damage severity. Trained model is tested on images acquired with three different tools, each providing images with a different resolution. The effect of image resolution on the performance of the custom object detection model is examined. The research affirms that the YOLOv7 model performs exceptionally well in identifying the most severe types of LEE damage, usually classified as Category 3, characterized by distinct visual features. However, the model's ability to detect less severe damage, namely Category 1 and 2, which are crucial for early detection and preventive measures, exhibits room for improvement. The findings point to a potential correlation between input image resolution and detection confidence in the context of wind turbine maintenance. These results stress the need for high-resolution images, leading to a discussion on the selection of appropriate imaging hardware and the creation of machine learning-ready datasets. The study thereby emphasizes the importance of industry-wide efforts to compile standardized image datasets and the potential impact of machine learning techniques on the efficiency of visual inspections and maintenance strategies. Future directions are proposed with the ultimate aim of enhancing the application of artificial intelligence in wind energy maintenance and management, enabling more efficient and effective operational procedures, and driving the industry towards a more sustainable future
- …