2,713 research outputs found
UAV-Sim: NeRF-based Synthetic Data Generation for UAV-based Perception
Tremendous variations coupled with large degrees of freedom in UAV-based
imaging conditions lead to a significant lack of data in adequately learning
UAV-based perception models. Using various synthetic renderers in conjunction
with perception models is prevalent to create synthetic data to augment the
learning in the ground-based imaging domain. However, severe challenges in the
austere UAV-based domain require distinctive solutions to image synthesis for
data augmentation. In this work, we leverage recent advancements in neural
rendering to improve static and dynamic novelview UAV-based image synthesis,
especially from high altitudes, capturing salient scene attributes. Finally, we
demonstrate a considerable performance boost is achieved when a state-ofthe-art
detection model is optimized primarily on hybrid sets of real and synthetic
data instead of the real or synthetic data separately.Comment: Video Link: https://www.youtube.com/watch?v=ucPzbPLqqp
A Comprehensive Review on Computer Vision Analysis of Aerial Data
With the emergence of new technologies in the field of airborne platforms and
imaging sensors, aerial data analysis is becoming very popular, capitalizing on
its advantages over land data. This paper presents a comprehensive review of
the computer vision tasks within the domain of aerial data analysis. While
addressing fundamental aspects such as object detection and tracking, the
primary focus is on pivotal tasks like change detection, object segmentation,
and scene-level analysis. The paper provides the comparison of various hyper
parameters employed across diverse architectures and tasks. A substantial
section is dedicated to an in-depth discussion on libraries, their
categorization, and their relevance to different domain expertise. The paper
encompasses aerial datasets, the architectural nuances adopted, and the
evaluation metrics associated with all the tasks in aerial data analysis.
Applications of computer vision tasks in aerial data across different domains
are explored, with case studies providing further insights. The paper
thoroughly examines the challenges inherent in aerial data analysis, offering
practical solutions. Additionally, unresolved issues of significance are
identified, paving the way for future research directions in the field of
aerial data analysis.Comment: 112 page
MOR-UAV: A Benchmark Dataset and Baselines for Moving Object Recognition in UAV Videos
Visual data collected from Unmanned Aerial Vehicles (UAVs) has opened a new
frontier of computer vision that requires automated analysis of aerial
images/videos. However, the existing UAV datasets primarily focus on object
detection. An object detector does not differentiate between the moving and
non-moving objects. Given a real-time UAV video stream, how can we both
localize and classify the moving objects, i.e. perform moving object
recognition (MOR)? The MOR is one of the essential tasks to support various UAV
vision-based applications including aerial surveillance, search and rescue,
event recognition, urban and rural scene understanding.To the best of our
knowledge, no labeled dataset is available for MOR evaluation in UAV videos.
Therefore, in this paper, we introduce MOR-UAV, a large-scale video dataset for
MOR in aerial videos. We achieve this by labeling axis-aligned bounding boxes
for moving objects which requires less computational resources than producing
pixel-level estimates. We annotate 89,783 moving object instances collected
from 30 UAV videos, consisting of 10,948 frames in various scenarios such as
weather conditions, occlusion, changing flying altitude and multiple camera
views. We assigned the labels for two categories of vehicles (car and heavy
vehicle). Furthermore, we propose a deep unified framework MOR-UAVNet for MOR
in UAV videos. Since, this is a first attempt for MOR in UAV videos, we present
16 baseline results based on the proposed framework over the MOR-UAV dataset
through quantitative and qualitative experiments. We also analyze the
motion-salient regions in the network through multiple layer visualizations.
The MOR-UAVNet works online at inference as it requires only few past frames.
Moreover, it doesn't require predefined target initialization from user.
Experiments also demonstrate that the MOR-UAV dataset is quite challenging
Transformers in Small Object Detection: A Benchmark and Survey of State-of-the-Art
Transformers have rapidly gained popularity in computer vision, especially in
the field of object recognition and detection. Upon examining the outcomes of
state-of-the-art object detection methods, we noticed that transformers
consistently outperformed well-established CNN-based detectors in almost every
video or image dataset. While transformer-based approaches remain at the
forefront of small object detection (SOD) techniques, this paper aims to
explore the performance benefits offered by such extensive networks and
identify potential reasons for their SOD superiority. Small objects have been
identified as one of the most challenging object types in detection frameworks
due to their low visibility. We aim to investigate potential strategies that
could enhance transformers' performance in SOD. This survey presents a taxonomy
of over 60 research studies on developed transformers for the task of SOD,
spanning the years 2020 to 2023. These studies encompass a variety of detection
applications, including small object detection in generic images, aerial
images, medical images, active millimeter images, underwater images, and
videos. We also compile and present a list of 12 large-scale datasets suitable
for SOD that were overlooked in previous studies and compare the performance of
the reviewed studies using popular metrics such as mean Average Precision
(mAP), Frames Per Second (FPS), number of parameters, and more. Researchers can
keep track of newer studies on our web page, which is available at
\url{https://github.com/arekavandi/Transformer-SOD}
- …