677 research outputs found
Primjena neuronskih mreža za otkrivanje i klasifikaciju topničkih ciljeva
Neural networks have been in use since the 1950s and are increasingly prevalent in various domains of human activity, including military applications. Notably, GoogLeNet and convolutional neural networks, when appropriately trained, are instrumental in identifying and detecting individual objects within a complex set. In military scenarios, neural networks play a crucial role in the fire support process, especially when receiving target descriptions from forward observers. These networks are trained on image datasets to recognize specific features of individual elements or military objects, such as vehicles. As a result of this training, when presented with a new image, the network can accurately determine the type of vehicle, expediting the targeting process and enhancing the ability to provide a suitable response. This paper describes the application of neural networks for detecting and classifying artillery targets. It presents a specific problem and proposes a scientific solution, including explaining the methodology used and the results obtained.Neuronske mreže u uporabi su od pedesetih godina prošlog stoljeća i sve su zastupljenije u različitim područjima ljudske aktivnosti, uključujući vojne primjene. Posebno, GoogleNet i konvolucijske mreže, kada su odgovarajuće utrenirane, ključne su u prepoznavanju i otkrivanju pojedinih objekata unutar složenog skupa. U vojnim scenarijima neuronske mreže imaju ključnu ulogu u postupku pružanja potpore vatrom, posebno kada primaju opise ciljeva od prednjih topničkih motritelja. Ove mreže trenirane su na slikovnim skupovima podataka kako bi prepoznale specifičnosti pojedinih elemenata ili vojnih objekata, kao što su vozila. Kao rezultat treniranja, kada se prikaže nova slika, mreža može točno odrediti tip vozila, ubrzati postupak ciljanja i poboljšati sposobnost pružanja prikladnog odgovora. U radu se opisuje primjena neuronskih mreža za otkrivanje i klasifikaciju topničkih ciljeva. Rad predstavlja poseban problem i predlaže rješenje primjenom znanstvenog pristupa, uključujući objašnjenje korištene metodologije i dobivene rezultate
Learning from Synthetic Humans
Estimating human pose, shape, and motion from images and videos are
fundamental challenges with many applications. Recent advances in 2D human pose
estimation use large amounts of manually-labeled training data for learning
convolutional neural networks (CNNs). Such data is time consuming to acquire
and difficult to extend. Moreover, manual labeling of 3D pose, depth and motion
is impractical. In this work we present SURREAL (Synthetic hUmans foR REAL
tasks): a new large-scale dataset with synthetically-generated but realistic
images of people rendered from 3D sequences of human motion capture data. We
generate more than 6 million frames together with ground truth pose, depth
maps, and segmentation masks. We show that CNNs trained on our synthetic
dataset allow for accurate human depth estimation and human part segmentation
in real RGB images. Our results and the new dataset open up new possibilities
for advancing person analysis using cheap and large-scale synthetic data.Comment: Appears in: 2017 IEEE Conference on Computer Vision and Pattern
Recognition (CVPR 2017). 9 page
Multi-level Feature Fusion-based CNN for Local Climate Zone Classification from Sentinel-2 Images: Benchmark Results on the So2Sat LCZ42 Dataset
As a unique classification scheme for urban forms and functions, the local
climate zone (LCZ) system provides essential general information for any
studies related to urban environments, especially on a large scale. Remote
sensing data-based classification approaches are the key to large-scale mapping
and monitoring of LCZs. The potential of deep learning-based approaches is not
yet fully explored, even though advanced convolutional neural networks (CNNs)
continue to push the frontiers for various computer vision tasks. One reason is
that published studies are based on different datasets, usually at a regional
scale, which makes it impossible to fairly and consistently compare the
potential of different CNNs for real-world scenarios. This study is based on
the big So2Sat LCZ42 benchmark dataset dedicated to LCZ classification. Using
this dataset, we studied a range of CNNs of varying sizes. In addition, we
proposed a CNN to classify LCZs from Sentinel-2 images, Sen2LCZ-Net. Using this
base network, we propose fusing multi-level features using the extended
Sen2LCZ-Net-MF. With this proposed simple network architecture and the highly
competitive benchmark dataset, we obtain results that are better than those
obtained by the state-of-the-art CNNs, while requiring less computation with
fewer layers and parameters. Large-scale LCZ classification examples of
completely unseen areas are presented, demonstrating the potential of our
proposed Sen2LCZ-Net-MF as well as the So2Sat LCZ42 dataset. We also
intensively investigated the influence of network depth and width and the
effectiveness of the design choices made for Sen2LCZ-Net-MF. Our work will
provide important baselines for future CNN-based algorithm developments for
both LCZ classification and other urban land cover land use classification
A Comprehensive Survey of Deep Learning in Remote Sensing: Theories, Tools and Challenges for the Community
In recent years, deep learning (DL), a re-branding of neural networks (NNs),
has risen to the top in numerous areas, namely computer vision (CV), speech
recognition, natural language processing, etc. Whereas remote sensing (RS)
possesses a number of unique challenges, primarily related to sensors and
applications, inevitably RS draws from many of the same theories as CV; e.g.,
statistics, fusion, and machine learning, to name a few. This means that the RS
community should be aware of, if not at the leading edge of, of advancements
like DL. Herein, we provide the most comprehensive survey of state-of-the-art
RS DL research. We also review recent new developments in the DL field that can
be used in DL for RS. Namely, we focus on theories, tools and challenges for
the RS community. Specifically, we focus on unsolved challenges and
opportunities as it relates to (i) inadequate data sets, (ii)
human-understandable solutions for modelling physical phenomena, (iii) Big
Data, (iv) non-traditional heterogeneous data sources, (v) DL architectures and
learning algorithms for spectral, spatial and temporal data, (vi) transfer
learning, (vii) an improved theoretical understanding of DL systems, (viii)
high barriers to entry, and (ix) training and optimizing the DL.Comment: 64 pages, 411 references. To appear in Journal of Applied Remote
Sensin
Towards Daily High-resolution Inundation Observations using Deep Learning and EO
Satellite remote sensing presents a cost-effective solution for synoptic
flood monitoring, and satellite-derived flood maps provide a computationally
efficient alternative to numerical flood inundation models traditionally used.
While satellites do offer timely inundation information when they happen to
cover an ongoing flood event, they are limited by their spatiotemporal
resolution in terms of their ability to dynamically monitor flood evolution at
various scales. Constantly improving access to new satellite data sources as
well as big data processing capabilities has unlocked an unprecedented number
of possibilities in terms of data-driven solutions to this problem.
Specifically, the fusion of data from satellites, such as the Copernicus
Sentinels, which have high spatial and low temporal resolution, with data from
NASA SMAP and GPM missions, which have low spatial but high temporal
resolutions could yield high-resolution flood inundation at a daily scale. Here
a Convolutional-Neural-Network is trained using flood inundation maps derived
from Sentinel-1 Synthetic Aperture Radar and various hydrological,
topographical, and land-use based predictors for the first time, to predict
high-resolution probabilistic maps of flood inundation. The performance of UNet
and SegNet model architectures for this task is evaluated, using flood masks
derived from Sentinel-1 and Sentinel-2, separately with 95 percent-confidence
intervals. The Area under the Curve (AUC) of the Precision Recall Curve
(PR-AUC) is used as the main evaluation metric, due to the inherently
imbalanced nature of classes in a binary flood mapping problem, with the best
model delivering a PR-AUC of 0.85
Deep Learning Training and Benchmarks for Earth Observation Images: Data Sets, Features, and Procedures
Deep learning methods are often used for image classification or local object segmentation. The corresponding test and validation data sets are an integral part of the learning process and also of the algorithm performance evaluation. High and particularly very high-resolution Earth observation (EO) applications based on satellite images primarily aim at the semantic labeling of land cover structures or objects as well as of temporal evolution classes. However, one of the main EO objectives is physical parameter retrievals such as temperatures, precipitation, and crop yield predictions. Therefore, we need reliably labeled data sets and tools to train the developed algorithms and to assess the performance of our deep learning paradigms. Generally, imaging sensors generate a visually understandable representation of the observed scene. However, this does not hold for many EO images, where the recorded images only depict a spectral subset of the scattered light field, thus generating an indirect signature of the imaged object. This spots the load of EO image understanding, as a new and particular challenge of Machine Learning (ML) and Artificial Intelligence (AI). This chapter reviews and analyses the new approaches of EO imaging leveraging the recent advances in physical process-based ML and AI methods and signal processing
- …