14 research outputs found
Vectorization of Large Amounts of Raster Satellite Images in a Distributed Architecture Using HIPI
Vectorization process focus on grouping pixels of a raster image into raw
line segments, and forming lines, polylines or poligons. To vectorize massive
raster images regarding resource and performane problems, weuse a distributed
HIPI image processing interface based on MapReduce approach. Apache Hadoop is
placed at the core of the framework. To realize such a system, we first define
mapper function, and then its input and output formats. In this paper, mappers
convert raster mosaics into vector counterparts. Reduc functions are not needed
for vectorization. Vector representations of raster images is expected to give
better performance in distributed computations by reducing the negative effects
of bandwidth problem and horizontal scalability analysis is done.Comment: In Turkish, Proceedings of International Artificial Intelligence and
Data Processing Symposium (IDAP) 201
DIFET: Distributed Feature Extraction Tool For High Spatial Resolution Remote Sensing Images
In this paper, we propose distributed feature extraction tool from high
spatial resolution remote sensing images. Tool is based on Apache Hadoop
framework and Hadoop Image Processing Interface. Two corner detection (Harris
and Shi-Tomasi) algorithms and five feature descriptors (SIFT, SURF, FAST,
BRIEF, and ORB) are considered. Robustness of the tool in the task of feature
extraction from LandSat-8 imageries are evaluated in terms of horizontal
scalability.Comment: Presented at 4th International GeoAdvances Worksho
Big Geospatial Data processing in the IQmulus Cloud
Remote sensing instruments are continuously evolving in terms of spatial, spectral and temporal resolutions and hence provide exponentially increasing amounts of raw data. These volumes increase significantly faster than computing speeds. All these techniques record lots of data, yet in different data models and representations; therefore, resulting datasets require harmonization and integration prior to deriving meaningful information from them. All in all, huge datasets are available but raw data is almost of no value if not processed, semantically enriched and quality checked. The derived information need to be transferred and published to all level of possible users (from decision makers to citizens). Up to now, there are only limited automatic procedures for this; thus, a wealth of information is latent in many datasets. This paper presents the first achievements of the IQmulus EU FP7 research and development project with respect to processing and analysis of big geospatial data in the context of flood and waterlogging detection
A HADOOP-BASED DISTRIBUTED FRAMEWORK FOR EFFICIENT MANAGING AND PROCESSING BIG REMOTE SENSING IMAGES
ABSTRACT: Various sensors from airborne and satellite platforms are producing large volumes of remote sensing images for mapping, environmental monitoring, disaster management, military intelligence, and others. However, it is challenging to efficiently storage, query and process such big data due to the data-and computing-intensive issues. In this paper, a Hadoop-based framework is proposed to manage and process the big remote sensing data in a distributed and parallel manner. Especially, remote sensing data can be directly fetched from other data platforms into the Hadoop Distributed File System (HDFS). The Orfeo toolbox, a ready-to-use tool for large image processing, is integrated into MapReduce to provide affluent image processing operations. With the integration of HDFS, Orfeo toolbox and MapReduce, these remote sensing images can be directly processed in parallel in a scalable computing environment. The experiment results show that the proposed framework can efficiently manage and process such big remote sensing data
A HADOOP-BASED DISTRIBUTED FRAMEWORK FOR EFFICIENT MANAGING AND PROCESSING BIG REMOTE SENSING IMAGES
Hadoop Optimization for Massive Image Processing: Case Study Face Detection
Face detection applications are widely used for searching, tagging and classifying people inside very large image databases. This type of applications requires processing of relatively small sized and large number of images. On the other hand, Hadoop Distributed File System (HDFS) is originally designed for storing and processing largesize files. Huge number of small-size images causes slowdown in HDFS by increasing total initialization time of jobs, scheduling overhead of tasks and memory usage of the file system manager (Namenode). The study in this paper presents two approaches to improve small image file processing performance of HDFS. These are (1) converting the images into single large-size file by merging and (2) combining many images for a single task without merging. We also introduce novel Hadoop file formats and record generation methods (for reading image content) in order to develop these technique