8,452 research outputs found
An Application of Combined Neural Networks to Remotely Sensed Images
Studies in the area of pattern recognition have indicated that in most cases a classifier performs
differently from one pattern class to another. This observation gave birth to the idea of combining the
individual results from different classifiers to derive a consensus decision. This work investigates the
potential of combining neural networks to remotely sensed images. A classifier system is built by
integrating the results of a plurarity of feed-forward neural networks, each of them designed to have the
best performance for one class. Fuzzy Integrals are used as the combining strategy. Experiments carried
out to evaluate the system, using a satellite image of an area undergoing a rapid degradation process, have
shown that the combination may yield a better performance than that of a single neural network
Recommended from our members
Estimation of physical variables from multichannel remotely sensed imagery using a neural network: Application to rainfall estimation
Satellite-based remotely sensed data have the potential to provide hydrologically relevant information about spatially and temporally varying physical variables. A methodology for estimating such variables from multichannel remotely sensed data is presented; the approach is based on a modified counterpropagation neural network (MCPN) and is both effective and efficient at building complex nonlinear input-output function mappings from large amounts of data. An application to high-resolution estimation of the spatial and temporal variation of surface rainfall using geostationary satellite infrared and visible imagery is presented. Test results also indicate that spatially and temporally sparse ground-based observations can be assimilated via an adaptive implementation of the MCPN method, thereby allowing on-line improvement of the estimates
A Neural Network Method for Land Use Change Classification, with Application to the Nile River Delta
Detecting and monitoring changes in conditions at the Earth's surface are essential for understanding human impact on the environment and for assessing the sustainability of development. In the next decade, NASA will gather high-resolution multi-spectral and multi-temporal data, which could be used for analyzing long-term changes, provided that available methods can keep pace with the accelerating flow of information. This paper introduces an automated technique for change identification, based on the ARTMAP neural network. This system overcomes some of the limitations of traditional change detection methods, and also produces a measure of confidence in classification accuracy. Landsat thematic mapper (TM) imagery of the Nile River delta provides a testbed for land use change classification methods. This dataset consists of a sequence of ten images acquired between 1984 and 1993 at various times of year. Field observations and photo interpretations have identified 358 sites as belonging to eight classes, three of which represent changes in land use over the ten-year period. Aparticular challenge posed by this database is the unequal representation of various land use categories: three classes, urban, agriculture in delta, and other, comprise 95% of pixels in labeled sites. A two-step sampling method enables unbiased training of the neural network system across sites.National Science Foundation (SBR 95-13889); Office of Naval Research (N00014-95-1-409, N00014-95-0657); Air Force Office of Scientific Research (F49620-01-1-0397, F49620-01-1-042
Increasing the spatial resolution of agricultural land cover maps using a Hopfield neural network
Land cover class composition of remotely sensed image pixels can be estimated using soft classification techniques increasingly available in many GIS packages. However, their output provides no indication of how such classes are distributed spatially within the instantaneous field of view represented by the pixel. Techniques that attempt to provide an improved spatial representation of land cover have been developed, but not tested on the difficult task of mapping from real satellite imagery. The authors investigated the use of a Hopfield neural network technique to map the spatial distributions of classes reliably using information of pixel composition determined from soft classification previously. The approach involved designing the energy function to produce a ‘best guess’ prediction of the spatial distribution of class components in each pixel. In previous studies, the authors described the application of the technique to target identification, pattern prediction and land cover mapping at the sub-pixel scale, but only for simulated imagery.We now show how the approach can be applied to Landsat Thematic Mapper (TM) agriculture imagery to derive accurate estimates of land cover and reduce the uncertainty inherent in such imagery. The technique was applied to Landsat TM imagery of small-scale agriculture in Greece and largescale agriculture near Leicester, UK. The resultant maps provided an accurate and improved representation of the land covers studied, with RMS errors for the Landsat imagery of the order of 0.1 in the new fine resolution map recorded. The results showed that the neural network represents a simple efficient tool formapping land cover from operational satellite sensor imagery and can deliver requisite results and improvements over traditional techniques for the GIS analysis of practical remotely sensed imagery at the sub pixel scale
Recommended from our members
PERSIANN-CNN: Precipitation Estimation from Remotely Sensed Information Using Artificial Neural Networks-Convolutional Neural Networks
Abstract
Accurate and timely precipitation estimates are critical for monitoring and forecasting natural disasters such as floods. Despite having high-resolution satellite information, precipitation estimation from remotely sensed data still suffers from methodological limitations. State-of-the-art deep learning algorithms, renowned for their skill in learning accurate patterns within large and complex datasets, appear well suited to the task of precipitation estimation, given the ample amount of high-resolution satellite data. In this study, the effectiveness of applying convolutional neural networks (CNNs) together with the infrared (IR) and water vapor (WV) channels from geostationary satellites for estimating precipitation rate is explored. The proposed model performances are evaluated during summer 2012 and 2013 over central CONUS at the spatial resolution of 0.08° and at an hourly time scale. Precipitation Estimation from Remotely Sensed Information Using Artificial Neural Networks (PERSIANN)–Cloud Classification System (CCS), which is an operational satellite-based product, and PERSIANN–Stacked Denoising Autoencoder (PERSIANN-SDAE) are employed as baseline models. Results demonstrate that the proposed model (PERSIANN-CNN) provides more accurate rainfall estimates compared to the baseline models at various temporal and spatial scales. Specifically, PERSIANN-CNN outperforms PERSIANN-CCS (and PERSIANN-SDAE) by 54% (and 23%) in the critical success index (CSI), demonstrating the detection skills of the model. Furthermore, the root-mean-square error (RMSE) of the rainfall estimates with respect to the National Centers for Environmental Prediction (NCEP) Stage IV gauge–radar data, for PERSIANN-CNN was lower than that of PERSIANN-CCS (PERSIANN-SDAE) by 37% (14%), showing the estimation accuracy of the proposed model
- …