84 research outputs found
Digital Image Processing
This book presents several recent advances that are related or fall under the umbrella of 'digital image processing', with the purpose of providing an insight into the possibilities offered by digital image processing algorithms in various fields. The presented mathematical algorithms are accompanied by graphical representations and illustrative examples for an enhanced readability. The chapters are written in a manner that allows even a reader with basic experience and knowledge in the digital image processing field to properly understand the presented algorithms. Concurrently, the structure of the information in this book is such that fellow scientists will be able to use it to push the development of the presented subjects even further
Generation of a combined dataset of simulated radar and electro-optical imagery
In the world of remote sensing there exist radar sensors and EO/IR sensors, both of which carry with them unique information useful to the imaging community. Radar has the capability of imaging through all types of weather, day or night. EO/IR produces radiance maps and frequently images at much finer resolution than radar. While each of these systems is valuable to imaging, there exists unknown territory in the imaging community as to the value added in combining the best of both these worlds. This work will begin to explore the challenges in simulating a scene in both a radar tool called Xpatch and an EO/IR tool called DIRSIG. The capabilities and limitations inherent to both radar and EO/IR are similar in the image simulation tools, so the work done in a simulated environment will carry over to the real-world environment as well. The synthetic data generated will be compared to existing measured data to demonstrate the validity of the experiment. Future work should explore registration and various types of fusion of the resulting images to demonstrate the synergistic value of the combined images
JERS-1 SAR and LANDSAT-5 TM image data fusion: An application approach for lithological mapping
Satellite image data fusion is an image processing set of procedures utilise either for image optimisation for visual photointerpretation, or for automated thematic classification with low error rate and high accuracy. Lithological mapping using remote sensing image data relies on the spectral and textural information of the rock units of the area to be mapped. These pieces of information can be derived from Landsat optical TM and JERS-1 SAR images respectively. Prior to extracting such information (spectral and textural) and fusing them together, geometric image co-registration between TM and the SAR, atmospheric correction of the TM, and SAR despeckling are required. In this thesis, an appropriate atmospheric model is developed and implemented utilising the dark pixel subtraction method for atmospheric correction. For SAR despeckling, an efficient new method is also developed to test whether the SAR filter used remove the textural information or not. For image optimisation for visual photointerpretation, a new method of spectral coding of the six bands of the optical TM data is developed. The new spectral coding method is used to produce efficient colour composite with high separability between the spectral classes similar to that if the whole six optical TM bands are used together. This spectral coded colour composite is used as a spectral component, which is then fused with the textural component represented by the despeckled JERS-1 SAR using the fusion tools, including the colour transform and the PCT. The Grey Level Cooccurrence Matrix (GLCM) technique is used to build the textural data set using the speckle filtered JERS-1 SAR data making seven textural GLCM measures. For automated thematic mapping and by the use of both the six TM spectral data and the seven textural GLCM measures, a new method of classification has been developed using the Maximum Likelihood Classifier (MLC). The method is named the sequential maximum likelihood classification and works efficiently by comparison the classified textural pixels, the classified spectral pixels, and the classified textural-spectral pixels, and gives the means of utilising the textural and spectral information for automated lithological mapping
Synthetic Aperture Radar (SAR) Meets Deep Learning
This reprint focuses on the application of the combination of synthetic aperture radars and depth learning technology. It aims to further promote the development of SAR image intelligent interpretation technology. A synthetic aperture radar (SAR) is an important active microwave imaging sensor, whose all-day and all-weather working capacity give it an important place in the remote sensing community. Since the United States launched the first SAR satellite, SAR has received much attention in the remote sensing community, e.g., in geological exploration, topographic mapping, disaster forecast, and traffic monitoring. It is valuable and meaningful, therefore, to study SAR-based remote sensing applications. In recent years, deep learning represented by convolution neural networks has promoted significant progress in the computer vision community, e.g., in face recognition, the driverless field and Internet of things (IoT). Deep learning can enable computational models with multiple processing layers to learn data representations with multiple-level abstractions. This can greatly improve the performance of various applications. This reprint provides a platform for researchers to handle the above significant challenges and present their innovative and cutting-edge research results when applying deep learning to SAR in various manuscript types, e.g., articles, letters, reviews and technical reports
Signal Processing for Synthetic Aperture Sonar Image Enhancement
This thesis contains a description of SAS processing algorithms, offering improvements in Fourier-based reconstruction, motion-compensation, and autofocus. Fourier-based image reconstruction is reviewed and improvements shown as the result of improved system modelling. A number of new algorithms based on the wavenumber algorithm for correcting second order effects are proposed. In addition, a new framework for describing multiple-receiver reconstruction in terms of the bistatic geometry is presented and is a useful aid to understanding. Motion-compensation techniques for allowing Fourier-based reconstruction in widebeam geometries suffering large-motion errors are discussed. A motion-compensation algorithm exploiting multiple receiver geometries is suggested and shown to provide substantial improvement in image quality. New motion compensation techniques for yaw correction using the wavenumber algorithm are discussed. A common framework for describing phase estimation is presented and techniques from a number of fields are reviewed within this framework. In addition a new proof is provided outlining the relationship between eigenvector-based autofocus phase estimation kernels and the phase-closure techniques used astronomical imaging. Micronavigation techniques are reviewed and extensions to the shear average single-receiver micronavigation technique result in a 3 - 4 fold performance improvement when operating on high-contrast images. The stripmap phase gradient autofocus (SPGA) algorithm is developed and extends spotlight SAR PGA to the wide-beam, wide-band stripmap geometries common in SAS imaging. SPGA supersedes traditional PGA-based stripmap autofocus algorithms such as mPGA and PCA - the relationships between SPGA and these algorithms is discussed. SPGA's operation is verified on simulated and field-collected data where it provides significant image improvement. SPGA with phase-curvature based estimation is shown and found to perform poorly compared with phase-gradient techniques. The operation of SPGA on data collected from Sydney Harbour is shown with SPGA able to improve resolution to near the diffraction-limit. Additional analysis of practical stripmap autofocus operation in presence of undersampling and space-invariant blurring is presented with significant comment regarding the difficulties inherent in autofocusing field-collected data. Field-collected data from trials in Sydney Harbour is presented along with associated autofocus results from a number of algorithms
Proof-of-Concept
Biometry is an area in great expansion and is considered as possible solution to cases where high
authentication parameters are required. Although this area is quite advanced in theoretical
terms, using it in practical terms still carries some problems. The systems available still depend
on a high cooperation level to achieve acceptable performance levels, which was the backdrop
to the development of the following project. By studying the state of the art, we propose the
creation of a new and less cooperative biometric system that reaches acceptable performance
levels.A constante necessidade de parâmetros mais elevados de segurança, nomeadamente ao nível
de autenticação, leva ao estudo biometria como possível solução. Actualmente os mecanismos
existentes nesta área tem por base o conhecimento de algo que se sabe ”password” ou algo
que se possui ”codigo Pin”. Contudo este tipo de informação é facilmente corrompida ou contornada.
Desta forma a biometria é vista como uma solução mais robusta, pois garante que a
autenticação seja feita com base em medidas físicas ou compartimentais que definem algo que
a pessoa é ou faz (”who you are” ou ”what you do”).
Sendo a biometria uma solução bastante promissora na autenticação de indivíduos, é cada vez
mais comum o aparecimento de novos sistemas biométricos. Estes sistemas recorrem a medidas
físicas ou comportamentais, de forma a possibilitar uma autenticação (reconhecimento) com
um grau de certeza bastante considerável. O reconhecimento com base no movimento do corpo
humano (gait), feições da face ou padrões estruturais da íris, são alguns exemplos de fontes
de informação em que os sistemas actuais se podem basear. Contudo, e apesar de provarem
um bom desempenho no papel de agentes de reconhecimento autónomo, ainda estão muito
dependentes a nível de cooperação exigida. Tendo isto em conta, e tudo o que já existe no
ramo do reconhecimento biometrico, esta área está a dar passos no sentido de tornar os seus
métodos o menos cooperativos poss??veis. Possibilitando deste modo alargar os seus objectivos
para além da mera autenticação em ambientes controlados, para casos de vigilância e controlo
em ambientes não cooperativos (e.g. motins, assaltos, aeroportos).
É nesta perspectiva que o seguinte projecto surge. Através do estudo do estado da arte, pretende
provar que é possível criar um sistema capaz de agir perante ambientes menos cooperativos,
sendo capaz de detectar e reconhecer uma pessoa que se apresente ao seu alcance.O
sistema proposto PAIRS (Periocular and Iris Recognition Systema) tal como nome indica, efectua
o reconhecimento através de informação extraída da íris e da região periocular (região circundante
aos olhos). O sistema é construído com base em quatro etapas: captura de dados,
pré-processamento, extração de características e reconhecimento. Na etapa de captura de
dados, foi montado um dispositivo de aquisição de imagens com alta resolução com a capacidade
de capturar no espectro NIR (Near-Infra-Red). A captura de imagens neste espectro tem
como principal linha de conta, o favorecimento do reconhecimento através da íris, visto que
a captura de imagens sobre o espectro visível seria mais sensível a variações da luz ambiente.
Posteriormente a etapa de pré-processamento implementada, incorpora todos os módulos do
sistema responsáveis pela detecção do utilizador, avaliação de qualidade de imagem e segmentação
da íris. O modulo de detecção é responsável pelo desencadear de todo o processo, uma
vez que esta é responsável pela verificação da exist?ncia de um pessoa em cena. Verificada
a sua exist?ncia, são localizadas as regiões de interesse correspondentes ? íris e ao periocular,
sendo também verificada a qualidade com que estas foram adquiridas. Concluídas estas
etapas, a íris do olho esquerdo é segmentada e normalizada. Posteriormente e com base em
vários descritores, é extraída a informação biométrica das regiões de interesse encontradas,
e é criado um vector de características biométricas. Por fim, é efectuada a comparação dos
dados biometricos recolhidos, com os já armazenados na base de dados, possibilitando a criação
de uma lista com os níveis de semelhança em termos biometricos, obtendo assim um resposta
final do sistema. Concluída a implementação do sistema, foi adquirido um conjunto de imagens capturadas através do sistema implementado, com a participação de um grupo de voluntários.
Este conjunto de imagens permitiu efectuar alguns testes de desempenho, verificar e afinar
alguns parâmetros, e proceder a optimização das componentes de extração de características e
reconhecimento do sistema. Analisados os resultados foi possível provar que o sistema proposto
tem a capacidade de exercer as suas funções perante condições menos cooperativas
SURFO Technical Report No. 18-01
The 2018 technical reports written by undergraduate students participating in the SURFO (Summer Undergraduate Research Fellowships in Oceanography) Program while at the University of Rhode Island
Earth Resources: A continuing bibliography with indexes
This bibliography lists 475 reports, articles and other documents introduced into the NASA scientific and technical information system between January 1 and March 31, 1984. Emphasis is placed on the use of remote sensing and geophysical instrumentation in spacecraft and aircraft to survey and inventory natural resources and urban areas. Subject matter is grouped according to agriculture and forestry, environmental changes and cultural resources, geodesy and cartography, geology and mineral resources, hydrology and water management, data processing and distribution systems, instrumentation and sensors, and economical analysis
- …