10,289 research outputs found
Algorithm for adaptive observation based on method of instrumental variables
When the input signal and the output value of the object of control cannot be measured accurately, the state vector is estimated. The instrumental variables (IVs) method is a commonly used parameter estimation method [1-10]. The task of adaptive observation is to create state observers containing parameter estimators. In adaptive observers, the matrices A and b or c (depending on the chosen canonical state-space representation form) are assumed to be unknown. In the monitoring process, parameter estimation is performed, the unknown matrices are determined, and then the state vector is calculated. The paper aims to present a non-recurrent adaptive observation algorithm for SISO linear time-invariant (LTI) discrete systems. The algorithm is based on the instrumental variables (IVs) method, and the adaptive state observer (ASO) estimates the parameters, the initial and the current state vectors of the discrete system. The algorithm's workability and effectiveness are proved by using simulation data in MATLAB/Simulink
Automatic Caption Generation for Aerial Images: A Survey
Aerial images have attracted attention from researcher community since long time. Generating a caption for an aerial image describing its content in comprehensive way is less studied but important task as it has applications in agriculture, defence, disaster management and many more areas. Though different approaches were followed for natural image caption generation, generating a caption for aerial image remains a challenging task due to its special nature. Use of emerging techniques from Artificial Intelligence (AI) and Natural Language Processing (NLP) domains have resulted in generation of accepted quality captions for aerial images. However lot needs to be done to fully utilize potential of aerial image caption generation task. This paper presents detail survey of the various approaches followed by researchers for aerial image caption generation task. The datasets available for experimentation, criteria used for performance evaluation and future directions are also discussed
RSGPT: A Remote Sensing Vision Language Model and Benchmark
The emergence of large-scale large language models, with GPT-4 as a prominent
example, has significantly propelled the rapid advancement of artificial
general intelligence and sparked the revolution of Artificial Intelligence 2.0.
In the realm of remote sensing (RS), there is a growing interest in developing
large vision language models (VLMs) specifically tailored for data analysis in
this domain. However, current research predominantly revolves around visual
recognition tasks, lacking comprehensive, large-scale image-text datasets that
are aligned and suitable for training large VLMs, which poses significant
challenges to effectively training such models for RS applications. In computer
vision, recent research has demonstrated that fine-tuning large vision language
models on small-scale, high-quality datasets can yield impressive performance
in visual and language understanding. These results are comparable to
state-of-the-art VLMs trained from scratch on massive amounts of data, such as
GPT-4. Inspired by this captivating idea, in this work, we build a high-quality
Remote Sensing Image Captioning dataset (RSICap) that facilitates the
development of large VLMs in the RS field. Unlike previous RS datasets that
either employ model-generated captions or short descriptions, RSICap comprises
2,585 human-annotated captions with rich and high-quality information. This
dataset offers detailed descriptions for each image, encompassing scene
descriptions (e.g., residential area, airport, or farmland) as well as object
information (e.g., color, shape, quantity, absolute position, etc). To
facilitate the evaluation of VLMs in the field of RS, we also provide a
benchmark evaluation dataset called RSIEval. This dataset consists of
human-annotated captions and visual question-answer pairs, allowing for a
comprehensive assessment of VLMs in the context of RS
DATA AUGMENTATION FOR SYNTHETIC APERTURE RADAR USING ALPHA BLENDING AND DEEP LAYER TRAINING
Human-based object detection in synthetic aperture RADAR (SAR) imagery is complex and technical, laboriously slow but time critical—the perfect application for machine learning (ML). Training an ML network for object detection requires very large image datasets with imbedded objects that are accurately and precisely labeled. Unfortunately, no such SAR datasets exist. Therefore, this paper proposes a method to synthesize wide field of view (FOV) SAR images by combining two existing datasets: SAMPLE, which is composed of both real and synthetic single-object chips, and MSTAR Clutter, which is composed of real wide-FOV SAR images. Synthetic objects are extracted from SAMPLE using threshold-based segmentation before being alpha-blended onto patches from MSTAR Clutter. To validate the novel synthesis method, individual object chips are created and classified using a simple convolutional neural network (CNN); testing is performed against the measured SAMPLE subset. A novel technique is also developed to investigate training activity in deep layers. The proposed data augmentation technique produces a 17% increase in the accuracy of measured SAR image classification. This improvement shows that any residual artifacts from segmentation and blending do not negatively affect ML, which is promising for future use in wide-area SAR synthesis.Outstanding ThesisMajor, United States Air ForceApproved for public release. Distribution is unlimited
Introduction to Facial Micro Expressions Analysis Using Color and Depth Images: A Matlab Coding Approach (Second Edition, 2023)
The book attempts to introduce a gentle introduction to the field of Facial
Micro Expressions Recognition (FMER) using Color and Depth images, with the aid
of MATLAB programming environment. FMER is a subset of image processing and it
is a multidisciplinary topic to analysis. So, it requires familiarity with
other topics of Artifactual Intelligence (AI) such as machine learning, digital
image processing, psychology and more. So, it is a great opportunity to write a
book which covers all of these topics for beginner to professional readers in
the field of AI and even without having background of AI. Our goal is to
provide a standalone introduction in the field of MFER analysis in the form of
theorical descriptions for readers with no background in image processing with
reproducible Matlab practical examples. Also, we describe any basic definitions
for FMER analysis and MATLAB library which is used in the text, that helps
final reader to apply the experiments in the real-world applications. We
believe that this book is suitable for students, researchers, and professionals
alike, who need to develop practical skills, along with a basic understanding
of the field. We expect that, after reading this book, the reader feels
comfortable with different key stages such as color and depth image processing,
color and depth image representation, classification, machine learning, facial
micro-expressions recognition, feature extraction and dimensionality reduction.
The book attempts to introduce a gentle introduction to the field of Facial
Micro Expressions Recognition (FMER) using Color and Depth images, with the aid
of MATLAB programming environment.Comment: This is the second edition of the boo
Estimation of the occurrence, severity, and volume of heartwood rot using airborne laser scanning and optical satellite data
Rot in commercial timber reduces the value of the wood substantially and estimating the occurrence, severity, and volume of heartwood rot would be a useful tool in decision-making to minimize economic losses. Remotely sensed data has recently been used for mapping rot on a single-tree level, and although the results have been relatively poor, some potential has been shown. This study applied area-based approaches to predict rot occurrence, rot severity, and rot volume , at an area level. Ground reference data were collected from harvester operations in 2019–2021. Predictor variables were calculated from multi-temporal remotely sensed data together with environmental variables. Response variables from the harvester data and predictor variables from remotely sensed data were aggregated to grid cells and to forest stands. Random Forest models were built for the different combinations of response variables and predictor subsets, and validated with both random- and spatial cross-validation. The results showed that it was not possible to estimate rot occurrence and rot severity with the applied modeling procedure (pR2: 0.00–0.16), without spatially close training data. The better performance of rot volume models (pR2: 0.12–0.37) was mainly due to the correlation between timber volume and rot volum
Synthetic Aperture Radar (SAR) Meets Deep Learning
This reprint focuses on the application of the combination of synthetic aperture radars and depth learning technology. It aims to further promote the development of SAR image intelligent interpretation technology. A synthetic aperture radar (SAR) is an important active microwave imaging sensor, whose all-day and all-weather working capacity give it an important place in the remote sensing community. Since the United States launched the first SAR satellite, SAR has received much attention in the remote sensing community, e.g., in geological exploration, topographic mapping, disaster forecast, and traffic monitoring. It is valuable and meaningful, therefore, to study SAR-based remote sensing applications. In recent years, deep learning represented by convolution neural networks has promoted significant progress in the computer vision community, e.g., in face recognition, the driverless field and Internet of things (IoT). Deep learning can enable computational models with multiple processing layers to learn data representations with multiple-level abstractions. This can greatly improve the performance of various applications. This reprint provides a platform for researchers to handle the above significant challenges and present their innovative and cutting-edge research results when applying deep learning to SAR in various manuscript types, e.g., articles, letters, reviews and technical reports
Evaluation of Multi-frequency Synthetic Aperture Radar for Subsurface Archaeological Prospection in Arid Environments
The discovery of the subsurface paleochannels in the Saharan Desert with the 1981 Shuttle Imaging Radar (SIR-A) sensor was hugely significant in the field of synthetic aperture radar (SAR) remote sensing. Although previous studies had indicated the ability of microwaves to penetrate the earth’s surface in arid environments, this was the first applicable instance of subsurface imaging using a spaceborne sensor. And the discovery of the ‘radar rivers’ with associated archaeological evidence in this inhospitable environment proved the existence of an earlier less arid paleoclimate that supported past populations.
Since the 1980’s SAR subsurface prospection in arid environments has progressed, albeit primarily in the fields of hydrology and geology, with archaeology being investigated to a lesser extent. Currently there is a lack of standardised methods for data acquisition and processing regarding subsurface imaging, difficulties in image interpretation and insufficient supporting quantitative verification. These barriers keep SAR technology from becoming as integral as other remote sensing techniques in archaeological practice
The main objective of this thesis is to undertake a multi-frequency SAR analysis across different site types in arid landscapes to evaluate and enhance techniques for analysing SAR within the context of archaeological subsurface prospection. The analysis and associated fieldwork aim to address the gap in the literature regarding field verification of SAR image interpretation and contribute to the understanding of SAR microwave penetration in arid environments.
The results presented in this thesis demonstrate successful subsurface imaging of subtle feature(s) at the site of ‘Uqdat al-Bakrah, Oman with X-band data. Because shorter wavelengths are often ignored due to their limited penetration depths as compared to the C-band or L-band data, the effectiveness of X-band sensors in archaeological prospection at this site is significant. In addition, the associated ground penetrating radar and excavation fieldwork undertaken at ‘Uqdat al-Bakrah confirm the image interpretation and support the quantitative information regarding microwave penetration
Autonomous Radar-based Gait Monitoring System
Features related to gait are fundamental metrics of human motion [1]. Human gait has been shown to be a valuable and feasible clinical marker to determine the risk of physical and mental functional decline [2], [3]. Technologies that detect changes in people’s gait patterns, especially older adults, could support the detection, evaluation, and monitoring of parameters related to changes in mobility, cognition, and frailty. Gait assessment has the potential to be leveraged as a clinical measurement as it is not limited to a specific health care discipline and is a consistent and sensitive test [4].
A wireless technology that uses electromagnetic waves (i.e., radar) to continually measure gait parameters at home or in a hospital without a clinician’s participation has been proposed as a suitable solution [3], [5]. This approach is based on the interaction between electromagnetic waves with humans and how their bodies impact the surrounding and scattered wireless signals. Since this approach uses wireless waves, people do not need to wear or carry a device on their bodies. Additionally, an electromagnetic wave wireless sensor has no privacy issues because there is no video-based camera.
This thesis presents the design and testing of a radar-based contactless system that can monitor people’s gait patterns and recognize their activities in a range of indoor environments frequently and accurately. In this thesis, the use of commercially available radars for gait monitoring is investigated, which offers opportunities to implement unobtrusive and contactless gait monitoring and activity recognition. A novel fast and easy-to-implement gait extraction algorithm that enables an individual’s spatiotemporal gait parameter extraction at each gait cycle using a single FMCW (Frequency Modulated Continuous Wave) radar is proposed. The proposed system detects changes in gait that may be the signs of changes in mobility, cognition, and frailty, particularly for older adults in individual’s homes, retirement homes and long-term care facilities retirement homes. One of the straightforward applications for gait monitoring using radars is in corridors and hallways, which are commonly available in most residential homes, retirement, and long-term care homes. However, walls in the hallway have a strong “clutter” impact, creating multipath due to the wide beam of commercially available radar antennas. The multipath reflections could result in an inaccurate gait measurement because gait extraction algorithms employ the assumption that the maximum reflected signals come from the torso of the walking person (rather than indirect reflections or multipath) [6].
To address the challenges of hallway gait monitoring, two approaches were used: (1) a novel signal processing method and (2) modifying the radar antenna using a hyperbolic lens. For the first approach, a novel algorithm based on radar signal processing, unsupervised learning, and a subject detection, association and tracking method is proposed. This proposed algorithm could be paired with any type of multiple-input multiple-output (MIMO) or single-input multiple-output (SIMO) FMCW radar to capture human gait in a highly cluttered environment without needing radar antenna alteration. The algorithm functionality was validated by capturing spatiotemporal gait values (e.g., speed, step points, step time, step length, and step count) of people walking in a hallway. The preliminary results demonstrate the promising potential of the algorithm to accurately monitor gait in hallways, which increases opportunities for its applications in institutional and home environments. For the second approach, an in-package hyperbola-based lens antenna was designed that can be integrated with a radar module package empowered by the fast and easy-to-implement gait extraction method. The system functionality was successfully validated by capturing the spatiotemporal gait values of people walking in a hallway filled with metallic cabinets. The results achieved in this work pave the way to explore the use of stand-alone radar-based sensors in long hallways for day-to-day long-term monitoring of gait parameters of older adults or other populations.
The possibility of the coexistence of multiple walking subjects is high, especially in long-term care facilities where other people, including older adults, might need assistance during walking. GaitRite and wearables are not able to assess multiple people’s gait at the same time using only one device [7], [8]. In this thesis, a novel radar-based algorithm is proposed that is capable of tracking multiple people or extracting walking speed of a participant with the coexistence of other people. To address the problem of tracking and monitoring multiple walking people in a cluttered environment, a novel iterative framework based on unsupervised learning and advanced signal processing was developed and tested to analyze the reflected radio signals and extract walking movements and trajectories in a hallway environment. Advanced algorithms were developed to remove multipath effects or ghosts created due to the interaction between walking subjects and stationary objects, to identify and separate reflected signals of two participants walking at a close distance, and to track multiple subjects over time. This method allows the extraction of walking speed in multiple closely-spaced subjects simultaneously, which is distinct from previous approaches where the speed of only one subject was obtained. The proposed multiple-people gait monitoring was assessed with 22 participants who participated in a bedrest (BR) study conducted at McGill University Health Centre (MUHC).
The system functionality also was assessed for in-home applications. In this regard, a cloud-based system is proposed for non-contact, real-time recognition and monitoring of physical activities and walking periods within a domestic environment. The proposed system employs standalone Internet of Things (IoT)-based millimeter wave radar devices and deep learning models to enable autonomous, free-living activity recognition and gait analysis. Range-Doppler maps generated from a dataset of real-life in-home activities are used to train deep learning models. The performance of several deep learning models was evaluated based on accuracy and prediction time, with the gated recurrent network (GRU) model selected for real-time deployment due to its balance of speed and accuracy compared to 2D Convolutional Neural Network Long Short-Term Memory (2D-CNNLSTM) and Long Short-Term Memory (LSTM) models. In addition to recognizing and differentiating various activities and walking periods, the system also records the subject’s activity level over time, washroom use frequency, sleep/sedentary/active/out-of-home durations, current state, and gait parameters. Importantly, the system maintains privacy by not requiring the subject to wear or carry any additional devices
- …