8 research outputs found

    Landmine detection using semi-supervised learning.

    Get PDF
    Landmine detection is imperative for the preservation of both military and civilian lives. While landmines are easy to place, they are relatively difficult to remove. The classic method of detecting landmines was by using metal-detectors. However, many present-day landmines are composed of little to no metal, necessitating the use of additional technologies. One of the most successful and widely employed technologies is Ground Penetrating Radar (GPR). In order to maximize efficiency of GPR-based landmine detection and minimize wasted effort caused by false alarms, intelligent detection methods such as machine learning are used. Many sophisticated algorithms are developed and employed to accomplish this. One such successful algorithm is K Nearest Neighbors (KNN) classification. Most of these algorithms, including KNN, are based on supervised learning, which requires labeling of known data. This process can be tedious. Semi-supervised learning leverages both labeled and unlabeled data in the training process, alleviating over-dependency on labeling. Semi-supervised learning has several advantages over supervised learning. For example, it applies well to large datasets because it uses the topology of unlabeled data to classify test data. Also, by allowing unlabeled data to influence classification, one set of training data can be adopted into varying test environments. In this thesis, we explore a graph-based learning method known as Label Propagation as an alternative classifier to KNN classification, and validate its use on vehicle-mounted and handheld GPR systems

    Advanced Feature Learning and Representation in Image Processing for Anomaly Detection

    Get PDF
    Techniques for improving the information quality present in imagery for feature extraction are proposed in this thesis. Specifically, two methods are presented: soft feature extraction and improved Evolution-COnstructed (iECO) features. Soft features comprise the extraction of image-space knowledge by performing a per-pixel weighting based on an importance map. Through soft features, one is able to extract features relevant to identifying a given object versus its background. Next, the iECO features framework is presented. The iECO features framework uses evolutionary computation algorithms to learn an optimal series of image transforms, specific to a given feature descriptor, to best extract discriminative information. That is, a composition of image transforms are learned from training data to present a given feature descriptor with the best opportunity to extract its information for the application at hand. The proposed techniques are applied to an automatic explosive hazard detection application and significant results are achieved

    Ensemble learning method for hidden markov models.

    Get PDF
    For complex classification systems, data are gathered from various sources and potentially have different representations. Thus, data may have large intra-class variations. In fact, modeling each data class with a single model might lead to poor generalization. The classification error can be more severe for temporal data where each sample is represented by a sequence of observations. Thus, there is a need for building a classification system that takes into account the variations within each class in the data. This dissertation introduces an ensemble learning method for temporal data that uses a mixture of Hidden Markov Model (HMM) classifiers. We hypothesize that the data are generated by K models, each of which reacts a particular trend in the data. Model identification could be achieved through clustering in the feature space or in the parameters space. However, this approach is inappropriate in the context of sequential data. The proposed approach is based on clustering in the log-likelihood space, and has two main steps. First, one HMM is fit to each of the N individual sequences. For each fitted model, we evaluate the log-likelihood of each sequence. This will result in an N-by-N log-likelihood distance matrix that will be partitioned into K groups using a relational clustering algorithm. In the second step, we learn the parameters of one HMM per group. We propose using and optimizing various training approaches for the different K groups depending on their size and homogeneity. In particular, we investigate the maximum likelihood (ML), the minimum classification error (MCE) based discriminative, and the Variational Bayesian (VB) training approaches. Finally, to test a new sequence, its likelihood is computed in all the models and a final confidence value is assigned by combining the multiple models outputs using a decision level fusion method such as an artificial neural network or a hierarchical mixture of experts. Our approach was evaluated on two real-world applications: (1) identification of Cardio-Pulmonary Resuscitation (CPR) scenes in video simulating medical crises; and (2) landmine detection using Ground Penetrating Radar (GPR). Results on both applications show that the proposed method can identify meaningful and coherent HMM mixture components that describe different properties of the data. Each HMM mixture component models a group of data that share common attributes. The results indicate that the proposed method outperforms the baseline HMM that uses one model for each class in the data

    Generalized multi-stream hidden Markov models.

    Get PDF
    For complex classification systems, data is usually gathered from multiple sources of information that have varying degree of reliability. In fact, assuming that the different sources have the same relevance in describing all the data might lead to an erroneous behavior. The classification error accumulates and can be more severe for temporal data where each sample is represented by a sequence of observations. Thus, there is compelling evidence that learning algorithms should include a relevance weight for each source of information (stream) as a parameter that needs to be learned. In this dissertation, we assumed that the multi-stream temporal data is generated by independent and synchronous streams. Using this assumption, we develop, implement, and test multi- stream continuous and discrete hidden Markov model (HMM) algorithms. For the discrete case, we propose two new approaches to generalize the baseline discrete HMM. The first one combines unsupervised learning, feature discrimination, standard discrete HMMs and weighted distances to learn the codebook with feature-dependent weights for each symbol. The second approach consists of modifying the HMM structure to include stream relevance weights, generalizing the standard discrete Baum-Welch learning algorithm, and deriving the necessary conditions to optimize all model parameters simultaneously. We also generalize the minimum classification error (MCE) discriminative training algorithm to include stream relevance weights. For the continuous HMM, we introduce a. new approach that integrates the stream relevance weights in the objective function. Our approach is based on the linearization of the probability density function. Two variations are proposed: the mixture and state level variations. As in the discrete case, we generalize the continuous Baum-Welch learning algorithm to accommodate these changes, and we derive the necessary conditions for updating the model parameters. We also generalize the MCE learning algorithm to derive the necessary conditions for the model parameters\u27 update. The proposed discrete and continuous HMM are tested on synthetic data sets. They are also validated on various applications including Australian Sign Language, audio classification, face classification, and more extensively on the problem of landmine detection using ground penetrating radar data. For all applications, we show that considerable improvement can be achieved compared to the baseline HMM and the existing multi-stream HMM algorithms

    Multiple Instance Learning: A Survey of Problem Characteristics and Applications

    Full text link
    Multiple instance learning (MIL) is a form of weakly supervised learning where training instances are arranged in sets, called bags, and a label is provided for the entire bag. This formulation is gaining interest because it naturally fits various problems and allows to leverage weakly labeled data. Consequently, it has been used in diverse application fields such as computer vision and document classification. However, learning from bags raises important challenges that are unique to MIL. This paper provides a comprehensive survey of the characteristics which define and differentiate the types of MIL problems. Until now, these problem characteristics have not been formally identified and described. As a result, the variations in performance of MIL algorithms from one data set to another are difficult to explain. In this paper, MIL problem characteristics are grouped into four broad categories: the composition of the bags, the types of data distribution, the ambiguity of instance labels, and the task to be performed. Methods specialized to address each category are reviewed. Then, the extent to which these characteristics manifest themselves in key MIL application areas are described. Finally, experiments are conducted to compare the performance of 16 state-of-the-art MIL methods on selected problem characteristics. This paper provides insight on how the problem characteristics affect MIL algorithms, recommendations for future benchmarking and promising avenues for research

    Machine learning methods for sign language recognition: a critical review and analysis.

    Get PDF
    Sign language is an essential tool to bridge the communication gap between normal and hearing-impaired people. However, the diversity of over 7000 present-day sign languages with variability in motion position, hand shape, and position of body parts making automatic sign language recognition (ASLR) a complex system. In order to overcome such complexity, researchers are investigating better ways of developing ASLR systems to seek intelligent solutions and have demonstrated remarkable success. This paper aims to analyse the research published on intelligent systems in sign language recognition over the past two decades. A total of 649 publications related to decision support and intelligent systems on sign language recognition (SLR) are extracted from the Scopus database and analysed. The extracted publications are analysed using bibliometric VOSViewer software to (1) obtain the publications temporal and regional distributions, (2) create the cooperation networks between affiliations and authors and identify productive institutions in this context. Moreover, reviews of techniques for vision-based sign language recognition are presented. Various features extraction and classification techniques used in SLR to achieve good results are discussed. The literature review presented in this paper shows the importance of incorporating intelligent solutions into the sign language recognition systems and reveals that perfect intelligent systems for sign language recognition are still an open problem. Overall, it is expected that this study will facilitate knowledge accumulation and creation of intelligent-based SLR and provide readers, researchers, and practitioners a roadmap to guide future direction

    Microwave NDT&E using open-ended waveguide probe for multilayered structures

    Get PDF
    Ph. D. Thesis.Microwave NDT&E has been proved to be suitable for inspecting of dielectric structures due to low attenuation in dielectric materials and free-space. However, the microwave responses from multilayered structures are complex as an interrogation of scattering electromagnetic waves among the layers and defects. In many practical applications, electromagnetic analysis based on analytic- and forward structural models cannot be generalised since the defect shape and properties are usually unknown and hidden beneath the surface layer. This research proposes the design and implementation of microwave NDT&E system for inspection of multilayered structures. Standard microwave open-ended rectangular waveguides in X, Ku and K bands (frequency range between 8-26.5 GHz) and vector network analyser (VNA) generating sweep frequency of wideband monochromatic waves have been used to obtain reflection coefficient responses over three types of challenging multilayered samples: (1) corrosion progression under coating, (2) woven carbon fibre reinforced polymer (CFRP) with impact damages, and (3) thermal coated glass fibre reinforced polymer (GFRP) pipe with inner flat-bottom holes. The obtained data are analysed by the selected feature extraction method extracting informative features and verify with the sample parameters (defect parameters). In addition, visualisation methods are utilised to improve the presentation of the defects and material structures resulting in a better interpretation for quantitative evaluation. The contributions of this project are summarised as follows: (1) implementation of microwave NDT&E scanning system using open-ended waveguide with the highest resolution of 0.1mm x 0.1 mm, based on the NDT applications for the three aforementioned samples; (2) corrosion stages of steel corrosion under coating have been successfully characterised by the principal component analysis (PCA) method; (3) A frequency selective based PCA feature has been used to visualise the impact damage at different impact energies with elimination of woven texture influences; (4) PCA and SAR (synthetic aperture radar) tomography together with time-offlight extraction, have been used for detection and quantitative evaluation of flat-bottom hole defects (i.e., location, size and depth). The results conclude that the proposed microwave NDT&E system can be used for detection and evaluation of multilayered structures, which its major contributions are follows. (1) The early stages (0-12month) of steel corrosion undercoating has been successfully characterised by mean of spectral responses from microwave opened rectangular waveguide probe and PCA. (2) The detection of low energy impact damages on CFRP as low as 4 Joules has been archived with microwave opened rectangular waveguide probe raster scan together with SAR imaging and PCA for feature extraction methods. (3) The inner flat-bottom holes beneath the thermal coated GFRP up to 11.5 mm depth has been successfully quantitative evaluated by open-ended waveguide raster scan using PCA and 3-D reconstruction based on SAR tomography techniques. The evaluation includes location, sizing and depth. Nevertheless, the major downside of feature quantities extracted from statistically based methods such as PCA, is it intensely relies on the correlation of the input dataset, and thus hardly link them with the physical parameters of the test sample, in particular, the complex composite architectures. Therefore, there are still challenges of feature extraction and quantitative evaluation to accurately determine the essential parameters from the samples. This can be achieved by a future investigation of multiple features fusion and complementary features.Ministry of Science and Technology of Royal Thai Government and Office of Educational Affairs, the Royal Thai Embass

    Modelling, Simulation and Data Analysis in Acoustical Problems

    Get PDF
    Modelling and simulation in acoustics is currently gaining importance. In fact, with the development and improvement of innovative computational techniques and with the growing need for predictive models, an impressive boost has been observed in several research and application areas, such as noise control, indoor acoustics, and industrial applications. This led us to the proposal of a special issue about “Modelling, Simulation and Data Analysis in Acoustical Problems”, as we believe in the importance of these topics in modern acoustics’ studies. In total, 81 papers were submitted and 33 of them were published, with an acceptance rate of 37.5%. According to the number of papers submitted, it can be affirmed that this is a trending topic in the scientific and academic community and this special issue will try to provide a future reference for the research that will be developed in coming years
    corecore