40 research outputs found

    Machine learning paradigms for modeling spatial and temporal information in multimedia data mining

    Get PDF
    Multimedia data mining and knowledge discovery is a fast emerging interdisciplinary applied research area. There is tremendous potential for effective use of multimedia data mining (MDM) through intelligent analysis. Diverse application areas are increasingly relying on multimedia under-standing systems. Advances in multimedia understanding are related directly to advances in signal processing, computer vision, machine learning, pattern recognition, multimedia databases, and smart sensors. The main mission of this special issue is to identify state-of-the-art machine learning paradigms that are particularly powerful and effective for modeling and combining temporal and spatial media cues such as audio, visual, and face information and for accomplishing tasks of multimedia data mining and knowledge discovery. These models should be able to bridge the gap between low-level audiovisual features which require signal processing and high-level semantics. A number of papers have been submitted to the special issue in the areas of imaging, artificial intelligence; and pattern recognition and five contributions have been selected covering state-of-the-art algorithms and advanced related topics. The first contribution by D. Xiang et al. “Evaluation of data quality and drought monitoring capability of FY-3A MERSI data” describes some basic parameters and major technical indicators of the FY-3A, and evaluates data quality and drought monitoring capability of the Medium-Resolution Imager (MERSI) onboard the FY-3A. The second contribution by A. Belatreche et al. “Computing with biologically inspired neural oscillators: application to color image segmentation” investigates the computing capabilities and potential applications of neural oscillators, a biologically inspired neural model, to gray scale and color image segmentation, an important task in image understanding and object recognition. The major contribution of this paper is the ability to use neural oscillators as a learning scheme for solving real world engineering problems. The third paper by A. Dargazany et al. entitled “Multibandwidth Kernel-based object tracking” explores new methods for object tracking using the mean shift (MS). A bandwidth-handling MS technique is deployed in which the tracker reach the global mode of the density function not requiring a specific staring point. It has been proven via experiments that the Gradual Multibandwidth Mean Shift tracking algorithm can converge faster than the conventional kernel-based object tracking (known as the mean shift). The fourth contribution by S. Alzu’bi et al. entitled “3D medical volume segmentation using hybrid multi-resolution statistical approaches” studies new 3D volume segmentation using multiresolution statistical approaches based on discrete wavelet transform and hidden Markov models. This system commonly reduced the percentage error achieved using the traditional 2D segmentation techniques by several percent. Furthermore, a contribution by G. Cabanes et al. entitled “Unsupervised topographic learning for spatiotemporal data mining” proposes a new unsupervised algorithm, suitable for the analysis of noisy spatiotemporal Radio Frequency Identification (RFID) data. The new unsupervised algorithm depicted in this article is an efficient data mining tool for behavioral studies based on RFID technology. It has the ability to discover and compare stable patterns in a RFID signal, and is appropriate for continuous learning. Finally, we would like to thank all those who helped to make this special issue possible, especially the authors and the reviewers of the articles. Our thanks go to the Hindawi staff and personnel, the journal Manager in bringing about the issue and giving us the opportunity to edit this special issue

    A statistical multiresolution approach for face recognition using structural hidden Markov models

    Get PDF
    This paper introduces a novel methodology that combines the multiresolution feature of the discrete wavelet transform (DWT) with the local interactions of the facial structures expressed through the structural hidden Markov model (SHMM). A range of wavelet filters such as Haar, biorthogonal 9/7, and Coiflet, as well as Gabor, have been implemented in order to search for the best performance. SHMMs perform a thorough probabilistic analysis of any sequential pattern by revealing both its inner and outer structures simultaneously. Unlike traditional HMMs, the SHMMs do not perform the state conditional independence of the visible observation sequence assumption. This is achieved via the concept of local structures introduced by the SHMMs. Therefore, the long-range dependency problem inherent to traditional HMMs has been drastically reduced. SHMMs have not previously been applied to the problem of face identification. The results reported in this application have shown that SHMM outperforms the traditional hidden Markov model with a 73% increase in accuracy

    Categorization of Emotion Based on Causality

    Get PDF
    Background: Emotions come in all shapes and forms. Some of them can be external, visible, and clearly comprehensible, while others can seemingly be coming out of thin air. Knowing what causes an emotion is crucial for better therapy and mental health. Therefore, in this manuscript, we address the problem of emotions causality. Methods: We propose a comparison of three traditional clustering models: Gaussian mixture model, HDBSCAN, and fuzzy c-means, to categorize each emotion described in the DEAP database. It contains over 1700 points, and has no prior label as to which type of stressor the subject’s emotion is generated from. This labelling task has been conducted by a psychiatrist. Results: The fuzzy c-means yields the highest results, with an accuracy of 57.13%, followed by the Gaussian mixture model at 39.49% and the HDBSCAN method with only 18.86%. Another score computed is the mutual information score which shows how homogenous the clusters are for each model. Conclusion: The data from DEAP is very heterogeneous and its density is stable, which may indicate that classification would be the better option, in terms of accuracy and homogeneity of the clusters

    Generating One Biometric Feature from Another: Faces from Fingerprints

    Get PDF
    This study presents a new approach based on artificial neural networks for generating one biometric feature (faces) from another (only fingerprints). An automatic and intelligent system was designed and developed to analyze the relationships among fingerprints and faces and also to model and to improve the existence of the relationships. The new proposed system is the first study that generates all parts of the face including eyebrows, eyes, nose, mouth, ears and face border from only fingerprints. It is also unique and different from similar studies recently presented in the literature with some superior features. The parameter settings of the system were achieved with the help of Taguchi experimental design technique. The performance and accuracy of the system have been evaluated with 10-fold cross validation technique using qualitative evaluation metrics in addition to the expanded quantitative evaluation metrics. Consequently, the results were presented on the basis of the combination of these objective and subjective metrics for illustrating the qualitative properties of the proposed methods as well as a quantitative evaluation of their performances. Experimental results have shown that one biometric feature can be determined from another. These results have once more indicated that there is a strong relationship between fingerprints and faces

    Estimating Sparse Events using Probabilistic Logic: Application to Word n-Grams

    No full text
    In several tasks from different fields, we are encountering sparse events. In order to provide with probabilities for such events, researchers commonly perform a maximum likelihood (ML) estimation. However, it is well-known that the ML estimator is sensitive to extreme values. In other words, configurations with low or high frequencies are respectively underestimated or overestimated and therefore nonreliable. In order to solve this problem and to better evaluate these probability values, we propose a novel approach based on the probabilistic logic (PL) paradigm. For a sake of illustration, we focuss on this paper on events such as word trigrams (w 3 ; w 1 ; w 2 ) or word/pos-tag trigrams ((w 3 ; t 3 ); (w 1 ; t 1 ); (w 2 ; t 2 )). These latter entities are the basic objects used in speech or handwriting recognition. In order to distinguish between for example: "replace the fun" and "replace the floor" an accurate estimation of these two trigrams is needed. The ML estimation is equival..

    A Thematic Knowledge Extraction in Text using a Markovian Random Field Approach

    No full text
    . We present a Markovian Random Field modeling for thematic knowledge extraction in text. An analogy is made between a flow of thematic investigations/textual fragments matching and statistical mechanics systems. The Markovian Field Knowledge Extraction machine (MAFKE) that we propose is based on a dynamical interaction between thematic queries and fragments composing a text. The representation of the textual knowledge system is submitted to state variations emerging from the flow of thematic queries. The MAFKE machine tries to satisfy the user thematic queries by changing the set of Units of Information (UNIFs) contained in a fragment. This change is computed with respect to the input thematic queries. Hence, MAFKE machine transists from one configuration state to another by changing the threshold assigned to the pertinency of UNIFs. For each state, a certain degradation of the system which depends on the thematic query index and this threshold is considered. The equivalence concept b..

    Incorporating Diverse Information Sources in Handwriting Recognition Postprocessing

    No full text
    This paper describes the proposed implementation of a new model for the linguistic postprocessing component of the Human Language Technology (HLT) project. The model was designed for handwriting recognition applications but can be used for other text recognition problems and speech recognition. We demonstrate here that the current implementation (the POS model) fails to incorporate new sources of information such as word n-grams, and further handles the recogniser's scores incorrectly. We propose an alternative approach (the SSS model) which remedies these shortcomings. We also show that the SSS algorithm has a direct interpretation as a Hidden Markov Model whose states correspond to words that have been tagged with their parts of speech, and whose observations are discretised recogniser confidences. The HMM interpretation has the added advantage that the approach can be naturally extended to handle error recovery of the recogniser. Preliminary results indicate that the SSS model is su..
    corecore