25 research outputs found

    Automatic Identity Recognition Using Speech Biometric

    Get PDF
    Biometric technology refers to the automatic identification of a person using physical or behavioral traits associated with him/her. This technology can be an excellent candidate for developing intelligent systems such as speaker identification, facial recognition, signature verification...etc. Biometric technology can be used to design and develop automatic identity recognition systems, which are highly demanded and can be used in banking systems, employee identification, immigration, e-commerce…etc. The first phase of this research emphasizes on the development of automatic identity recognizer using speech biometric technology based on Artificial Intelligence (AI) techniques provided in MATLAB. For our phase one, speech data is collected from 20 (10 male and 10 female) participants in order to develop the recognizer. The speech data include utterances recorded for the English language digits (0 to 9), where each participant recorded each digit 3 times, which resulted in a total of 600 utterances for all participants. For our phase two, speech data is collected from 100 (50 male and 50 female) participants in order to develop the recognizer. The speech data is divided into text-dependent and text-independent data, whereby each participant selected his/her full name and recorded it 30 times, which makes up the text-independent data. On the other hand, the text-dependent data is represented by a short Arabic language story that contains 16 sentences, whereby every sentence was recorded by every participant 5 times. As a result, this new corpus contains 3000 (30 utterances * 100 speakers) sound files that represent the text-independent data using their full names and 8000 (16 sentences * 5 utterances * 100 speakers) sound files that represent the text-dependent data using the short story. For the purpose of our phase one of developing the automatic identity recognizer using speech, the 600 utterances have undergone the feature extraction and feature classification phases. The speech-based automatic identity recognition system is based on the most dominating feature extraction technique, which is known as the Mel-Frequency Cepstral Coefficient (MFCC). For feature classification phase, the system is based on the Vector Quantization (VQ) algorithm. Based on our experimental results, the highest accuracy achieved is 76%. The experimental results have shown acceptable performance, but can be improved further in our phase two using larger speech data size and better performance classification techniques such as the Hidden Markov Model (HMM)

    Automatic identity recognition systems : a review

    Get PDF
    Rapidly changed computer technology and fast growth of communication ways, makes everyday work easy and managed. Technology takes place everywhere, in business, education, market, security... etc. However, communication between human and these technologies become the main concern of many research areas, especially for developing automatic identity recognition systems. However, biometric technologies are among the most important technologies used in this area. Biometric technology refers to the automatic identity recognition using physical or behavioral traits associated with him/her. Using biometrics, it is possible to establish physiological-based systems that depend on physiological characteristics such as fingerprint, face recognition, DNA... etc, or behavioral-based systems that depend on behavioral characteristics such as gait, voice ...etc, or even combining both of them in one system. Therefore, biometrics technologies can be excellent candidates for developing intelligent systems such as speaker identification, facial recognition, signature verification...etc. In addition, biometric technologies are flexible enough to be combined with other tools to produce more secure and easier to use verification system

    Arabic Speaker-Independent Continuous Automatic Speech Recognition Based on a Phonetically Rich and Balanced Speech Corpus

    Get PDF
    This paper describes and proposes an efficient and effective framework for the design and development of a speaker-independent continuous automatic Arabic speech recognition system based on a phonetically rich and balanced speech corpus. The speech corpus contains a total of 415 sentences recorded by 40 (20 male and 20 female) Arabic native speakers from 11 different Arab countries representing the three major regions (Levant, Gulf, and Africa) in the Arab world. The proposed Arabic speech recognition system is based on the Carnegie Mellon University (CMU) Sphinx tools, and the Cambridge HTK tools were also used at some testing stages. The speech engine uses 3-emitting state Hidden Markov Models (HMM) for tri-phone based acoustic models. Based on experimental analysis of about 7 hours of training speech data, the acoustic model is best using continuous observation’s probability model of 16 Gaussian mixture distributions and the state distributions were tied to 500 senones. The language model contains both bi-grams and tri-grams. For similar speakers but different sentences, the system obtained a word recognition accuracy of 92.67% and 93.88% and a Word Error Rate (WER) of 11.27% and 10.07% with and without diacritical marks respectively. For different speakers with similar sentences, the system obtained a word recognition accuracy of 95.92% and 96.29% and a WER of 5.78% and 5.45% with and without diacritical marks respectively. Whereas different speakers and different sentences, the system obtained a word recognition accuracy of 89.08% and 90.23% and a WER of 15.59% and 14.44% with and without diacritical marks respectively

    Arabic automatic continuous speech recognition systems

    Get PDF
    MSA is the current formal linguistic standard of Arabic language, which is widely taught in schools and universities, and often used in the office and the media. MSA is also considered as the only acceptable form of Arabic language for all native speakers [I]. As recently, the research community has witnessed an improvement in the performance of ASR systems, there is an increasingly widespread use of this technology for several languages of the world. Similarly, research interests have grown significantly in the past few years for Arabic ASR research. It is noticed that Arabic ASR research is not only conducted and investigated by researchers in the Arab world, but also by many others located in different parts of the \vorld especially the western countries

    Al-Quran learning using mobile speech recognition:an overview

    Get PDF
    The usage of mobile application in various aspects has been worldwide accepted and there are variety of mobile applications which developed to cater the usage of different background of the user. In this paper, a short survey which includes questionnaire is distributed to find the interest of user whom using application for learning Quran and concept of mobile speech apps. The main interest of this survey is to find the acceptance of user and explanation on the proposed usage of mobile speech recognition with feature of learning apps. Factors of mobile speech recognition and mobile learning are listed to support the results from the short survey

    Acoustic echo cancellation using adaptive filtering algorithms for quranic accents (Qiraat) identification

    Get PDF
    Echoed parts of Quranic accent (Qiraat) signals are exposed to reverberation of signals especially if they are listened to in a conference room or the Quranic recordings found in different media such as the web. Quranic verse rules identification/Tajweed are prone to additive noise and may reduce classification results. This research work aims to present our work towards Quranic accents (Qiraat) identification, which emphasizes on acoustic echo cancellation (AEC) of all echoed Quranic signals during the preprocessing phase of the system development. In order to conduct the AEC, three adaptive algorithms known as affine projection (AP), least mean square (LMS), and recursive least squares (RLS) are used during the preprocessing phase. Once clean Quranic signals are produced, they undergo feature extraction and pattern classification phases. The Mel Frequency Cepstral Coefficients is the most widely used technique for feature extraction and is adopted in this research work, whereas probabilities principal component analysis (PPCA), K-nearest neighbor (KNN) and gaussian mixture model (GMM) are used for pattern classification. In order to verify our methodology, audio files have been collected for Surat Ad-Duhaa for five different Quranic accents (Qiraat), namely: (1) Ad-Duri, (2) Al-Kisaie, (3) Hafs an A’asem, (4) IbnWardan, and (5) Warsh. Based on our experimental results, the AP algorithm achieved 93.9 % accuracy rate against all pattern classification techniques including PPCA, KNN, and GMM. For LMS and RLS, the achieved accuracy rates are different for PPCA, KNN, and GMM, whereby LMS with PPCA and GMM achieved the same accuracy rate of 96.9 %; however, LMS with KNN achieved 84.8 %. In addition, RLS with PPCA and GMM achieved the same accuracy rate of 90.9 %; however, RLS with KNN achieved 78.8 %. Therefore, the AP adaptive algorithm is able to reduce the echo of Quranic accents (Qiraat) signals in a consistent manner against all pattern classification techniques

    Speaker’s variabilities, technology and language issues that affect automatic speech and speaker recognition systems

    Get PDF
    Automatic Speech Recognition (ASR) is gammg its importance due to the vast growth generally in technology and computing in specific. From industrial perspective, computers, laptops, and mobile devices nowadays have the ASR support embedded into the operating system. From academia on the other hand, there are many research efforts being conducted addressing this technology in order to contribute to its state-of-the-art. On the other hand, speaker recognition systems are also growing due to various threats, therefore, these systems are mostly meant for security purpose

    English digits speech recognition system based on Hidden Markov Models

    No full text
    This paper aims to design and implement English digits speech recognition system using Matlab (GUI). This work was based on the Hidden Markov Model (HMM), which provides a highly reliable way for recognizing speech. The system is able to recognize the speech waveform by translating the speech waveform into a set of feature vectors using Mel Frequency Cepstral Coefficients (MFCC) technique This paper focuses on all English digits from (Zero through Nine), which is based on isolated words structure. Two modules were developed, namely the isolated words speech recognition and the continuous speech recognition. Both modules were tested in both clean and noisy environments and showed a successful recognition rates. In clean environment and isolated words speech recognition module, the multi-speaker mode achieved 99.5% whereas the speaker-independent mode achieved 79.5%. In clean environment and continuous speech recognition module, the multi-speaker mode achieved 72.5% whereas the speaker-independent mode achieved 56.25%. However in noisy environment and isolated words speech recognition module, the multi-speaker mode achieved 88% whereas the speaker-independent mode achieved 67%. In noisy environment and continuous speech recognition module, the multi-speaker mode achieved 82.5% whereas the speaker-independent mode achieved 76.67%. These recognition rates are relatively successful if compared to similar systems

    Implicit aspect extraction in sentiment analysis: Review, taxonomy, oppportunities, and open challenges

    No full text
    Sentiment analysis is a text classification branch, which is defined as the process of extracting sentiment terms (i.e. feature/aspect, or opinion) and determining their opinion semantic orientation. At aspect level, aspect extraction is the core task for sentiment analysis which can either be implicit or explicit aspects. The growth of sentiment analysis has resulted in the emergence of various techniques for both explicit and implicit aspect extraction. However, majority of the research attempts targeted explicit aspect extraction, which indicates that there is a lack of research on implicit aspect extraction. This research provides a review of implicit aspect/features extraction techniques from different perspectives. The first perspective is making a comparison analysis for the techniques available for implicit term extraction with a brief summary of each technique. The second perspective is classifying and comparing the performance, datasets, language used, and shortcomings of the available techniques. In this study, over 50 articles have been reviewed, however, only 45 articles on implicit aspect extraction that span from 2005 to 2016 were analyzed and discussed. Majority of the researchers on implicit aspects extraction rely heavily on unsupervised methods in their research, which makes about 64% of the 45 articles, followed by supervised methods of about 27%, and lastly semi-supervised of 9%. In addition, 25 articles conducted the research work solely on product reviews, and 5 articles conducted their research work using product reviews jointly with other types of data, which makes product review datasets the most frequently used data type compared to other types. Furthermore, research on implicit aspect features extraction has focused on English and Chinese languages compared to other languages. Finally, this review also provides recommendations for future research directions and open problems

    Automatic person identification system using handwritten signatures

    No full text
    This paper reports the design, implementation, and evaluation of a research work for developing an automatic person identification system using hand signatures biometric. The developed automatic person identification system mainly used toolboxes provided by MATLAB environment. . In order to train and test the developed automatic person identification system, an in-house hand signatures database is created, which contains hand signatures of 100 persons (50 males and 50 females) each of which is repeated 30 times. Therefore, a total of 3000 hand signatures are collected. The collected hand signatures have gone through pre-processing steps such as producing a digitized version of the signatures using a scanner, converting input images type to a standard binary images type, cropping, normalizing images size, and reshaping in order to produce a ready-to-use hand signatures database for training and testing the automatic person identification system. Global features such as signature height, image area, pure width, and pure height are then selected to be used in the system, which reflect information about the structure of the hand signature image. For features training and classification, the Multi-Layer Perceptron (MLP) architecture of Artificial Neural Network (ANN) is used. This paper also investigates the effect of the persons’ gender on the overall performance of the system. For performance optimization, the effect of modifying values of basic parameters in ANN such as the number of hidden neurons and the number of epochs are investigated in this work. The handwritten signature data collected from male persons outperformed those collected from the female persons, whereby the system obtained average recognition rates of 76.20% and74.20% for male and female persons, respectively. Overall, the handwritten signatures based system obtained an average recognition rate of 75.20% for all persons
    corecore