83 research outputs found

    Stable Electromyographic Sequence Prediction During Movement Transitions using Temporal Convolutional Networks

    Full text link
    Transient muscle movements influence the temporal structure of myoelectric signal patterns, often leading to unstable prediction behavior from movement-pattern classification methods. We show that temporal convolutional network sequential models leverage the myoelectric signal's history to discover contextual temporal features that aid in correctly predicting movement intentions, especially during interclass transitions. We demonstrate myoelectric classification using temporal convolutional networks to effect 3 simultaneous hand and wrist degrees-of-freedom in an experiment involving nine human-subjects. Temporal convolutional networks yield significant (p<0.001)(p<0.001) performance improvements over other state-of-the-art methods in terms of both classification accuracy and stability.Comment: 4 pages, 5 figures, accepted for Neural Engineering (NER) 2019 Conferenc

    Segmentation of Speech and Humming in Vocal Input

    Get PDF
    Non-verbal vocal interaction (NVVI) is an interaction method in which sounds other than speech produced by a human are used, such as humming. NVVI complements traditional speech recognition systems with continuous control. In order to combine the two approaches (e.g. "volume up, mmm") it is necessary to perform a speech/NVVI segmentation of the input sound signal. This paper presents two novel methods of speech and humming segmentation. The first method is based on classification of MFCC and RMS parameters using a neural network (MFCC method), while the other method computes volume changes in the signal (IAC method). The two methods are compared using a corpus collected from 13 speakers. The results indicate that the MFCC method outperforms IAC in terms of accuracy, precision, and recall

    Speech Recognition using Surface Electromyography

    Get PDF

    Frame-Based Phone Classification Using EMG Signals

    Get PDF
    This paper evaluates the impact of inter-speaker and inter-session variability on the development of a silent speech interface (SSI) based on electromyographic (EMG) signals from the facial muscles. The final goal of the SSI is to provide a communication tool for Spanish-speaking laryngectomees by generating audible speech from voiceless articulation. However, before moving on to such a complex task, a simpler phone classification task in different modalities regarding speaker and session dependency is performed for this study. These experiments consist of processing the recorded utterances into phone-labeled segments and predicting the phonetic labels using only features obtained from the EMG signals. We evaluate and compare the performance of each model considering the classification accuracy. Results show that the models are able to predict the phonetic label best when they are trained and tested using data from the same session. The accuracy drops drastically when the model is tested with data from a different session, although it improves when more data are added to the training data. Similarly, when the same model is tested on a session from a different speaker, the accuracy decreases. This suggests that using larger amounts of data could help to reduce the impact of inter-session variability, but more research is required to understand if this approach would suffice to account for inter-speaker variability as well.This research was funded by Agencia Estatal de Investigación grant number ref.PID2019-108040RB-C21/AEI/10.13039/50110001103

    Correlation analysis of electromyogram signals for multiuser myoelectric interfaces

    Full text link
    An inability to adapt myoelectric interfaces to a novel user's unique style of hand motion, or even to adapt to the motion style of an opposite limb upon which the interface is trained, are important factors inhibiting the practical application of myoelectric interfaces. This is mainly attributed to the individual differences in the exhibited electromyogram (EMG) signals generated by the muscles of different limbs. We propose in this paper a multiuser myoelectric interface which easily adapts to novel users and maintains good movement recognition performance. The main contribution is a framework for implementing style-independent feature transformation by using canonical correlation analysis (CCA) in which different users' data is projected onto a unified-style space. The proposed idea is summarized into three steps: 1) train a myoelectric pattern classifier on the set of style-independent features extracted from multiple users using the proposed CCA-based mapping; 2) create a new set of features describing the movements of a novel user during a quick calibration session; and 3) project the novel user's features onto a lower dimensional unified-style space with features maximally correlated with training data and classify accordingly. The proposed method has been validated on a set of eight intact-limbed subjects, left-and-right handed, performing ten classes of bilateral synchronous fingers movements with four electrodes on each forearm. The method was able to overcome individual differences through the style-independent framework with accuracies of >83% across multiple users. Testing was also performed on a set of ten intact-limbed and six below-elbow amputee subjects as they performed finger and thumb movements. The proposed framework allowed us to train the classifier on a normal subject's data while subsequently testing it on an amputee's data after calibration with a performance of >82% on average across all amputees. © 2001-2011 IEEE

    EMG-to-Speech: Direct Generation of Speech from Facial Electromyographic Signals

    Get PDF
    The general objective of this work is the design, implementation, improvement and evaluation of a system that uses surface electromyographic (EMG) signals and directly synthesizes an audible speech output: EMG-to-speech

    Principal Component Analysis Applied to Surface Electromyography: A Comprehensive Review

    Get PDF
    © 2016 IEEE. Surface electromyography (sEMG) records muscle activities from the surface of muscles, which offers a wealth of information concerning muscle activation patterns in both research and clinical settings. A key principle underlying sEMG analyses is the decomposition of the signal into a number of motor unit action potentials (MUAPs) that capture most of the relevant features embedded in a low-dimensional space. Toward this, the principal component analysis (PCA) has extensively been sought after, whereby the original sEMG data are translated into low-dimensional MUAP components with a reduced level of redundancy. The objective of this paper is to disseminate the role of PCA in conjunction with the quantitative sEMG analyses. Following the preliminaries on the sEMG methodology and a statement of PCA algorithm, an exhaustive collection of PCA applications related to sEMG data is in order. Alongside the technical challenges associated with the PCA-based sEMG processing, the envisaged research trend is also discussed

    Towards a Multimodal Silent Speech Interface for European Portuguese

    Get PDF
    Automatic Speech Recognition (ASR) in the presence of environmental noise is still a hard problem to tackle in speech science (Ng et al., 2000). Another problem well described in the literature is the one concerned with elderly speech production. Studies (Helfrich, 1979) have shown evidence of a slower speech rate, more breaks, more speech errors and a humbled volume of speech, when comparing elderly with teenagers or adults speech, on an acoustic level. This fact makes elderly speech hard to recognize, using currently available stochastic based ASR technology. To tackle these two problems in the context of ASR for HumanComputer Interaction, a novel Silent Speech Interface (SSI) in European Portuguese (EP) is envisioned.info:eu-repo/semantics/acceptedVersio
    corecore