141 research outputs found

    Leveraging Smartphone Sensor Data for Human Activity Recognition

    Get PDF
    Using smartphones for human activity recognition (HAR) has a wide range of applications including healthcare, daily fitness recording, and anomalous situations alerting. This study focuses on human activity recognition based on smartphone embedded sensors. The proposed human activity recognition system recognizes activities including walking, running, sitting, going upstairs, and going downstairs. Embedded sensors (a tri-axial accelerometer and a gyroscope sensor) are employed for motion data collection. Both time-domain and frequency-domain features are extracted and analyzed. Our experiment results show that time-domain features are good enough to recognize basic human activities. The system is implemented in an Android smartphone platform. While the focus has been on human activity recognition systems based on a supervised learning approach, an incremental clustering algorithm is investigated. The proposed unsupervised (clustering) activity detection scheme works in an incremental manner, which contains two stages. In the first stage, streamed sensor data will be processed. A single-pass clustering algorithm is used to generate pre-clustered results for the next stage. In the second stage, pre-clustered results will be refined to form the final clusters, which means the clusters are built incrementally by adding one cluster at a time. Experiments on smartphone sensor data of five basic human activities show that the proposed scheme can get comparable results with traditional clustering algorithms but working in a streaming and incremental manner. In order to develop more accurate activity recognition systems independent of smartphone models, effects of sensor differences across various smartphone models are investigated. We present the impairments of different smartphone embedded sensor models on HAR applications. Outlier removal, interpolation, and filtering in pre-processing stage are proposed as mitigating techniques. Based on datasets collected from four distinct smartphones, the proposed mitigating techniques show positive effects on 10-fold cross validation, device-to-device validation, and leave-one-out validation. Improved performance for smartphone based human activity recognition is observed. With the efforts of developing human activity recognition systems based on supervised learning approach, investigating a clustering based incremental activity recognition system with its potential applications, and applying techniques for alleviating sensor difference effects, a robust human activity recognition system can be trained in either supervised or unsupervised way and can be adapted to multiple devices with being less dependent on different sensor specifications

    Advanced Internet of Things for Personalised Healthcare System: A Survey

    Get PDF
    As a new revolution of the Internet, Internet of Things (IoT) is rapidly gaining ground as a new research topic in many academic and industrial disciplines, especially in healthcare. Remarkably, due to the rapid proliferation of wearable devices and smartphone, the Internet of Things enabled technology is evolving healthcare from conventional hub based system to more personalised healthcare system (PHS). However, empowering the utility of advanced IoT technology in PHS is still significantly challenging in the area considering many issues, like shortage of cost-effective and accurate smart medical sensors, unstandardized IoT system architectures, heterogeneity of connected wearable devices, multi-dimensionality of data generated and high demand for interoperability. In an effect to understand advance of IoT technologies in PHS, this paper will give a systematic review on advanced IoT enabled PHS. It will review the current research of IoT enabled PHS, and key enabling technologies, major IoT enabled applications and successful case studies in healthcare, and finally point out future research trends and challenges

    Deep learning approaches for human activity recognition using wearable technology

    Get PDF
    The need for long-term monitoring of individuals in their natural environment has initiated the development of a various number of wearable healthcare sensors for a wide range of applications: medical monitoring in clinical or home environments, physical activity assessment of athletes and recreators, baby monitoring in maternity hospitals and homes etc. Neural networks (NN) are data-driven type of modelling. Neural networks learn from experience, without knowledge about the model of phenomenon, but knowing the desired 'output' data for the training 'input' data. The most promising concept of machine learning that involves NN is the deep learning (DL) approach. The focus of this review is on approaches of DL for physiological activity recognition or human movement analysis purposes, using wearable technologies. This review shows that deep learning techniques are useful tools for health condition prediction or overall monitoring of data, streamed by wearable systems. Despite the considerable progress and wide field of applications, there are still some limitations and room for improvement of DL approaches for wearable healthcare systems, which may lead to more robust and reliable technology for personalized healthcare

    Optimized Biosignals Processing Algorithms for New Designs of Human Machine Interfaces on Parallel Ultra-Low Power Architectures

    Get PDF
    The aim of this dissertation is to explore Human Machine Interfaces (HMIs) in a variety of biomedical scenarios. The research addresses typical challenges in wearable and implantable devices for diagnostic, monitoring, and prosthetic purposes, suggesting a methodology for tailoring such applications to cutting edge embedded architectures. The main challenge is the enhancement of high-level applications, also introducing Machine Learning (ML) algorithms, using parallel programming and specialized hardware to improve the performance. The majority of these algorithms are computationally intensive, posing significant challenges for the deployment on embedded devices, which have several limitations in term of memory size, maximum operative frequency, and battery duration. The proposed solutions take advantage of a Parallel Ultra-Low Power (PULP) architecture, enhancing the elaboration on specific target architectures, heavily optimizing the execution, exploiting software and hardware resources. The thesis starts by describing a methodology that can be considered a guideline to efficiently implement algorithms on embedded architectures. This is followed by several case studies in the biomedical field, starting with the analysis of a Hand Gesture Recognition, based on the Hyperdimensional Computing algorithm, which allows performing a fast on-chip re-training, and a comparison with the state-of-the-art Support Vector Machine (SVM); then a Brain Machine Interface (BCI) to detect the respond of the brain to a visual stimulus follows in the manuscript. Furthermore, a seizure detection application is also presented, exploring different solutions for the dimensionality reduction of the input signals. The last part is dedicated to an exploration of typical modules for the development of optimized ECG-based applications

    Examining sensor-based physical activity recognition and monitoring for healthcare using Internet of Things: A systematic review.

    Get PDF
    Due to importantly beneficial effects on physical and mental health and strong association with many rehabilitation programs, Physical Activity Recognition and Monitoring (PARM) have been considered as a key paradigm for smart healthcare. Traditional methods for PARM focus on controlled environments with the aim of increasing the types of identifiable activity subjects complete and improving recognition accuracy and system robustness by means of novel body-worn sensors or advanced learning algorithms. The emergence of the Internet of Things (IoT) enabling technology is transferring PARM studies to open and connected uncontrolled environments by connecting heterogeneous cost-effective wearable devices and mobile apps. Little is currently known about whether traditional PARM technologies can tackle the new challenges of IoT environments and how to effectively harness and improve these technologies. In an effort to understand the use of IoT technologies in PARM studies, this paper will give a systematic review, critically examining PARM studies from a typical IoT layer-based perspective. It will firstly summarize the state-of-the-art in traditional PARM methodologies as used in the healthcare domain, including sensory, feature extraction and recognition techniques. The paper goes on to identify some new research trends and challenges of PARM studies in the IoT environments, and discusses some key enabling techniques for tackling them. Finally, this paper consider some of the successful case studies in the area and look at the possible future industrial applications of PARM in smart healthcare

    A CSI-Based Human Activity Recognition Using Deep Learning

    Get PDF
    The Internet of Things (IoT) has become quite popular due to advancements in Information and Communications technologies and has revolutionized the entire research area in Human Activity Recognition (HAR). For the HAR task, vision-based and sensor-based methods can present better data but at the cost of users’ inconvenience and social constraints such as privacy issues. Due to the ubiquity of WiFi devices, the use of WiFi in intelligent daily activity monitoring for elderly persons has gained popularity in modern healthcare applications. Channel State Information (CSI) as one of the characteristics ofWiFi signals, can be utilized to recognize different human activities. We have employed a Raspberry Pi 4 to collect CSI data for seven different human daily activities, and converted CSI data to images and then used these images as inputs of a 2D Convolutional Neural Network (CNN) classifier. Our experiments have shown that the proposed CSI-based HAR outperforms other competitor methods including 1D-CNN, Long Short-Term Memory (LSTM), and Bi-directional LSTM, and achieves an accuracy of around 95% for seven activities

    Designing a Sensor-Based Wearable Computing System for Custom Hand Gesture Recognition Using Machine Learning

    Get PDF
    This thesis investigates how assistive technology can be made to facilitate communication for people that are unable to or have difficulty communicating via vocal speech, and how this technology can be made more universal and compatible with the many different types of sign language that they use. Through this research, a fully customisable and stand-alone wearable device was developed, that employs machine learning techniques to recognise individual hand gestures and translate them into text, images and speech. The device can recognise and translate custom hand gestures by training a personal classifier for each user, relying on a small training sample size, that works online on an embedded system or mobile device, with a classification accuracy rate of up to 99%. This was achieved through a series of iterative case studies, with user testing carried out by real users in their every day environments and in public spaces

    A context aware approach for enhancing gesture recognition accuracy on handheld devices

    Get PDF
    Ankara : The Department of Computer Engineering and the Institute of Engineering and Science of Bilkent University, 2010.Thesis (Master's) -- Bilkent University, 2010.Includes bibliographical references leaves 75-84.Input capabilities (e.g. joystick, keypad) of handheld devices allow users to interact with the user interface to access the information and mobile services. However, these input capabilities are very limited because of the mobile convenience. New input devices and interaction techniques are needed for handheld devices. Gestural interaction with accelerometer sensor is one of the newest interaction techniques on mobile computing. In this thesis, we introduce solutions that can be used for automatically enhancing the gesture recognition accuracy of accelerometer sensor, and as a standardized gesture library for gestural interaction on touch screen and accelerometer sensor. In this novel solution, we propose a framework that decides on suitable signal processing techniques for acceleration sensor data for a given context of the user. First system recognizes the context of the user using pattern recognition algorithm. Then, system automatically chooses signal ltering techniques for recognized context, and recognizes gestures. Gestures are also standardized for better usage. In this work, we also present several experiments which show the feasibility and e ectiveness of our automated gesture recognition enhancement system.Yıldırım, Hacı MehmetM.S
    corecore