342 research outputs found

    Hand Motion Tracking System using Inertial Measurement Units and Infrared Cameras

    Get PDF
    This dissertation presents a novel approach to develop a system for real-time tracking of the position and orientation of the human hand in three-dimensional space, using MEMS inertial measurement units (IMUs) and infrared cameras. This research focuses on the study and implementation of an algorithm to correct the gyroscope drift, which is a major problem in orientation tracking using commercial-grade IMUs. An algorithm to improve the orientation estimation is proposed. It consists of: 1.) Prediction of the bias offset error while the sensor is static, 2.) Estimation of a quaternion orientation from the unbiased angular velocity, 3.) Correction of the orientation quaternion utilizing the gravity vector and the magnetic North vector, and 4.) Adaptive quaternion interpolation, which determines the final quaternion estimate based upon the current conditions of the sensor. The results verified that the implementation of the orientation correction algorithm using the gravity vector and the magnetic North vector is able to reduce the amount of drift in orientation tracking and is compatible with position tracking using infrared cameras for real-time human hand motion tracking. Thirty human subjects participated in an experiment to validate the performance of the hand motion tracking system. The statistical analysis shows that the error of position tracking is, on average, 1.7 cm in the x-axis, 1.0 cm in the y-axis, and 3.5 cm in the z-axis. The Kruskal-Wallis tests show that the orientation correction algorithm using gravity vector and magnetic North vector can significantly reduce the errors in orientation tracking in comparison to fixed offset compensation. Statistical analyses show that the orientation correction algorithm using gravity vector and magnetic North vector and the on-board Kalman-based orientation filtering produced orientation errors that were not significantly different in the Euler angles, Phi, Theta and Psi, with the p-values of 0.632, 0.262 and 0.728, respectively. The proposed orientation correction algorithm represents a contribution to the emerging approaches to obtain reliable orientation estimates from MEMS IMUs. The development of a hand motion tracking system using IMUs and infrared cameras in this dissertation enables future improvements in natural human-computer interactions within a 3D virtual environment

    Sign language recognition using wearable electronics: Implementing K-nearest neighbors with dynamic time warping and convolutional neural network algorithms

    Get PDF
    We propose a sign language recognition system based on wearable electronics and two different classification algorithms. The wearable electronics were made of a sensory glove and inertial measurement units to gather fingers, wrist, and arm/forearm movements. The classifiers were k-Nearest Neighbors with Dynamic Time Warping (that is a non-parametric method) and Convolutional Neural Networks (that is a parametric method). Ten sign-words were considered from the Italian Sign Language: cose, grazie, maestra, together with words with international meaning such as google, internet, jogging, pizza, television, twitter, and ciao. The signs were repeated one-hundred times each by seven people, five male and two females, aged 29โ€“54 y ยฑ 10.34 (SD). The adopted classifiers performed with an accuracy of 96.6% ยฑ 3.4 (SD) for the k-Nearest Neighbors plus the Dynamic Time Warping and of 98.0% ยฑ 2.0 (SD) for the Convolutional Neural Networks. Our system was made of wearable electronics among the most complete ones, and the classifiers top performed in comparison with other relevant works reported in the literature

    Novel smart glove technology as a biomechanical monitoring tool

    Get PDF
    Developments in Virtual Reality (VR) technology and its overall market have been occurring since the 1960s when Ivan Sutherland created the worldโ€™s first tracked head-mounted display (HMD) โ€“ a goggle type head gear. In society today, consumers are expecting a more immersive experience and associated tools to bridge the cyber-physical divide. This paper presents the development of a next generation smart glove microsystem to facilitate Human Computer Interaction through the integration of sensors, processors and wireless technology. The objective of the glove is to measure the range of hand joint movements, in real time and empirically in a quantitative manner. This includes accurate measurement of flexion, extension, adduction and abduction of the metacarpophalangeal (MCP), Proximal interphalangeal (PIP) and Distal interphalangeal (DIP) joints of the fingers and thumb in degrees, together with thumb-index web space movement. This system enables full real-time monitoring of complex hand movements. Commercially available gloves are not fitted with sufficient sensors for full data capture, and require calibration for each glove wearer. Unlike these current state-of-the-art data gloves, the UU / Tyndall Inertial Measurement Unit (IMU) glove uses a combination of novel stretchable substrate material and 9 degree of freedom (DOF) inertial sensors in conjunction with complex data analytics to detect joint movement. Our novel IMU data glove requires minimal calibration and is therefore particularly suited to multiple application domains such as Human Computer interfacing, Virtual reality, the healthcare environment

    An Instrumented Glove to Assess Manual Dexterity in Simulation-Based Neurosurgical Education

    Get PDF
    ABSTRACT: The traditional neurosurgical apprenticeship scheme includes the assessment of traineeโ€™s manual skills carried out by experienced surgeons. However, the introduction of surgical simulation technology presents a new paradigm where residents can refine surgical techniques on a simulator before putting them into practice in real patients. Unfortunately, in this new scheme, an experienced surgeon will not always be available to evaluate traineeโ€™s performance. For this reason, it is necessary to develop automatic mechanisms to estimate metrics for assessing manual dexterity in a quantitative way. Authors have proposed some hardware-software approaches to evaluate manual dexterity on surgical simulators. This paper presents IGlove, a wearable device that uses inertial sensors embedded on an elastic glove to capture hand movements. Metrics to assess manual dexterity are estimated from sensors signals using data processing and information analysis algorithms. It has been designed to be used with a neurosurgical simulator called Daubara NS Trainer, but can be easily adapted to another benchtop- and manikin-based medical simulators. The system was tested with a sample of 14 volunteers who performed a test that was designed to simultaneously evaluate their fine motor skills and the IGloveโ€™s functionalities. Metrics obtained by each of the participants are presented as results in this work; it is also shown how these metrics are used to automatically evaluate the level of manual dexterity of each volunteer

    ์ธ๊ฐ„ ๊ธฐ๊ณ„ ์ƒํ˜ธ์ž‘์šฉ์„ ์œ„ํ•œ ๊ฐ•๊ฑดํ•˜๊ณ  ์ •ํ™•ํ•œ ์†๋™์ž‘ ์ถ”์  ๊ธฐ์ˆ  ์—ฐ๊ตฌ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ๊ธฐ๊ณ„ํ•ญ๊ณต๊ณตํ•™๋ถ€, 2021.8. ์ด๋™์ค€.Hand-based interface is promising for realizing intuitive, natural and accurate human machine interaction (HMI), as the human hand is main source of dexterity in our daily activities. For this, the thesis begins with the human perception study on the detection threshold of visuo-proprioceptive conflict (i.e., allowable tracking error) with or without cutantoues haptic feedback, and suggests tracking error specification for realistic and fluidic hand-based HMI. The thesis then proceeds to propose a novel wearable hand tracking module, which, to be compatible with the cutaneous haptic devices spewing magnetic noise, opportunistically employ heterogeneous sensors (IMU/compass module and soft sensor) reflecting the anatomical properties of human hand, which is suitable for specific application (i.e., finger-based interaction with finger-tip haptic devices). This hand tracking module however loses its tracking when interacting with, or being nearby, electrical machines or ferromagnetic materials. For this, the thesis presents its main contribution, a novel visual-inertial skeleton tracking (VIST) framework, that can provide accurate and robust hand (and finger) motion tracking even for many challenging real-world scenarios and environments, for which the state-of-the-art technologies are known to fail due to their respective fundamental limitations (e.g., severe occlusions for tracking purely with vision sensors; electromagnetic interference for tracking purely with IMUs (inertial measurement units) and compasses; and mechanical contacts for tracking purely with soft sensors). The proposed VIST framework comprises a sensor glove with multiple IMUs and passive visual markers as well as a head-mounted stereo camera; and a tightly-coupled filtering-based visual-inertial fusion algorithm to estimate the hand/finger motion and auto-calibrate hand/glove-related kinematic parameters simultaneously while taking into account the hand anatomical constraints. The VIST framework exhibits good tracking accuracy and robustness, affordable material cost, light hardware and software weights, and ruggedness/durability even to permit washing. Quantitative and qualitative experiments are also performed to validate the advantages and properties of our VIST framework, thereby, clearly demonstrating its potential for real-world applications.์† ๋™์ž‘์„ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•œ ์ธํ„ฐํŽ˜์ด์Šค๋Š” ์ธ๊ฐ„-๊ธฐ๊ณ„ ์ƒํ˜ธ์ž‘์šฉ ๋ถ„์•ผ์—์„œ ์ง๊ด€์„ฑ, ๋ชฐ์ž…๊ฐ, ์ •๊ตํ•จ์„ ์ œ๊ณตํ•ด์ค„ ์ˆ˜ ์žˆ์–ด ๋งŽ์€ ์ฃผ๋ชฉ์„ ๋ฐ›๊ณ  ์žˆ๊ณ , ์ด๋ฅผ ์œ„ํ•ด ๊ฐ€์žฅ ํ•„์ˆ˜์ ์ธ ๊ธฐ์ˆ  ์ค‘ ํ•˜๋‚˜๊ฐ€ ์† ๋™์ž‘์˜ ๊ฐ•๊ฑดํ•˜๊ณ  ์ •ํ™•ํ•œ ์ถ”์  ๊ธฐ์ˆ  ์ด๋‹ค. ์ด๋ฅผ ์œ„ํ•ด ๋ณธ ํ•™์œ„๋…ผ๋ฌธ์—์„œ๋Š” ๋จผ์ € ์‚ฌ๋žŒ ์ธ์ง€์˜ ๊ด€์ ์—์„œ ์† ๋™์ž‘ ์ถ”์  ์˜ค์ฐจ์˜ ์ธ์ง€ ๋ฒ”์œ„๋ฅผ ๊ทœ๋ช…ํ•œ๋‹ค. ์ด ์˜ค์ฐจ ์ธ์ง€ ๋ฒ”์œ„๋Š” ์ƒˆ๋กœ์šด ์† ๋™์ž‘ ์ถ”์  ๊ธฐ์ˆ  ๊ฐœ๋ฐœ ์‹œ ์ค‘์š”ํ•œ ์„ค๊ณ„ ๊ธฐ์ค€์ด ๋  ์ˆ˜ ์žˆ์–ด ์ด๋ฅผ ํ”ผํ—˜์ž ์‹คํ—˜์„ ํ†ตํ•ด ์ •๋Ÿ‰์ ์œผ๋กœ ๋ฐํžˆ๊ณ , ํŠนํžˆ ์†๋ ์ด‰๊ฐ ์žฅ๋น„๊ฐ€ ์žˆ์„๋•Œ ์ด ์ธ์ง€ ๋ฒ”์œ„์˜ ๋ณ€ํ™”๋„ ๋ฐํžŒ๋‹ค. ์ด๋ฅผ ํ† ๋Œ€๋กœ, ์ด‰๊ฐ ํ”ผ๋“œ๋ฐฑ์„ ์ฃผ๋Š” ๊ฒƒ์ด ๋‹ค์–‘ํ•œ ์ธ๊ฐ„-๊ธฐ๊ณ„ ์ƒํ˜ธ์ž‘์šฉ ๋ถ„์•ผ์—์„œ ๋„๋ฆฌ ์—ฐ๊ตฌ๋˜์–ด ์™”์œผ๋ฏ€๋กœ, ๋จผ์ € ์†๋ ์ด‰๊ฐ ์žฅ๋น„์™€ ํ•จ๊ป˜ ์‚ฌ์šฉํ•  ์ˆ˜ ์žˆ๋Š” ์† ๋™์ž‘ ์ถ”์  ๋ชจ๋“ˆ์„ ๊ฐœ๋ฐœํ•œ๋‹ค. ์ด ์†๋ ์ด‰๊ฐ ์žฅ๋น„๋Š” ์ž๊ธฐ์žฅ ์™ธ๋ž€์„ ์ผ์œผ์ผœ ์ฐฉ์šฉํ˜• ๊ธฐ์ˆ ์—์„œ ํ”ํžˆ ์‚ฌ์šฉ๋˜๋Š” ์ง€์ž๊ธฐ ์„ผ์„œ๋ฅผ ๊ต๋ž€ํ•˜๋Š”๋ฐ, ์ด๋ฅผ ์ ์ ˆํ•œ ์‚ฌ๋žŒ ์†์˜ ํ•ด๋ถ€ํ•™์  ํŠน์„ฑ๊ณผ ๊ด€์„ฑ ์„ผ์„œ/์ง€์ž๊ธฐ ์„ผ์„œ/์†Œํ”„ํŠธ ์„ผ์„œ์˜ ์ ์ ˆํ•œ ํ™œ์šฉ์„ ํ†ตํ•ด ํ•ด๊ฒฐํ•œ๋‹ค. ์ด๋ฅผ ํ™•์žฅํ•˜์—ฌ ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š”, ์ด‰๊ฐ ์žฅ๋น„ ์ฐฉ์šฉ ์‹œ ๋ฟ ์•„๋‹ˆ๋ผ ๋ชจ๋“  ์žฅ๋น„ ์ฐฉ์šฉ / ํ™˜๊ฒฝ / ๋ฌผ์ฒด์™€์˜ ์ƒํ˜ธ์ž‘์šฉ ์‹œ์—๋„ ์‚ฌ์šฉ ๊ฐ€๋Šฅํ•œ ์ƒˆ๋กœ์šด ์† ๋™์ž‘ ์ถ”์  ๊ธฐ์ˆ ์„ ์ œ์•ˆํ•œ๋‹ค. ๊ธฐ์กด์˜ ์† ๋™์ž‘ ์ถ”์  ๊ธฐ์ˆ ๋“ค์€ ๊ฐ€๋ฆผ ํ˜„์ƒ (์˜์ƒ ๊ธฐ๋ฐ˜ ๊ธฐ์ˆ ), ์ง€์ž๊ธฐ ์™ธ๋ž€ (๊ด€์„ฑ/์ง€์ž๊ธฐ ์„ผ์„œ ๊ธฐ๋ฐ˜ ๊ธฐ์ˆ ), ๋ฌผ์ฒด์™€์˜ ์ ‘์ด‰ (์†Œํ”„ํŠธ ์„ผ์„œ ๊ธฐ๋ฐ˜ ๊ธฐ์ˆ ) ๋“ฑ์œผ๋กœ ์ธํ•ด ์ œํ•œ๋œ ํ™˜๊ฒฝ์—์„œ ๋ฐ–์— ์‚ฌ์šฉํ•˜์ง€ ๋ชปํ•œ๋‹ค. ์ด๋ฅผ ์œ„ํ•ด ๋งŽ์€ ๋ฌธ์ œ๋ฅผ ์ผ์œผํ‚ค๋Š” ์ง€์ž๊ธฐ ์„ผ์„œ ์—†์ด ์ƒ๋ณด์ ์ธ ํŠน์„ฑ์„ ์ง€๋‹ˆ๋Š” ๊ด€์„ฑ ์„ผ์„œ์™€ ์˜์ƒ ์„ผ์„œ๋ฅผ ์œตํ•ฉํ•˜๊ณ , ์ด๋•Œ ์ž‘์€ ๊ณต๊ฐ„์— ๋‹ค ์ž์œ ๋„์˜ ์›€์ง์ž„์„ ๊ฐ–๋Š” ์† ๋™์ž‘์„ ์ถ”์ ํ•˜๊ธฐ ์œ„ํ•ด ๋‹ค์ˆ˜์˜ ๊ตฌ๋ถ„๋˜์ง€ ์•Š๋Š” ๋งˆ์ปค๋“ค์„ ์‚ฌ์šฉํ•œ๋‹ค. ์ด ๋งˆ์ปค์˜ ๊ตฌ๋ถ„ ๊ณผ์ • (correspondence search)๋ฅผ ์œ„ํ•ด ๊ธฐ์กด์˜ ์•ฝ๊ฒฐํ•ฉ (loosely-coupled) ๊ธฐ๋ฐ˜์ด ์•„๋‹Œ ๊ฐ•๊ฒฐํ•ฉ (tightly-coupled ๊ธฐ๋ฐ˜ ์„ผ์„œ ์œตํ•ฉ ๊ธฐ์ˆ ์„ ์ œ์•ˆํ•˜๊ณ , ์ด๋ฅผ ํ†ตํ•ด ์ง€์ž๊ธฐ ์„ผ์„œ ์—†์ด ์ •ํ™•ํ•œ ์† ๋™์ž‘์ด ๊ฐ€๋Šฅํ•  ๋ฟ ์•„๋‹ˆ๋ผ ์ฐฉ์šฉํ˜• ์„ผ์„œ๋“ค์˜ ์ •ํ™•์„ฑ/ํŽธ์˜์„ฑ์— ๋ฌธ์ œ๋ฅผ ์ผ์œผํ‚ค๋˜ ์„ผ์„œ ๋ถ€์ฐฉ ์˜ค์ฐจ / ์‚ฌ์šฉ์ž์˜ ์† ๋ชจ์–‘ ๋“ฑ์„ ์ž๋™์œผ๋กœ ์ •ํ™•ํžˆ ๋ณด์ •ํ•œ๋‹ค. ์ด ์ œ์•ˆ๋œ ์˜์ƒ-๊ด€์„ฑ ์„ผ์„œ ์œตํ•ฉ ๊ธฐ์ˆ  (Visual-Inertial Skeleton Tracking (VIST)) ์˜ ๋›ฐ์–ด๋‚œ ์„ฑ๋Šฅ๊ณผ ๊ฐ•๊ฑด์„ฑ์ด ๋‹ค์–‘ํ•œ ์ •๋Ÿ‰/์ •์„ฑ ์‹คํ—˜์„ ํ†ตํ•ด ๊ฒ€์ฆ๋˜์—ˆ๊ณ , ์ด๋Š” VIST์˜ ๋‹ค์–‘ํ•œ ์ผ์ƒํ™˜๊ฒฝ์—์„œ ๊ธฐ์กด ์‹œ์Šคํ…œ์ด ๊ตฌํ˜„ํ•˜์ง€ ๋ชปํ•˜๋˜ ์† ๋™์ž‘ ์ถ”์ ์„ ๊ฐ€๋Šฅ์ผ€ ํ•จ์œผ๋กœ์จ, ๋งŽ์€ ์ธ๊ฐ„-๊ธฐ๊ณ„ ์ƒํ˜ธ์ž‘์šฉ ๋ถ„์•ผ์—์„œ์˜ ๊ฐ€๋Šฅ์„ฑ์„ ๋ณด์—ฌ์ค€๋‹ค.1 Introduction 1 1.1. Motivation 1 1.2. Related Work 5 1.3. Contribution 12 2 Detection Threshold of Hand Tracking Error 16 2.1. Motivation 16 2.2. Experimental Environment 20 2.2.1. Hardware Setup 21 2.2.2. Virtual Environment Rendering 23 2.2.3. HMD Calibration 23 2.3. Identifying the Detection Threshold of Tracking Error 26 2.3.1. Experimental Setup 27 2.3.2. Procedure 27 2.3.3. Experimental Result 31 2.4. Enlarging the Detection Threshold of Tracking Error by Haptic Feedback 31 2.4.1. Experimental Setup 31 2.4.2. Procedure 32 2.4.3. Experimental Result 34 2.5. Discussion 34 3 Wearable Finger Tracking Module for Haptic Interaction 38 3.1. Motivation 38 3.2. Development of Finger Tracking Module 42 3.2.1. Hardware Setup 42 3.2.2. Tracking algorithm 45 3.2.3. Calibration method 48 3.3. Evaluation for VR Haptic Interaction Task 50 3.3.1. Quantitative evaluation of FTM 50 3.3.2. Implementation of Wearable Cutaneous Haptic Interface 51 3.3.3. Usability evaluation for VR peg-in-hole task 53 3.4. Discussion 57 4 Visual-Inertial Skeleton Tracking for Human Hand 59 4.1. Motivation 59 4.2. Hardware Setup and Hand Models 62 4.2.1. Human Hand Model 62 4.2.2. Wearable Sensor Glove 62 4.2.3. Stereo Camera 66 4.3. Visual Information Extraction 66 4.3.1. Marker Detection in Raw Images 68 4.3.2. Cost Function for Point Matching 68 4.3.3. Left-Right Stereo Matching 69 4.4. IMU-Aided Correspondence Search 72 4.5. Filtering-based Visual-Inertial Sensor Fusion 76 4.5.1. EKF States for Hand Tracking and Auto-Calibration 78 4.5.2. Prediction with IMU Information 79 4.5.3. Correction with Visual Information 82 4.5.4. Correction with Anatomical Constraints 84 4.6. Quantitative Evaluation for Free Hand Motion 87 4.6.1. Experimental Setup 87 4.6.2. Procedure 88 4.6.3. Experimental Result 90 4.7. Quantitative and Comparative Evaluation for Challenging Hand Motion 95 4.7.1. Experimental Setup 95 4.7.2. Procedure 96 4.7.3. Experimental Result 98 4.7.4. Performance Comparison with Existing Methods for Challenging Hand Motion 101 4.8. Qualitative Evaluation for Real-World Scenarios 105 4.8.1. Visually Complex Background 105 4.8.2. Object Interaction 106 4.8.3. Wearing Fingertip Cutaneous Haptic Devices 109 4.8.4. Outdoor Environment 111 4.9. Discussion 112 5 Conclusion 116 References 124 Abstract (in Korean) 139 Acknowledgment 141๋ฐ•

    Hand Pose Estimation with Mems-Ultrasonic Sensors

    Full text link
    Hand tracking is an important aspect of human-computer interaction and has a wide range of applications in extended reality devices. However, current hand motion capture methods suffer from various limitations. For instance, visual-based hand pose estimation is susceptible to self-occlusion and changes in lighting conditions, while IMU-based tracking gloves experience significant drift and are not resistant to external magnetic field interference. To address these issues, we propose a novel and low-cost hand-tracking glove that utilizes several MEMS-ultrasonic sensors attached to the fingers, to measure the distance matrix among the sensors. Our lightweight deep network then reconstructs the hand pose from the distance matrix. Our experimental results demonstrate that this approach is both accurate, size-agnostic, and robust to external interference. We also show the design logic for the sensor selection, sensor configurations, circuit diagram, as well as model architecture

    Review of Wearable Devices and Data Collection Considerations for Connected Health

    Get PDF
    Wearable sensor technology has gradually extended its usability into a wide range of well-known applications. Wearable sensors can typically assess and quantify the wearerโ€™s physiology and are commonly employed for human activity detection and quantified self-assessment. Wearable sensors are increasingly utilised to monitor patient health, rapidly assist with disease diagnosis, and help predict and often improve patient outcomes. Clinicians use various self-report questionnaires and well-known tests to report patient symptoms and assess their functional ability. These assessments are time consuming and costly and depend on subjective patient recall. Moreover, measurements may not accurately demonstrate the patientโ€™s functional ability whilst at home. Wearable sensors can be used to detect and quantify specific movements in different applications. The volume of data collected by wearable sensors during long-term assessment of ambulatory movement can become immense in tuple size. This paper discusses current techniques used to track and record various human body movements, as well as techniques used to measure activity and sleep from long-term data collected by wearable technology devices
    • โ€ฆ
    corecore