190 research outputs found

    Development of Kinectᵀᴿ applications for assembly simulation and ergonomic analysis

    Get PDF
    Marker-less motion capture technology has been harnessed for several years to track human movements for developing various applications. Recently, with the launch of Microsoft Kinect, researchers have been keenly interested in developing applications using this device. Since Kinect is very inexpensive (only $110 at the time of writing this thesis), it is a low-cost and a promising substitute for the comparatively expensive marker-based motion capture systems. Though it is principally designed for home entertainment, numerous applications can be developed with the capabilities of Kinect. The skeleton data of a human being tracked by a single Kinect device is enough to simulate the human movements, in some cases. However, it is highly desirable to develop a multiple Kinect system to enhance the tracking volume and to address an issue of occlusions. This thesis presents a novel approach for addressing the issue of interference of infrared light patterns while using multiple Kinect devices for human motion capture without lowering the frame rate. This research also presents a software solution to obtain skeleton data from multiple Kinect devices using Kinect for Windows SDK. It also discusses the development of an application involving auto scaling of a human model in digital human modeling software by Siemens Jack and human motion simulation using skeleton tracking data from Kinect to assist the industries with a flexible tool for ergonomic analysis. Further, the capability of this application for obtaining assembly simulations of fastening operations on an aircraft fuselage is also presented. --Abstract, page iii

    Real-time Body Tracking and Projection Mapping in the Interactive Arts

    Get PDF
    Projection mapping, a subtopic of augmented reality, displays computer-generated light visualizations from projectors onto the real environment. A challenge for projection mapping in performing interactive arts is dynamic body movements. Accuracy and speed are key components for an immersive application of body projection mapping and dependent on scanning and processing time. This thesis presents a novel technique to achieve real-time body projection mapping utilizing a state of the art body tracking device, Microsoft’s Azure Kinect DK, by using an array of trackers for error minimization and movement prediction. The device\u27s Sensor and Bodytracking SDKs allow multiple device synchronization. We combine our tracking results from this feature with motion prediction to provide an accurate approximation for body joint tracking. Using the new joint approximations and the depth information from the Kinect, we create a silhouette and map textures and animations to it before projecting it back onto the user. Our implementation of gesture detection provides interaction between the user and the projected images. Our results decreased the lag time created from the devices, code, and projector to create a realistic real-time body projection mapping. Our end goal was to display it in an art show. This thesis was presented at Burning Man 2019 and Delfines de San Carlos 2020 as interactive art installations

    A mixed reality telepresence system for collaborative space operation

    Get PDF
    This paper presents a Mixed Reality system that results from the integration of a telepresence system and an application to improve collaborative space exploration. The system combines free viewpoint video with immersive projection technology to support non-verbal communication, including eye gaze, inter-personal distance and facial expression. Importantly, these can be interpreted together as people move around the simulation, maintaining natural social distance. The application is a simulation of Mars, within which the collaborators must come to agreement over, for example, where the Rover should land and go. The first contribution is the creation of a Mixed Reality system supporting contextualization of non-verbal communication. Tw technological contributions are prototyping a technique to subtract a person from a background that may contain physical objects and/or moving images, and a light weight texturing method for multi-view rendering which provides balance in terms of visual and temporal quality. A practical contribution is the demonstration of pragmatic approaches to sharing space between display systems of distinct levels of immersion. A research tool contribution is a system that allows comparison of conventional authored and video based reconstructed avatars, within an environment that encourages exploration and social interaction. Aspects of system quality, including the communication of facial expression and end-to-end latency are reported

    Context-aware gestural interaction in the smart environments of the ubiquitous computing era

    Get PDF
    A thesis submitted to the University of Bedfordshire in partial fulfilment of the requirements for the degree of Doctor of PhilosophyTechnology is becoming pervasive and the current interfaces are not adequate for the interaction with the smart environments of the ubiquitous computing era. Recently, researchers have started to address this issue introducing the concept of natural user interface, which is mainly based on gestural interactions. Many issues are still open in this emerging domain and, in particular, there is a lack of common guidelines for coherent implementation of gestural interfaces. This research investigates gestural interactions between humans and smart environments. It proposes a novel framework for the high-level organization of the context information. The framework is conceived to provide the support for a novel approach using functional gestures to reduce the gesture ambiguity and the number of gestures in taxonomies and improve the usability. In order to validate this framework, a proof-of-concept has been developed. A prototype has been developed by implementing a novel method for the view-invariant recognition of deictic and dynamic gestures. Tests have been conducted to assess the gesture recognition accuracy and the usability of the interfaces developed following the proposed framework. The results show that the method provides optimal gesture recognition from very different view-points whilst the usability tests have yielded high scores. Further investigation on the context information has been performed tackling the problem of user status. It is intended as human activity and a technique based on an innovative application of electromyography is proposed. The tests show that the proposed technique has achieved good activity recognition accuracy. The context is treated also as system status. In ubiquitous computing, the system can adopt different paradigms: wearable, environmental and pervasive. A novel paradigm, called synergistic paradigm, is presented combining the advantages of the wearable and environmental paradigms. Moreover, it augments the interaction possibilities of the user and ensures better gesture recognition accuracy than with the other paradigms

    Computational Multimedia for Video Self Modeling

    Get PDF
    Video self modeling (VSM) is a behavioral intervention technique in which a learner models a target behavior by watching a video of oneself. This is the idea behind the psychological theory of self-efficacy - you can learn or model to perform certain tasks because you see yourself doing it, which provides the most ideal form of behavior modeling. The effectiveness of VSM has been demonstrated for many different types of disabilities and behavioral problems ranging from stuttering, inappropriate social behaviors, autism, selective mutism to sports training. However, there is an inherent difficulty associated with the production of VSM material. Prolonged and persistent video recording is required to capture the rare, if not existed at all, snippets that can be used to string together in forming novel video sequences of the target skill. To solve this problem, in this dissertation, we use computational multimedia techniques to facilitate the creation of synthetic visual content for self-modeling that can be used by a learner and his/her therapist with a minimum amount of training data. There are three major technical contributions in my research. First, I developed an Adaptive Video Re-sampling algorithm to synthesize realistic lip-synchronized video with minimal motion jitter. Second, to denoise and complete the depth map captured by structure-light sensing systems, I introduced a layer based probabilistic model to account for various types of uncertainties in the depth measurement. Third, I developed a simple and robust bundle-adjustment based framework for calibrating a network of multiple wide baseline RGB and depth cameras

    Enhancing the measurement of clinical outcomes using Microsoft Kinect

    Get PDF
    There is a growing body of applications leveraging Microsoft Kinect and the associated Windows Software Development Kit in health and wellness. In particular, this platform has been valuable in developing interactive solutions for rehabilitation including creating more engaging exercise regimens and ensuring that exercises are performed correctly for optimal outcomes. Clinical trials rely upon robust and validated methodologies to measure health status and to detect treatment-related changes over time to enable the efficacy and safety of new drug treatments to be assessed and measured. In many therapeutic areas, traditional outcome measures rely on subjective investigator and patient ratings. Subjective ratings are not always sensitive to detecting small improvements, are subject to inter- and intra-rater variability and limited in their ability to record detailed or subtle aspects of movement and mobility. For these reasons, objective measurements may provide greater sensitivity to detect treatment-related changes where they exist. In this review paper, we explore the use of the Kinect platform to develop low-cost approaches to objectively measure aspects of movement. We consider published applications that measure aspects of gait and balance, upper extremity movement, chest wall motion and facial analysis. In each case, we explore the utility of the approach for clinical trials, and the precision and accuracy of estimates derived from the Kinect output. We conclude that the use of games platforms such as Microsoft Kinect to measure clinical outcomes offer a versatile, easy to use and low-cost approach that may add significant value and utility to clinical drug development, in particular in replacing conventional subjective measures and providing richer information about movement than previously possible in large scale clinical trials, especially in the measurement of gross spatial movements. Regulatory acceptance of clinical outcomes collected in this way will be subject to comprehensive assessment of validity and clinical relevance, and this will require good quality peer-reviewed publications of scientific evidence

    Automatic and adaptable registration of live RGBD video streams sharing partially overlapping views

    Get PDF
    In this thesis, we introduce DeReEs-4v, an algorithm for unsupervised and automatic registration of two video frames captured depth-sensing cameras. DeReEs-4V receives two RGBD video streams from two depth-sensing cameras arbitrary located in an indoor space that share a minimum amount of 25% overlap between their captured scenes. The motivation of this research is to employ multiple depth-sensing cameras to enlarge the field of view and acquire a more complete and accurate 3D information of the environment. A typical way to combine multiple views from different cameras is through manual calibration. However, this process is time-consuming and may require some technical knowledge. Moreover, calibration has to be repeated when the location or position of the cameras change. In this research, we demonstrate how DeReEs-4V registration can be used to find the transformation of the view of one camera with respect to the other at interactive rates. Our algorithm automatically finds the 3D transformation to match the views from two cameras, requires no human interference, and is robust to camera movements while capturing. To validate this approach, a thorough examination of the system performance under different scenarios is presented. The system presented here supports any application that might benefit from the wider field-of-view provided by the combined scene from both cameras, including applications in 3D telepresence, gaming, people tracking, videoconferencing and computer vision

    A Low-cost, Wearable Opto-inertial 6 DOF Hand Pose Tracking System for VR

    Get PDF
    In this paper, a low cost, wearable six Degree of Freedom (6-DOF) hand pose tracking system is proposed for Virtual Reality applications. It is designed for use with an integrated hand exoskeleton system for kinesthetic haptic feedback. The tracking system consists of an Infrared (IR) based optical tracker with low cost mono-camera and inertial and magnetic measurement unit. Image processing is done on LabVIEW software to extract the 3-DOF position from two IR targets and Magdwick filter has been implemented on Mbed LPC1768 board to obtain orientation data. Six DOF hand tracking outputs filtered and synchronized on LabVIEW software are then sent to the Unity Virtual environment via User Datagram Protocol (UDP) stream. Experimental results show that this low cost and compact system has a comparable performance of minimal Jitter with position and orientation Root Mean Square Error (RMSE) of less than 0.2 mm and 0.15 degrees, respectively. Total Latency of the system is also less than 40 ms
    • …
    corecore