20,899 research outputs found

    Penerapan Teknik Motion Tracking Pada Pembuatan Video Music “ Salah Mencintai ” Lifely Band

    Full text link
    The development of the digital age more rapidly, especially in the making of the music video. A variety of models and design presented in the making of the music video. Analogue to digital era changes also adds to the ease in making music videos. The use of a live shoot and motion tracking is a technique that will be used in the making of this music video.According to the theme of the Band applying Lifely story that was in the music video, will produce an interesting product. The concept of the music video is set in a light and fun to eat. The concept of vintage retro-style story that made this music video is richer. This music video tell me about the experience of personnel in the band reaching lifely adulation, but nothing like what is expected and ultimately wrong love. It is supported by a shooting that is filled with variety. The cinematography technique of shooting is also very important. The addition of motion tracking on this music video will add to the richness of the story's in this music video. Taking pictures using the camera DSLR (Digital Single Lens Reflect) which has the resolution of HD Video (High Difination).Tools used in the production of the music video includes cameras, lighting, slider, and other support tools. Adobe Premiere CC and Adobe After effects CC as well as Adobe Photoshop is software editing and composite on the creation of this music video

    MilliSonic: Pushing the Limits of Acoustic Motion Tracking

    Full text link
    Recent years have seen interest in device tracking and localization using acoustic signals. State-of-the-art acoustic motion tracking systems however do not achieve millimeter accuracy and require large separation between microphones and speakers, and as a result, do not meet the requirements for many VR/AR applications. Further, tracking multiple concurrent acoustic transmissions from VR devices today requires sacrificing accuracy or frame rate. We present MilliSonic, a novel system that pushes the limits of acoustic based motion tracking. Our core contribution is a novel localization algorithm that can provably achieve sub-millimeter 1D tracking accuracy in the presence of multipath, while using only a single beacon with a small 4-microphone array.Further, MilliSonic enables concurrent tracking of up to four smartphones without reducing frame rate or accuracy. Our evaluation shows that MilliSonic achieves 0.7mm median 1D accuracy and a 2.6mm median 3D accuracy for smartphones, which is 5x more accurate than state-of-the-art systems. MilliSonic enables two previously infeasible interaction applications: a) 3D tracking of VR headsets using the smartphone as a beacon and b) fine-grained 3D tracking for the Google Cardboard VR system using a small microphone array

    Vehicle Motion Tracking

    Get PDF
    A number of challenges are associated with vehicles and their operation in the driving environment. As one example, vehicles may be involved in a crash or accident which might injure the vehicle passengers. As another example, road conditions such as potholes, roadwork, or other complications may develop and change over time. It would be useful for both vehicle operators and for entities charged with maintaining road conditions (e.g., a “Department of Transportation”) to have real time, up-to-date knowledge of the issues described above

    A Conceptual Framework for Motion Based Music Applications

    Get PDF
    Imaginary projections are the core of the framework for motion based music applications presented in this paper. Their design depends on the space covered by the motion tracking device, but also on the musical feature involved in the application. They can be considered a very powerful tool because they allow not only to project in the virtual environment the image of a traditional acoustic instrument, but also to express any spatially defined abstract concept. The system pipeline starts from the musical content and, through a geometrical interpretation, arrives to its projection in the physical space. Three case studies involving different motion tracking devices and different musical concepts will be analyzed. The three examined applications have been programmed and already tested by the authors. They aim respectively at musical expressive interaction (Disembodied Voices), tonal music knowledge (Harmonic Walk) and XX century music composition (Hand Composer)

    Survey of Motion Tracking Methods Based on Inertial Sensors: A Focus on Upper Limb Human Motion

    Get PDF
    Motion tracking based on commercial inertial measurements units (IMUs) has been widely studied in the latter years as it is a cost-effective enabling technology for those applications in which motion tracking based on optical technologies is unsuitable. This measurement method has a high impact in human performance assessment and human-robot interaction. IMU motion tracking systems are indeed self-contained and wearable, allowing for long-lasting tracking of the user motion in situated environments. After a survey on IMU-based human tracking, five techniques for motion reconstruction were selected and compared to reconstruct a human arm motion. IMU based estimation was matched against motion tracking based on the Vicon marker-based motion tracking system considered as ground truth. Results show that all but one of the selected models perform similarly (about 35 mm average position estimation error)

    Skeleton Driven Non-rigid Motion Tracking and 3D Reconstruction

    Full text link
    This paper presents a method which can track and 3D reconstruct the non-rigid surface motion of human performance using a moving RGB-D camera. 3D reconstruction of marker-less human performance is a challenging problem due to the large range of articulated motions and considerable non-rigid deformations. Current approaches use local optimization for tracking. These methods need many iterations to converge and may get stuck in local minima during sudden articulated movements. We propose a puppet model-based tracking approach using skeleton prior, which provides a better initialization for tracking articulated movements. The proposed approach uses an aligned puppet model to estimate correct correspondences for human performance capture. We also contribute a synthetic dataset which provides ground truth locations for frame-by-frame geometry and skeleton joints of human subjects. Experimental results show that our approach is more robust when faced with sudden articulated motions, and provides better 3D reconstruction compared to the existing state-of-the-art approaches.Comment: Accepted in DICTA 201

    Specialized CNT-based Sensor Framework for Advanced Motion Tracking

    Get PDF
    In this work, we discuss the design and development of an advanced framework for high-fidelity finger motion tracking based on Specialized Carbon Nanotube (CNT) stretchable sensors developed at our research facilities. Earlier versions of the CNT sensors have been employed in the high-fidelity finger motion tracking Data Glove commercialized by Yamaha, Japan. The framework presented in this paper encompasses our continuing research and development of more advanced CNT-based sensors and the implementation of novel high-fidelity motion tracking products based on them. The CNT sensor production and communication framework components are considered in detail and wireless motion tracking experiments with the developed hardware and software components integrated with the Yamaha Data Glove are reported
    • 

    corecore