1 research outputs found

    (The original publication is available at www.springerlink.com) Analysis of Composite Gestures with a Coherent Probabilistic Graphical Model

    No full text
    Abstract Traditionally, gesture-based interaction in virtual environments is composed of either static, posture-based gesture primitives or temporally analyzed dynamic primitives. However, it would be ideal to incorporate both static and dynamic gestures to fully utilize the potential of gesture-based interaction. To that end, we propose a probabilistic framework that incorporates both static and dynamic gesture primitives. We call these primitives Gesture Words (GWords). Using a probabilistic graphical model (PGM), we integrate these heterogeneous GWords and a high-level language model in a coherent fashion. Composite gestures are represented as stochastic paths through the PGM. A gesture is analyzed by finding the path that maximizes the likelihood on the PGM with respect to the video sequence. To facilitate online computation, we propose a greedy algorithm for performing inference on the PGM. The parameters of the PGM can be learned via three different methods: supervised, unsupervised, and hybrid. We implemented the PGM model for a gesture set of 10 GWords with 6 composite gestures. The experimental results show that the PGM can accurately recognize composite gestures. Key words human computer interaction – gesture recognition – hand postures – vision-based interaction – probabilistic graphical model
    corecore