143,569 research outputs found

    Revealing Daily Human Activity Pattern using Process Mining Approach

    Get PDF
    In  the  last  few  years,  with  the  emergence  of ambient assisted living, the study of human behavioral pattern took  a  wide  interest  from  research  communities  around  the world. In many literatures, pattern recognition was widely adopted approach to implements in human behavior study from computing perspective. Pattern recognition brings a promising results in terms of accuracy for modeling human behavior. But the problem with this approach is the complexity of knowledge representation  which  formulated  in  mathematical  model.  In turns, a correction by the experts is hardly conducted. In another hand, gathering a graphical insight is  not a trivial task. This paper  investigate the use of process mining technology to gives an alternative to such problems. Process mining is data-driven approach to infer a graphical representation of any kind of process. In terms of human behavior, process can be defined as sequences of activities performed by human on daily basis. From the conducted experiments process mining was shown a potential use to infer a human daily activity pattern in a graphical representation

    Action recognition in depth videos using nonparametric probabilistic graphical models

    Get PDF
    Action recognition involves automatically labelling videos that contain human motion with action classes. It has applications in diverse areas such as smart surveillance, human computer interaction and content retrieval. The recent advent of depth sensing technology that produces depth image sequences has offered opportunities to solve the challenging action recognition problem. The depth images facilitate robust estimation of a human skeleton’s 3D joint positions and a high level action can be inferred from a sequence of these joint positions. A natural way to model a sequence of joint positions is to use a graphical model that describes probabilistic dependencies between the observed joint positions and some hidden state variables. A problem with these models is that the number of hidden states must be fixed a priori even though for many applications this number is not known in advance. This thesis proposes nonparametric variants of graphical models with the number of hidden states automatically inferred from data. The inference is performed in a full Bayesian setting by using the Dirichlet Process as a prior over the model’s infinite dimensional parameter space. This thesis describes three original constructions of nonparametric graphical models that are applied in the classification of actions in depth videos. Firstly, the action classes are represented by a Hidden Markov Model (HMM) with an unbounded number of hidden states. The formulation enables information sharing and discriminative learning of parameters. Secondly, a hierarchical HMM with an unbounded number of actions and poses is used to represent activities. The construction produces a simplified model for activity classification by using logistic regression to capture the relationship between action states and activity labels. Finally, the action classes are modelled by a Hidden Conditional Random Field (HCRF) with the number of intermediate hidden states learned from data. Tractable inference procedures based on Markov Chain Monte Carlo (MCMC) techniques are derived for all these constructions. Experiments with multiple benchmark datasets confirm the efficacy of the proposed approaches for action recognition

    Activity Representation from Video Using Statistical Models on Shape Manifolds

    Get PDF
    Activity recognition from video data is a key computer vision problem with applications in surveillance, elderly care, etc. This problem is associated with modeling a representative shape which contains significant information about the underlying activity. In this dissertation, we represent several approaches for view-invariant activity recognition via modeling shapes on various shape spaces and Riemannian manifolds. The first two parts of this dissertation deal with activity modeling and recognition using tracks of landmark feature points. The motion trajectories of points extracted from objects involved in the activity are used to build deformation shape models for each activity, and these models are used for classification and detection of unusual activities. In the first part of the dissertation, these models are represented by the recovered 3D deformation basis shapes corresponding to the activity using a non-rigid structure from motion formulation. We use a theory for estimating the amount of deformation for these models from the visual data. We study the special case of ground plane activities in detail because of its importance in video surveillance applications. In the second part of the dissertation, we propose to model the activity by learning an affine invariant deformation subspace representation that captures the space of possible body poses associated with the activity. These subspaces can be viewed as points on a Grassmann manifold. We propose several statistical classification models on Grassmann manifold that capture the statistical variations of the shape data while following the intrinsic Riemannian geometry of these manifolds. The last part of this dissertation addresses the problem of recognizing human gestures from silhouette images. We represent a human gesture as a temporal sequence of human poses, each characterized by a contour of the associated human silhouette. The shape of a contour is viewed as a point on the shape space of closed curves and, hence, each gesture is characterized and modeled as a trajectory on this shape space. We utilize the Riemannian geometry of this space to propose a template-based and a graphical-based approaches for modeling these trajectories. The two models are designed in such a way to account for the different invariance requirements in gesture recognition, and also capture the statistical variations associated with the contour data

    A novel approach towards usability studies for visual search tasks in graphical user interface applications using the activity theory approach

    Get PDF
    The field of Human Computer Interaction still strives for a generalized model of visual search tasks (icon search, menu search, text search, label search, search through hypertext and feature recognition). The existing models of visual search, in spite of being impressive, are limited under certain perspectives due to lack of generality. The thesis tries to provide a holistic approach for the modeling of visual search tasks in graphical user interfaces from the Activity Theory (AT) perspective with the aim of rendering a theoretical bridge between HCI and Psychology. A detailed review of literature from the variegated discipline contributing to the study of Visual Search revealed the presence of gray areas, which can be partially addressed by the Activity Theory approach. The case study uses thinking aloud Protocol Analysis technique for analyzing the complex interaction of behavior, cognition and motor action, which manifest in these tasks. The results have been analyzed and possible modifications have been identified. Interestingly, it is observed that Activity Theory can provide substantial theoretical support to aid Usability Testing Techniques

    Evaluation of Wearable Sensor Tag Data Segmentation Approaches for Real Time Activity Classification in Elderly

    Get PDF
    Abstract. The development of human activity monitoring has allowed the creation of multiple applications, among them is the recognition of high falls risk activities of older people for the mitigation of falls occurrences. In this study, we apply a graphical model based classification technique(conditional random field) to evaluate various sliding window based techniques for the real time prediction of activities in older subjects wearing a passive (batteryless) sensor enabled RFID tag. The system achieved maximum overall real time activity prediction accuracy of 95% using a time weighted windowing technique to aggregate contextual information to input sensor data
    corecore