333,643 research outputs found

    A new framework of human interaction recognition based on multiple stage probability fusion

    Get PDF
    Visual-based human interactive behavior recognition is a challenging research topic in computer vision. There exist some important problems in the current interaction recognition algorithms, such as very complex feature representation and inaccurate feature extraction induced by wrong human body segmentation. In order to solve these problems, a novel human interaction recognition method based on multiple stage probability fusion is proposed in this paper. According to the human body’s contact in interaction as a cut-off point, the process of the interaction can be divided into three stages: start stage, execution stage and end stage. Two persons’ motions are respectively extracted and recognizes in the start stage and the finish stage when there is no contact between those persons. The two persons’ motion is extracted as a whole and recognized in the execution stage. In the recognition process, the final recognition results are obtained by the weighted fusing these probabilities in different stages. The proposed method not only simplifies the extraction and representation of features, but also avoids the wrong feature extraction caused by occlusion. Experiment results on the UT-interaction dataset demonstrated that the proposed method results in a better performance than other recent interaction recognition methods

    Does M31 result from an ancient major merger?

    Full text link
    The numerous streams in the M31 halo are currently assumed to be due to multiple minor mergers. Here we use the GADGET2 simulation code to test whether M31 could have experienced a major merger in its past history. It results that a 3+/-0.5:1 gaseous rich merger with r(per)=25+/-5 kpc and a polar orbit can explain many properties of M31 and of its halo. The interaction and the fusion may have begun 8.75+/-0.35 Gyr and 5.5 +/-0.5 Gyr ago, respectively. With an almost quiescent star formation history before the fusion we retrieve fractions of bulge, thin and thick disks as well as relative fractions of intermediate age and old stars in both the thick disk and the Giant Stream. The Giant Stream is caused by returning stars from a tidal tail previously stripped from the satellite prior to the fusion. These returning stars are trapped into elliptical orbits or loops for almost a Hubble time period. Large loops are also predicted and they scale rather well with the recently discovered features in the M31 outskirts. We demonstrate that a single merger could explain first-order (intensity and size), morphological and kinematical properties of the disk, thick disk, bulge and streams in the halo of M31, as well as the distribution of stellar ages, and perhaps metallicities. It challenges scenarios assuming one minor merger per feature in the disk (10 kpc ring) or at the outskirts (numerous streams & thick disk). Further constraints will help to properly evaluate the impact of such a major event to the Local Group.Comment: accepted in Astrophysical Journal, 29 September, 2010 ; proof-edited version; 1st column of Table 3 correcte

    Mining multimodal sequential patterns : a case study on affect detection

    Get PDF
    Temporal data from multimodal interaction such as speech and bio-signals cannot be easily analysed without a preprocessing phase through which some key characteristics of the signals are extracted. Typically, standard statistical signal features such as average values are calculated prior to the analysis and, subsequently, are presented either to a multimodal fusion mechanism or a computational model of the interaction. This paper proposes a feature extraction methodology which is based on frequent sequence mining within and across multiple modalities of user input. The proposed method is applied for the fusion of physiological signals and gameplay information in a game survey dataset. The obtained sequences are analysed and used as predictors of user affect resulting in computational models of equal or higher accuracy compared to the models built on standard statistical features.peer-reviewe

    Deep Affordance-grounded Sensorimotor Object Recognition

    Full text link
    It is well-established by cognitive neuroscience that human perception of objects constitutes a complex process, where object appearance information is combined with evidence about the so-called object "affordances", namely the types of actions that humans typically perform when interacting with them. This fact has recently motivated the "sensorimotor" approach to the challenging task of automatic object recognition, where both information sources are fused to improve robustness. In this work, the aforementioned paradigm is adopted, surpassing current limitations of sensorimotor object recognition research. Specifically, the deep learning paradigm is introduced to the problem for the first time, developing a number of novel neuro-biologically and neuro-physiologically inspired architectures that utilize state-of-the-art neural networks for fusing the available information sources in multiple ways. The proposed methods are evaluated using a large RGB-D corpus, which is specifically collected for the task of sensorimotor object recognition and is made publicly available. Experimental results demonstrate the utility of affordance information to object recognition, achieving an up to 29% relative error reduction by its inclusion.Comment: 9 pages, 7 figures, dataset link included, accepted to CVPR 201

    A Deep Spatio-Temporal Fuzzy Neural Network for Passenger Demand Prediction

    Get PDF
    In spite of its importance, passenger demand prediction is a highly challenging problem, because the demand is simultaneously influenced by the complex interactions among many spatial and temporal factors and other external factors such as weather. To address this problem, we propose a Spatio-TEmporal Fuzzy neural Network (STEF-Net) to accurately predict passenger demands incorporating the complex interactions of all known important factors. We design an end-to-end learning framework with different neural networks modeling different factors. Specifically, we propose to capture spatio-temporal feature interactions via a convolutional long short-term memory network and model external factors via a fuzzy neural network that handles data uncertainty significantly better than deterministic methods. To keep the temporal relations when fusing two networks and emphasize discriminative spatio-temporal feature interactions, we employ a novel feature fusion method with a convolution operation and an attention layer. As far as we know, our work is the first to fuse a deep recurrent neural network and a fuzzy neural network to model complex spatial-temporal feature interactions with additional uncertain input features for predictive learning. Experiments on a large-scale real-world dataset show that our model achieves more than 10% improvement over the state-of-the-art approaches.Comment: https://epubs.siam.org/doi/abs/10.1137/1.9781611975673.1
    corecore