88 research outputs found

    Encoding Feature Maps of CNNs for Action Recognition

    Get PDF
    CVPR International Workshop and Competition on Action Recognition with a Large Number of ClassesWe describe our approach for action classification in the THUMOS Challenge 2015. Our approach is based on two types of features, improved dense trajectories and CNN features. For trajectory features, we extract HOG, HOF, MBHx, and MBHy descriptors and apply Fisher vector encoding. For CNN features, we utilize a recent deep CNN model, VGG19, to capture appearance features and use VLAD encoding to encode/pool convolutional feature maps which shows better performance than average pooling of feature maps and full-connected activation features. After concatenating them, we train a linear SVM classifier for each class in a one-vs-all scheme

    Self-Ensemling for 3D Point Cloud Domain Adaption

    Full text link
    Recently 3D point cloud learning has been a hot topic in computer vision and autonomous driving. Due to the fact that it is difficult to manually annotate a qualitative large-scale 3D point cloud dataset, unsupervised domain adaptation (UDA) is popular in 3D point cloud learning which aims to transfer the learned knowledge from the labeled source domain to the unlabeled target domain. However, the generalization and reconstruction errors caused by domain shift with simply-learned model are inevitable which substantially hinder the model's capability from learning good representations. To address these issues, we propose an end-to-end self-ensembling network (SEN) for 3D point cloud domain adaption tasks. Generally, our SEN resorts to the advantages of Mean Teacher and semi-supervised learning, and introduces a soft classification loss and a consistency loss, aiming to achieve consistent generalization and accurate reconstruction. In SEN, a student network is kept in a collaborative manner with supervised learning and self-supervised learning, and a teacher network conducts temporal consistency to learn useful representations and ensure the quality of point clouds reconstruction. Extensive experiments on several 3D point cloud UDA benchmarks show that our SEN outperforms the state-of-the-art methods on both classification and segmentation tasks. Moreover, further analysis demonstrates that our SEN also achieves better reconstruction results

    AU-Supervised Convolutional Vision Transformers for Synthetic Facial Expression Recognition

    Full text link
    The paper describes our proposed methodology for the six basic expression classification track of Affective Behavior Analysis in-the-wild (ABAW) Competition 2022. In Learing from Synthetic Data(LSD) task, facial expression recognition (FER) methods aim to learn the representation of expression from the artificially generated data and generalise to real data. Because of the ambiguous of the synthetic data and the objectivity of the facial Action Unit (AU), we resort to the AU information for performance boosting, and make contributions as follows. First, to adapt the model to synthetic scenarios, we use the knowledge from pre-trained large-scale face recognition data. Second, we propose a conceptually-new framework, termed as AU-Supervised Convolutional Vision Transformers (AU-CVT), which clearly improves the performance of FER by jointly training auxiliary datasets with AU or pseudo AU labels. Our AU-CVT achieved F1 score as 0.68630.6863, accuracy as 0.74330.7433 on the validation set. The source code of our work is publicly available online: https://github.com/msy1412/ABAW

    A fully asymptotic preserving decomposed multi-group method for the frequency-dependent radiative transfer equations

    Full text link
    The opacity of FRTE depends on not only the material temperature but also the frequency, whose values may vary several orders of magnitude for different frequencies. The gray radiation diffusion and frequency-dependent diffusion equations are two simplified models that can approximate the solution to FRTE in the thick opacity regime. The frequency discretization for the two limit models highly affects the numerical accuracy. However, classical frequency discretization for FRTE considers only the absorbing coefficient. In this paper, we propose a new decomposed multi-group method for frequency discretization that is not only AP in both gray radiation diffusion and frequency-dependent diffusion limits, but also the frequency discretization of the limiting models can be tuned. Based on the decomposed multi-group method, a full AP scheme in frequency, time, and space is proposed. Several numerical examples are used to verify the performance of the proposed scheme.Comment: 36 pages, 14 figure
    corecore