105 research outputs found

    Articulated Pose Estimation by a Graphical Model with Image Dependent Pairwise Relations

    Full text link
    We present a method for estimating articulated human pose from a single static image based on a graphical model with novel pairwise relations that make adaptive use of local image measurements. More precisely, we specify a graphical model for human pose which exploits the fact the local image measurements can be used both to detect parts (or joints) and also to predict the spatial relationships between them (Image Dependent Pairwise Relations). These spatial relationships are represented by a mixture model. We use Deep Convolutional Neural Networks (DCNNs) to learn conditional probabilities for the presence of parts and their spatial relationships within image patches. Hence our model combines the representational flexibility of graphical models with the efficiency and statistical power of DCNNs. Our method significantly outperforms the state of the art methods on the LSP and FLIC datasets and also performs very well on the Buffy dataset without any training.Comment: NIPS 2014 Camera Read

    Detect What You Can: Detecting and Representing Objects using Holistic Models and Body Parts

    Get PDF
    Detecting objects becomes difficult when we need to deal with large shape deformation, occlusion and low resolution. We propose a novel approach to i) handle large deformations and partial occlusions in animals (as examples of highly deformable objects), ii) describe them in terms of body parts, and iii) detect them when their body parts are hard to detect (e.g., animals depicted at low resolution). We represent the holistic object and body parts separately and use a fully connected model to arrange templates for the holistic object and body parts. Our model automatically decouples the holistic object or body parts from the model when they are hard to detect. This enables us to represent a large number of holistic object and body part combinations to better deal with different “detectability” patterns caused by deformations, occlusion and/or low resolution. We apply our method to the six animal categories in the PASCAL VOC dataset and show that our method significantly improves state-of-the-art (by 4.1% AP) and provides a richer representation for objects. During training we use annotations for body parts (e.g., head, torso, etc), making use of a new dataset of fully annotated object parts for PASCAL VOC 2010, which provides a mask for each part.This material is based upon work supported by the Center for Minds, Brains and Machines (CBMM), funded by NSF STC award CCF-1231216

    Uncertainty-informed Mutual Learning for Joint Medical Image Classification and Segmentation

    Full text link
    Classification and segmentation are crucial in medical image analysis as they enable accurate diagnosis and disease monitoring. However, current methods often prioritize the mutual learning features and shared model parameters, while neglecting the reliability of features and performances. In this paper, we propose a novel Uncertainty-informed Mutual Learning (UML) framework for reliable and interpretable medical image analysis. Our UML introduces reliability to joint classification and segmentation tasks, leveraging mutual learning with uncertainty to improve performance. To achieve this, we first use evidential deep learning to provide image-level and pixel-wise confidences. Then, an Uncertainty Navigator Decoder is constructed for better using mutual features and generating segmentation results. Besides, an Uncertainty Instructor is proposed to screen reliable masks for classification. Overall, UML could produce confidence estimation in features and performance for each link (classification and segmentation). The experiments on the public datasets demonstrate that our UML outperforms existing methods in terms of both accuracy and robustness. Our UML has the potential to explore the development of more reliable and explainable medical image analysis models. We will release the codes for reproduction after acceptance.Comment: 13 page

    A SVM-based method for identifying fracture modes of rock using WVD spectrogram features of AE signals

    Get PDF
    In order to achieve the highly efficient and accurate identification of fracture modes including tension or shear fractures during rock failure, an intelligent identification method based on Wigner-Ville distribution (WVD) spectrogram features of acoustic emission (AE) signals was proposed. This method was mainly constructed by the following steps: Firstly, AE hits corre-sponding to tension and shear fractures were obtained through conducting the Brazilian disc test (tension fracture) and direct shear test (shear fracture) of limestone. Secondly, the WVD spectro-grams of these tensile-type and shear-type AE hits were respectively extracted and then trans-formed into the image features of relatively low-dimension as the sample set based on the gray-level cooccurrence matrix (GLCM) and histogram of oriented gradient (HOG). Finally, on the basis of the processed and classified sample set of the WVD spectrogram features, an identifica-tion model of rock fracture modes was established by a support vector machine (SVM) learning algorithm. To verify this method, the fracture modes of limestone subjected to biaxial compres-sion were identified by the method. The results showed that the method not only can greatly re-veal the fracture modes change from tension-dominated to shear-dominated fractures, but also has advantages over the RA-AF value method, such as applicability, accuracy and practicality
    corecore