327 research outputs found

    Color-to-Grayscale: Does the Method Matter in Image Recognition?

    Get PDF
    In image recognition it is often assumed the method used to convert color images to grayscale has little impact on recognition performance. We compare thirteen different grayscale algorithms with four types of image descriptors and demonstrate that this assumption is wrong: not all color-to-grayscale algorithms work equally well, even when using descriptors that are robust to changes in illumination. These methods are tested using a modern descriptor-based image recognition framework, on face, object, and texture datasets, with relatively few training instances. We identify a simple method that generally works best for face and object recognition, and two that work well for recognizing textures

    Background suppressing Gabor energy filtering

    Get PDF
    In the field of facial emotion recognition, early research advanced with the use of Gabor filters. However, these filters lack generalization and result in undesirably large feature vector size. In recent work, more attention has been given to other local appearance features. Two desired characteristics in a facial appearance feature are generalization capability, and the compactness of representation. In this paper, we propose a novel texture feature inspired by Gabor energy filters, called background suppressing Gabor energy filtering. The feature has a generalization component that removes background texture. It has a reduced feature vector size due to maximal representation and soft orientation histograms, and it is awhite box representation. We demonstrate improved performance on the non-trivial Audio/Visual Emotion Challenge 2012 grand-challenge dataset by a factor of 7.17 over the Gabor filter on the development set. We also demonstrate applicability of our approach beyond facial emotion recognition which yields improved classification rate over the Gabor filter for four bioimaging datasets by an average of 8.22%

    Clinical feasibility of quantitative ultrasound texture analysis: A robustness study using fetal lung ultrasound images

    Get PDF
    OBJECTIVES: To compare the robustness of several methods based on quantitative ultrasound (US) texture analysis to evaluate its feasibility for extracting features from US images to use as a clinical diagnostic tool. METHODS: We compared, ranked, and validated the robustness of 5 texture-based methods for extracting textural features from US images acquired under different conditions. For comparison and ranking purposes, we used 13,171 non-US images from widely known available databases (OUTEX [University of Oulu, Oulu, Finland] and PHOTEX [Texture Lab, Heriot-Watt University, Edinburgh, Scotland]), which were specifically acquired under different controlled parameters (illumination, resolution, and rotation) from 103 textures. The robustness of those methods with better results from the non-US images was validated by using 666 fetal lung US images acquired from singleton pregnancies. In this study, 2 similarity measurements (correlation and Chebyshev distances) were used to evaluate the repeatability of the features extracted from the same tissue images. RESULTS: Three of the 5 methods (gray-level co-occurrence matrix, local binary patterns, and rotation-invariant local phase quantization) had favorably robust performance when using the non-US database. In fact, these methods showed similarity values close to 0 for the acquisition variations and delineations. Results from the US database confirmed robustness for all of the evaluated methods (gray-level co-occurrence matrix, local binary patterns, and rotation-invariant local phase quantization) when comparing the same texture obtained from different regions of the image (proximal/distal lungs and US machine brand stratification). CONCLUSIONS: Our results confirmed that texture analysis can be robust (high similarity for different condition acquisitions) with potential to be included as a clinical tool

    A Structural Based Feature Extraction for Detecting the Relation of Hidden Substructures in Coral Reef Images

    Get PDF
    In this paper, we present an efficient approach to extract local structural color texture features for classifying coral reef images. Two local texture descriptors are derived from this approach. The first one, based on Median Robust Extended Local Binary Pattern (MRELBP), is called Color MRELBP (CMRELBP). CMRELBP is very accurate and can capture the structural information from color texture images. To reduce the dimensionality of the feature vector, the second descriptor, co-occurrence CMRELBP (CCMRELBP) is introduced. It is constructed by applying the Integrative Co-occurrence Matrix (ICM) on the Color MRELBP images. This way we can detect and extract the relative relations between structural texture patterns. Moreover, we propose a multiscale LBP based approach with these two schemes to capture microstructure and macrostructure texture information. The experimental results on coral reef (EILAT, EILAT2, RSMAS, and MLC) and four well-known texture datasets (OUTEX, KTH-TIPS, CURET, and UIUCTEX) show that the proposed scheme is quite effective in designing an accurate, robust to noise, rotation and illumination invariant texture classification system. Moreover, it makes an admissible tradeoff between accuracy and number of features

    Binary Patterns Encoded Convolutional Neural Networks for Texture Recognition and Remote Sensing Scene Classification

    Full text link
    Designing discriminative powerful texture features robust to realistic imaging conditions is a challenging computer vision problem with many applications, including material recognition and analysis of satellite or aerial imagery. In the past, most texture description approaches were based on dense orderless statistical distribution of local features. However, most recent approaches to texture recognition and remote sensing scene classification are based on Convolutional Neural Networks (CNNs). The d facto practice when learning these CNN models is to use RGB patches as input with training performed on large amounts of labeled data (ImageNet). In this paper, we show that Binary Patterns encoded CNN models, codenamed TEX-Nets, trained using mapped coded images with explicit texture information provide complementary information to the standard RGB deep models. Additionally, two deep architectures, namely early and late fusion, are investigated to combine the texture and color information. To the best of our knowledge, we are the first to investigate Binary Patterns encoded CNNs and different deep network fusion architectures for texture recognition and remote sensing scene classification. We perform comprehensive experiments on four texture recognition datasets and four remote sensing scene classification benchmarks: UC-Merced with 21 scene categories, WHU-RS19 with 19 scene classes, RSSCN7 with 7 categories and the recently introduced large scale aerial image dataset (AID) with 30 aerial scene types. We demonstrate that TEX-Nets provide complementary information to standard RGB deep model of the same network architecture. Our late fusion TEX-Net architecture always improves the overall performance compared to the standard RGB network on both recognition problems. Our final combination outperforms the state-of-the-art without employing fine-tuning or ensemble of RGB network architectures.Comment: To appear in ISPRS Journal of Photogrammetry and Remote Sensin

    Multi-Directional Multi-Level Dual-Cross Patterns for Robust Face Recognition

    Full text link
    © 1979-2012 IEEE. To perform unconstrained face recognition robust to variations in illumination, pose and expression, this paper presents a new scheme to extract 'Multi-Directional Multi-Level Dual-Cross Patterns' (MDML-DCPs) from face images. Specifically, the MDML-DCPs scheme exploits the first derivative of Gaussian operator to reduce the impact of differences in illumination and then computes the DCP feature at both the holistic and component levels. DCP is a novel face image descriptor inspired by the unique textural structure of human faces. It is computationally efficient and only doubles the cost of computing local binary patterns, yet is extremely robust to pose and expression variations. MDML-DCPs comprehensively yet efficiently encodes the invariant characteristics of a face image from multiple levels into patterns that are highly discriminative of inter-personal differences but robust to intra-personal variations. Experimental results on the FERET, CAS-PERL-R1, FRGC 2.0, and LFW databases indicate that DCP outperforms the state-of-the-art local descriptors (e.g., LBP, LTP, LPQ, POEM, tLBP, and LGXP) for both face identification and face verification tasks. More impressively, the best performance is achieved on the challenging LFW and FRGC 2.0 databases by deploying MDML-DCPs in a simple recognition scheme

    Convolutional Neural Network on Three Orthogonal Planes for Dynamic Texture Classification

    Get PDF
    Dynamic Textures (DTs) are sequences of images of moving scenes that exhibit certain stationarity properties in time such as smoke, vegetation and fire. The analysis of DT is important for recognition, segmentation, synthesis or retrieval for a range of applications including surveillance, medical imaging and remote sensing. Deep learning methods have shown impressive results and are now the new state of the art for a wide range of computer vision tasks including image and video recognition and segmentation. In particular, Convolutional Neural Networks (CNNs) have recently proven to be well suited for texture analysis with a design similar to a filter bank approach. In this paper, we develop a new approach to DT analysis based on a CNN method applied on three orthogonal planes x y , xt and y t . We train CNNs on spatial frames and temporal slices extracted from the DT sequences and combine their outputs to obtain a competitive DT classifier. Our results on a wide range of commonly used DT classification benchmark datasets prove the robustness of our approach. Significant improvement of the state of the art is shown on the larger datasets.Comment: 19 pages, 10 figure
    • 

    corecore