2,787 research outputs found

    ANALYZING BIG DATA WITH DECISION TREES

    Get PDF
    ANALYZING BIG DATA WITH DECISION TREE

    Automatic Construction of Accurate Image Processing using AdaBoost

    Get PDF
    Image processing and recognition technologies are required to solve various problems. We have already proposed the system which automatically constructs image processing with Genetic Programming (GP), Automatic Construction of Tree-structural Image Transformation (ACTIT). However, it is difficult to construct an accurate image processing for all training image sets if they have various characteristics. In this paper, we propose ACTIT-Boost which automatically constructs an accurate image processing by employing Adaptive Boosting (AdaBoost) to ACTIT. It learns training image sets and their areas which are difficultly approximated to target images in particular. We show experimentally that ACTIT-Boost is more effective in comparison with ordinary ACTIT

    The Role of the Superior Order GLCM in the Characterization and Recognition of the Liver Tumors from Ultrasound Images

    Get PDF
    The hepatocellular carcinoma (HCC) is the most frequent malignant liver tumor. It often has a similar visual aspect with the cirrhotic parenchyma on which it evolves and with the benign liver tumors. The golden standard for HCC diagnosis is the needle biopsy, but this is an invasive, dangerous method. We aim to develop computerized,noninvasive techniques for the automatic diagnosis of HCC, based on information obtained from ultrasound images. The texture is an important property of the internal organs tissue, able to provide subtle information about the pathology. We previously defined the textural model of HCC, consisting in the exhaustive set of the relevant textural features, appropriate for HCC characterization and in the specific values of these features. In this work, we analyze the role that the superior order Grey Level Cooccurrence Matrices (GLCM) and the associated parameters have in the improvement of HCC characterization and automatic diagnosis. We also determine the best spatial relations between the pixels that lead to the highest performances, for the third, fifth and seventh order GLCM. The following classes will be considered: HCC, cirrhotic liver parenchyma on which it evolves and benign liver tumors

    Labeling the Features Not the Samples: Efficient Video Classification with Minimal Supervision

    Full text link
    Feature selection is essential for effective visual recognition. We propose an efficient joint classifier learning and feature selection method that discovers sparse, compact representations of input features from a vast sea of candidates, with an almost unsupervised formulation. Our method requires only the following knowledge, which we call the \emph{feature sign}---whether or not a particular feature has on average stronger values over positive samples than over negatives. We show how this can be estimated using as few as a single labeled training sample per class. Then, using these feature signs, we extend an initial supervised learning problem into an (almost) unsupervised clustering formulation that can incorporate new data without requiring ground truth labels. Our method works both as a feature selection mechanism and as a fully competitive classifier. It has important properties, low computational cost and excellent accuracy, especially in difficult cases of very limited training data. We experiment on large-scale recognition in video and show superior speed and performance to established feature selection approaches such as AdaBoost, Lasso, greedy forward-backward selection, and powerful classifiers such as SVM.Comment: arXiv admin note: text overlap with arXiv:1411.771

    Asymmetric bagging and random subspace for support vector machines-based relevance feedback in image retrieval

    Get PDF
    Relevance feedback schemes based on support vector machines (SVM) have been widely used in content-based image retrieval (CBIR). However, the performance of SVM-based relevance feedback is often poor when the number of labeled positive feedback samples is small. This is mainly due to three reasons: 1) an SVM classifier is unstable on a small-sized training set, 2) SVM's optimal hyperplane may be biased when the positive feedback samples are much less than the negative feedback samples, and 3) overfitting happens because the number of feature dimensions is much higher than the size of the training set. In this paper, we develop a mechanism to overcome these problems. To address the first two problems, we propose an asymmetric bagging-based SVM (AB-SVM). For the third problem, we combine the random subspace method and SVM for relevance feedback, which is named random subspace SVM (RS-SVM). Finally, by integrating AB-SVM and RS-SVM, an asymmetric bagging and random subspace SVM (ABRS-SVM) is built to solve these three problems and further improve the relevance feedback performance
    • 

    corecore