4 research outputs found

    Data Mining Feature Subset Weighting and Selection Using Genetic Algorithms

    Get PDF
    We present a simple genetic algorithm (sGA), which is developed under Genetic Rule and Classifier Construction Environment (GRaCCE) to solve feature subset selection and weighting problem to have better classification accuracy on k-nearest neighborhood (KNN) algorithm. Our hypotheses are that weighting the features will affect the performance of the KNN algorithm and will cause better classification accuracy rate than that of binary classification. The weighted-sGA algorithm uses real-value chromosomes to find the weights for features and binary-sGA uses integer-value chromosomes to select the subset of features from original feature set. A Repair algorithm is developed for weighted-sGA algorithm to guarantee the feasibility of chromosomes. By feasibility we mean that the sum of values of each gene in a chromosome must be equal to 1. To calculate the fitness values for each chromosome in the population, we use K Nearest Neighbor Algorithm (KNN) as our fitness function. The Euclidean distance from one individual to other individuals is calculated on the d-dimensional feature space to classify an unknown instance. GRaCCE searches for good feature subsets and their associated weights. These feature weights are then multiplied with normalized feature values and these new values are used to calculate the distance between features

    Flexibility and accuracy enhancement techniques for neural networks

    Get PDF
    Master'sMASTER OF ENGINEERIN

    Performing Effective Feature Selection by Investigating the Deep Structure of the Data

    No full text
    This paper introduces ADHOC (Automatic Discov-erer of Higher-Order Correlation), an algorithm that combines the advantages of both filter and feedback models to enhance the understanding of the given data and to increase the efficiency of the feature se-lection process. ADHOC partitions the observed features into a number of groups, called factors, that reflect the major dimensions of the phenomenon un-der consideration. The set of learned factors define the starting point of the search of the best performing feature subset. A genetic algorithm is used to explore the feature space originated by the factors and to de-termine the set of most informative feature configu-rations. The feature subset evaluation function is the performance of the induction algorithm. This ap-proach offers three main advantages: (i) the likeli-hood of selecting good performing features grows; (ii) the complexity of search diminishes consistently; (iii) the possibility of selecting a bad feature subset due to over-fitting problems decreases. Extensive ex-periments on real-world data have been conducted to demonstrate the effectiveness of ADHOC as data re-duction technique as well as feature selection method
    corecore