13 research outputs found
Application of multiobjective genetic programming to the design of robot failure recognition systems
We present an evolutionary approach using multiobjective genetic programming (MOGP) to derive optimal feature extraction preprocessing stages for robot failure detection. This data-driven machine learning method is compared both with conventional (nonevolutionary) classifiers and a set of domain-dependent feature extraction methods. We conclude MOGP is an effective and practical design method for failure recognition systems with enhanced recognition accuracy over conventional classifiers, independent of domain knowledge
The influence of mutation on population dynamics in multiobjective genetic programming
Using multiobjective genetic programming with a complexity objective to overcome tree bloat is usually very successful but can sometimes lead to undesirable collapse of the population to all single-node trees. In this paper we report a detailed examination of why and when collapse occurs. We have used different types of crossover and mutation operators (depth-fair and sub-tree), different evolutionary approaches (generational and steady-state), and different datasets (6-parity Boolean and a range of benchmark machine learning problems) to strengthen our conclusion. We conclude that mutation has a vital role in preventing population collapse by counterbalancing parsimony pressure and preserving population diversity. Also, mutation controls the size of the generated individuals which tends to dominate the time needed for fitness evaluation and therefore the whole evolutionary process. Further, the average size of the individuals in a GP population depends on the evolutionary approach employed. We also demonstrate that mutation has a wider role than merely culling single-node individuals from the population; even within a diversity-preserving algorithm such as SPEA2 mutation has a role in preserving diversity
A generic optimising feature extraction method using multiobjective genetic programming
In this paper, we present a generic, optimising feature extraction method using multiobjective genetic programming. We re-examine the feature extraction problem and show that effective feature extraction can significantly enhance the performance of pattern recognition systems with simple classifiers. A framework is presented to evolve optimised feature extractors that transform an input pattern space into a decision space in which maximal class separability is obtained. We have applied this method to real world datasets from the UCI Machine Learning and StatLog databases to verify our approach and compare our proposed method with other reported results. We conclude that our algorithm is able to produce classifiers of superior (or equivalent) performance to the conventional classifiers examined, suggesting removal of the need to exhaustively evaluate a large family of conventional classifiers on any new problem. (C) 2010 Elsevier B.V. All rights reserved
The use of vicinal-risk minimization for training decision trees
We propose the use of Vapnik's vicinal risk minimization (VRM) for training decision trees to approximately maximize decision margins. We implement VRM by propagating uncertainties in the input attributes into the labeling decisions. In this way, we perform a global regularization over the decision tree structure. During a training phase, a decision tree is constructed to minimize the total probability of misclassifying the labeled training examples, a process which approximately maximizes the margins of the resulting classifier. We perform the necessary minimization using an appropriate meta-heuristic (genetic programming) and present results over a range of synthetic and benchmark real datasets. We demonstrate the statistical superiority of VRM training over conventional empirical risk minimization (ERM) and the well-known C4.5 algorithm, for a range of synthetic and real datasets. We also conclude that there is no statistical difference between trees trained by ERM and using C4.5. Training with VRM is shown to be more stable and repeatable than by ERM
Tikhonov Regularization as a Complexity Measure in Multiobjective Genetic Programming
© 1997-2012 IEEE. In this paper, we propose the use of Tikhonov regularization in conjunction with node count as a general complexity measure in multiobjective genetic programming. We demonstrate that employing this general complexity yields mean squared test error measures over a range of regression problems, which are typically superior to those from conventional node count (but never statistically worse). We also analyze the reason that our new method outperforms the conventional complexity measure and conclude that it forms a decision mechanism that balances both syntactic and semantic information
Pruning of Genetic Programming Trees Using Permutation Tests
We present a novel approach based on statistical permutation tests for pruning redundant
subtrees from genetic programming (GP) trees. We observe that over a range of regression problems,
median tree sizes are reduced by around 20% largely independent of test function, and that while
some large subtrees are removed, the median pruned subtree comprises just three nodes; most
take the form of an exact algebraic simplification. Our statistically-based pruning technique has
allowed us to explore the hypothesis that a given subtree can be replaced with a constant if this
substitution results in no statistical change to the behaviour of the parent tree—what we term
approximate simplification. In the eventuality, we infer that &95% of the pruned subtrees are the
result of algebraic simplifications, which provides some practical insight into the scope of removing
redundancies in GP trees
Comparison of semantic-based local search methods for multiobjective genetic programming
We report a series of experiments that use semantic-based local search within a multiobjective genetic programming (GP) framework. We compare various ways of selecting target subtrees for local search as well as different methods for performing that search; we have also made comparison with the random desired operator of Pawlak et al. using statistical hypothesis testing. We find that a standard steady state or generational GP followed by a carefully-designed single-objective GP implementing semantic-based local search produces models that are mode accurate and with statistically smaller (or equal) tree size than those generated by the corresponding baseline GP algorithms. The depth fair selection strategy of Ito et al. is found to perform best compared with other subtree selection methods in the model refinement
GPML: An XML-based Standard for the Interchange of Genetic Programming Trees
We propose a Genetic Programming Markup Language (GPML), an XML-based standard for
the interchange of genetic programming trees, and outline the benefits such a format would bring.
We present a formal definition of this standard and describe details of an implementation
Semantic-based local search in multiobjective genetic programming
We report a series of experiments within a multiobjective genetic programming (GP) framework using semantic-based local GP search. We have made comparison with the Random Desired Operator (RDO) of Pawlak et al. and find that a standard generational GP followed by a carefully-designed single-objective GP implementing semantic-based local search yields results statistically comparable to those obtained with the RDO operator. The trees obtained with our GP-based local search technique are, however, around half the size of the trees resulting from the use of the RDO