12 research outputs found

    Linearized biogeography-based optimization with re-initialization and local search

    Get PDF
    Biogeography-based optimization (BBO) is an evolutionary optimization algorithm that uses migration to share information among candidate solutions. One limitation of BBO is that it changes only one independent variable at a time in each candidate solution. In this paper, a linearized version of BBO, called LBBO, is proposed to reduce rotational variance. The proposed method is combined with periodic re-initialization and local search operators to obtain an algorithm for global optimization in a continuous search space. Experiments have been conducted on 45 benchmarks from the 2005 and 2011 Congress on Evolutionary Computation, and LBBO performance is compared with the results published in those conferences. The results show that LBBO provides competitive performance with state-of-the-art evolutionary algorithms. In particular, LBBO performs particularly well for certain types of multimodal problems, including high-dimensional real-world problems. Also, LBBO is insensitive to whether or not the solution lies on the search domain boundary, in a wide or narrow basin, and within or outside the initialization domain

    Student performance assessment using clustering techniques

    Get PDF
    The application of informatics in the university system management allows managers to count with a great amount of data which, rationally treated, can offer significant help for the student programming monitoring. This research proposes the use of clustering techniques as a useful tool of management strategy to evaluate the progression of the students’ behavior by dividing the population into homogeneous groups according to their characteristics and skills. These applications can help both the teacher and the student to improve the quality of education. The selected method is the data grouping analysis by means of fuzzy logic using the Fuzzy C-means algorithm to achieve a standard indicator called Grade, through an expert system to enable segmentation.Universidad de la Costa, 2 Universidad Nacional Experimental Politécnica “Antonio José de Sucre”, Universidad Simón Bolívar, Corporación Universitaria Latinoamericana, Corporación Universitaria Minuto de Dios

    Particle Swarm Optimization Methods for Pattern Recognition and Image Processing

    Get PDF
    Pattern recognition has as its objective to classify objects into different categories and classes. It is a fundamental component of artificial intelligence and computer vision. This thesis investigates the application of an efficient optimization method, known as Particle Swarm Optimization (PSO), to the field of pattern recognition and image processing. First a clustering method that is based on PSO is proposed. The application of the proposed clustering algorithm to the problem of unsupervised classification and segmentation of images is investigated. A new automatic image generation tool tailored specifically for the verification and comparison of various unsupervised image classification algorithms is then developed. A dynamic clustering algorithm which automatically determines the "optimum" number of clusters and simultaneously clusters the data set with minimal user interference is then developed. Finally, PSO-based approaches are proposed to tackle the color image quantization and spectral unmixing problems. In all the proposed approaches, the influence of PSO parameters on the performance of the proposed algorithms is evaluated.Thesis (PhD)--University of Pretoria, 2006.Computer Scienceunrestricte

    Dynamic clustering using Particle Swarm Optimization with application in image segmentation

    No full text
    Abstract—A new dynamic clustering approach (DCPSO), based on Particle Swarm Optimization, is proposed. This approach is applied to unsupervised image classification. The proposed approach automatically determines the "optimum " number of clusters and simultaneously clusters the data set with minimal user interference. The algorithm starts by partitioning the data set into a relatively large number of clusters to reduce the effects of initial conditions. Using binary particle swarm optimization the "best " number of clusters is selected. The centers of the chosen clusters is then refined via the Kmeans clustering algorithm. The experiments conducted show that the proposed approach generally found the "optimum " number of clusters on the tested images

    Differential evolution methods for unsupervised image classification

    No full text
    Abstract- A clustering method that is based o

    Permutation Tests for Metaheuristic Algorithms

    No full text
    Many metaheuristic approaches are inherently stochastic. In order to compare such methods, statistical tests are needed. However, choosing an appropriate test is not trivial, given that each test has some assumptions about the distribution of the underlying data that must be true before it can be used. Permutation tests (P-Tests) are statistical tests with minimal number of assumptions. These tests are simple, intuitive and nonparametric. In this paper, we argue researchers in the field of metaheuristics to adopt P-Tests to compare their algorithms. We define two statistic tests and then present an algorithm that uses them to compute the p-value. The proposed process is used to compare 5 metaheuristic algorithms on 10 benchmark functions. The resulting p-values are compared with the p-values of two widely used statistical tests. The results show that the proposed P-test is generally consistent with the classical tests, but more conservative in few cases

    Permutation Tests for Metaheuristic Algorithms

    No full text
    Many metaheuristic approaches are inherently stochastic. In order to compare such methods, statistical tests are needed. However, choosing an appropriate test is not trivial, given that each test has some assumptions about the distribution of the underlying data that must be true before it can be used. Permutation tests (P-Tests) are statistical tests with minimal number of assumptions. These tests are simple, intuitive and nonparametric. In this paper, we argue researchers in the field of metaheuristics to adopt P-Tests to compare their algorithms. We define two statistic tests and then present an algorithm that uses them to compute the p-value. The proposed process is used to compare 5 metaheuristic algorithms on 10 benchmark functions. The resulting p-values are compared with the p-values of two widely used statistical tests. The results show that the proposed P-test is generally consistent with the classical tests, but more conservative in few cases
    corecore