3 research outputs found

    Improving the efficiency of Bayesian Network Based EDAs and their application in Bioinformatics

    Get PDF
    Estimation of distribution algorithms (EDAs) is a relatively new trend of stochastic optimizers which have received a lot of attention during last decade. In each generation, EDAs build probabilistic models of promising solutions of an optimization problem to guide the search process. New sets of solutions are obtained by sampling the corresponding probability distributions. Using this approach, EDAs are able to provide the user a set of models that reveals the dependencies between variables of the optimization problems while solving them. In order to solve a complex problem, it is necessary to use a probabilistic model which is able to capture the dependencies. Bayesian networks are usually used for modeling multiple dependencies between variables. Learning Bayesian networks, especially for large problems with high degree of dependencies among their variables is highly computationally expensive which makes it the bottleneck of EDAs. Therefore introducing efficient Bayesian learning algorithms in EDAs seems necessary in order to use them for large problems. In this dissertation, after comparing several Bayesian network learning algorithms, we propose an algorithm, called CMSS-BOA, which uses a recently introduced heuristic called max-min parent children (MMPC) in order to constrain the model search space. This algorithm does not consider a fixed and small upper bound on the order of interaction between variables and is able solve problems with large numbers of variables efficiently. We compare the efficiency of CMSS-BOA with the standard Bayesian network based EDA for solving several benchmark problems and finally we use it to build a predictor for predicting the glycation sites in mammalian proteins

    Model complexity vs. performance in the Bayesian Optimization Algorithm

    No full text
    The Bayesian Optimization Algorithm (BOA) uses a Bayesian network to estimate the probability distribution of promising solutions to a given optimization problem. This distribution is then used to generate new candidate solutions. The objective is to improve the population of candidate solutions by learning and sampling from good solutions. A Bayesian network (BN) is a graphical representation of a probability distribution over a set of variables of a given problem domain. The number of topological states that a BN can create depends on a parameter called maximum allowed indegree. We show that the value of the maximum allowed indegree given to the Bayesian network used by the BOA strongly affects the performance of this algorithm. Furthermore, there is a limited set of values for this parameter for which the performance of the BOA is maximized
    corecore