7,627 research outputs found
Recommended from our members
EASe : integrating search with learned episodes
Weak methods are insufficient to solve complex problems. Constrained weak methods, like hill-climbing, search too little of the problem space. Unconstrained weak methods, like breadth-first search, are intractable. Fortunately, through the integration of multiple weak methods more powerful problem solvers can be created. We demonstrate that augmenting a weak constrained search method with episodes provides a tractable method for solving a large class of problems. We demonstrate that these episodes can be generated using an unconstrained weak method while solving simple problems from a domain. We provide an analytical model of our approach and empirical results from the logic synthesis domain of VLSI design as well as the classic tile-sliding domain
Dueling CSP representations: Local search in the primal versus dual constraint graph.
Constraint Satisfaction Problems (CSPs) can be used to represent and solve many problems in Artificial Intelligence and the real world. When solving Constraint Satisfaction Problems, many of the methods developed and studied have focused only on the solution of binary CSPs while a large portion of real life problems are naturally modeled as non-binary CSPs. In this thesis we have designed an empirical study to investigate the behaviour of several local search methods in primal and dual constraint graph representations when solving non-binary CSPs. Local search methods tend to find a solution quickly since they generally give up the guarantee of completeness for polynomial time performance. Such local search methods include simple hill-climbing, steepest ascent hill-climbing and min-conflicts heuristics hill-climbing. We evaluate the performance of these three algorithms in each representation for a variety of parameter settings and we compare the search time cost means of two groups to support the comparison. Our comparison shows that we can use local search to solve a CSP with tight constraints in its dual representation and gain a better performance than using it in its primal representation. When constraints are getting looser, using local search in primal representation is a better choice. Among the three local search methods used in our empirical study, min-conflicts heuristics hill-climbing always gain the best performance while steepest ascent hill-climbing tends to have the worst performance and simple hill climbing is in the middle or sometimes it is the best. Paper copy at Leddy Library: Theses & Major Papers - Basement, West Bldg. / Call Number: Thesis2004 .H824. Source: Masters Abstracts International, Volume: 43-01, page: 0237. Adviser: Scott Goodwin. Thesis (M.Sc.)--University of Windsor (Canada), 2004
ADAPTIVE SEARCH AND THE PRELIMINARY DESIGN OF GAS TURBINE BLADE COOLING SYSTEMS
This research concerns the integration of Adaptive Search (AS) technique such as the
Genetic Algorithms (GA) with knowledge based software to develop a research prototype
of an Adaptive Search Manager (ASM). The developed approach allows to utilise both
quantitative and qualitative information in engineering design decision making. A Fuzzy
Expert System manipulates AS software within the design environment concerning the
preliminary design of gas turbine blade cooling systems. Steady state cooling hole geometry
models have been developed for the project in collaboration with Rolls Royce plc. The
research prototype of ASM uses a hybrid of Adaptive Restricted Tournament Selection
(ARTS) and Knowledge Based Hill Climbing (KBHC) to identify multiple "good" design
solutions as potential design options. ARTS is a GA technique that is particularly suitable
for real world problems having multiple sub-optima. KBHC uses information gathered
during the ARTS search as well as information from the designer to perform a deterministic
hill climbing. Finally, a local stochastic hill climbing fine tunes the "good" designs. Design
solution sensitivity, design variable sensitivities and constraint sensitivities are calculated
following Taguchi's methodology, which extracts sensitivity information with a very small
number of model evaluations. Each potential design option is then qualitatively evaluated
separately for manufacturability, choice of materials and some designer's special preferences
using the knowledge of domain experts. In order to guarantee that the qualitative evaluation
module can evaluate any design solution from the entire design space with a reasonably
small number of rules, a novel knowledge representation technique is developed. The
knowledge is first separated in three categories: inter-variable knowledge, intra-variable
knowledge and heuristics. Inter-variable knowledge and intra-variable knowledge are then
integrated using a concept of compromise. Information about the "good" design solutions is
presented to the designer through a designer's interface for decision support.Rolls Royce plc., Bristol (UK
The Optimisation of Stochastic Grammars to Enable Cost-Effective Probabilistic Structural Testing
The effectiveness of probabilistic structural testing depends on the characteristics of the probability distribution from which test inputs are sampled at random. Metaheuristic search has been shown to be a practical method of optimis- ing the characteristics of such distributions. However, the applicability of the existing search-based algorithm is lim- ited by the requirement that the software’s inputs must be a fixed number of numeric values. In this paper we relax this limitation by means of a new representation for the probability distribution. The repre- sentation is based on stochastic context-free grammars but incorporates two novel extensions: conditional production weights and the aggregation of terminal symbols represent- ing numeric values. We demonstrate that an algorithm which combines the new representation with hill-climbing search is able to effi- ciently derive probability distributions suitable for testing software with structurally-complex input domains
Efficient computational strategies to learn the structure of probabilistic graphical models of cumulative phenomena
Structural learning of Bayesian Networks (BNs) is a NP-hard problem, which is
further complicated by many theoretical issues, such as the I-equivalence among
different structures. In this work, we focus on a specific subclass of BNs,
named Suppes-Bayes Causal Networks (SBCNs), which include specific structural
constraints based on Suppes' probabilistic causation to efficiently model
cumulative phenomena. Here we compare the performance, via extensive
simulations, of various state-of-the-art search strategies, such as local
search techniques and Genetic Algorithms, as well as of distinct regularization
methods. The assessment is performed on a large number of simulated datasets
from topologies with distinct levels of complexity, various sample size and
different rates of errors in the data. Among the main results, we show that the
introduction of Suppes' constraints dramatically improve the inference
accuracy, by reducing the solution space and providing a temporal ordering on
the variables. We also report on trade-offs among different search techniques
that can be efficiently employed in distinct experimental settings. This
manuscript is an extended version of the paper "Structural Learning of
Probabilistic Graphical Models of Cumulative Phenomena" presented at the 2018
International Conference on Computational Science
Metaheuristic Approaches to the Placement of Suicide Bomber Detectors.
Suicide bombing is an infamous form of terrorism that is becoming increasingly prevalent in the current era of global terror warfare. We consider the case of targeted attacks of this kind, and the use of detectors distributed over the area under threat as a protective countermeasure. Such detectors are non-fully reliable, and must be strategically placed in order to maximize the chances of detecting the attack, hence minimizing the expected number of casualties. To this end, different metaheuristic approaches based on local search and on population-based search (such as a hill climber, different Greedy randomized adaptive search procedures, an evolutionary algorithm and several estimation of distribution algorithms) are considered and benchmarked against a powerful greedy heuristic from the literature. We conduct an extensive empirical evaluation on synthetic instances featuring very diverse properties. Most metaheuristics outperform the greedy algorithm, and a hill-climber is shown to be superior to remaining approaches. This hill-climber is subsequently subject to a sensitivity analysis to determine which problem features make it stand above the greedy approach, and is finally deployed on a number of problem instances built after realistic scenarios, corroborating the good performance of the heuristic.Spanish Ministry of Economy and Competitiveness and European Regional Development Fund (FEDER) under project EphemeCH (TIN2014-56494-C4-1-P)
The GRT Planning System: Backward Heuristic Construction in Forward State-Space Planning
This paper presents GRT, a domain-independent heuristic planning system for
STRIPS worlds. GRT solves problems in two phases. In the pre-processing phase,
it estimates the distance between each fact and the goals of the problem, in a
backward direction. Then, in the search phase, these estimates are used in
order to further estimate the distance between each intermediate state and the
goals, guiding so the search process in a forward direction and on a best-first
basis. The paper presents the benefits from the adoption of opposite directions
between the preprocessing and the search phases, discusses some difficulties
that arise in the pre-processing phase and introduces techniques to cope with
them. Moreover, it presents several methods of improving the efficiency of the
heuristic, by enriching the representation and by reducing the size of the
problem. Finally, a method of overcoming local optimal states, based on domain
axioms, is proposed. According to it, difficult problems are decomposed into
easier sub-problems that have to be solved sequentially. The performance
results from various domains, including those of the recent planning
competitions, show that GRT is among the fastest planners
A hybrid algorithm for Bayesian network structure learning with application to multi-label learning
We present a novel hybrid algorithm for Bayesian network structure learning,
called H2PC. It first reconstructs the skeleton of a Bayesian network and then
performs a Bayesian-scoring greedy hill-climbing search to orient the edges.
The algorithm is based on divide-and-conquer constraint-based subroutines to
learn the local structure around a target variable. We conduct two series of
experimental comparisons of H2PC against Max-Min Hill-Climbing (MMHC), which is
currently the most powerful state-of-the-art algorithm for Bayesian network
structure learning. First, we use eight well-known Bayesian network benchmarks
with various data sizes to assess the quality of the learned structure returned
by the algorithms. Our extensive experiments show that H2PC outperforms MMHC in
terms of goodness of fit to new data and quality of the network structure with
respect to the true dependence structure of the data. Second, we investigate
H2PC's ability to solve the multi-label learning problem. We provide
theoretical results to characterize and identify graphically the so-called
minimal label powersets that appear as irreducible factors in the joint
distribution under the faithfulness condition. The multi-label learning problem
is then decomposed into a series of multi-class classification problems, where
each multi-class variable encodes a label powerset. H2PC is shown to compare
favorably to MMHC in terms of global classification accuracy over ten
multi-label data sets covering different application domains. Overall, our
experiments support the conclusions that local structural learning with H2PC in
the form of local neighborhood induction is a theoretically well-motivated and
empirically effective learning framework that is well suited to multi-label
learning. The source code (in R) of H2PC as well as all data sets used for the
empirical tests are publicly available.Comment: arXiv admin note: text overlap with arXiv:1101.5184 by other author
- …