946 research outputs found
Plan stability: replanning versus plan repair
The ultimate objective in planning is to construct plans for execution. However, when a plan is executed in a real environment it can encounter differences between the expected and actual context of execution. These differences can manifest as divergences between the expected and observed states of the world, or as a change in the goals to be achieved by the plan. In both cases, the old plan must be replaced with a new one. In replacing the plan an important consideration is plan stability. We compare two alternative strategies for achieving the {em stable} repair of a plan: one is simply to replan from scratch and the other is to adapt the existing plan to the new context. We present arguments to support the claim that plan stability is a valuable property. We then propose an implementation, based on LPG, of a plan repair strategy that adapts a plan to its new context. We demonstrate empirically that our plan repair strategy achieves more stability than replanning and can produce repaired plans more efficiently than replanning
Neural Networks for Predicting Algorithm Runtime Distributions
Many state-of-the-art algorithms for solving hard combinatorial problems in
artificial intelligence (AI) include elements of stochasticity that lead to
high variations in runtime, even for a fixed problem instance. Knowledge about
the resulting runtime distributions (RTDs) of algorithms on given problem
instances can be exploited in various meta-algorithmic procedures, such as
algorithm selection, portfolios, and randomized restarts. Previous work has
shown that machine learning can be used to individually predict mean, median
and variance of RTDs. To establish a new state-of-the-art in predicting RTDs,
we demonstrate that the parameters of an RTD should be learned jointly and that
neural networks can do this well by directly optimizing the likelihood of an
RTD given runtime observations. In an empirical study involving five algorithms
for SAT solving and AI planning, we show that neural networks predict the true
RTDs of unseen instances better than previous methods, and can even do so when
only few runtime observations are available per training instance
A hybrid shifting bottleneck-tabu search heuristic for the job shop total weighted tardiness problem
In this paper, we study the job shop scheduling problem with the objective of minimizing the total weighted tardiness. We propose a hybrid shifting bottleneck - tabu search (SB-TS) algorithm by replacing the reoptimization step in the shifting bottleneck (SB) algorithm by a tabu search (TS). In terms of the shifting bottleneck heuristic, the proposed tabu search optimizes the total weighted tardiness for partial schedules in which some machines are currently assumed to have infinite capacity. In the context of tabu search, the shifting bottleneck heuristic features a long-term memory which helps to diversify the local search. We exploit this synergy to develop a state-of-the-art algorithm for the job shop total weighted tardiness problem (JS-TWT). The computational
effectiveness of the algorithm is demonstrated on standard benchmark instances from the literature
Progress in AI Planning Research and Applications
Planning has made significant progress since its inception in the 1970s, in terms both of the efficiency and sophistication of its algorithms and representations and its potential for application to real problems. In this paper we sketch the foundations of planning as a sub-field of Artificial Intelligence and the history of its development over the past three decades. Then some of the recent achievements within the field are discussed and provided some experimental data demonstrating the progress that has been made in the application of general planners to realistic and complex problems. The paper concludes by identifying some of the open issues that remain as important challenges for future research in planning
Taming Numbers and Durations in the Model Checking Integrated Planning System
The Model Checking Integrated Planning System (MIPS) is a temporal least
commitment heuristic search planner based on a flexible object-oriented
workbench architecture. Its design clearly separates explicit and symbolic
directed exploration algorithms from the set of on-line and off-line computed
estimates and associated data structures. MIPS has shown distinguished
performance in the last two international planning competitions. In the last
event the description language was extended from pure propositional planning to
include numerical state variables, action durations, and plan quality objective
functions. Plans were no longer sequences of actions but time-stamped
schedules. As a participant of the fully automated track of the competition,
MIPS has proven to be a general system; in each track and every benchmark
domain it efficiently computed plans of remarkable quality. This article
introduces and analyzes the most important algorithmic novelties that were
necessary to tackle the new layers of expressiveness in the benchmark problems
and to achieve a high level of performance. The extensions include critical
path analysis of sequentially generated plans to generate corresponding optimal
parallel plans. The linear time algorithm to compute the parallel plan bypasses
known NP hardness results for partial ordering by scheduling plans with respect
to the set of actions and the imposed precedence relations. The efficiency of
this algorithm also allows us to improve the exploration guidance: for each
encountered planning state the corresponding approximate sequential plan is
scheduled. One major strength of MIPS is its static analysis phase that grounds
and simplifies parameterized predicates, functions and operators, that infers
knowledge to minimize the state description length, and that detects domain
object symmetries. The latter aspect is analyzed in detail. MIPS has been
developed to serve as a complete and optimal state space planner, with
admissible estimates, exploration engines and branching cuts. In the
competition version, however, certain performance compromises had to be made,
including floating point arithmetic, weighted heuristic search exploration
according to an inadmissible estimate and parameterized optimization
Planning through Automatic Portfolio Configuration: The PbP Approach
In the field of domain-independent planning, several powerful planners implementing different techniques have been developed. However, no one of these systems outperforms all others in every known benchmark domain. In this work, we propose a multi-planner approach that automatically configures a portfolio of planning techniques for each given domain. The configuration process for a given domain uses a set of training instances to: (i) compute and analyze some alternative sets of macro-actions for each planner in the portfolio identifying a (possibly empty) useful set, (ii) select a cluster of planners, each one with the identified useful set of macro-actions, that is expected to perform best, and (iii) derive some additional information for configuring the execution scheduling of the selected planners at planning time. The resulting planning system, called PbP (Portfolio- based Planner), has two variants focusing on speed and plan quality. Different versions of PbP entered and won the learning track of the sixth and seventh International Planning Competitions. In this paper, we experimentally analyze PbP considering planning speed and plan quality in depth. We provide a collection of results that help to understand PbP�s behavior, and demonstrate the effectiveness of our approach to configuring a portfolio of planners with macro-actions
Creating planning portfolios with predictive models
Mención Internacional en el título de doctorSequential planning portfolios are very powerful in exploiting the complementary
strength of different automated planners: for each planning task
there are one or more base planners that obtain the best solution. Therefore,
the main challenge when building a planning portfolio is to ensure that
a suitable planner be chosen and that it gets enough planning time. To solve
this problem we need firstly to define three elements. The first is the settings
or planning conditions: time, memory, or other constraints. The second one
is the set of base planners. And finally, a benchmark that provides us with
knowledge on how the base planners will behave under the given settings,
following some kind of inductive process. Ideally, if the previous elements
are correctly defined, when a new planning task arrives, an oracle will be
able to tell which base planner to run and for how long. In practice, since
no oracle exists, the challenge to choose a sub-set of base planners, is assigning
them a running time and deciding the order in which they are run
to optimize a planning metric under the predefined settings. Many state-of-the-
art portfolios might never achieve an optimal performance because they
do not select different planners for the different planning tasks. In addition,
these static techniques typically assign a fixed running time to the selected
set of planners, independently of the task. besides, the old-fashioned dynamic
portfolios present a poor characterization of the planning task and
do not have enough knowledge to predict an accurate portfolio configuration
in many cases. The aforementioned drawbacks are intensified by the
fact that there is an increasing number of planners available to choose from,
although many of them are designed following similar approaches, so they
are expected to behave similarly.
This dissertation is built on two main hypotheses. Firstly that the space
of the base planners can be reduced just by selecting a subset of diverse or
complementary planners; e.g. that there is a minimal set of planners that
ensure that the optimal portfolio can be computed. Secondly, that planning
tasks can be characterized, and that the difficulty in solving them can be
modelled as a function of these features. To evaluate the first hypothesis,
we analyze different metrics that could be used to filter the initial set of base
planners. Classical metrics such as coverage, quality or execution time have
been chosen by different portfolios in the past. We demonstrate that these
selection methods may reduce the diversity of the portfolios, and propose
an alternative method based on the Pareto dominance. We then carry out
a profound analysis on previous planning task characterizations and show how we could exploit them in current planning paradigms.
A group of very informative features are proposed to improve the current feature definition of the planning tasks. These features have enough knowledge to differentiate
planning tasks with similar \a priori" complexity. In this thesis we
demonstrate that the implicit knowledge can be exploited in the construction
of predictive models. These models estimate whether a base planner
will be able to solve a given problem and, if so, how long it will take. Nevertheless,
the predictive models are not perfect and sometimes provide wrong
(or inaccurate) predictions. To solve this kind of problems, we propose different
portfolio strategies to combine the number of selected base planners
and their times. These strategies take into account the predefined settings
and the knowledge learned in previous phases.
In conclusion, this thesis sets out a profound analysis of three different
mechanisms or steps to create planning portfolios with predictive models,
including new proposals for developing: planner filtering, planning task
featuring, learning predictive models and portfolio construction strategies.
One of the proposed portfolios was the winner of the Sequential Satisficing
Track of the International Planning Competition held in 2014Los portfolios de planificadores tienen un gran potencial ya que pueden
aprovecharse de los diferentes planificadores automáticos, consiguiendo mejorar
el rendimiento de un único planificador. Sin embargo, la creación de un
portfolio no es una tarea sencilla, ya que para poder crear uno lo suficientemente
bueno, hay que tratar tres problemas fundamentales. El primero de
ellos es encontrar qué planificadores hay que seleccionar como componentes
del mismo. La segunda es el tiempo que hay que asignar a cada planificador
y, la última y no menos importante el orden en el que se tienen que ejecutar.
Actualmente en el estado del arte, estas configuraciones, se realizan a
partir de los resultados obtenidos por los planificadores en una fase previa
de entrenamiento con un conjunto de problemas y restricciones prefijado
(tiempo, memoria, etc), consiguiendo una configuración específica tratando
de optimizar una métrica. Idealmente, la mejor configuración posible consiste
en asignar el tiempo suficiente al mejor planificador para cada tarea
de planificación. Sin embargo, esta configuración no siempre es posible, y
hay que recurrir a otras aproximaciones como asignar un tiempo fijo a una
selección de planificadores. Ésta no es la única simplificación utilizada, existen
otras técnicas más cercanas a la óptima, en las cuales se selecciona un
planificador o varios en función de la tarea a resolver. Sin embargo, estos
sistemas, denominados dinámicos, incluyen una escasa caracterización de
las tareas de planificación.
En esta tesis se parte de dos hipótesis. La primera de ellas es que existe un
conjunto reducido de planificadores que maximiza la diversidad. La segunda
de ellas consiste en la posibilidad de crear un conjunto de descriptivos lo
suficientemente bueno para caracterizar la tarea de planificación. La caracterización de las tareas de planificación puede estar basada en sus distintas
representaciones, así como en sus paradigmas. La primera tarea es seleccionar
un conjunto de planificadores; realizando un análisis basado en las
métricas clásicas de planificación, como son problemas resueltos, calidad
y tiempo para seleccionar un subconjunto de planificadores. Adicionalmente,
proponemos como alternativa a estas métricas, una técnica multiobjetivo.
Este criterio está basado en la dominancia de Pareto combinando
las métricas de tiempo y calidad. Continuando con nuestras hip_otesis es
necesario crear un conjunto de características bien informado para la tarea
de planificación. Estas características deben ser capaces de diferenciar adecuadamente
por problema y para ello sería necesario basarse en los distintos
paradigmas de la planificación automática. Este grupo de características
tienen que ser úutiles para crear modelos predictivos. Estos modelos podrán
darnos además de una selección de planificadores, una aproximación del
tiempo asignado a cada componente y el orden de los mismos. Adicionalmente
se presentarán una serie de estrategias para explotar el conocimiento
obtenido con los modelos predictivos.
En conclusión, se plantea y desarrolla un sistema para configurar porfolios
de planificadores usando modelos predictivos en tres fases distintas. Una
instanciación de este sistema fue el ganador de la competición internacional
de planificación en el áarea de satisfacibilidad en el año 2014.Programa Oficial de Doctorado en Ciencia y Tecnología InformáticaPresidente: María Araceli Sanchís de Miguel.- Secretario: Álvaro Torralba Arias de Reyna.- Vocal: Alessandro Saett
- …