2,228 research outputs found

    Interval-Based Relaxation for General Numeric Planning

    Get PDF
    We generalise the interval-based relaxation to sequential numeric planning problems with non-linear conditions and effects, and cyclic dependencies. This effectively removes all the limitations on the problem placed in previous work on numeric planning heuristics, and even allows us to extend the planning language with a wider set of mathematical functions. Heuristics obtained from the generalised relaxation are pruning-safe. We derive one such heuristic and use it to solve discrete-time control-like planning problems with autonomous processes. Few planners can solve such problems, and search with our new heuristic compares favourably with the

    Surrogate Search As a Way to Combat Harmful Effects of Ill-behaved Evaluation Functions

    Full text link
    Recently, several researchers have found that cost-based satisficing search with A* often runs into problems. Although some "work arounds" have been proposed to ameliorate the problem, there has been little concerted effort to pinpoint its origin. In this paper, we argue that the origins of this problem can be traced back to the fact that most planners that try to optimize cost also use cost-based evaluation functions (i.e., f(n) is a cost estimate). We show that cost-based evaluation functions become ill-behaved whenever there is a wide variance in action costs; something that is all too common in planning domains. The general solution to this malady is what we call a surrogatesearch, where a surrogate evaluation function that doesn't directly track the cost objective, and is resistant to cost-variance, is used. We will discuss some compelling choices for surrogate evaluation functions that are based on size rather that cost. Of particular practical interest is a cost-sensitive version of size-based evaluation function -- where the heuristic estimates the size of cheap paths, as it provides attractive quality vs. speed tradeoffsComment: arXiv admin note: substantial text overlap with arXiv:1103.368

    Creating planning portfolios with predictive models

    Get PDF
    Mención Internacional en el título de doctorSequential planning portfolios are very powerful in exploiting the complementary strength of different automated planners: for each planning task there are one or more base planners that obtain the best solution. Therefore, the main challenge when building a planning portfolio is to ensure that a suitable planner be chosen and that it gets enough planning time. To solve this problem we need firstly to define three elements. The first is the settings or planning conditions: time, memory, or other constraints. The second one is the set of base planners. And finally, a benchmark that provides us with knowledge on how the base planners will behave under the given settings, following some kind of inductive process. Ideally, if the previous elements are correctly defined, when a new planning task arrives, an oracle will be able to tell which base planner to run and for how long. In practice, since no oracle exists, the challenge to choose a sub-set of base planners, is assigning them a running time and deciding the order in which they are run to optimize a planning metric under the predefined settings. Many state-of-the- art portfolios might never achieve an optimal performance because they do not select different planners for the different planning tasks. In addition, these static techniques typically assign a fixed running time to the selected set of planners, independently of the task. besides, the old-fashioned dynamic portfolios present a poor characterization of the planning task and do not have enough knowledge to predict an accurate portfolio configuration in many cases. The aforementioned drawbacks are intensified by the fact that there is an increasing number of planners available to choose from, although many of them are designed following similar approaches, so they are expected to behave similarly. This dissertation is built on two main hypotheses. Firstly that the space of the base planners can be reduced just by selecting a subset of diverse or complementary planners; e.g. that there is a minimal set of planners that ensure that the optimal portfolio can be computed. Secondly, that planning tasks can be characterized, and that the difficulty in solving them can be modelled as a function of these features. To evaluate the first hypothesis, we analyze different metrics that could be used to filter the initial set of base planners. Classical metrics such as coverage, quality or execution time have been chosen by different portfolios in the past. We demonstrate that these selection methods may reduce the diversity of the portfolios, and propose an alternative method based on the Pareto dominance. We then carry out a profound analysis on previous planning task characterizations and show how we could exploit them in current planning paradigms. A group of very informative features are proposed to improve the current feature definition of the planning tasks. These features have enough knowledge to differentiate planning tasks with similar \a priori" complexity. In this thesis we demonstrate that the implicit knowledge can be exploited in the construction of predictive models. These models estimate whether a base planner will be able to solve a given problem and, if so, how long it will take. Nevertheless, the predictive models are not perfect and sometimes provide wrong (or inaccurate) predictions. To solve this kind of problems, we propose different portfolio strategies to combine the number of selected base planners and their times. These strategies take into account the predefined settings and the knowledge learned in previous phases. In conclusion, this thesis sets out a profound analysis of three different mechanisms or steps to create planning portfolios with predictive models, including new proposals for developing: planner filtering, planning task featuring, learning predictive models and portfolio construction strategies. One of the proposed portfolios was the winner of the Sequential Satisficing Track of the International Planning Competition held in 2014Los portfolios de planificadores tienen un gran potencial ya que pueden aprovecharse de los diferentes planificadores automáticos, consiguiendo mejorar el rendimiento de un único planificador. Sin embargo, la creación de un portfolio no es una tarea sencilla, ya que para poder crear uno lo suficientemente bueno, hay que tratar tres problemas fundamentales. El primero de ellos es encontrar qué planificadores hay que seleccionar como componentes del mismo. La segunda es el tiempo que hay que asignar a cada planificador y, la última y no menos importante el orden en el que se tienen que ejecutar. Actualmente en el estado del arte, estas configuraciones, se realizan a partir de los resultados obtenidos por los planificadores en una fase previa de entrenamiento con un conjunto de problemas y restricciones prefijado (tiempo, memoria, etc), consiguiendo una configuración específica tratando de optimizar una métrica. Idealmente, la mejor configuración posible consiste en asignar el tiempo suficiente al mejor planificador para cada tarea de planificación. Sin embargo, esta configuración no siempre es posible, y hay que recurrir a otras aproximaciones como asignar un tiempo fijo a una selección de planificadores. Ésta no es la única simplificación utilizada, existen otras técnicas más cercanas a la óptima, en las cuales se selecciona un planificador o varios en función de la tarea a resolver. Sin embargo, estos sistemas, denominados dinámicos, incluyen una escasa caracterización de las tareas de planificación. En esta tesis se parte de dos hipótesis. La primera de ellas es que existe un conjunto reducido de planificadores que maximiza la diversidad. La segunda de ellas consiste en la posibilidad de crear un conjunto de descriptivos lo suficientemente bueno para caracterizar la tarea de planificación. La caracterización de las tareas de planificación puede estar basada en sus distintas representaciones, así como en sus paradigmas. La primera tarea es seleccionar un conjunto de planificadores; realizando un análisis basado en las métricas clásicas de planificación, como son problemas resueltos, calidad y tiempo para seleccionar un subconjunto de planificadores. Adicionalmente, proponemos como alternativa a estas métricas, una técnica multiobjetivo. Este criterio está basado en la dominancia de Pareto combinando las métricas de tiempo y calidad. Continuando con nuestras hip_otesis es necesario crear un conjunto de características bien informado para la tarea de planificación. Estas características deben ser capaces de diferenciar adecuadamente por problema y para ello sería necesario basarse en los distintos paradigmas de la planificación automática. Este grupo de características tienen que ser úutiles para crear modelos predictivos. Estos modelos podrán darnos además de una selección de planificadores, una aproximación del tiempo asignado a cada componente y el orden de los mismos. Adicionalmente se presentarán una serie de estrategias para explotar el conocimiento obtenido con los modelos predictivos. En conclusión, se plantea y desarrolla un sistema para configurar porfolios de planificadores usando modelos predictivos en tres fases distintas. Una instanciación de este sistema fue el ganador de la competición internacional de planificación en el áarea de satisfacibilidad en el año 2014.Programa Oficial de Doctorado en Ciencia y Tecnología InformáticaPresidente: María Araceli Sanchís de Miguel.- Secretario: Álvaro Torralba Arias de Reyna.- Vocal: Alessandro Saett

    The 2011 International Planning Competition

    Get PDF
    After a 3 years gap, the 2011 edition of the IPC involved a total of 55 planners, some of them versions of the same planner, distributed among four tracks: the sequential satisficing track (27 planners submitted out of 38 registered), the sequential multicore track (8 planners submitted out of 12 registered), the sequential optimal track (12 planners submitted out of 24 registered) and the temporal satisficing track (8 planners submitted out of 14 registered). Three more tracks were open to participation: temporal optimal, preferences satisficing and preferences optimal. Unfortunately the number of submitted planners did not allow these tracks to be finally included in the competition. A total of 55 people were participating, grouped in 31 teams. Participants came from Australia, Canada, China, France, Germany, India, Israel, Italy, Spain, UK and USA. For the sequential tracks 14 domains, with 20 problems each, were selected, while the temporal one had 12 domains, also with 20 problems each. Both new and past domains were included. As in previous competitions, domains and problems were unknown for participants and all the experimentation was carried out by the organizers. To run the competition a cluster of eleven 64-bits computers (Intel XEON 2.93 Ghz Quad core processor) using Linux was set up. Up to 1800 seconds, 6 GB of RAM memory and 750 GB of hard disk were available for each planner to solve a problem. This resulted in 7540 computing hours (about 315 days), plus a high number of hours devoted to preliminary experimentation with new domains, reruns and bugs fixing. The detailed results of the competition, the software used for automating most tasks, the source code of all the participating planners and the description of domains and problems can be found at the competition’s web page: http://www.plg.inf.uc3m.es/ipc2011-deterministicThis booklet summarizes the participants on the Deterministic Track of the International Planning Competition (IPC) 2011. Papers describing all the participating planners are included

    Optimal Planning with State Constraints

    Get PDF
    In the classical planning model, state variables are assigned values in the initial state and remain unchanged unless explicitly affected by action effects. However, some properties of states are more naturally modelled not as direct effects of actions but instead as derived, in each state, from the primary variables via a set of rules. We refer to those rules as state constraints. The two types of state constraints that will be discussed here are numeric state constraints and logical rules that we will refer to as axioms. When using state constraints we make a distinction between primary variables, whose values are directly affected by action effects, and secondary variables, whose values are determined by state constraints. While primary variables have finite and discrete domains, as in classical planning, there is no such requirement for secondary variables. For example, using numeric state constraints allows us to have secondary variables whose values are real numbers. We show that state constraints are a construct that lets us combine classical planning methods with specialised solvers developed for other types of problems. For example, introducing numeric state constraints enables us to apply planning techniques in domains involving interconnected physical systems, such as power networks. To solve these types of problems optimally, we adapt commonly used methods from optimal classical planning, namely state-space search guided by admissible heuristics. In heuristics based on monotonic relaxation, the idea is that in a relaxed state each variable assumes a set of values instead of just a single value. With state constraints, the challenge becomes to evaluate the conditions, such as goals and action preconditions, that involve secondary variables. We employ consistency checking tools to evaluate whether these conditions are satisfied in the relaxed state. In our work with numerical constraints we use linear programming, while with axioms we use answer set programming and three value semantics. This allows us to build a relaxed planning graph and compute constraint-aware version of heuristics based on monotonic relaxation. We also adapt pattern database heuristics. We notice that an abstract state can be thought of as a state in the monotonic relaxation in which the variables in the pattern hold only one value, while the variables not in the pattern simultaneously hold all the values in their domains. This means that we can apply the same technique for evaluating conditions on secondary variables as we did for the monotonic relaxation and build pattern databases similarly as it is done in classical planning. To make better use of our heuristics, we modify the A* algorithm by combining two techniques that were previously used independently – partial expansion and preferred operators. Our modified algorithm, which we call PrefPEA, is most beneficial in cases where heuristic is expensive to compute, but accurate, and states have many successors

    Efficient Automated Planning with New Formulations

    Get PDF
    Problem solving usually strongly relies on how the problem is formulated. This fact also applies to automated planning, a key field in artificial intelligence research. Classical planning used to be dominated by STRIPS formulation, a simple model based on propositional logic. In the recently introduced SAS+ formulation, the multi-valued variables naturally depict certain invariants that are missed in STRIPS, make SAS+ have many favorable features. Because of its rich structural information SAS+ begins to attract lots of research interest. Existing works, however, are mostly limited to one single thing: to improve heuristic functions. This is in sharp contrast with the abundance of planning models and techniques in the field. On the other hand, although heuristic is a key part for search, its effectiveness is limited. Recent investigations have shown that even if we have almost perfect heuristics, the number of states to visit is still exponential. Therefore, there is a barrier between the nice features of SAS+ and its applications in planning algorithms. In this dissertation, we have recasted two major planning paradigms: state space search and planning as Satisfiability: SAT), with three major contributions. First, we have utilized SAS+ for a new hierarchical state space search model by taking advantage of the decomposable structure within SAS+. This algorithm can greatly reduce the time complexity for planning. Second, planning as Satisfiability is a major planning approach, but it is traditionally based on STRIPS. We have developed a new SAS+ based SAT encoding scheme: SASE) for planning. The state space modeled by SASE shows a decomposable structure with certain components independent to others, showing promising structure that STRIPS based encoding does not have. Third, the expressiveness of planning is important for real world scenarios, thus we have also extended the planning as SAT to temporally expressive planning and planning with action costs, two advanced features beyond classical planning. The resulting planner is competitive to state-of-the-art planners, in terms of both quality and performance. Overall, our work strongly suggests a shifting trend of planning from STRIPS to SAS+, and shows the power of formulating planning problems as Satisfiability. Given the important roles of both classical planning and temporal planning, our work will inspire new developments in other advanced planning problem domains

    Short Term Unit Commitment as a Planning Problem

    Get PDF
    ‘Unit Commitment’, setting online schedules for generating units in a power system to ensure supply meets demand, is integral to the secure, efficient, and economic daily operation of a power system. Conflicting desires for security of supply at minimum cost complicate this. Sustained research has produced methodologies within a guaranteed bound of optimality, given sufficient computing time. Regulatory requirements to reduce emissions in modern power systems have necessitated increased renewable generation, whose output cannot be directly controlled, increasing complex uncertainties. Traditional methods are thus less efficient, generating more costly schedules or requiring impractical increases in solution time. Meta-Heuristic approaches are studied to identify why this large body of work has had little industrial impact despite continued academic interest over many years. A discussion of lessons learned is given, and should be of interest to researchers presenting new Unit Commitment approaches, such as a Planning implementation. Automated Planning is a sub-field of Artificial Intelligence, where a timestamped sequence of predefined actions manipulating a system towards a goal configuration is sought. This differs from previous Unit Commitment formulations found in the literature. There are fewer times when a unit’s online status switches, representing a Planning action, than free variables in a traditional formulation. Efficient reasoning about these actions could reduce solution time, enabling Planning to tackle Unit Commitment problems with high levels of renewable generation. Existing Planning formulations for Unit Commitment have not been found. A successful formulation enumerating open challenges would constitute a good benchmark problem for the field. Thus, two models are presented. The first demonstrates the approach’s strength in temporal reasoning over numeric optimisation. The second balances this but current algorithms cannot handle it. Extensions to an existing algorithm are proposed alongside a discussion of immediate challenges and possible solutions. This is intended to form a base from which a successful methodology can be developed
    corecore