788 research outputs found
Dynamic Controllability of Temporally-flexible Reactive Programs
In this paper we extend dynamic controllability of temporally-flexible plans to temporally-flexible reactive programs. We consider three reactive programming language constructs whose behavior depends on runtime observations; conditional execution, iteration, and exception handling. Temporally-flexible reactive programs are distinguished from temporally-flexible plans in that program execution is conditioned on the runtime state of the world. In addition, exceptions are thrown and caught at runtime in response to violated timing constraints, and handled exceptions are considered successful program executions. Dynamic controllability corresponds to a guarantee that a program will execute to completion, despite runtime constraint violations and uncertainty in runtime state. An algorithm is developed which frames the dynamic controllability problem as an AND/OR search tree over possible program executions. A key advantage of this approach is the ability to enumerate only a subset of possible program executions that guarantees dynamic controllability, framed as an AND/OR solution subtree
Optimal Ensemble Control of Loads in Distribution Grids with Network Constraints
Flexible loads, e.g. thermostatically controlled loads (TCLs), are
technically feasible to participate in demand response (DR) programs. On the
other hand, there is a number of challenges that need to be resolved before it
can be implemented in practice en masse. First, individual TCLs must be
aggregated and operated in sync to scale DR benefits. Second, the uncertainty
of TCLs needs to be accounted for. Third, exercising the flexibility of TCLs
needs to be coordinated with distribution system operations to avoid
unnecessary power losses and compliance with power flow and voltage limits.
This paper addresses these challenges. We propose a network-constrained,
open-loop, stochastic optimal control formulation. The first part of this
formulation represents ensembles of collocated TCLs modelled by an aggregated
Markov Process (MP), where each MP state is associated with a given power
consumption or production level. The second part extends MPs to a multi-period
distribution power flow optimization. In this optimization, the control of TCL
ensembles is regulated by transition probability matrices and physically
enabled by local active and reactive power controls at TCL locations. The
optimization is solved with a Spatio-Temporal Dual Decomposition (ST-D2)
algorithm. The performance of the proposed formulation and algorithm is
demonstrated on the IEEE 33-bus distribution model.Comment: 7 pages, 6 figures, accepted PSCC 201
Risk-minimizing program execution in robotic domains
Thesis (Sc. D.)--Massachusetts Institute of Technology, Dept. of Aeronautics and Astronautics, 2012.Cataloged from PDF version of thesis.Includes bibliographical references (p. 153-161).In this thesis, we argue that autonomous robots operating in hostile and uncertain environments can improve robustness by computing and reasoning explicitly about risk. Autonomous robots with a keen sensitivity to risk can be trusted with critical missions, such as exploring deep space and assisting on the battlefield. We introduce a novel, risk-minimizing approach to program execution that utilizes program flexibility and estimation of risk in order to make runtime decisions that minimize the probability of program failure. Our risk-minimizing executive, called Murphy, utilizes two forms of program flexibility, 1) flexible scheduling of activity timing, and 2) redundant choice between subprocedures, in order to minimize two forms of program risk, 1) exceptions arising from activity failures, and 2) exceptions arising from timing constraint violations in a program. Murphy takes two inputs, a program written in a nondeterministic variant of the Reactive Model-based Programming Language (RMPL) and a set of stochastic activity failure models, one for each activity in a program, and computes two outputs, a risk-minimizing decision policy and value function. The decision policy informs Murphy which decisions to make at runtime in order to minimize risk, while the value function quantifies risk. In order to execute with low latency, Murphy computes the decision policy and value function offline, as a compilation step prior to program execution. In this thesis, we develop three approaches to RMPL program execution. First, we develop an approach that is guaranteed to minimize risk. For this approach, we reason probabilistically about risk by framing program execution as a Markov Decision Process (MDP). Next, we develop an approach that avoids risk altogether. For this approach, we frame program execution as a novel form of constraint-based temporal reasoning. Finally, we develop an execution approach that trades optimality in risk avoidance for tractability. For this approach, we leverage prior work in hierarchical decomposition of MDPs in order to mitigate complexity. We benchmark the tractability of each approach on a set of representative RMPL programs, and we demonstrate the applicability of the approach on a humanoid robot simulator.by Robert T. Effinger, IV.Sc.D
Optimising Flexibility of Temporal Problems with Uncertainty
Temporal networks have been applied in many autonomous systems.
In real situations, we cannot ignore the uncertain factors when
using those autonomous systems. Achieving robust schedules and
temporal plans by optimising flexibility to tackle the
uncertainty is the motivation of the thesis.
This thesis focuses on the optimisation problems of temporal
networks with uncertainty and controllable options in the field
of Artificial Intelligence Planning and Scheduling. The goal of
this thesis is to construct flexibility and robustness metrics
for temporal networks under the constraints of different levels
of controllability. Furthermore, optimising flexibility for
temporal plans and schedules to achieve robust solutions with
flexible executions.
When solving temporal problems with uncertainty, postponing
decisions according to the observations of uncertain events
enables flexible strategies as the solutions instead of fixed
schedules or plans. Among the three levels of controllability of
the Simple Temporal Problem with Uncertainty (STPU), a problem is
dynamically controllable if there is a successful dynamic
strategy such that every decision in it is made according to the
observations of past events.
In the thesis, we make the following contributions. (1) We
introduce an optimisation model for STPU based on the existing
dynamic controllability checking algorithms. Some flexibility and
robustness measures are introduced based on the model. (2) We
extend the definition and verification algorithm of dynamic
controllability to temporal problems with controllable discrete
variables and uncertainty, which is called Controllable
Conditional Temporal Problems with Uncertainty (CCTPU). An
entirely dynamically controllable strategy of CCTPU consists of
both temporal scheduling and variable assignments being
dynamically decided, which maximize the flexibility of the
execution. (3) We introduce optimisation models of CCTPU under
fully dynamic controllability. The optimisation models aim to
answer the questions how flexible, robust or controllable a
schedule or temporal plan is. The experiments show that making
decisions dynamically can achieve better objective values than
doing statically.
The thesis also contributes to the field of AI planning and
scheduling by introducing robustness metrics of temporal
networks, proposing an envelope-based algorithm that can check
dynamic controllability of temporal networks with uncertainty and
controllable discrete decisions, evaluating improvements from
making decisions strongly controllable to temporally dynamically
controllable and fully dynamically controllable and comparing the
runtime of different implementations to present the scalability
of dynamically controllable strategies
Dynamic Controllability Made Simple
Simple Temporal Networks with Uncertainty (STNUs) are a well-studied model for representing temporal constraints, where some intervals (contingent links) have an unknown but bounded duration, discovered only during execution. An STNU is dynamically controllable (DC) if there exists a strategy to execute its time-points satisfying all the constraints, regardless of the actual duration of contingent links revealed during execution.
In this work we present a new system of constraint propagation rules for STNUs, which is sound-and-complete for DC checking. Our system comprises just three rules which, differently from the ones proposed in all previous works, only generate unconditioned constraints. In particular, after applying our sound rules, the network remains an STNU in all respects. Moreover, our completeness proof is short and non-algorithmic, based on the explicit construction of a valid execution strategy. This is a substantial simplification of the theory which underlies all the polynomial-time algorithms for DC-checking.
Our analysis also shows: (1) the existence of late execution strategies for STNUs, (2) the equivalence of several variants of the notion of DC, (3) the existence of a fast algorithm for real-time execution of STNUs, which runs in O(KN) total time in a network with K contingent links and N time points, considerably improving the previous O(N^3)-time bound
LTLf and LDLf Synthesis under Partial Observability
In this paper, we study synthesis under partial observability for logical specifications over finite traces expressed in LTLf/LDLf. This form of synthesis can be seen as a generalization of planning under partial observability in nondeterministic domains, which is known to be 2EXPTIME-complete. We start by showing that the usual "belief-state construction" used in planning under partial observability works also for general LTLf/LDLf synthesis, though with a jump in computational complexity from 2EXPTIME to 3EXPTIME. Then we show that the belief-state construction can be avoided in favor of a direct automata construction which exploits projection to hide unobservable propositions. This allow us to prove that the problem remains 2EXPTIME-complete. The new synthesis technique proposed is effective and readily implementable
Temporal and Hierarchical Models for Planning and Acting in Robotics
The field of AI planning has seen rapid progress over the last decade and planners are now able to find plan with hundreds of actions in a matter of seconds. Despite those important progresses, robotic systems still tend to have a reactive architecture with very little deliberation on the course of the plan they might follow. In this thesis, we argue that a successful integration with a robotic system requires the planner to have capacities for both temporal and hierarchical reasoning. The former is indeed a universal resource central in many robot activities while the latter is a critical component for the integration of reasoning capabilities at different abstraction levels, typically starting with a high level view of an activity that is iteratively refined down to motion primitives. As a first step to carry out this vision, we present a model for temporal planning unifying the generative and hierarchical approaches. At the center of the model are temporal action templates, similar to those of PDDL complemented with a specification of the initial state as well as the expected evolution of the environment over time. In addition, our model allows for the specification of hierarchical knowledge possibly with a partial coverage. Consequently, our model generalizes the existing generative and HTN approaches together with an explicit time representation. In the second chapter, we introduce a planning procedure suitable for our planning model. In order to support hierarchical features, we extend the existing Partial-Order Causal Link approach used in many constraintbased planners, with the notions of task and decomposition. We implement it in FAPE (Flexible Acting and Planning Environment) together with automated problem analysis techniques used for search guidance. We show FAPE to have performance similar to state of the art temporal planners when used in a generative setting. The addition of hierarchical information leads to further performance gain and allows us to outperform traditional planners. In the third chapter, we study the usual methods used to reason on temporal uncertainty while planning. We relax the usual assumption of total observability and instead provide techniques to reason on the observations needed to maintain a plan dispatchable. We show how such needed observations can be detected at planning time and incrementally dealt with by considering the appropriate sensing actions. In a final chapter, we discuss the place of the proposed planning system as a central component for the control of a robotic actor. We demonstrate how the explicit time representation facilitates plan monitoring and action dispatching when dealing with contingent events that require observation. We take advantage of the constraint-based and hierarchical representation to facilitate both plan-repair procedures as well opportunistic plan refinement at acting time
- …