74,534 research outputs found
Neural Feedback Scheduling of Real-Time Control Tasks
Many embedded real-time control systems suffer from resource constraints and
dynamic workload variations. Although optimal feedback scheduling schemes are
in principle capable of maximizing the overall control performance of
multitasking control systems, most of them induce excessively large
computational overheads associated with the mathematical optimization routines
involved and hence are not directly applicable to practical systems. To
optimize the overall control performance while minimizing the overhead of
feedback scheduling, this paper proposes an efficient feedback scheduling
scheme based on feedforward neural networks. Using the optimal solutions
obtained offline by mathematical optimization methods, a back-propagation (BP)
neural network is designed to adapt online the sampling periods of concurrent
control tasks with respect to changes in computing resource availability.
Numerical simulation results show that the proposed scheme can reduce the
computational overhead significantly while delivering almost the same overall
control performance as compared to optimal feedback scheduling.Comment: To appear in International Journal of Innovative Computing,
Information and Contro
Optimal control of nonlinear partially-unknown systems with unsymmetrical input constraints and its applications to the optimal UAV circumnavigation problem
Aimed at solving the optimal control problem for nonlinear systems with
unsymmetrical input constraints, we present an online adaptive approach for
partially unknown control systems/dynamics. The designed algorithm converges
online to the optimal control solution without the knowledge of the internal
system dynamics. The optimality of the obtained control policy and the
stability for the closed-loop dynamic optimality are proved theoretically. The
proposed method greatly relaxes the assumption on the form of the internal
dynamics and input constraints in previous works. Besides, the control design
framework proposed in this paper offers a new approach to solve the optimal
circumnavigation problem involving a moving target for a fixed-wing unmanned
aerial vehicle (UAV). The control performance of our method is compared with
that of the existing circumnavigation control law in a numerical simulation and
the simulation results validate the effectiveness of our algorithm
Certified Reinforcement Learning with Logic Guidance
This paper proposes the first model-free Reinforcement Learning (RL)
framework to synthesise policies for unknown, and continuous-state Markov
Decision Processes (MDPs), such that a given linear temporal property is
satisfied. We convert the given property into a Limit Deterministic Buchi
Automaton (LDBA), namely a finite-state machine expressing the property.
Exploiting the structure of the LDBA, we shape a synchronous reward function
on-the-fly, so that an RL algorithm can synthesise a policy resulting in traces
that probabilistically satisfy the linear temporal property. This probability
(certificate) is also calculated in parallel with policy learning when the
state space of the MDP is finite: as such, the RL algorithm produces a policy
that is certified with respect to the property. Under the assumption of finite
state space, theoretical guarantees are provided on the convergence of the RL
algorithm to an optimal policy, maximising the above probability. We also show
that our method produces ''best available'' control policies when the logical
property cannot be satisfied. In the general case of a continuous state space,
we propose a neural network architecture for RL and we empirically show that
the algorithm finds satisfying policies, if there exist such policies. The
performance of the proposed framework is evaluated via a set of numerical
examples and benchmarks, where we observe an improvement of one order of
magnitude in the number of iterations required for the policy synthesis,
compared to existing approaches whenever available.Comment: This article draws from arXiv:1801.08099, arXiv:1809.0782
- …