14,095 research outputs found

    Automating Vehicles by Deep Reinforcement Learning using Task Separation with Hill Climbing

    Full text link
    Within the context of autonomous driving a model-based reinforcement learning algorithm is proposed for the design of neural network-parameterized controllers. Classical model-based control methods, which include sampling- and lattice-based algorithms and model predictive control, suffer from the trade-off between model complexity and computational burden required for the online solution of expensive optimization or search problems at every short sampling time. To circumvent this trade-off, a 2-step procedure is motivated: first learning of a controller during offline training based on an arbitrarily complicated mathematical system model, before online fast feedforward evaluation of the trained controller. The contribution of this paper is the proposition of a simple gradient-free and model-based algorithm for deep reinforcement learning using task separation with hill climbing (TSHC). In particular, (i) simultaneous training on separate deterministic tasks with the purpose of encoding many motion primitives in a neural network, and (ii) the employment of maximally sparse rewards in combination with virtual velocity constraints (VVCs) in setpoint proximity are advocated.Comment: 10 pages, 6 figures, 1 tabl

    Norm Optimal Iterative Learning Control with Application to Problems in Accelerator based Free Electron Lasers and Rehabilitation Robotics

    No full text
    This paper gives an overview of the theoretical basis of the norm optimal approach to iterative learning control followed by results that describe more recent work which has experimentally benchmarking the performance that can be achieved. The remainder of then paper then describes its actual application to a physical process and a very novel application in stroke rehabilitation

    Robust nonlinear control of vectored thrust aircraft

    Get PDF
    An interdisciplinary program in robust control for nonlinear systems with applications to a variety of engineering problems is outlined. Major emphasis will be placed on flight control, with both experimental and analytical studies. This program builds on recent new results in control theory for stability, stabilization, robust stability, robust performance, synthesis, and model reduction in a unified framework using Linear Fractional Transformations (LFT's), Linear Matrix Inequalities (LMI's), and the structured singular value micron. Most of these new advances have been accomplished by the Caltech controls group independently or in collaboration with researchers in other institutions. These recent results offer a new and remarkably unified framework for all aspects of robust control, but what is particularly important for this program is that they also have important implications for system identification and control of nonlinear systems. This combines well with Caltech's expertise in nonlinear control theory, both in geometric methods and methods for systems with constraints and saturations

    Distributed Estimation with Information-Seeking Control in Agent Network

    Get PDF
    We introduce a distributed, cooperative framework and method for Bayesian estimation and control in decentralized agent networks. Our framework combines joint estimation of time-varying global and local states with information-seeking control optimizing the behavior of the agents. It is suited to nonlinear and non-Gaussian problems and, in particular, to location-aware networks. For cooperative estimation, a combination of belief propagation message passing and consensus is used. For cooperative control, the negative posterior joint entropy of all states is maximized via a gradient ascent. The estimation layer provides the control layer with probabilistic information in the form of sample representations of probability distributions. Simulation results demonstrate intelligent behavior of the agents and excellent estimation performance for a simultaneous self-localization and target tracking problem. In a cooperative localization scenario with only one anchor, mobile agents can localize themselves after a short time with an accuracy that is higher than the accuracy of the performed distance measurements.Comment: 17 pages, 10 figure

    New control strategies for neuroprosthetic systems

    Get PDF
    The availability of techniques to artificially excite paralyzed muscles opens enormous potential for restoring both upper and lower extremity movements with\ud neuroprostheses. Neuroprostheses must stimulate muscle, and control and regulate the artificial movements produced. Control methods to accomplish these tasks include feedforward (open-loop), feedback, and adaptive control. Feedforward control requires a great deal of information about the biomechanical behavior of the limb. For the upper extremity, an artificial motor program was developed to provide such movement program input to a neuroprosthesis. In lower extremity control, one group achieved their best results by attempting to meet naturally perceived gait objectives rather than to follow an exact joint angle trajectory. Adaptive feedforward control, as implemented in the cycleto-cycle controller, gave good compensation for the gradual decrease in performance observed with open-loop control. A neural network controller was able to control its system to customize stimulation parameters in order to generate a desired output trajectory in a given individual and to maintain tracking performance in the presence of muscle fatigue. The authors believe that practical FNS control systems must\ud exhibit many of these features of neurophysiological systems
    corecore