30 research outputs found

    Distributed optimization for control and learning

    Get PDF
    Large scale multi-agent networked systems are becoming increasingly popular in industry and academia as they can be applied to represent systems in diverse application areas, such as intelligent surveillance and reconnaissance, mobile robotics, transportation networks and complex buildings. In such systems, issues related to control and learning have been significant technical challenges to affect system performance and overall cost. While centralized optimization approaches have been widely used by the engineering and computer science community, advanced and effective distributed optimization techniques have not been explored sufficiently and thoroughly in this regard. This study explores various categories of centralized and distributed optimization methods that have been applied or may be applicable for diverse engineering and science problems. The performance of centralized or distributed optimization schemes significantly depends on various factors including the types of objective functions, constraints, step sizes, and communication networks, etc. In this context, the focus of this dissertation is towards developing novel distributed optimization algorithms in order to solve challenging control and learning problems in various domains such as large-scale building energy systems and robotic networks. Specifically, we develop a generalized gossip-based subgradient method for solving distributed optimization problems in large-scale networked systems, e.g., larger-scale commercial building energy systems. Different from previous work, a user-defined control parameter is introduced to control a spectrum from globally optimal solution to suboptimal solutions and the trade-off between the solution accuracy and temporal convergence. We test and validate our proposed algorithm on a real testbed involving multiple zones incorporating a distributed control and sensing platform. In addition, we extend the distributed optimization to the deep learning area for solving an emerging topic, i.e., distributed deep learning, in fixed topology networks. While some previous work exists on this topic, the data parallelism and distributed computation are still not sufficiently explored. Therefore, we propose a class of distributed deep learning methods to tackle such issues by combining the consensus protocol and stochastic gradient descent approach. Moreover, to address the consensus-optimality trade-offs in distributed convex and nonconvex optimization, especially in deep learning when the training datasets for agents are non-balanced (non-iid), we propose and develop new approaches in this research, namely, incremental consensus-based distributed stochastic gradient descent and generalized consensus-based distributed (stochastic) gradient descent approach

    Simulation-Based Approximate Policy Gradient and Its Building Control Application

    Get PDF
    The goal of this paper is to study the potential applicability of a stochastic approximationbased policy gradient method for optimal office building HVAC (Heating, Ventilation, and Air Conditioning) control systems. A real-world building thermal dynamics with occupant interactions is the main focus of this paper. It is a complex stochastic system in the sense that its statistical properties depend on its state variables. In this case, existing approaches, for instance, stochastic model predictive control methods, cannot be applied to optimal control designs. As a remedy, we approximate the gradient of the cost function using simulations and use a gradient descent type algorithm to design a suboptimal control policy. We assess its performance through a simulation study of building HVAC systems

    Toward optimal operation of multienergy home-microgrids for power balancing in distribution networks: a model predictive control approach

    Get PDF
    The energy policy objectives of the German government regarding renewable energy sources and energy efficiency will lead to a significantly increase in the share of photovoltaics, storage systems, CHP plants, and heat pumps, especially at the distribution grid level. In the future, inside a household, such systems must be coordinated in such a way that they can respond to variable network conditions as a single flexible unit. This dissertation defines home-microgrid as a residential building with integrated distributed energy resources, and follows a bottom-up approach, based on the cellular approach, which aims at improving local balancing in low-voltage grids by using the flexibilities of home-microgrids. For this purpose, the dissertation develops optimization-based strategies for the coordination of multienergy home-microgrids, focusing on the use of model predictive control. The main core of the work is the formulation of the underlying optimization problems and the investigation of coordination strategies for interconnected home-microgrids. In this context, the work presents the use of the dual decomposition and the alternating direction method of multipliers for hierarchical-distributed coordination strategies. Finally, this dissertation introduces a framework for the co-simulation of electrical networks with penetration of multienergy home-microgrids

    Co-design for Multi-subsystem and Vehicle Routing-and-Control Problems

    Get PDF
    Co-design refers to the process of integrating optimization of the physical design with a controller for a system. The challenge in co-design is that the optimization is simultaneously applied to both static/time-invariant (e.g., physical plant design) variables and dynamic/time-variant (e.g., state and control) variables, which can be coupled with each other. The objective of this dissertation is to explore new formulations and approaches in co-design for multi-subsystem and vehicle routing-and-control problems. Specifically, four research questions are considered and resolved. In Research Question 1 (RQ1), the critical issue is how to formulate a class of multi-subsystem co-design problems with convex physical design subproblems and linear quadratic regulator control subproblems, and construct a decentralized solution approach for such problems. In Research Question 2 (RQ2), solution methods for a broader class of multi-subsystem co-design problems than those considered in RQ1 are investigated. In Research Question 3 (RQ3), the question is whether, in the context of co-design, the combined routing and control costs of a fleet of vehicles can be improved if optimal control is introduced into the routing. Finally, an extension of RQ3 is considered in Research Question 4 (RQ4), where the possibility of constructing an integrated vehicle routing-and-control problem with load-dependent dynamics is investigated. Beyond the articles published by the author of this dissertation, the proposed research questions, models and methods presented have not been considered elsewhere in the literature

    Provision of energy and regulation reserve services by buildings

    Full text link
    Power consumption and generation in the electrical grid must be balanced at all times. This balance is maintained by the grid operator through the procurement of energy and regulation reserve services in the wholesale electricity market. Traditionally, these services could only be procured from generation resources. However, helped by the advances in the computational and communication infrastructure, the demand resources are increasingly being leveraged in this regard. In particular, the Heating, Ventilation and Air-Conditioning (HVAC) systems of buildings are gaining traction due to the consumption flexibility provided by their energy-shifting and fast-response capabilities. The provision of energy and regulation reserve services in the wholesale market, from the perspective of a typical building’s HVAC system, can be construed in terms of two synergistic problems: an hourly deterministic optimization problem, referred to as Scheduling Problem, and a real-time (seconds timescale) stochastic control problem, referred to as Deployment Problem. So far, the existing literature has synthesized the solutions of these two problems in a simplistic, sequential/iterative manner without employing an integrated approach that captures explicitly their cost and constraint interactions. Moreover, the deployment problem has only been solved with classical controllers which are not optimal, whereas the non-convexities in the scheduling problem have been addressed with methods that are sensitive to initialization. The current approaches therefore do not fully optimize the decisions of the two problems either individually or collectively, and hence do not fully exploit the HVAC resource. The goal of the proposed research is to have optimal decision-making across both the scheduling and deployment problems. Our approach proposes deriving an optimal control policy for the deployment problem, and expressing the corresponding expected sum of deployment costs over the hour (called ‘expected intra-hour costs’) as a closed-form analytic function of the scheduling decisions. The inclusion of these expected intra-hour costs into the scheduling problem allows the optimization of the hourly scheduling decisions, pursuant to the real-time use of the optimal deployment control policy. Thus, our approach captures the interaction of the two problems and optimizes decisions across timescales yielding a truly integrated solution. We investigate the estimation of the expected intra-hour costs (based on a myopic policy optimizing instantaneous tracking error and utility loss), and solve the integrated problem with tight relaxations, demonstrating the value and applicability of the approach. Further, we investigate the derivation of the optimal control policy for the deployment problem, formulating and solving it as discounted-cost infinite horizon Dynamic Program (DP) and Reinforcement Learning (RL) problems. We also characterize the optimal policy as a closed-form analytic mapping from state-space to action-space, and obtain the corresponding expected cost-to-go (a.k.a. expected intra-hour costs) as a closed-form analytic function of the scheduling decisions. We further illustrate that the optimal policy better captures the deployment costs compared to existing approaches. As such, our work represents a structured, interpretable and automated way for the end-to-end consideration of energy and regulation reserve market participation, and can be extended to other demand side resources

    Leveraging deep reinforcement learning in the smart grid environment

    Full text link
    L’apprentissage statistique moderne démontre des résultats impressionnants, où les or- dinateurs viennent à atteindre ou même à excéder les standards humains dans certaines applications telles que la vision par ordinateur ou les jeux de stratégie. Pourtant, malgré ces avancées, force est de constater que les applications fiables en déploiement en sont encore à leur état embryonnaire en comparaison aux opportunités qu’elles pourraient apporter. C’est dans cette perspective, avec une emphase mise sur la théorie de décision séquentielle et sur les recherches récentes en apprentissage automatique, que nous démontrons l’applica- tion efficace de ces méthodes sur des cas liés au réseau électrique et à l’optimisation de ses acteurs. Nous considérons ainsi des instances impliquant des unités d’emmagasinement éner- gétique ou des voitures électriques, jusqu’aux contrôles thermiques des bâtiments intelligents. Nous concluons finalement en introduisant une nouvelle approche hybride qui combine les performances modernes de l’apprentissage profond et de l’apprentissage par renforcement au cadre d’application éprouvé de la recherche opérationnelle classique, dans le but de faciliter l’intégration de nouvelles méthodes d’apprentissage statistique sur différentes applications concrètes.While modern statistical learning is achieving impressive results, as computers start exceeding human baselines in some applications like computer vision, or even beating pro- fessional human players at strategy games without any prior knowledge, reliable deployed applications are still in their infancy compared to what these new opportunities could fathom. In this perspective, with a keen focus on sequential decision theory and recent statistical learning research, we demonstrate efficient application of such methods on instances involving the energy grid and the optimization of its actors, from energy storage and electric cars to smart buildings and thermal controls. We conclude by introducing a new hybrid approach combining the modern performance of deep learning and reinforcement learning with the proven application framework of operations research, in the objective of facilitating seamlessly the integration of new statistical learning-oriented methodologies in concrete applications
    corecore