119 research outputs found

    OPTIMIZATION MODELS AND METHODOLOGIES TO SUPPORT EMERGENCY PREPAREDNESS AND POST-DISASTER RESPONSE

    Get PDF
    This dissertation addresses three important optimization problems arising during the phases of pre-disaster emergency preparedness and post-disaster response in time-dependent, stochastic and dynamic environments. The first problem studied is the building evacuation problem with shared information (BEPSI), which seeks a set of evacuation routes and the assignment of evacuees to these routes with the minimum total evacuation time. The BEPSI incorporates the constraints of shared information in providing on-line instructions to evacuees and ensures that evacuees departing from an intermediate or source location at a mutual point in time receive common instructions. A mixed-integer linear program is formulated for the BEPSI and an exact technique based on Benders decomposition is proposed for its solution. Numerical experiments conducted on a mid-sized real-world example demonstrate the effectiveness of the proposed algorithm. The second problem addressed is the network resilience problem (NRP), involving an indicator of network resilience proposed to quantify the ability of a network to recover from randomly arising disruptions resulting from a disaster event. A stochastic, mixed integer program is proposed for quantifying network resilience and identifying the optimal post-event course of action to take. A solution technique based on concepts of Benders decomposition, column generation and Monte Carlo simulation is proposed. Experiments were conducted to illustrate the resilience concept and procedure for its measurement, and to assess the role of network topology in its magnitude. The last problem addressed is the urban search and rescue team deployment problem (USAR-TDP). The USAR-TDP seeks an optimal deployment of USAR teams to disaster sites, including the order of site visits, with the ultimate goal of maximizing the expected number of saved lives over the search and rescue period. A multistage stochastic program is proposed to capture problem uncertainty and dynamics. The solution technique involves the solution of a sequence of interrelated two-stage stochastic programs with recourse. A column generation-based technique is proposed for the solution of each problem instance arising as the start of each decision epoch over a time horizon. Numerical experiments conducted on an example of the 2010 Haiti earthquake are presented to illustrate the effectiveness of the proposed approach

    Scaling reinforcement learning to the unconstrained multi-agent domain

    Get PDF
    Reinforcement learning is a machine learning technique designed to mimic the way animals learn by receiving rewards and punishment. It is designed to train intelligent agents when very little is known about the agent’s environment, and consequently the agent’s designer is unable to hand-craft an appropriate policy. Using reinforcement learning, the agent’s designer can merely give reward to the agent when it does something right, and the algorithm will craft an appropriate policy automatically. In many situations it is desirable to use this technique to train systems of agents (for example, to train robots to play RoboCup soccer in a coordinated fashion). Unfortunately, several significant computational issues occur when using this technique to train systems of agents. This dissertation introduces a suite of techniques that overcome many of these difficulties in various common situations. First, we show how multi-agent reinforcement learning can be made more tractable by forming coalitions out of the agents, and training each coalition separately. Coalitions are formed by using information-theoretic techniques, and we find that by using a coalition-based approach, the computational complexity of reinforcement-learning can be made linear in the total system agent count. Next we look at ways to integrate domain knowledge into the reinforcement learning process, and how this can signifi-cantly improve the policy quality in multi-agent situations. Specifically, we find that integrating domain knowledge into a reinforcement learning process can overcome training data deficiencies and allow the learner to converge to acceptable solutions when lack of training data would have prevented such convergence without domain knowledge. We then show how to train policies over continuous action spaces, which can reduce problem complexity for domains that require continuous action spaces (analog controllers) by eliminating the need to finely discretize the action space. Finally, we look at ways to perform reinforcement learning on modern GPUs and show how by doing this we can tackle significantly larger problems. We find that by offloading some of the RL computation to the GPU, we can achieve almost a 4.5 speedup factor in the total training process

    Engineering coordination : eine Methodologie für die Koordination von Planungssystemen

    Get PDF
    Planning problems, like real-world planning and scheduling problems, are complex tasks. As an efficient strategy for handing such problems is the ‘divide and conquer’ strategy has been identified. Each sub problem is then solved independently. Typically the sub problems are solved in a linear way. This approach enables the generation of sub-optimal plans for a number of real world problems. Today, this approach is widely accepted and has been established e.g. in the organizational structure of companies. But existing interdependencies between the sub problems are not sufficiently regarded, as each problem are solved sequentially and no feedback information is given. The field of coordination has been covered by a number of academic fields, like the distributed artificial intelligence, economics or game theory. An important result is, that there exist no method that leads to optimal results in any given coordination problem. Consequently, a suitable coordination mechanism has to be identified for each single coordination problem. Up to now, there exists no process for the selection of a coordination mechanism, neither in the engineering of distributed systems nor in agent oriented software engineering. Within the scope of this work the ECo process is presented, that address exactly this selection problem. The Eco process contains the following five steps. • Modeling of the coordination problem • Defining the coordination requirements • Selection / Design of the coordination mechanism • Implementation • Evaluation Each of these steps is detailed in the thesis. The modeling has to be done to enable a systemic analysis of the coordination problem. Coordination mechanisms have to respect the given situation and the context in which the coordination has to be done. The requirements imposed by the context of the coordination problem are formalized in the coordination requirements. The selection process is driven by these coordination requirements. Using the requirements as a distinction for the selection of a coordination mechanism is a central aspect of this thesis. Additionally these requirements can be used for documentation of design decisions. Therefore, it is reasonable to annotate the coordination mechanisms with the coordination requirements they fulfill and fail to ease the selection process, for a given situation. For that reason we present a new classification scheme for coordination methods within this thesis that classifies existing coordination methods according to a set of criteria that has been identified as important for the distinction between different coordination methods. The implementation phase of the ECo process is supported by the CoPS process and CoPS framework that has been developed within this thesis, as well. The CoPS process structures the design making that has to be done during the implementation phase. The CoPS framework provides a set of basic features software agents need for realizing the selected coordination method. Within the CoPS process techniques are presented for the design and implementation of conversations between agents that can be applied not only within the context of the coordination of planning systems, but for multiagent systems in general. The ECo-CoPS approach has been successfully validated in two case studies from the logistic domain.Reale Planungsprobleme, wie etwa die Produktionsplanung in einer Supply Chain, sind komplex Planungsprobleme. Eine übliche Strategie derart komplexen Problemen zu lösen, ist es diese Probleme in einfachere Teilprobleme zu zerlegen und diese dann separat, meist sequentiell, zu lösen (divide-and-conquer Strategie). Dieser Ansatz erlaubt die Erstellung von (suboptimalen) Plänen für eine Reihe von realen Anwendungen, und ist heute in den Organisationsstrukturen von größeren Unternehmen institutionalisiert worden. Allerdings werden Abhängigkeiten zwischen den Teilproblemen nicht ausreichend berücksichtigt, da die Partialprobleme sequentiell ohne Feedback gelöst werden. Die erstellten Teillösungen müssen deswegen oft nachträglich koordiniert werden. Das Gebiet der Koordination wird in verschiedenen Forschungsgebieten, wie etwa der verteilten Künstlichen Intelligenz, den Wirtschaftswissenschaften oder der Spieltheorie untersucht. Ein zentrales Ergebnis dieser Forschung ist, dass es keinen für alle Situationen geeigneten Koordinationsmechanismus gibt. Es stellt sich also die Aufgabe aus den zahlreichen vorgeschlagenen Koordinationsmechanismen eine Auswahl zu treffen, die für die aktuelle Situation den geeigneten Mechanismus identifiziert. Für die Auswahl eines solchen Mechanismus existiert bisher jedoch kein strukturiertes Verfahren für die Entwicklung von verteilten Systems und insbesondere im Bereich der Agenten orientierter Softwareentwicklung. Im Rahmen dieser Arbeit wird genau hierfür ein Verfahren vorgestellt, der ECo-Prozess. Mit Hilfe dieses Prozesses wird der Auswahlprozess in die folgenden Schritte eingeteilt: • Modellierung der Problemstellung und des relevante Kontextes • Formulierung von Anforderungen an einen Koordinationsmechanismus (coordination requirements) • Auswahl/Entwurf eines Koordinationsmechanismuses • Implementierung des Koordinationsverfahrens • Evaluation des Koordinationsverfahrens Diese Schritte werden im Rahmen der vorliegenden Arbeit detailliert beschrieben. Die Modellierung der Problemstellung stellt dabei den ersten Schritt dar, um die Problemstellung analytisch zugänglich zu machen. Koordinationsverfahren müssen die Gegebenheiten, den Kontext und die Domäne, in der sie angewendet werden sollen hinreichend berücksichtigen um anwendbar zu sein. Dieses kann über Anforderungen an den Koordinationsprozess formalisiert werden. Der von den Anforderungen getrieben Auswahlprozess ist ein Kernstück der hier vorgestellten Arbeit. Durch die Formulierung der Anforderungen und der Annotation eines Koordinationsmechanismus bezüglich der erfüllten und nicht erfüllten Anforderungen werden die Motive für Designentscheidungen dieses Verfahren expliziert. Wenn Koordinationsverfahren anhand dieser Anforderungen klassifiziert werden können, ist es weiterhin möglich den Auswahlprozess (unabhängig vom ECo-Ansatz) zu vereinfachen und zu beschleunigen. Im Rahmen dieser Arbeit wird eine Klassifikation von Koordinationsansätzen anhand von allgemeinen Kriterien vorgestellt, die die Identifikation von geeigneten Kandidaten erleichtern. Diese Kandidaten können dann detaillierter untersucht werden. Dies wurde in den vorgestellten Fallstudien erfolgreich demonstriert. Für die Unterstützung der Implementierung eines Koordinationsansatzes wird in dieser Arbeit zusätzlich der CoPS Prozess vorgeschlagen. Der CoPS Prozess erlaubt einen ganzheitlichen systematischen Ansatz für den Entwurf und die Implementierung eines Koordinationsverfahrens. Unterstürzt wird der CoPS Prozess durch das CoPS Framework, das die Implementierung erleichtert, indem es als eine Plattform mit Basisfunktionalität eines Agenten bereitstellt, der für die Koordination von Planungssystemen verantwortlich ist. Im Rahmen des CoPS Verfahrens werden Techniken für den Entwurf und die Implementierung von Konversation im Kontext des agenten-orientiertem Software Engineerings ausführlich behandelt. Der Entwurf von Konversationen geht dabei weit über Fragestellung der Formatierung von Nachrichten hinaus, wie dies etwa in den FIPA Standards geregelt ist, und ist für die Implementierung von agentenbasierten Systemen im Allgemeinen von Bedeutung. Die Funktionsweise des ECo-CoPS Ansatzes wird anhand von zweierfolgreich durchgeführten Fallstudien aus dem betriebswirtschaftlichen Kontext vorgestellt

    Working Notes from the 1992 AAAI Workshop on Automating Software Design. Theme: Domain Specific Software Design

    Get PDF
    The goal of this workshop is to identify different architectural approaches to building domain-specific software design systems and to explore issues unique to domain-specific (vs. general-purpose) software design. Some general issues that cut across the particular software design domain include: (1) knowledge representation, acquisition, and maintenance; (2) specialized software design techniques; and (3) user interaction and user interface

    Tools and Algorithms for the Construction and Analysis of Systems

    Get PDF
    This open access book constitutes the proceedings of the 28th International Conference on Tools and Algorithms for the Construction and Analysis of Systems, TACAS 2022, which was held during April 2-7, 2022, in Munich, Germany, as part of the European Joint Conferences on Theory and Practice of Software, ETAPS 2022. The 46 full papers and 4 short papers presented in this volume were carefully reviewed and selected from 159 submissions. The proceedings also contain 16 tool papers of the affiliated competition SV-Comp and 1 paper consisting of the competition report. TACAS is a forum for researchers, developers, and users interested in rigorously based tools and algorithms for the construction and analysis of systems. The conference aims to bridge the gaps between different communities with this common interest and to support them in their quest to improve the utility, reliability, exibility, and efficiency of tools and algorithms for building computer-controlled systems

    Design of resource to backbone transmission for a high wind penetration future

    Get PDF
    In a high wind penetration future, transmission must be designed to integrate groups of new wind farms with a high capacity inter-regional ``backbone transmission system. A design process is described which begins by identifying feasible sites for future wind farms, identifies an optimal set of those wind farms for a specified future, and designs a reliable low-cost ``resource to backbone collector transmission network to connect each individual wind farm to the backbone transmission network. A model of the transmission and generation system in the state of Iowa is used to test these methods, and to make observations about the nature of these resource to backbone networks

    Tools and Algorithms for the Construction and Analysis of Systems

    Get PDF
    This open access book constitutes the proceedings of the 28th International Conference on Tools and Algorithms for the Construction and Analysis of Systems, TACAS 2022, which was held during April 2-7, 2022, in Munich, Germany, as part of the European Joint Conferences on Theory and Practice of Software, ETAPS 2022. The 46 full papers and 4 short papers presented in this volume were carefully reviewed and selected from 159 submissions. The proceedings also contain 16 tool papers of the affiliated competition SV-Comp and 1 paper consisting of the competition report. TACAS is a forum for researchers, developers, and users interested in rigorously based tools and algorithms for the construction and analysis of systems. The conference aims to bridge the gaps between different communities with this common interest and to support them in their quest to improve the utility, reliability, exibility, and efficiency of tools and algorithms for building computer-controlled systems

    Consumer load modeling and fair mechanisms in the efficient transactive energy market

    Get PDF
    Doctor of PhilosophyDepartment of Electrical and Computer EngineeringSanjoy DasTwo significant and closely related issues pertaining to the grid-constrained transactive distribution system market are investigated in this research. At first, the problem of spatial fairness in the allocation of energy among energy consumers is addressed, where consumer agents that are located at large distances from the substation – in terms of grid layout, are charged at higher rates than those close to it. This phenomenon, arising from the grid’s voltage and flow limits is aggravated during demand peaks. Using the Jain’s index to quantify fairness, two auction mechanisms are proposed. Both approaches are bilevel, with aggregators acting as interface agents between the consumers and the upstream distribution system operator (DSO). Furthermore, in spite of maximizing social welfare, neither mechanism makes use of the agents’ utility functions. The first mechanism is cost-setting, with the DSO determining unit costs. It implements the Jain’s index as a second term to the social welfare. Next, a power setting auction mechanism is put forth where the DSO’s role is to allocate energy in response to market equilibrium unit costs established at each aggregator from an iterative bidding process among its consumers. The Augmented Lagrangian Multigradient Approach (ALMA), which is based on vector gradient descent, is proposed in this research for implementation at the upper level. The mechanism’s lower level comprises of multiple auctions realized by the aggregators. The quasi-concavity of the Jain’s index is theoretically established, and it has been shown that ALMA converges to the Pareto front representing tradeoffs between social welfare and fairness. The effectiveness of both mechanisms is established through simulations carried out using a modified IEEE 37-bus system platform. The issue of extracting patterns of energy usage from time series energy use profiles of individual consumers is the focus of the second phase of this research. Two novel approaches for non-intrusive load disaggregation based on non-negative matrix factorization (NMF), are proposed. Both algorithms distinguish between fixed and shiftable load classes, with the latter being characterized by binary OFF and ON cycles. Fixed loads are represented as linear combinations of a set of basis vectors that are learned by NMF. One approach imposes L0 normed constraints on each shiftable load using a new method called binary load decomposition. The other approach models shiftable loads as Gaussian mixture models (GMM), therefore using expectation-maximization for unsupervised learning. This hybrid NMF-GMM algorithm enjoys the theoretical advantage of being interpretable as a maximum-likelihood procedure within a probabilistic framework. Numerical studies with real load profiles demonstrate that both algorithms can effectively disaggregate total loads into energy used by individual appliances. Using disaggregated loads, a maximum-margin regression approach to derive more elaborate, temperature-dependent utility functions of the consumers, is proposed. The research concludes by identifying the various ways gleaning such information can lead to more effective auction mechanisms for multi-period operation

    Collective Adaptive Systems: Qualitative and Quantitative Modelling and Analysis (Dagstuhl Seminar 14512)

    Get PDF
    This report documents the program and the outcomes of Dagstuhl Seminar 14512 "Collective Adaptive Systems: Qualitative and Quantitative Modelling and Analysis". Besides presentations on current work in the area, the seminar focused on the following topics: (i) Modelling techniques and languages for collective adaptive systems based on the above formalisms. (ii) Verification of collective adaptive systems. (iii) Humans-in-the-loop in collective adaptive systems
    • …
    corecore