1,087 research outputs found

    On the Traveling Salesman Problem in Nautical Environments: an Evolutionary Computing Approach to Optimization of Tourist Route Paths in Medulin, Croatia

    Get PDF
    The Traveling salesman problem (TSP) defines the problem of finding the optimal path between multiple points, connected by paths of a certain cost. This paper applies that problem formulation in the maritime environment, specifically a path planning problem for a tour boat visiting popular tourist locations in Medulin, Croatia. The problem is solved using two evolutionary computing methods โ€“ the genetic algorithm (GA) and the simulated annealing (SA) - and comparing the results (are compared) by an extensive search of the solution space. The results show that evolutionary computing algorithms provide comparable results to an extensive search in a shorter amount of time, with SA providing better results of the two

    ๊ณ„์ธต ๊ฐ•ํ™” ํ•™์Šต์—์„œ์˜ ํƒํ—˜์  ํ˜ผํ•ฉ ํƒ์ƒ‰

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ „๊ธฐยท์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2020. 8. ๋ฌธ๋ณ‘๋กœ.Balancing exploitation and exploration is a great challenge in many optimization problems. Evolutionary algorithms, such as evolutionary strategies and genetic algorithms, are algorithms inspired by biological evolution. They have been used for various optimization problems, such as combinatorial optimization and continuous optimization. However, evolutionary algorithms lack fine-tuning near local optima; in other words, they lack exploitation power. This drawback can be overcome by hybridization. Hybrid genetic algorithms, or memetic algorithms, are successful examples of hybridization. Although the solution space is exponentially vast in some optimization problems, these algorithms successfully find satisfactory solutions. In the deep learning era, the problem of exploitation and exploration has been relatively neglected. In deep reinforcement learning problems, however, balancing exploitation and exploration is more crucial than that in problems with supervision. Many environments in the real world have an exponentially wide state space that must be explored by agents. Without sufficient exploration power, agents only reveal a small portion of the state space and end up with seeking only instant rewards. In this thesis, a hybridization method is proposed which contains both gradientbased policy optimization with strong exploitation power and evolutionary policy optimization with strong exploration power. First, the gradient-based policy optimization and evolutionary policy optimization are analyzed in various environments. The results demonstrate that evolutionary policy optimization is robust for sparse rewards but weak for instant rewards, whereas gradient-based policy optimization is effective for instant rewards but weak for sparse rewards. This difference between the two optimizations reveals the potential of hybridization in policy optimization. Then, a hybrid search is suggested in the framework of hierarchical reinforcement learning. The results demonstrate that the hybrid search finds an effective agent for complex environments with sparse rewards thanks to its balanced exploitation and exploration.๋งŽ์€ ์ตœ์ ํ™” ๋ฌธ์ œ์—์„œ ํƒ์‚ฌ์™€ ํƒํ—˜์˜ ๊ท ํ˜•์„ ๋งž์ถ”๋Š” ๊ฒƒ์€ ๋งค์šฐ ์ค‘์š”ํ•œ ๋ฌธ์ œ์ด๋‹ค. ์ง„ํ™” ์ „๋žต๊ณผ ์œ ์ „ ์•Œ๊ณ ๋ฆฌ์ฆ˜๊ณผ ๊ฐ™์€ ์ง„ํ™” ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์ž์—ฐ์—์„œ์˜ ์ง„ํ™”์—์„œ ์˜๊ฐ์„ ์–ป์€ ๋ฉ”ํƒ€ํœด๋ฆฌ์Šคํ‹ฑ ์•Œ๊ณ ๋ฆฌ์ฆ˜์ด๋‹ค. ์ด๋“ค์€ ์กฐํ•ฉ ์ตœ์ ํ™”, ์—ฐ์† ์ตœ์ ํ™”์™€ ๊ฐ™์€ ๋‹ค์–‘ํ•œ ์ตœ์ ํ™” ๋ฌธ์ œ๋ฅผ ํ’€๊ธฐ ์œ„ํ•ด ์‚ฌ์šฉ๋˜์—ˆ๋‹ค. ํ•˜์ง€๋งŒ ์ง„ํ™” ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์ง€์—ญ ์ตœ์ ํ•ด ๊ทผ์ฒ˜์—์„œ์˜ ๋ฏธ์„ธ ์กฐ์ •, ์ฆ‰ ํƒ์‚ฌ์— ์•ฝํ•œ ํŠน์„ฑ์ด ์žˆ๋‹ค. ์ด๋Ÿฌํ•œ ๊ฒฐ์ ํ•จ์€ ํ˜ผํ•ฉํ™”๋ฅผ ํ†ตํ•ด ๊ทน๋ณตํ•  ์ˆ˜ ์žˆ๋‹ค. ํ˜ผํ•ฉ ์œ ์ „ ์•Œ๊ณ ๋ฆฌ์ฆ˜, ํ˜น์€ ๋ฏธ๋ฏธํ‹ฑ ์•Œ๊ณ ๋ฆฌ์ฆ˜์ด ์„ฑ๊ณต์ ์ธ ํ˜ผํ•ฉํ™”์˜ ์‚ฌ๋ก€์ด๋‹ค. ์ด๋Ÿฌํ•œ ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ์ตœ์ ํ™” ๋ฌธ์ œ์˜ ํ•ด ๊ณต๊ฐ„์ด ๊ธฐํ•˜๊ธ‰์ˆ˜์ ์œผ๋กœ ๋„“๋”๋ผ๋„ ์„ฑ๊ณต์ ์œผ๋กœ ๋งŒ์กฑ์Šค๋Ÿฌ์šด ํ•ด๋ฅผ ์ฐพ์•„๋‚ธ๋‹ค. ํ•œํŽธ ์‹ฌ์ธต ํ•™์Šต์˜ ์‹œ๋Œ€์—์„œ, ํƒ์‚ฌ์™€ ํƒํ—˜์˜ ๊ท ํ˜•์„ ๋งž์ถ”๋Š” ๋ฌธ์ œ๋Š” ์ข…์ข… ๋ฌด์‹œ๋˜์—ˆ๋‹ค. ํ•˜์ง€๋งŒ ์‹ฌ์ธต ๊ฐ•ํ™”ํ•™์Šต์—์„œ๋Š” ํƒ์‚ฌ์™€ ํƒํ—˜์˜ ๊ท ํ˜•์„ ๋งž์ถ”๋Š” ์ผ์€ ์ง€๋„ํ•™์Šต์—์„œ๋ณด๋‹ค ํ›จ์”ฌ ๋” ์ค‘์š”ํ•˜๋‹ค. ๋งŽ์€ ์‹ค์ œ ์„ธ๊ณ„์˜ ํ™˜๊ฒฝ์€ ๊ธฐํ•˜๊ธ‰์ˆ˜์ ์œผ๋กœ ํฐ ์ƒํƒœ ๊ณต๊ฐ„์„ ๊ฐ€์ง€๊ณ  ์žˆ๊ณ  ์—์ด์ „ํŠธ๋Š” ์ด๋ฅผ ํƒํ—˜ํ•ด์•ผ๋งŒ ํ•œ๋‹ค. ์ถฉ๋ถ„ํ•œ ํƒํ—˜ ๋Šฅ๋ ฅ์ด ์—†์œผ๋ฉด ์—์ด์ „ํŠธ๋Š” ์ƒํƒœ ๊ณต๊ฐ„์˜ ๊ทนํžˆ ์ผ๋ถ€๋งŒ์„ ๋ฐํ˜€๋‚ด์–ด ๊ฒฐ๊ตญ ์ฆ‰๊ฐ์ ์ธ ๋ณด์ƒ๋งŒ ํƒํ•˜๊ฒŒ ๋  ๊ฒƒ์ด๋‹ค. ๋ณธ ํ•™์œ„๋…ผ๋ฌธ์—์„œ๋Š” ๊ฐ•ํ•œ ํƒ์‚ฌ ๋Šฅ๋ ฅ์„ ๊ฐ€์ง„ ๊ทธ๋ ˆ๋””์–ธํŠธ ๊ธฐ๋ฐ˜ ์ •์ฑ… ์ตœ์ ํ™”์™€ ๊ฐ•ํ•œ ํƒํ—˜ ๋Šฅ๋ ฅ์„ ๊ฐ€์ง„ ์ง„ํ™”์  ์ •์ฑ… ์ตœ์ ํ™”๋ฅผ ํ˜ผํ•ฉํ•˜๋Š” ๊ธฐ๋ฒ•์„ ์ œ์‹œํ•  ๊ฒƒ์ด๋‹ค. ์šฐ์„  ๊ทธ๋ ˆ๋””์–ธํŠธ ๊ธฐ๋ฐ˜ ์ •์ฑ… ์ตœ์ ํ™”์™€ ์ง„ํ™”์  ์ •์ฑ… ์ตœ์ ํ™”๋ฅผ ๋‹ค์–‘ํ•œ ํ™˜๊ฒฝ์—์„œ ๋ถ„์„ํ•œ๋‹ค. ๊ฒฐ๊ณผ์ ์œผ๋กœ ๊ทธ๋ ˆ๋””์–ธํŠธ ๊ธฐ๋ฐ˜ ์ •์ฑ… ์ตœ์ ํ™”๋Š” ์ฆ‰๊ฐ์  ๋ณด์ƒ์— ํšจ๊ณผ์ ์ด์ง€๋งŒ ๋ณด์ƒ์˜ ๋ฐ€๋„๊ฐ€ ๋‚ฎ์„๋•Œ ์ทจ์•ฝํ•œ ๋ฐ˜๋ฉด ์ง„ํ™”์  ์ •์ฑ… ์ตœ์ ํ™”๊ฐ€ ๋ฐ€๋„๊ฐ€ ๋‚ฎ์€ ๋ณด์ƒ์— ๋Œ€ํ•ด ๊ฐ•ํ•˜์ง€๋งŒ ์ฆ‰๊ฐ์ ์ธ ๋ณด์ƒ์— ๋Œ€ํ•ด ์ทจ์•ฝํ•˜๋‹ค๋Š” ๊ฒƒ์„ ์•Œ ์ˆ˜ ์žˆ๋‹ค. ๋‘ ๊ฐ€์ง€ ์ตœ์ ํ™”์˜ ํŠน์ง• ์ƒ ์ฐจ์ด์ ์ด ํ˜ผํ•ฉ์  ์ •์ฑ… ์ตœ์ ํ™”์˜ ๊ฐ€๋Šฅ์„ฑ์„ ๋ณด์—ฌ์ค€๋‹ค. ๊ทธ๋ฆฌ๊ณ  ๊ณ„์ธต์  ๊ฐ•ํ™” ํ•™์Šต ํ”„๋ ˆ์ž„์›Œํฌ์—์„œ์˜ ํ˜ผํ•ฉ ํƒ์ƒ‰ ๊ธฐ๋ฒ•์„ ์ œ์‹œํ•œ๋‹ค. ๊ทธ ๊ฒฐ๊ณผ ํ˜ผํ•ฉ ํƒ์ƒ‰ ๊ธฐ๋ฒ•์ด ๊ท ํ˜•์žกํžŒ ํƒ์‚ฌ์™€ ํƒํ—˜ ๋•๋ถ„์— ๋ฐ€๋„๊ฐ€ ๋‚ฎ์€ ๋ณด์ƒ์„ ์ฃผ๋Š” ๋ณต์žกํ•œ ํ™˜๊ฒฝ์—์„œ ํšจ๊ณผ์ ์ธ ์—์ด์ „ํŠธ๋ฅผ ์ฐพ์•„๋‚ธ ๋‹ค๋Š” ๊ฒƒ์„ ๋ณด์—ฌ์ค€๋‹ค.I. Introduction 1 II. Background 6 2.1 Evolutionary Computations 6 2.1.1 Hybrid Genetic Algorithm 7 2.1.2 Evolutionary Strategy 9 2.2 Hybrid Genetic Algorithm Example: Brick Layout Problem 10 2.2.1 Problem Statement 11 2.2.2 Hybrid Genetic Algorithm 11 2.2.3 Experimental Results 14 2.2.4 Discussion 15 2.3 Reinforcement Learning 16 2.3.1 Policy Optimization 19 2.3.2 Proximal Policy Optimization 21 2.4 Neuroevolution for Reinforcement Learning 23 2.5 Hierarchical Reinforcement Learning 25 2.5.1 Option-based HRL 26 2.5.2 Goal-based HRL 27 2.5.3 Exploitation versus Exploration 27 III. Understanding Features of Evolutionary Policy Optimizations 29 3.1 Experimental Setup 31 3.2 Feature Analysis 32 3.2.1 Convolution Filter Inspection 32 3.2.2 Saliency Map 36 3.3 Discussion 40 3.3.1 Behavioral Characteristics 40 3.3.2 ES Agent without Inputs 42 IV. Hybrid Search for Hierarchical Reinforcement Learning 44 4.1 Method 45 4.2 Experimental Setup 47 4.2.1 Environment 47 4.2.2 Network Architectures 50 4.2.3 Training 50 4.3 Results 51 4.3.1 Comparison 51 4.3.2 Experimental Results 53 4.3.3 Behavior of Low-Level Policy 54 4.4 Conclusion 55 V. Conclusion 56 5.1 Summary 56 5.2 Future Work 57 Bibliography 58Docto

    Evolutionary Computation

    Get PDF
    This book presents several recent advances on Evolutionary Computation, specially evolution-based optimization methods and hybrid algorithms for several applications, from optimization and learning to pattern recognition and bioinformatics. This book also presents new algorithms based on several analogies and metafores, where one of them is based on philosophy, specifically on the philosophy of praxis and dialectics. In this book it is also presented interesting applications on bioinformatics, specially the use of particle swarms to discover gene expression patterns in DNA microarrays. Therefore, this book features representative work on the field of evolutionary computation and applied sciences. The intended audience is graduate, undergraduate, researchers, and anyone who wishes to become familiar with the latest research work on this field

    A Unified Framework for Solving Multiagent Task Assignment Problems

    Get PDF
    Multiagent task assignment problem descriptors do not fully represent the complex interactions in a multiagent domain, and algorithmic solutions vary widely depending on how the domain is represented. This issue is compounded as related research fields contain descriptors that similarly describe multiagent task assignment problems, including complex domain interactions, but generally do not provide the mechanisms needed to solve the multiagent aspect of task assignment. This research presents a unified approach to representing and solving the multiagent task assignment problem for complex problem domains. Ideas central to multiagent task allocation, project scheduling, constraint satisfaction, and coalition formation are combined to form the basis of the constrained multiagent task scheduling (CMTS) problem. Basic analysis reveals the exponential size of the solution space for a CMTS problem, approximated by O(2n(m+n)) based on the number of agents and tasks involved in a problem. The shape of the solution space is shown to contain numerous discontinuous regions due to the complexities involved in relational constraints defined between agents and tasks. The CMTS descriptor represents a wide range of classical and modern problems, such as job shop scheduling, the traveling salesman problem, vehicle routing, and cooperative multi-object tracking. Problems using the CMTS representation are solvable by a suite of algorithms, with varying degrees of suitability. Solution generating methods range from simple random scheduling to state-of-the-art biologically inspired approaches. Techniques from classical task assignment solvers are extended to handle multiagent task problems where agents can also multitask. Additional ideas are incorporated from constraint satisfaction, project scheduling, evolutionary algorithms, dynamic coalition formation, auctioning, and behavior-based robotics to highlight how different solution generation strategies apply to the complex problem space
    • โ€ฆ
    corecore