4,696 research outputs found
Reinforcement Learning for Automatic Test Case Prioritization and Selection in Continuous Integration
Testing in Continuous Integration (CI) involves test case prioritization,
selection, and execution at each cycle. Selecting the most promising test cases
to detect bugs is hard if there are uncertainties on the impact of committed
code changes or, if traceability links between code and tests are not
available. This paper introduces Retecs, a new method for automatically
learning test case selection and prioritization in CI with the goal to minimize
the round-trip time between code commits and developer feedback on failed test
cases. The Retecs method uses reinforcement learning to select and prioritize
test cases according to their duration, previous last execution and failure
history. In a constantly changing environment, where new test cases are created
and obsolete test cases are deleted, the Retecs method learns to prioritize
error-prone test cases higher under guidance of a reward function and by
observing previous CI cycles. By applying Retecs on data extracted from three
industrial case studies, we show for the first time that reinforcement learning
enables fruitful automatic adaptive test case selection and prioritization in
CI and regression testing.Comment: Spieker, H., Gotlieb, A., Marijan, D., & Mossige, M. (2017).
Reinforcement Learning for Automatic Test Case Prioritization and Selection
in Continuous Integration. In Proceedings of 26th International Symposium on
Software Testing and Analysis (ISSTA'17) (pp. 12--22). AC
Addressing Function Approximation Error in Actor-Critic Methods
In value-based reinforcement learning methods such as deep Q-learning,
function approximation errors are known to lead to overestimated value
estimates and suboptimal policies. We show that this problem persists in an
actor-critic setting and propose novel mechanisms to minimize its effects on
both the actor and the critic. Our algorithm builds on Double Q-learning, by
taking the minimum value between a pair of critics to limit overestimation. We
draw the connection between target networks and overestimation bias, and
suggest delaying policy updates to reduce per-update error and further improve
performance. We evaluate our method on the suite of OpenAI gym tasks,
outperforming the state of the art in every environment tested.Comment: Accepted at ICML 201
Local ant system for allocating robot swarms to time-constrained tasks
We propose a novel application of the Ant Colony Optimization algorithm to efficiently allocate a swarm of homogeneous robots to a set of tasks that need to be accomplished by specific deadlines. We exploit the local communication between robots to periodically evaluate the quality of the allocation solutions, and agents select independently among the high-quality alternatives. The evaluation is performed using pheromone trails to favor allocations which minimize the execution time of the tasks. Our approach is validated in both static and dynamic environments (i.e. the task availability changes over time) using different sets of physics-based simulations. (C) 2018 Elsevier B.V. All rights reserved
κ°λ―Έμκ³ λ¦¬μ¦μ μ΄μ©ν λλ‘ μ μ μ€ κ²½λ‘ μ΅μ ν
νμλ
Όλ¬Έ(μμ¬) -- μμΈλνκ΅λνμ : 곡과λν 건μ€ν경곡νλΆ, 2022.2. κΉλκ·.Drones can overcome the limitation of ground vehicles by replacing the congestion time and allowing rapid service. For sudden snowfall with climate change, a quickly deployed drone can be a flexible alternative considering the deadhead route and the labor costs. The goal of this study is to optimize a drone arc routing problem (D-ARP), servicing the required roads for snow removal. A D-ARP creates computational burden especially in large network. The D-ARP has a large search space due to its exponentially increased candidate route, arc direction decision, and continuous arc space. To reduce the search space, we developed the auxiliary transformation method in ACO algorithm and adopted the random walk method. The contribution of the work is introducing a new problem and optimization approach of D-ARP in snow removal operation and reduce its search space. The optimization results confirmed that the drone travels shorter distance compared to the truck with a reduction of 5% to 22%. Furthermore, even under the length constraint model, the drone shows 4% reduction compared to the truck. The result of the test sets demonstrated that the adopted heuristic algorithm performs well in the large size networks in reasonable time. Based on the results, introducing a drone in snow removal is expected to save the operation cost in practical terms.λλ‘ μ νΌμ‘μκ°λλ₯Ό λ체νκ³ λΉ λ₯Έ μλΉμ€λ₯Ό κ°λ₯νκ² ν¨μΌλ‘μ¨ μ§μμ°¨λμ νκ³λ₯Ό 극볡ν μ μλ€. μ΅κ·Ό κΈ°νλ³νμ λ°λ₯Έ κ°μμ€λ° κ°μ€μ κ²½μ°μ, λλ‘ κ³Ό κ°μ΄ λΉ λ₯΄κ² ν¬μ
ν μ μλ μλΉμ€λ μ΄ν κ²½λ‘μ λ
ΈλλΉμ©μ κ³ λ €νμ λλ μ μ°ν μ΄μ μ΅μ
μ΄ λ μ μλ€. λ³Έ μ°κ΅¬μ λͺ©μ μ λλ‘ μν¬ λΌμ°ν
(D-ARP)μ μ΅μ ννλ κ²μ΄λ©°, μ΄λ μ μ€μ νμν λλ‘λ₯Ό μλΉμ€νλ κ²½λ‘λ₯Ό νμνλ κ²μ΄λ€. λλ‘ μν¬ λΌμ°ν
μ νΉν ν° λ€νΈμν¬μμ μ»΄ν¨ν° λΆνλ₯Ό μμ±νλ€. λ€μ λ§ν΄D-ARPλ ν° κ²μ곡κ°μ νμλ‘ νλ©°, μ΄λ κΈ°νκΈμμ μΌλ‘ μ¦κ°νλ ν보 κ²½λ‘ λ° νΈμ λ°©ν₯ κ²°μ κ·Έλ¦¬κ³ μ°μμ μΈ νΈμ 곡κ°μΌλ‘λΆν° κΈ°μΈνλ€. κ²μ곡κ°μ μ€μ΄κΈ° μν΄, μ°λ¦¬λ κ°λ―Έμκ³ λ¦¬μ¦μ 보쑰λ³νλ°©λ²μ μ μ©νλ λ°©μμ λμ
νμμΌλ©° λν λλ€μν¬ κΈ°λ²μ μ±ννμλ€. λ³Έ μ°κ΅¬μ κΈ°μ¬λ μ μ€ μ΄μμ μμ΄ D-ARPλΌλ μλ‘μ΄ λ¬Έμ λ₯Ό μ€μ νκ³ μ΅μ ν μ κ·Όλ²μ λμ
νμμΌλ©° κ²μ곡κ°μ μ΅μνν κ²μ΄λ€. μ΅μ ν κ²°κ³Ό, λλ‘ μ μ§μνΈλμ λΉν΄ μ½ 5% ~ 22%μ κ²½λ‘ λΉμ© κ°μλ₯Ό 보μλ€. λμκ° κΈΈμ΄ μ μ½ λͺ¨λΈμμλ λλ‘ μ 4%μ λΉμ© κ°μλ₯Ό 보μλ€. λν μ€νκ²°κ³Όλ μ μ©ν ν΄λ¦¬μ€ν± μκ³ λ¦¬μ¦μ΄ ν° λ€νΈμν¬μμλ ν©λ¦¬μ μκ° λ΄μ μ΅μ ν΄λ₯Ό μ°Ύμμ μ
μ¦νμλ€. μ΄λ¬ν κ²°κ³Όλ₯Ό λ°νμΌλ‘, λλ‘ μ μ μ€μ λμ
νλ κ²μ λ―Έλμ μ μ€ μ΄μ λΉμ©μ μ€μ§μ μΌλ‘ κ°μμν¬ κ²μΌλ‘ κΈ°λλλ€.Chapter 1. Introduction 4
1.1. Study Background 4
1.2. Purpose of Research 6
Chapter 2. Literature Review 7
2.1. Drone Arc Routing problem 7
2.2. Snow Removal Routing Problem 8
2.3. The Classic ARPs and Algorithms 9
2.4. Large Search Space and Arc direction 11
Chapter 3. Method 13
3.1. Problem Statement 13
3.2. Formulation 16
Chapter 4. Algorithm 17
4.1. Overview 17
4.2. Auxilary Transformation Method 18
4.3. Ant Colony Optimization (ACO) 20
4.4. Post Process for Arc Direction Decision 23
4.5. Length Constraint and Random Walk 24
Chapter 5. Results 27
5.1. Application in Toy Network 27
5.2. Application in Real-world Networks 29
5.3. Application of the Refill Constraint in Seoul 31
Chapter 6. Conclusion 34
References 35
Acknowledgment 40μ
Hopscotch: Robust Multi-agent Search
The task of searching a space is critical to a wide range of diverse applications such as land mine clearing and planetary exploration. Because applications frequently require searching remote or hazardous locations, and because the task is easily divisible, it is natural to consider the use of multi-robot teams to accomplish the search task. An important topic of research in this area is the division of the task among robot agents. Interrelated with subtask assignment is failure handling, in the sense that, when an agent fails, its part of the task must then be performed by other agents. This thesis describes Hopscotch, a multi-agent search strategy that divides the search area into a grid of lots. Each agent is assigned responsibility to search one lot at a time, and upon completing the search of that lot the agent is assigned a new lot. Assignment occurs in real time using a simple contract net. Because lots that have been previously searched are skipped, the order of search from the point of view of a particular agent is reminiscent of the progression of steps in the playground game of Hopscotch. Decomposition of the search area is a common approach to multi-agent search, and auction-based contract net strategies have appeared in recent literature as a method of task allocation in multi-agent systems. The Hopscotch strategy combines the two, with a strong focus on robust tolerance of agent failures. Contract nets typically divide all known tasks among available resources. In contrast, Hopscotch limits each agent to one assigned lot at a time, so that failure of an agent compels re-allocation of only one lot search task. Furthermore, the contract net is implemented in an unconventional manner that empowers each agent with responsibility for contract management. This novel combination of real-time assignment and decentralized management allows Hopscotch to resiliently cope with agent failures. The Hopscotch strategy was modeled and compared to other multi-agent strate- gies that tackle the search task in a variety of ways. Simulation results show that Hopscotch is failure-tolerant and very effective in comparison to the other approaches in terms of both search time and search efficiency. Although the search task modeled here is a basic one, results from simulations show the promise of using this strategy for more complicated scenarios, and with actual robot agents
- β¦