1,738 research outputs found
μλμ μμ λΉνμμκ³Ό μλμ λ²μ€νμΈ λ€νμμ μ΄μ©ν λ€μ μΏΌλλ‘ν°μ κ²½λ‘ κ³ν
νμλ
Όλ¬Έ (μμ¬) -- μμΈλνκ΅ λνμ : 곡과λν κΈ°κ³ν곡곡νλΆ, 2020. 8. κΉνμ§.Multi-agent systems consisting of unmanned aerial vehicles (UAVs) are receiving attention from
many industrial domains due to their mobility, and applicability. To safely operate these multiagent
systems, path planning algorithm that can generate safe, dynamically feasible trajectory is
required. However, existing multi-agent trajectory planning methods may fail to generate multiagent
trajectory in obstacle-dense environment due to deadlock or optimization failure caused
by infeasible collision constraints. In this paper, we presents a new e client algorithm which
guarantees a solution for a class of multi-agent trajectory planning problems in obstacle-dense
environments. Our algorithm combines the advantages of both grid-based and optimization-based
approaches, and generates safe, dynamically feasible trajectories without su ering from an erroneous
optimization setup such as imposing infeasible collision constraints. We adopt a sequential
optimization method with dummy agents to improve the scalability of the algorithm, and utilize
the convex hull property of Bernstein polynomial to replace non-convex collision avoidance constraints
to convex ones. We validate the proposed algorithm through the comparison with our
previous work and SCP-based method. The proposed method reduces more than 50% of the objective
cost compared to our previous work, and reduces more than 75% of the computation time
compared to SCP-based method. Furthermore, the proposed method can compute the trajectory
for 64 agents on average 6.36 seconds with Intel Core i7-7700 @ 3.60GHz CPU and 16G RAM.무μΈλΉν체(UAV)λ‘ κ΅¬μ±λ λ€μ€ μμ΄μ νΈ μμ€ν
μ λμ κΈ°λμ± λ° μμ© κ°λ₯μ±μΌλ‘ λ§μ μ°μ
λΆμΌμμ κ΄μ¬μ λ°κ³ μλ€. μ΄λ¬ν λ€μ€ μμ΄μ νΈ μμ€ν
μ μμ νκ² μ΄μ©νλ €λ©΄ μμ νκ³ λμ μΌλ‘ μ€ν κ°λ₯ κ²½λ‘λ₯Ό μμ±ν μ μλ κ²½λ‘ κ³ν μκ³ λ¦¬μ¦μ΄ νμνλ€. κ·Έλ¬λ κΈ°μ‘΄μ λ€μ€ μμ΄μ νΈ κ²½λ‘ κ³ν λ°©λ²μ μ₯μ λ¬Ό νκ²½μμ κ΅μ°© μνλ λΆμ μ ν μΆ©λ ννΌ μ‘°κ±΄μΌλ‘ μΈν μ΅μ ν μ€ν¨κ° μΌμ΄λ μ μλ€λ νκ³κ° μλ€.
λ³Έ λ
Όλ¬Έμμλ μ₯μ λ¬Ό νκ²½μμ ν΄μ μ‘΄μ¬λ₯Ό 보μ₯νλλ‘ λ€μ€ μμ΄μ νΈ κ²½λ‘ κ³ν λ¬Έμ λ₯Ό λ³νν λ€ μ΄λ₯Ό ν¨μ¨μ μΌλ‘ νμ΄λΌ μ μλ μλ‘μ΄ κ²½λ‘ κ³ν μκ³ λ¦¬μ¦μ μ μνλ€. μ΄ μκ³ λ¦¬μ¦μ 그리λ κΈ°λ° μ κ·Όλ²κ³Ό μ΅μ ν κΈ°λ° μ κ·Όλ²μ μ₯μ μ λͺ¨λ κ°μ§λλ‘ μ€κ³λμμΌλ©°, λΆκ°λ₯ν μΆ©λ ꡬμ쑰건μ λΆκ³Όνμ§ μκ³ μμ νκ³ λμ μΌλ‘ μ€ν κ°λ₯ν κΆ€μ μ μμ±ν μ μλ€. μ΄ μκ³ λ¦¬μ¦μ λλ―Έ μμ΄μ νΈ(dummy agents)μ μ΄μ©ν μμ°¨ μ΅μ ν λ°©λ²μ μ¬μ©νμ¬ μκ³ λ¦¬μ¦μ νμ₯μ±(scalability)μ λμμΌλ©°, λ²μ€νμΈ(Bernstein) λ€νμμ λ³Όλ‘ κ»μ§(convex hull) μ±μ§μ νμ©νμ¬ λ³Όλ‘νμ§ μμ μΆ©λ ννΌ μ μ½ μ‘°κ±΄μ λ³Όλ‘ννμλ€.
μ μλ μκ³ λ¦¬μ¦μ μ±λ₯μ μ ν μ°κ΅¬μ SCP κΈ°λ° λ°©λ²κ³Όμ λΉκ΅λ₯Ό ν΅ν΄ κ²μ¦λμλ€. μ μλ λ°©λ²μ μ ν μ°κ΅¬μ λΉν΄ λͺ©ν λΉμ©μ 50% μ΄μ μ κ°νμμΌλ©°, SCP κΈ°λ° λ°©λ²μ λΉν΄ κ³μ° μκ°μ 75% μ΄μ κ°μνμλ€. λν μ μλ λ°©λ²μ μΈν
μ½μ΄ i7-7700 @ 3.60GHz CPU λ° 16G RAM νκ²½μμ 64κ° μμ΄μ νΈμ κΆ€μ μ κ³μ°νλλ° νκ· 6.36μ΄κ° μμλλ€.1 Introduction 1
1.1 Literature review 2
1.2 Thesis contribution 3
1.3 Thesis outline 3
2 Bernstein polynomial 4
2.1 Definition 4
2.2 Properties 5
2.2.1 Convex hull property 5
2.2.2 Endpoint interpolation property 5
2.2.3 Arithmetic operations and derivatives 6
3 Multi-agent trajectory optimization 7
3.1 Problem formulation 7
3.1.1 Assumption 7
3.1.2 Trajectory Representation 8
3.1.3 Objective function 9
3.1.4 Convex constraints 9
3.1.5 Non-convex collision avoidance constraints 10
3.2 Collision constraints construction 11
3.2.1 Initial trajectory planning 12
3.2.2 Safe flight corridor 14
3.2.3 Relative safe flight corridor 16
3.3 Trajectory optimization 18
4 Sequential optimization with dummy agents 20
5 Experimental results 24
5.1 Comparison with the previous work 24
5.1.1 Success rate 25
5.1.2 Solution quality 26
5.1.3 Scalability analysis 26
5.2 Comparison with SCP-based method 27
5.3 Flight test 29
6 Conclusion 31Maste
Reinforcement Learning and Planning for Preference Balancing Tasks
Robots are often highly non-linear dynamical systems with many degrees of freedom, making solving motion problems computationally challenging. One solution has been reinforcement learning (RL), which learns through experimentation to automatically perform the near-optimal motions that complete a task. However, high-dimensional problems and task formulation often prove challenging for RL. We address these problems with PrEference Appraisal Reinforcement Learning (PEARL), which solves Preference Balancing Tasks (PBTs). PBTs define a problem as a set of preferences that the system must balance to achieve a goal. The method is appropriate for acceleration-controlled systems with continuous state-space and either discrete or continuous action spaces with unknown system dynamics. We show that PEARL learns a sub-optimal policy on a subset of states and actions, and transfers the policy to the expanded domain to produce a more refined plan on a class of robotic problems. We establish convergence to task goal conditions, and even when preconditions are not verifiable, show that this is a valuable method to use before other more expensive approaches. Evaluation is done on several robotic problems, such as Aerial Cargo Delivery, Multi-Agent Pursuit, Rendezvous, and Inverted Flying Pendulum both in simulation and experimentally. Additionally, PEARL is leveraged outside of robotics as an array sorting agent. The results demonstrate high accuracy and fast learning times on a large set of practical applications
Optimal Guidance and Control with Nonlinear Dynamics Using Sequential Convex Programming
This paper presents a novel method for expanding the use of sequential convex programming (SCP) to the domain of optimal guidance and control problems with nonlinear dynamics constraints. SCP is a useful tool in obtaining real-time solutions to direct optimal control, but it is unable to adequately model nonlinear dynamics due to the linearization and discretization required. As nonlinear program solvers are not yet functioning in real-time, a tool is needed to bridge the gap between satisfying the nonlinear dynamics and completing execution fast enough to be useful. Two methods are proposed, sequential convex programming with nonlinear dynamics correction (SCPn) and modified SCPn (M-SCPn), which mixes SCP and SCPn to reduce runtime and improve algorithmic robustness. Both methods are proven to generate optimal state and control trajectories that satisfy the nonlinear dynamics. Simulations are presented to validate the efficacy of the methods as compared to SCP
Proximal operators for multi-agent path planning
We address the problem of planning collision-free paths for multiple agents
using optimization methods known as proximal algorithms. Recently this approach
was explored in Bento et al. 2013, which demonstrated its ease of
parallelization and decentralization, the speed with which the algorithms
generate good quality solutions, and its ability to incorporate different
proximal operators, each ensuring that paths satisfy a desired property.
Unfortunately, the operators derived only apply to paths in 2D and require that
any intermediate waypoints we might want agents to follow be preassigned to
specific agents, limiting their range of applicability. In this paper we
resolve these limitations. We introduce new operators to deal with agents
moving in arbitrary dimensions that are faster to compute than their 2D
predecessors and we introduce landmarks, space-time positions that are
automatically assigned to the set of agents under different optimality
criteria. Finally, we report the performance of the new operators in several
numerical experiments.Comment: See movie at http://youtu.be/gRnsjd_ocx
- β¦