302 research outputs found

    A robust, distributed task allocation algorithm for time-critical, multi agent systems operating in uncertain environments

    Get PDF
    The aim of this work is to produce and test a robust, distributed, multi-agent task allocation algorithm, as these are scarce and not well-documented in the literature. The vehicle used to create the robust system is the Performance Impact algorithm (PI), as it has previously shown good performance. Three different variants of PI are designed to improve its robustness, each using Monte Carlo sampling to approximate Gaussian distributions. Variant A uses the expected value of the task completion times, variant B uses the worst-case scenario metric and variant C is a hybrid that implements a combination of these. The paper shows that, in simulated trials, baseline PI does not han-dle uncertainty well; the task-allocation success rate tends to decrease linear-ly as degree of uncertainty increases. Variant B demonstrates a worse per-formance and variant A improves the failure rate only slightly. However, in comparison, the hybrid variant C exhibits a very low failure rate, even under high uncertainty. Furthermore, it demonstrates a significantly better mean ob-jective function value than the baseline.EPSR

    Addressing robustness in time-critical, distributed, task allocation algorithms.

    Get PDF
    The aim of this work is to produce and test a robustness module (ROB-M) that can be generally applied to distributed, multi-agent task allocation algorithms, as robust versions of these are scarce and not well-documented in the literature. ROB-M is developed using the Performance Impact (PI) algorithm, as this has previously shown good results in deterministic trials. Different candidate versions of the module are thus bolted on to the PI algorithm and tested using two different task allocation problems under simulated uncertain conditions, and results are compared with baseline PI. It is shown that the baseline does not handle uncertainty well; the task-allocation success rate tends to decrease linearly as degree of uncertainty increases. However, when PI is run with one of the candidate robustness modules, the failure rate becomes very low for both problems, even under high simulated uncertainty, and so its architecture is adopted for ROB-M and also applied to MIT’s baseline Consensus Based Bundle Algorithm (CBBA) to demonstrate its flexibility. Strong evidence is provided to show that ROB-M can work effectively with CBBA to improve performance under simulated uncertain conditions, as long as the deterministic versions of the problems can be solved with baseline CBBA. Furthermore, the use of ROB-M does not appear to increase mean task completion time in either algorithm, and only 100 Monte Carlo samples are required compared to 10,000 in MIT’s robust version of the CBBA algorithm. PI with ROB-M is also tested directly against MIT’s robust algorithm and demonstrates clear superiority in terms of mean numbers of solved tasks.N/

    A Survey on Aerial Swarm Robotics

    Get PDF
    The use of aerial swarms to solve real-world problems has been increasing steadily, accompanied by falling prices and improving performance of communication, sensing, and processing hardware. The commoditization of hardware has reduced unit costs, thereby lowering the barriers to entry to the field of aerial swarm robotics. A key enabling technology for swarms is the family of algorithms that allow the individual members of the swarm to communicate and allocate tasks amongst themselves, plan their trajectories, and coordinate their flight in such a way that the overall objectives of the swarm are achieved efficiently. These algorithms, often organized in a hierarchical fashion, endow the swarm with autonomy at every level, and the role of a human operator can be reduced, in principle, to interactions at a higher level without direct intervention. This technology depends on the clever and innovative application of theoretical tools from control and estimation. This paper reviews the state of the art of these theoretical tools, specifically focusing on how they have been developed for, and applied to, aerial swarms. Aerial swarms differ from swarms of ground-based vehicles in two respects: they operate in a three-dimensional space and the dynamics of individual vehicles adds an extra layer of complexity. We review dynamic modeling and conditions for stability and controllability that are essential in order to achieve cooperative flight and distributed sensing. The main sections of this paper focus on major results covering trajectory generation, task allocation, adversarial control, distributed sensing, monitoring, and mapping. Wherever possible, we indicate how the physics and subsystem technologies of aerial robots are brought to bear on these individual areas

    LSAR: Multi-UAV Collaboration for Search and Rescue Missions

    Get PDF
    In this paper, we consider the use of a team of multiple unmanned aerial vehicles (UAVs) to accomplish a search and rescue (SAR) mission in the minimum time possible while saving the maximum number of people. A novel technique for the SAR problem is proposed and referred to as the layered search and rescue (LSAR) algorithm. The novelty of LSAR involves simulating real disasters to distribute SAR tasks among UAVs. The performance of LSAR is compared, in terms of percentage of rescued survivors and rescue and execution times, with the max-sum, auction-based, and locust-inspired approaches for multi UAV task allocation (LIAM) and opportunistic task allocation (OTA) schemes. The simulation results show that the UAVs running the LSAR algorithm on average rescue approximately 74% of the survivors, which is 8% higher than the next best algorithm (LIAM). Moreover, this percentage increases with the number of UAVs, almost linearly with the least slope, which means more scalability and coverage is obtained in comparison to other algorithms. In addition, the empirical cumulative distribution function of LSAR results shows that the percentages of rescued survivors clustered around the [78% 100%] range under an exponential curve, meaning most results are above 50%. In comparison, all the other algorithms have almost equal distributions of their percentage of rescued survivor results. Furthermore, because the LSAR algorithm focuses on the center of the disaster, it nds more survivors and rescues them faster than the other algorithms, with an average of 55% 77%. Moreover, most registered times to rescue survivors by LSAR are bounded by a time of 04:50:02 with 95% con dence for a one-month mission time.info:eu-repo/semantics/publishedVersio

    How good are distributed allocation algorithms for solving urban search and rescue problems? A comparative study with centralized algorithms

    Get PDF
    In this paper, a modified centralized algorithm based on particle swarm optimization (MCPSO) is presented to solve the task allocation problem in the search and rescue domain. The reason for this paper is to provide a benchmark against distributed algorithms in search and rescue application area. The hypothesis of this paper is that a centralized algorithm should perform better than distributed algorithms because it has all the available information at hand to solve the problem. Therefore, the centralized approach will provide a benchmark for evaluating how well the distributed algorithms are working and how much improvement can still be gained. Among the distributed algorithms, the consensus-based bundle algorithm (CBBA) is a relatively recent method based on the market auction mechanism, which is receiving considerable attention. Other distributed algorithms, such as PI and PI with softmax, have shown to perform better than CBBA. Therefore, in this paper, the three distributed algorithms mentioned earlier are compared against three centralized algorithms. They are particle swarm optimization, MCPSO, described in this paper, and genetic algorithms. Two experiments were conducted. The first involved comparing all the above-mentioned algorithms, both centralized and distributed, using the same set of application scenarios. It is found that MCPSO always outperforms the other five algorithms in time cost. Due to the high failure rate of CBBA and the other two centralized methods, the second experiment focused on carrying out more tests to compare MCPSO against PI and PI with softmax. All the results are shown and analyzed to determine the performance gaps between the distributed algorithms and the MCPSO

    Reliable, distributed scheduling and rescheduling for time-critical, multiagent systems

    Get PDF
    This paper addresses two main problems with many heuristic task allocation approaches – solution trapping in local minima and static structure. The existing distributed task allocation algorithm known as PI (Performance Impact) is used as the vehicle for developing solutions to these problems as it has been shown to out-perform the state-of-the-art Consensus Based Bundle Algorithm (CBBA) for time-critical problems with tight deadlines, but is both static and sub-optimal with a tendency towards trapping in local minima. The paper describes two additional modules that are easily integrated with PI. The first extends the algorithm to permit dynamic online rescheduling in real time, and the second boosts performance by introducing an additional soft max action selection procedure that increases the algorithm’s exploratory properties. The paper demonstrates the effectiveness of the dynamic rescheduling module and shows that the average time taken to perform tasks can be reduced by up to 9% when the soft max module is used. In addition, the solution of some problems that baseline PI cannot handle is enabled by the second module. These developments represent a significant advance in the state-of-the-art for multi-agent, time-critical task assignment.EPSR

    Energy-aware Graph Job Allocation in Software Defined Air-Ground Integrated Vehicular Networks

    Full text link
    The software defined air-ground integrated vehicular (SD-AGV) networks have emerged as a promising paradigm, which realize the flexible on-ground resource sharing to support innovative applications for UAVs with heavy computational overhead. In this paper, we investigate a vehicular cloud-assisted graph job allocation problem in SD-AGV networks, where the computation-intensive jobs carried by UAVs, and the vehicular cloud are modeled as graphs. To map each component of the graph jobs to a feasible vehicle, while achieving the trade-off among minimizing UAVs' job completion time, energy consumption, and the data exchange cost among vehicles, we formulate the problem as a mixed-integer non-linear programming problem, which is Np-hard. Moreover, the constraint associated with preserving job structures poses addressing the subgraph isomorphism problem, that further complicates the algorithm design. Motivated by which, we propose an efficient decoupled approach by separating the template (feasible mappings between components and vehicles) searching from the transmission power allocation. For the former, we present an efficient algorithm of searching for all the subgraph isomorphisms with low computation complexity. For the latter, we introduce a power allocation algorithm by applying convex optimization techniques. Extensive simulations demonstrate that the proposed approach outperforms the benchmark methods considering various problem sizes.Comment: 14 pages, 7 figure
    • …
    corecore