11,788 research outputs found

    Artificial Intelligence and Systems Theory: Applied to Cooperative Robots

    Full text link
    This paper describes an approach to the design of a population of cooperative robots based on concepts borrowed from Systems Theory and Artificial Intelligence. The research has been developed under the SocRob project, carried out by the Intelligent Systems Laboratory at the Institute for Systems and Robotics - Instituto Superior Tecnico (ISR/IST) in Lisbon. The acronym of the project stands both for "Society of Robots" and "Soccer Robots", the case study where we are testing our population of robots. Designing soccer robots is a very challenging problem, where the robots must act not only to shoot a ball towards the goal, but also to detect and avoid static (walls, stopped robots) and dynamic (moving robots) obstacles. Furthermore, they must cooperate to defeat an opposing team. Our past and current research in soccer robotics includes cooperative sensor fusion for world modeling, object recognition and tracking, robot navigation, multi-robot distributed task planning and coordination, including cooperative reinforcement learning in cooperative and adversarial environments, and behavior-based architectures for real time task execution of cooperating robot teams

    Synthesized cooperative strategies for intelligent multi-robots in a real-time distributed environment : a thesis presented in partial fulfillment of the requirements for the degree of Master of Science in Computer Science at Massey University, Albany, New Zealand

    Get PDF
    In the robot soccer domain, real-time response usually curtails the development of more complex Al-based game strategies, path-planning and team cooperation between intelligent agents. In light of this problem, distributing computationally intensive algorithms between several machines to control, coordinate and dynamically assign roles to a team of robots, and allowing them to communicate via a network gives rise to real-time cooperation in a multi-robotic team. This research presents a myriad of algorithms tested on a distributed system platform that allows for cooperating multi- agents in a dynamic environment. The test bed is an extension of a popular robot simulation system in the public domain developed at Carnegie Mellon University, known as TeamBots. A low-level real-time network game protocol using TCP/IP and UDP were incorporated to allow for a conglomeration of multi-agent to communicate and work cohesively as a team. Intelligent agents were defined to take on roles such as game coach agent, vision agent, and soccer player agents. Further, team cooperation is demonstrated by integrating a real-time fuzzy logic-based ball-passing algorithm and a fuzzy logic algorithm for path planning. Keywords Artificial Intelligence, Ball Passing, the coaching system, Collaborative, Distributed Multi-Agent, Fuzzy Logic, Role Assignmen

    Integrating mobile robotics and vision with undergraduate computer science

    Get PDF
    This paper describes the integration of robotics education into an undergraduate Computer Science curriculum. The proposed approach delivers mobile robotics as well as covering the closely related field of Computer Vision, and is directly linked to the research conducted at the authors’ institution. The paper describes the most relevant details of the module content and assessment strategy, paying particular attention to the practical sessions using Rovio mobile robots. The specific choices are discussed that were made with regard to the mobile platform, software libraries and lab environment. The paper also presents a detailed qualitative and quantitative analysis of student results, including the correlation between student engagement and performance, and discusses the outcomes of this experience

    Motion Planning Among Dynamic, Decision-Making Agents with Deep Reinforcement Learning

    Full text link
    Robots that navigate among pedestrians use collision avoidance algorithms to enable safe and efficient operation. Recent works present deep reinforcement learning as a framework to model the complex interactions and cooperation. However, they are implemented using key assumptions about other agents' behavior that deviate from reality as the number of agents in the environment increases. This work extends our previous approach to develop an algorithm that learns collision avoidance among a variety of types of dynamic agents without assuming they follow any particular behavior rules. This work also introduces a strategy using LSTM that enables the algorithm to use observations of an arbitrary number of other agents, instead of previous methods that have a fixed observation size. The proposed algorithm outperforms our previous approach in simulation as the number of agents increases, and the algorithm is demonstrated on a fully autonomous robotic vehicle traveling at human walking speed, without the use of a 3D Lidar
    • …
    corecore