2,145 research outputs found

    Multi-rover testbed for teleconducted and autonomous surveillance, reconnaissance, and exploration

    Get PDF
    At Caltech's Visual and Autonomous Exploration Systems Research Laboratory (http://autonomy.caltech.edu) an outdoor multi-rover testbed has been developed that allows for near real-time interactive or automatic control from anywhere in the world via the Internet. It enables the implementation, field-testing, and validation of algorithms/software and strategies for navigation, exploration, feature extraction, anomaly detection, and target prioritization with applications in planetary exploration, security surveillance, reconnaissance of disaster areas, military reconnaissance, and delivery of lethal force such as explosives for urban warfare. Several rover platforms have been developed, enabling testing of cooperative multi-rover scenarios (e.g., inter-rover communication/coordination) and distributed exploration of operational areas

    Low-cost, multi-agent systems for planetary surface exploration

    Get PDF
    The use of off-the-shelf consumer electronics combined with top-down design methodologies have made small and inexpensive satellites, such as CubeSats, emerge as viable, low-cost and attractive space-based platforms that enable a range of new and exciting mission scenarios. In addition, to overcome some of the resource limitation issues encountered with these platforms, distributed architectures have emerged to enable complex tasks through the use of multiple low complexity units. The low-cost characteristics of such systems coupled with the distributed architecture allows for an increase in the size of the system beyond what would have been feasible with a monolithic system, hence widening the operational capabilities without significantly increasing the control complexity of the system. These ideas are not new for Earth orbiting devices, but excluding some distributed remote sensing architectures they are yet to be applied for the purpose of planetary exploration. Experience gained through large rovers demonstrates the value of in-situ exploration, which is however limited by the associated high-cost and risk. The loss of a rover can and has happened because of a number of possible failures: besides the hazards directly linked to the launch and journey to the target-body, hard landing and malfunctioning of parts are all threats to the success of the mission. To overcome these issues this paper introduces the concept of using off-the-shelf consumer electronics to deploy a low-cost multi-rover system for future planetary surface exploration. It is shown that such a system would significantly reduce the programmatic-risk of the mission (for example catastrophic failure of a single rover), while exploiting the inherent advantages of cooperative behaviour. These advantages are analysed with a particular emphasis put upon the guidance, navigation and control of such architectures using the method of artificial potential field. Laboratory tests on multi-agent robotic systems support the analysis. Principal features of the system are identified and the underlying advantages over a monolithic single-agent system highlighted

    Explainable shared control in assistive robotics

    Get PDF
    Shared control plays a pivotal role in designing assistive robots to complement human capabilities during everyday tasks. However, traditional shared control relies on users forming an accurate mental model of expected robot behaviour. Without this accurate mental image, users may encounter confusion or frustration whenever their actions do not elicit the intended system response, forming a misalignment between the respective internal models of the robot and human. The Explainable Shared Control paradigm introduced in this thesis attempts to resolve such model misalignment by jointly considering assistance and transparency. There are two perspectives of transparency to Explainable Shared Control: the human's and the robot's. Augmented reality is presented as an integral component that addresses the human viewpoint by visually unveiling the robot's internal mechanisms. Whilst the robot perspective requires an awareness of human "intent", and so a clustering framework composed of a deep generative model is developed for human intention inference. Both transparency constructs are implemented atop a real assistive robotic wheelchair and tested with human users. An augmented reality headset is incorporated into the robotic wheelchair and different interface options are evaluated across two user studies to explore their influence on mental model accuracy. Experimental results indicate that this setup facilitates transparent assistance by improving recovery times from adverse events associated with model misalignment. As for human intention inference, the clustering framework is applied to a dataset collected from users operating the robotic wheelchair. Findings from this experiment demonstrate that the learnt clusters are interpretable and meaningful representations of human intent. This thesis serves as a first step in the interdisciplinary area of Explainable Shared Control. The contributions to shared control, augmented reality and representation learning contained within this thesis are likely to help future research advance the proposed paradigm, and thus bolster the prevalence of assistive robots.Open Acces

    Adaptive Robot Framework: Providing Versatility and Autonomy to Manufacturing Robots Through FSM, Skills and Agents

    Get PDF
    207 p.The main conclusions that can be extracted from an analysis of the current situation and future trends of the industry,in particular manufacturing plants, are the following: there is a growing need to provide customization of products, ahigh variation of production volumes and a downward trend in the availability of skilled operators due to the ageingof the population. Adapting to this new scenario is a challenge for companies, especially small and medium-sizedenterprises (SMEs) that are suffering first-hand how their specialization is turning against them.The objective of this work is to provide a tool that can serve as a basis to face these challenges in an effective way.Therefore the presented framework, thanks to its modular architecture, allows focusing on the different needs of eachparticular company and offers the possibility of scaling the system for future requirements. The presented platform isdivided into three layers, namely: interface with robot systems, the execution engine and the application developmentlayer.Taking advantage of the provided ecosystem by this framework, different modules have been developed in order toface the mentioned challenges of the industry. On the one hand, to address the need of product customization, theintegration of tools that increase the versatility of the cell are proposed. An example of such tools is skill basedprogramming. By applying this technique a process can be intuitively adapted to the variations or customizations thateach product requires. The use of skills favours the reuse and generalization of developed robot programs.Regarding the variation of the production volumes, a system which permits a greater mobility and a faster reconfigurationis necessary. If in a certain situation a line has a production peak, mechanisms for balancing the loadwith a reasonable cost are required. In this respect, the architecture allows an easy integration of different roboticsystems, actuators, sensors, etc. In addition, thanks to the developed calibration and set-up techniques, the system canbe adapted to new workspaces at an effective time/cost.With respect to the third mentioned topic, an agent-based monitoring system is proposed. This module opens up amultitude of possibilities for the integration of auxiliary modules of protection and security for collaboration andinteraction between people and robots, something that will be necessary in the not so distant future.For demonstrating the advantages and adaptability improvement of the developed framework, a series of real usecases have been presented. In each of them different problematic has been resolved using developed skills,demonstrating how are adapted easily to the different casuistic

    Collaborative Trolley Transportation System with Autonomous Nonholonomic Robots

    Full text link
    Cooperative object transportation using multiple robots has been intensively studied in the control and robotics literature, but most approaches are either only applicable to omnidirectional robots or lack a complete navigation and decision-making framework that operates in real time. This paper presents an autonomous nonholonomic multi-robot system and an end-to-end hierarchical autonomy framework for collaborative luggage trolley transportation. This framework finds kinematic-feasible paths, computes online motion plans, and provides feedback that enables the multi-robot system to handle long lines of luggage trolleys and navigate obstacles and pedestrians while dealing with multiple inherently complex and coupled constraints. We demonstrate the designed collaborative trolley transportation system through practical transportation tasks, and the experiment results reveal their effectiveness and reliability in complex and dynamic environments

    Virtual Reality Games for Motor Rehabilitation

    Get PDF
    This paper presents a fuzzy logic based method to track user satisfaction without the need for devices to monitor users physiological conditions. User satisfaction is the key to any product’s acceptance; computer applications and video games provide a unique opportunity to provide a tailored environment for each user to better suit their needs. We have implemented a non-adaptive fuzzy logic model of emotion, based on the emotional component of the Fuzzy Logic Adaptive Model of Emotion (FLAME) proposed by El-Nasr, to estimate player emotion in UnrealTournament 2004. In this paper we describe the implementation of this system and present the results of one of several play tests. Our research contradicts the current literature that suggests physiological measurements are needed. We show that it is possible to use a software only method to estimate user emotion
    corecore