9,829 research outputs found

    On the convergence of autonomous agent communities

    Get PDF
    This is the post-print version of the final published paper that is available from the link below. Copyright @ 2010 IOS Press and the authors.Community is a common phenomenon in natural ecosystems, human societies as well as artificial multi-agent systems such as those in web and Internet based applications. In many self-organizing systems, communities are formed evolutionarily in a decentralized way through agents' autonomous behavior. This paper systematically investigates the properties of a variety of the self-organizing agent community systems by a formal qualitative approach and a quantitative experimental approach. The qualitative formal study by applying formal specification in SLABS and Scenario Calculus has proven that mature and optimal communities always form and become stable when agents behave based on the collective knowledge of the communities, whereas community formation does not always reach maturity and optimality if agents behave solely based on individual knowledge, and the communities are not always stable even if such a formation is achieved. The quantitative experimental study by simulation has shown that the convergence time of agent communities depends on several parameters of the system in certain complicated patterns, including the number of agents, the number of community organizers, the number of knowledge categories, and the size of the knowledge in each category

    Active Perception in Adversarial Scenarios using Maximum Entropy Deep Reinforcement Learning

    Full text link
    We pose an active perception problem where an autonomous agent actively interacts with a second agent with potentially adversarial behaviors. Given the uncertainty in the intent of the other agent, the objective is to collect further evidence to help discriminate potential threats. The main technical challenges are the partial observability of the agent intent, the adversary modeling, and the corresponding uncertainty modeling. Note that an adversary agent may act to mislead the autonomous agent by using a deceptive strategy that is learned from past experiences. We propose an approach that combines belief space planning, generative adversary modeling, and maximum entropy reinforcement learning to obtain a stochastic belief space policy. By accounting for various adversarial behaviors in the simulation framework and minimizing the predictability of the autonomous agent's action, the resulting policy is more robust to unmodeled adversarial strategies. This improved robustness is empirically shown against an adversary that adapts to and exploits the autonomous agent's policy when compared with a standard Chance-Constraint Partially Observable Markov Decision Process robust approach

    Understanding the Role of Trust in Human-Autonomy Teaming

    Get PDF
    This study aims to better understand trust in human-autonomy teams, finding that trust is related to team performance. A wizard of oz methodology was used in an experiment to simulate an autonomous agent as a team member in a remotely piloted aircraft system environment. Specific focuses of the study were team performance and team social behaviors (specifically trust) of human-autonomy teams. Results indicate 1) that there are lower levels of trust in the autonomous agent in low performing teams than both medium and high performing teams, 2) there is a loss of trust in the autonomous agent across low, medium, and high performing teams over time, and 3) that in addition to the human team members indicating low levels of trust in the autonomous agent, both low and medium performing teams also indicated lower levels of trust in their human team members

    Response knowledge learning of autonomous agent

    Get PDF
    In robot applications, the performance of a robot agent is measured by the quantity of award received from its response. Many literatures [1-5] define the response as either a state diagram or a neural network. Due to the absence of a desired response, neither of them is applicable to an unstructural environment. In this paper, a novel Response Knowledge Learning algorithm is proposed to handle this domain. By using a set of experiences, the algorithm can extract the contributed experiences to construct the response function. Two sets of environments are provided to illustrate the performance of the proposed algorithm. The results show that it can effectively construct the response function that receives an award which is very close to the true maximum.published_or_final_versio

    Information theory and representation in associative word learning

    Get PDF
    A significant portion of early language learning can be viewed as an associative learning problem. We investigate the use of associative language learning based on the principle that words convey Shannon information about the environment. We discuss the shortcomings in representation used by previous associative word learners and propose a functional representation that not only denotes environmental categories, but serves as the basis for activities and interaction with the environment. We present experimental results with an autonomous agent acquiring language

    Integrating planning, execution, and learning

    Get PDF
    To achieve the goal of building an autonomous agent, the usually disjoint capabilities of planning, execution, and learning must be used together. An architecture, called MAX, within which cognitive capabilities can be purposefully and intelligently integrated is described. The architecture supports the codification of capabilities as explicit knowledge that can be reasoned about. In addition, specific problem solving, learning, and integration knowledge is developed
    • 

    corecore