6 research outputs found

    Performance protection strategies for resilient cockpits : the case of fatigue in an unexpected problem-solving situation

    No full text
    Résumé confidentielRésumé confidentie

    Integrating Aircrew Resources Variability into the Design of Future Cockpit

    Get PDF
    Cockpit design may have major consequences on future pilot’s tasks. Functions needed to ensure flight safety are shared between pilots and aircraft systems. Today, cockpits are designed considering a theoretical minimum level of crew resources availability although it is widely acknowledged that availability of crew resources may vary from time to time because of aircrew internal state. Therefore, it is crucial to ensure that future cockpits are designed while taking into consideration this variability. This work aims at developing a methodology enabling designer to systematically integrate crew resources availability in the design process. In order to assess the impact of different sources of variability, the principles of the systemic model FRAM is used. The present work is analyzing the impact of crew resources availability on several use cases using FRAM principles. The data are collected by the means of focus group with operational and human factors experts. Some preliminary results are presented and discussed

    Stratégies de protection de la performance pour la conception de cockpits résilients (le cas de la fatigue en situation inattendue de résolution de problème)

    No full text
    Résumé confidentielRésumé confidentielPARIS5-Bibliotheque electronique (751069902) / SudocSudocFranceF

    Explaining Aha! moments in artificial agents through IKE-XAI: Implicit Knowledge Extraction for eXplainable AI

    No full text
    International audienceDuring the learning process, a child develops a mental representation of the task he or she is learning. A Machine Learning algorithm develops a latent representation of the task it learns. We investigate the development of the knowledge construction of an artificial agent (AA) by getting inspiration from the one of children, through the analysis of its behavior, i.e., its sequences of moves. We focus on the Tower of Hanoï (TOH) task, a well-known transformation problem in the field of problem-solving as well as one of the fundamental tasks to study children’s knowledge construction about their world and the Aha! phenomenon. We define knowledge here as a set of facts, information, and skills acquired through experience by the AA that contribute to gaining a theoretical or practical understanding of a subject or the world. The main contribution of our work is to propose a 3-step end-to-end methodology for knowledge extraction from AA named Implicit Knowledge Extraction with eXplainable Artificial Intelligence (IKE-XAI). IKE-XAI extracts the AA implicit knowledge in form of an automaton, encoded during its learning. We showcase this technique to solve and explain the TOH task when researchers have only access to moves that represent observational behavior as in human-machine interaction. The 3 steps of IKE-XAI are: first, a Q-learning agent that learns to perform the TOH task; second, a trained recurrent neural network with LSTM units that encodes an implicit representation of the TOH task; and third, an XAI process using a post-hoc implicit rule extraction algorithm to extract graph representations (Finite State Automata, FSA) as visual and explicit explanations of the behavior of the Q-learning agent. This methodology blends neural and symbolic (in our case FSA) components to provide more interpretable model outcomes.At the experimental level, we demonstrate that it is possible to extract the vision of the AA of a simple task (TOH with N=3 disks) and complex one (TOH with N=4 disks and N =6 disks), in the form of FSA that represents AA’s problem-solving strategies, for their explainability. In parallel to the decrease in the average number of movements required to complete a task, namely the acquisition of expertise, we also observed a change in the FSAs extracted at different moments of this acquisition of expertise. The analysis of the characteristics of the FSAs shows a change in the number of nodes and the weights of the transitions. Regarding the Aha! moment, in the 3 experimental contexts, the analyses carried out allowed us to conclude that the Aha! moment for an AA occurs when it changes its behavior in a noticeable way, which translates into a significant change in the extracted FSAs and a stabilization of these. Our experiments show that the IKE-XAI approach helps to understand the development of the Q-learning agent behavior by providing a global explanation of its knowledge evolution during learning. IKE-XAI also allows researchers to identify the agent's Aha! moment by determining from what moment the knowledge representation stabilizes and the agent no longer learns.As a conclusion, we showed that IKE-XAI makes it possible to elucidate the evolution of knowledge acquisition of a learning AA through the study of its behavior over time in terms of an extracted, synthesizing FSA. This allows us to convey, in a symbolic manner, a more explainable vision of AAs. This work also brings a light on the subject of the Aha! moment for autonomous agents and beyond, it leads to a reflection on the question of the definition of insight for an autonomous artificial agent. The convergence of models is thus interesting for the study of this phenomenon in autonomous artificial agents, and more globally for the question of explainability

    Explaining Aha! moments in artificial agents through IKE-XAI: Implicit Knowledge Extraction for eXplainable AI

    No full text
    International audienceDuring the learning process, a child develops a mental representation of the task he or she is learning. A Machine Learning algorithm develops also a latent representation of the task it learns. We investigate the development of the knowledge construction of an artificial agent through the analysis of its behavior, i.e., its sequences of moves while learning to perform the Tower of Hanoï (TOH) task. The TOH is a wellknown task in experimental contexts to study the problem-solving processes and one of the fundamental processes of children's knowledge construction about their world. We position ourselves in the field of explainable reinforcement learning for developmental robotics, at the crossroads of cognitive modeling and explainable AI. Our main contribution proposes a 3-step methodology named Implicit Knowledge Extraction with eXplainable Artificial Intelligence (IKE-XAI) to extract the implicit knowledge, in form of an automaton, encoded by an artificial agent during its learning. We showcase this technique to solve and explain the TOH task when researchers have only access to moves that represent observational behavior as in human-machine interaction. Therefore, to extract the agent acquired knowledge at different stages of its training, our approach combines: first, a Q-learning agent that learns to perform the TOH task; second, a trained recurrent neural network that encodes an implicit representation of the TOH task; and third, an XAI process using a post-hoc implicit rule extraction algorithm to extract finite state automata. We propose using graph representations as visual and explicit explanations of the behavior of the Q-learning agent. Our experiments show that the IKE-XAI approach helps understanding the development of the Q-learning agent behavior by providing a global explanation of its knowledge evolution during learning. IKE-XAI also allows researchers to identify the agent's Aha! moment by determining from what moment the knowledge representation stabilizes and the agent no longer learns

    Annuaire 2005-2006

    No full text
    corecore