3,646 research outputs found

    Deep Shape Matching

    Full text link
    We cast shape matching as metric learning with convolutional networks. We break the end-to-end process of image representation into two parts. Firstly, well established efficient methods are chosen to turn the images into edge maps. Secondly, the network is trained with edge maps of landmark images, which are automatically obtained by a structure-from-motion pipeline. The learned representation is evaluated on a range of different tasks, providing improvements on challenging cases of domain generalization, generic sketch-based image retrieval or its fine-grained counterpart. In contrast to other methods that learn a different model per task, object category, or domain, we use the same network throughout all our experiments, achieving state-of-the-art results in multiple benchmarks.Comment: ECCV 201

    Landmark-based approaches for goal recognition as planning

    Get PDF
    This article is a revised and extended version of two papers published at AAAI 2017 (Pereira et al., 2017b) and ECAI 2016 (Pereira and Meneguzzi, 2016). We thank the anonymous reviewers that helped improve the research in this article. The authors thank Shirin Sohrabi for discussing the way in which the algorithms of Sohrabi et al. (2016) should be configured, and Yolanda Escudero-Martın for providing code for the approach of E-Martın et al. (2015) and engaging with us. We also thank Miquel Ramırez and Mor Vered for various discussions, and Andre Grahl Pereira for a discussion of properties of our algorithm. Felipe thanks CNPq for partial financial support under its PQ fellowship, grant number 305969/2016-1.Peer reviewedPostprin

    Planning Landmark Based Goal Recognition Revisited: Does Using Initial State Landmarks Make Sense?

    Full text link
    Goal recognition is an important problem in many application domains (e.g., pervasive computing, intrusion detection, computer games, etc.). In many application scenarios, it is important that goal recognition algorithms can recognize goals of an observed agent as fast as possible. However, many early approaches in the area of Plan Recognition As Planning, require quite large amounts of computation time to calculate a solution. Mainly to address this issue, recently, Pereira et al. developed an approach that is based on planning landmarks and is much more computationally efficient than previous approaches. However, the approach, as proposed by Pereira et al., also uses trivial landmarks (i.e., facts that are part of the initial state and goal description are landmarks by definition). In this paper, we show that it does not provide any benefit to use landmarks that are part of the initial state in a planning landmark based goal recognition approach. The empirical results show that omitting initial state landmarks for goal recognition improves goal recognition performance.Comment: Will be presented at KI 202

    Appearance-Based Gaze Estimation in the Wild

    Full text link
    Appearance-based gaze estimation is believed to work well in real-world settings, but existing datasets have been collected under controlled laboratory conditions and methods have been not evaluated across multiple datasets. In this work we study appearance-based gaze estimation in the wild. We present the MPIIGaze dataset that contains 213,659 images we collected from 15 participants during natural everyday laptop use over more than three months. Our dataset is significantly more variable than existing ones with respect to appearance and illumination. We also present a method for in-the-wild appearance-based gaze estimation using multimodal convolutional neural networks that significantly outperforms state-of-the art methods in the most challenging cross-dataset evaluation. We present an extensive evaluation of several state-of-the-art image-based gaze estimation algorithms on three current datasets, including our own. This evaluation provides clear insights and allows us to identify key research challenges of gaze estimation in the wild

    Explainable Goal Recognition: A Framework Based on Weight of Evidence

    Full text link
    We introduce and evaluate an eXplainable Goal Recognition (XGR) model that uses the Weight of Evidence (WoE) framework to explain goal recognition problems. Our model provides human-centered explanations that answer why? and why not? questions. We computationally evaluate the performance of our system over eight different domains. Using a human behavioral study to obtain the ground truth from human annotators, we further show that the XGR model can successfully generate human-like explanations. We then report on a study with 60 participants who observe agents playing Sokoban game and then receive explanations of the goal recognition output. We investigate participants' understanding obtained by explanations through task prediction, explanation satisfaction, and trust.Comment: 11 pages, 5 figure

    L'apprentissage profond, une puissante alternative pour la reconnaissance d'intention

    Get PDF
    Ce mémoire s'inscrit dans la lignée d'une avancée de connaissances en reconnaissance d'intention, une discipline de recherche en intelligence artificielle visant à inférer les buts poursuivis par un individu à l'aide d'observations de son comportement. Ce problème, du fait de sa complexité, reste irrésolu dans les domaines réels: les voitures autonomes, les instruments de détection d'intrusion, les conseillers virtuels par messagerie et tant d'autres profiteraient encore actuellement d'une capacité de reconnaissance d'intention. Longtemps abordé sous l'angle de considérations symboliques spécifiées par des experts humains, le problème commence à être résolu par des approches récentes usant d'algorithmes d'apprentissage dans des contextes simples. Nous nous inspirons ici des progrès de l'apprentissage profond dans des domaines connexes pour en faire usage à des fins de reconnaissance de but à long-terme. Encore sous-exploité pour cette catégorie de problèmes, nous l'avons mis à l'épreuve pour résoudre les problèmes traités dans la littérature et cherchons à améliorer les performances de l'état de l'art. Pour ce faire, nous présentons trois articles de recherche. Le premier, accepté au workshop PAIR (Plan, Activity and Intent Recognition) lors de la conférence AAAI 2018 (Association for the Advancement of Artificial Intelligence), propose une comparaison expérimentale entre différentes architectures d'apprentissage profond et les méthodes symboliques de l'état de l'art. Nous montrons de ce fait que nos meilleurs résultats surpassent ces méthodes symboliques dans les domaines considérés. Le deuxième, publié sur arXiv, introduit une méthode pour permettre à un réseau de neurones de généraliser rapidement à plusieurs environnements grâce à une projection des données sur un espace intermédiaire et en s'inspirant des progrès du few-shot transfer learning. Enfin, le troisième, soumis à ICAPS 2020 (International Conference on Automated Planning and Scheduling), améliore encore les résultats précédents en fournissant aux réseaux des caractéristiques supplémentaires leur permettant de se projeter dans le futur avec une capacité d'imagination et de résoudre le principal défaut inhérent aux approches symboliques de l'état de l'art, à savoir la dépendance à une représentation approximée de l'environnement
    corecore