57,627 research outputs found

    How does the design of landmarks on a mobile map influence wayfinding experts’ spatial learning during a real-world navigation task?

    Full text link
    Humans increasingly rely on GPS-enabled mobile maps to navigate novel environments. However, this reliance can negatively affect spatial learning, which can be detrimental even for expert navigators such as search and rescue personnel. Landmark visualization has been shown to improve spatial learning in general populations by facilitating object identification between the map and the environment. How landmark visualization supports expert users’ spatial learning during map-assisted navigation is still an open research question. We thus conducted a real-world study with wayfinding experts in an unknown residential neighborhood. We aimed to assess how two different landmark visualization styles (abstract 2D vs. realistic 3D buildings) would affect experts’ spatial learning in a map-assisted navigation task during an emergency scenario. Using a between-subjects design, we asked Swiss military personnel to follow a given route using a mobile map, and to identify five task-relevant landmarks along the route. We recorded experts’ gaze behavior while navigating and examined their spatial learning after the navigation task. We found that experts’ spatial learning improved when they focused their visual attention on the environment, but the direction of attention between the map and the environment was not affected by the landmark visualization style. Further, there was no difference in spatial learning between the 2D and 3D groups. Contrary to previous research with general populations, this study suggests that the landmark visualization style does not enhance expert navigators’ navigation or spatial learning abilities, thus highlighting the need for population-specific mobile map design solutions

    Smart Geographic object: Toward a new understanding of GIS Technology in Ubiquitous Computing

    Get PDF
    One of the fundamental aspects of ubiquitous computing is the instrumentation of the real world by smart devices. This instrumentation constitutes an opportunity to rethink the interactions between human beings and their environment on the one hand, and between the components of this environment on the other. In this paper we discuss what this understanding of ubiquitous computing can bring to geographic science and particularly to GIS technology. Our main idea is the instrumentation of the geographic environment through the instrumentation of geographic objects composing it. And then investigate how this instrumentation can meet the current limitations of GIS technology, and offers a new stage of rapprochement between the earth and its abstraction. As result, the current research work proposes a new concept we named Smart Geographic Object SGO. The latter is a convergence point between the smart objects and geographic objects, two concepts appertaining respectively to

    Using visual analytics to develop situation awareness in astrophysics

    Get PDF
    We present a novel collaborative visual analytics application for cognitively overloaded users in the astrophysics domain. The system was developed for scientists who need to analyze heterogeneous, complex data under time pressure, and make predictions and time-critical decisions rapidly and correctly under a constant influx of changing data. The Sunfall Data Taking system utilizes several novel visualization and analysis techniques to enable a team of geographically distributed domain specialists to effectively and remotely maneuver a custom-built instrument under challenging operational conditions. Sunfall Data Taking has been in production use for 2 years by a major international astrophysics collaboration (the largest data volume supernova search currently in operation), and has substantially improved the operational efficiency of its users. We describe the system design process by an interdisciplinary team, the system architecture and the results of an informal usability evaluation of the production system by domain experts in the context of Endsley's three levels of situation awareness

    Fireground location understanding by semantic linking of visual objects and building information models

    Get PDF
    This paper presents an outline for improved localization and situational awareness in fire emergency situations based on semantic technology and computer vision techniques. The novelty of our methodology lies in the semantic linking of video object recognition results from visual and thermal cameras with Building Information Models (BIM). The current limitations and possibilities of certain building information streams in the context of fire safety or fire incident management are addressed in this paper. Furthermore, our data management tools match higher-level semantic metadata descriptors of BIM and deep-learning based visual object recognition and classification networks. Based on these matches, estimations can be generated of camera, objects and event positions in the BIM model, transforming it from a static source of information into a rich, dynamic data provider. Previous work has already investigated the possibilities to link BIM and low-cost point sensors for fireground understanding, but these approaches did not take into account the benefits of video analysis and recent developments in semantics and feature learning research. Finally, the strengths of the proposed approach compared to the state-of-the-art is its (semi -)automatic workflow, generic and modular setup and multi-modal strategy, which allows to automatically create situational awareness, to improve localization and to facilitate the overall fire understanding

    Smartphone Augmented Reality Applications for Tourism

    Get PDF
    Invisible, attentive and adaptive technologies that provide tourists with relevant services and information anytime and anywhere may no longer be a vision from the future. The new display paradigm, stemming from the synergy of new mobile devices, context-awareness and AR, has the potential to enhance tourists’ experiences and make them exceptional. However, effective and usable design is still in its infancy. In this publication we present an overview of current smartphone AR applications outlining tourism-related domain-specific design challenges. This study is part of an ongoing research project aiming at developing a better understanding of the design space for smartphone context-aware AR applications for tourists

    Supervised Autonomous Locomotion and Manipulation for Disaster Response with a Centaur-like Robot

    Full text link
    Mobile manipulation tasks are one of the key challenges in the field of search and rescue (SAR) robotics requiring robots with flexible locomotion and manipulation abilities. Since the tasks are mostly unknown in advance, the robot has to adapt to a wide variety of terrains and workspaces during a mission. The centaur-like robot Centauro has a hybrid legged-wheeled base and an anthropomorphic upper body to carry out complex tasks in environments too dangerous for humans. Due to its high number of degrees of freedom, controlling the robot with direct teleoperation approaches is challenging and exhausting. Supervised autonomy approaches are promising to increase quality and speed of control while keeping the flexibility to solve unknown tasks. We developed a set of operator assistance functionalities with different levels of autonomy to control the robot for challenging locomotion and manipulation tasks. The integrated system was evaluated in disaster response scenarios and showed promising performance.Comment: In Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, October 201

    Weighted simplicial complex reconstruction from mobile laser scanning using sensor topology

    Full text link
    We propose a new method for the reconstruction of simplicial complexes (combining points, edges and triangles) from 3D point clouds from Mobile Laser Scanning (MLS). Our method uses the inherent topology of the MLS sensor to define a spatial adjacency relationship between points. We then investigate each possible connexion between adjacent points, weighted according to its distance to the sensor, and filter them by searching collinear structures in the scene, or structures perpendicular to the laser beams. Next, we create and filter triangles for each triplet of self-connected edges and according to their local planarity. We compare our results to an unweighted simplicial complex reconstruction.Comment: 8 pages, 11 figures, CFPT 2018. arXiv admin note: substantial text overlap with arXiv:1802.0748
    • …
    corecore