892 research outputs found

    A multimodal smartphone interface for active perception by visually impaired

    Get PDF
    The diffuse availability of mobile devices, such as smartphones and tablets, has the potential to bring substantial benefits to the people with sensory impairments. The solution proposed in this paper is part of an ongoing effort to create an accurate obstacle and hazard detector for the visually impaired, which is embedded in a hand-held device. In particular, it presents a proof of concept for a multimodal interface to control the orientation of a smartphone's camera, while being held by a person, using a combination of vocal messages, 3D sounds and vibrations. The solution, which is to be evaluated experimentally by users, will enable further research in the area of active vision with human-in-the-loop, with potential application to mobile assistive devices for indoor navigation of visually impaired people

    Exploring the Use of Wearables to Enable Indoor Navigation for Blind Users

    Get PDF
    One of the challenges that people with visual impairments (VI) have to have to confront daily, is navigating independently through foreign or unfamiliar spaces.Navigating through unfamiliar spaces without assistance is very time consuming and leads to lower mobility. Especially in the case of indoor environments where the use of GPS is impossible, this task becomes even harder.However, advancements in mobile and wearable computing pave the path to new cheap assistive technologies that can make the lives of people with VI easier.Wearable devices have great potential for assistive applications for users who are blind as they typically feature a camera and support hands and eye free interaction. Smart watches and heads up displays (HUDs), in combination with smartphones, can provide a basis for development of advanced algorithms, capable of providing inexpensive solutions for navigation in indoor spaces. New interfaces are also introduced making the interaction between users who are blind and mo-bile devices more intuitive.This work presents a set of new systems and technologies created to help users with VI navigate indoor environments. The first system presented is an indoor navigation system for people with VI that operates by using sensors found in mo-bile devices and virtual maps of the environment. The second system presented helps users navigate large open spaces with minimum veering. Next a study is conducted to determine the accuracy of pedometry based on different body placements of the accelerometer sensors. Finally, a gesture detection system is introduced that helps communication between the user and mobile devices by using sensors in wearable devices

    Comparative analysis of computer-vision and BLE technology based indoor navigation systems for people with visual impairments

    Get PDF
    Background: Considerable number of indoor navigation systems has been proposed to augment people with visual impairments (VI) about their surroundings. These systems leverage several technologies, such as computer-vision, Bluetooth low energy (BLE), and other techniques to estimate the position of a user in indoor areas. Computer-vision based systems use several techniques including matching pictures, classifying captured images, recognizing visual objects or visual markers. BLE based system utilizes BLE beacons attached in the indoor areas as the source of the radio frequency signal to localize the position of the user. Methods: In this paper, we examine the performance and usability of two computer-vision based systems and BLE-based system. The first system is computer-vision based system, called CamNav that uses a trained deep learning model to recognize locations, and the second system, called QRNav, that utilizes visual markers (QR codes) to determine locations. A field test with 10 blindfolded users has been conducted while using the three navigation systems. Results: The obtained results from navigation experiment and feedback from blindfolded users show that QRNav and CamNav system is more efficient than BLE based system in terms of accuracy and usability. The error occurred in BLE based application is more than 30% compared to computer vision based systems including CamNav and QRNav. Conclusions: The developed navigation systems are able to provide reliable assistance for the participants during real time experiments. Some of the participants took minimal external assistance while moving through the junctions in the corridor areas. Computer vision technology demonstrated its superiority over BLE technology in assistive systems for people with visual impairments. - 2019 The Author(s).Scopu

    NavCog: turn-by-turn smartphone navigation assistant for people with visual impairments or blindness

    Get PDF
    NavCog is a novel smartphone navigation system for people with visual impairments or blindness, capable of assisting the users during autonomous mobility in complex and unfamiliar indoor/outdoor environments. The accurate localization achieved by NavCog is used for precise turn-by-turn way-finding assistance as the first step, but the ultimate goal is to present a variety of location based information to the user, such as points of interest gathered from social media and online geografic information services

    Inclusive Landmark based Pedestrian Wayfinding via Multi-modal Directions

    Get PDF
    Navigational skills are fundamental to travelling from place to place, personal independence and community integration [2]. Current research in pedestrian wayfinding suggests that people vary significantly in their choice of navigation modalities [6, 7, 25]. In addition, pedestrians with learning disabilities find it difficult to recall routes travelled daily and stay oriented while enroute to unknown locations. This paper proposes a wayfinding interface that has 2 components: 1) temporary poly-coated cardboard signage along with imprinted information indicating a specific destination, minutes by foot, directional arrow and a QR code; 2) online interactive website to provide additional contextualized navigation instructions for pedestrians through various modalities. The University of Toronto Scarborough campus (UTSC) is being used as the physical environment to implement and test the proposed wayfinding interface. The QR code tags link the cardboard signage to the online interface and generate streaming of route instructions in the modes of panoramic video, photographs, aerial map, audio or text. The goal of the proposed wayfinding system is to aid UTSC pedestrians - especially those with learning disabilities - to orient themselves and navigate to their destination through multi-modal landmark-based, turn-by-turn directions

    Wayfinding and Navigation for People with Disabilities Using Social Navigation Networks

    Get PDF
    To achieve safe and independent mobility, people usually depend on published information, prior experience, the knowledge of others, and/or technology to navigate unfamiliar outdoor and indoor environments. Today, due to advances in various technologies, wayfinding and navigation systems and services are commonplace and are accessible on desktop, laptop, and mobile devices. However, despite their popularity and widespread use, current wayfinding and navigation solutions often fail to address the needs of people with disabilities (PWDs). We argue that these shortcomings are primarily due to the ubiquity of the compute-centric approach adopted in these systems and services, where they do not benefit from the experience-centric approach. We propose that following a hybrid approach of combining experience-centric and compute-centric methods will overcome the shortcomings of current wayfinding and navigation solutions for PWDs
    corecore