396 research outputs found
Using remote vision: The effects of video image frame rate on visual object recognition performance
This is the author's accepted manuscript. The final published article is available from the link below. Copyright @ 2010 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other users, including reprinting/ republishing this material for advertising or promotional purposes, creating new collective works for resale or redistribution to servers or lists, or reuse of any copyrighted components of this work in other works.The process of using remote vision was simulated in order to determine the effects of video image frame rate on the performance in visual recognition of stationary environmental hazards in the dynamic video footage of the pedestrian travel environment. The recognition performance was assessed against two different video image frame rate variations: 25 and 2 fps. The assessment included a range of objective and subjective criteria. The obtained results show that the effects of the frame rate variations on the performance are statistically insignificant. This paper belongs to the process of development of a novel system for navigation of visually impaired pedestrians. The navigation system includes a remote vision facility, and the visual recognition of the environmental hazards by the sighted human guide is a basic activity in aiding the visually impaired user of the system in mobility
Recommended from our members
Interface design for a remote guidance system for the blind: Using dual-screen displays
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The mobility for the visually impaired people is one of the main challenges that researchers are still facing around the world. Although some projects have been conducted to improve the mobility of visually impaired people, further research is still needed. One of these projects is Brunel Remote Guidance System (BRGS). BRGS is aimed to assist visually impaired users in avoiding obstacles and reaching their destinations safely by providing online instructions via a remote sighted guide.
This study comes as continuation of the development process of BRGS; the main aim that has been achieved of this research is the optimisation of the interface design for the system guide terminal. This helps the sighted guide to assist the VIUs to avoid obstacles safely and comfortably in the micro-navigation, as well as to keep them on the right track to reach their destination in the macro-navigation. After using the content analysis, the performance factors and their assessments method were identified in each BRGSâ element, which concluded that there is a lack of research on the guide terminal setup and the assessment method for the sighted guide performance. Furthermore, there are no model to assist the sighted guide performance and two-screen displays used in the literature review and similar projects. A model was designed as a platform to conduct the evaluation on sighted guide performance. Based on this model, the computer-based simulation was established and tested, which made the simulation is ready for next task; the evaluation of the sighted guide performance. The conducted study determined the effects of the two-screen displays on the recognition performance of the 80 participants in the guide terminal. The performance was measured with the context of four different resolution conditions. The study was based on a simulation technique, which is consisted of two key performance elements in order to examine the sighted guide performance; the macro-navigation element and the micro-navigation element. The results show that the two-screen displays have an effect on the performance of the sighted guide. The optimum setup for the two-screen displays for the guide terminal consisted of a big digital map screen display (4CIF [704p x 576p]) and a small video image screen display (CIF [352p x 288p]), which one of the four different resolutions. This interface design has been recommended as a final setup in the guide terminal
Interface design for a remote guidance system for the blind : using dual-screen displays
The mobility for the visually impaired people is one of the main challenges that researchers are still facing around the world. Although some projects have been conducted to improve the mobility of visually impaired people, further research is still needed. One of these projects is Brunel Remote Guidance System (BRGS). BRGS is aimed to assist visually impaired users in avoiding obstacles and reaching their destinations safely by providing online instructions via a remote sighted guide. This study comes as continuation of the development process of BRGS; the main aim that has been achieved of this research is the optimisation of the interface design for the system guide terminal. This helps the sighted guide to assist the VIUs to avoid obstacles safely and comfortably in the micro-navigation, as well as to keep them on the right track to reach their destination in the macro-navigation. After using the content analysis, the performance factors and their assessments method were identified in each BRGSâ element, which concluded that there is a lack of research on the guide terminal setup and the assessment method for the sighted guide performance. Furthermore, there are no model to assist the sighted guide performance and two-screen displays used in the literature review and similar projects. A model was designed as a platform to conduct the evaluation on sighted guide performance. Based on this model, the computer-based simulation was established and tested, which made the simulation is ready for next task; the evaluation of the sighted guide performance. The conducted study determined the effects of the two-screen displays on the recognition performance of the 80 participants in the guide terminal. The performance was measured with the context of four different resolution conditions. The study was based on a simulation technique, which is consisted of two key performance elements in order to examine the sighted guide performance; the macro-navigation element and the micro-navigation element. The results show that the two-screen displays have an effect on the performance of the sighted guide. The optimum setup for the two-screen displays for the guide terminal consisted of a big digital map screen display (4CIF [704p x 576p]) and a small video image screen display (CIF [352p x 288p]), which one of the four different resolutions. This interface design has been recommended as a final setup in the guide terminal.EThOS - Electronic Theses Online ServiceGBUnited Kingdo
Recommended from our members
Mobile assistive technologies for the visually impaired
There are around 285 million visually impaired people worldwide, and around 370,000 people are registered as blind or partially sighted in the UK. Ongoing advances in information technology (IT) are increasing the scope for IT-based mobile assistive technologies to facilitate the independence, safety, and improved quality of life of the visually impaired. Research is being directed at making mobile phones and other handheld devices accessible via our haptic (touch) and audio sensory channels. We review research and innovation within the field of mobile assistive technology for the visually impaired and, in so doing, highlight the need for successful collaboration between clinical expertise, computer science, and domain users to realize fully the potential benefits of such technologies. We initially reflect on research that has been conducted to make mobile phones more accessible to people with vision loss. We then discuss innovative assistive applications designed for the visually impaired that are either delivered via mainstream devices and can be used while in motion (e.g., mobile phones) or are embedded within an environment that may be in motion (e.g., public transport) or within which the user may be in motion (e.g., smart homes)
Recommended from our members
Precise positioning in real-time using GPS-RTK signal for visually impaired people navigation system
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University, 24/9/2010.This thesis presents the research carried out to investigate and achieve highly reliable and accurate navigation system of guidance for visually impaired pedestrians. The main aim with this PhD project has been to identify the limits and insufficiencies in utilising Network Real-Time Kinematic Global Navigation Satellite Systems (NRTK GNSS) and its augmentation techniques within the frame of pedestrian applications in a variety of environments and circumstances. Moreover, the system can be used in many other applications, including unmanned vehicles, military applications, police, etc. NRTK GNSS positioning is considered to be a superior solution in comparison to the conventional standalone Global Positioning System (GPS) technique whose accuracy is highly affected by the distance dependent errors such as satellite orbital and atmospheric biases.
Nevertheless, NRTK GNSS positioning is particularly constrained by wireless data link coverage, delays of correction and transmission and completeness, GPS and GLONASS signal availability, etc., which could downgrade the positioning quality of the NRTK results.
This research is based on the dual frequency NRTK GNSS (GPS and GLONASS). Additionally, it is incorporated into several positioning and communication methods responsible for data correction while providing the position solutions, in which all identified contextual factors and application requirements are accounted.
The positioning model operates through client-server based architecture consisted of a Navigation Service Centre (NSC) and a Mobile Navigation Unit (MNU). Hybrid functional approaches were consisting of several processing procedures allowing the positioning model to operate in position determination modes. NRTK GNSS and augmentation service is used if enough navigation information was available at the MNU using its local positioning device (GPS/GLONASS receiver).The positioning model at MNU was experimentally evaluated and centimetric accuracy was generally attained during both static and kinematic tests in various environments (urban, suburban and rural). This high accuracy was merely affected by some level of unavailability mainly caused by GPS and GLONASS signal blockage. Additionally, the influence of the number of satellites in view, dilution of precision (DOP) and age corrections (AoC) over the accuracy and stability of the NRTK GNSS solution was also investigated during this research and presented in the thesis.
This positioning performance has outperformed the existing GPS service. In addition, utilising a simulation evaluation facility the positioning model at MNU performance was quantified with reference to a hybrid positioning service that will be offered by future Galileo Open Service (OS) along with GPS. However, a significant difference in terms of the service availability for the advantage of the hybrid system was experienced in all remaining scenarios and environments more especially the urban areas due to surrounding obstacles and conditions.
As an outcome of this research a new and precise positioning model was proposed. The adaptive framework is understood as approaching an integration of the available positioning technology into the context of surrounding wireless communication for a maintainable performance. The positioning model has the capability of delivering indeed accurate, precise and consistent position solutions, and thus is fulfilling the requirements of visually impaired people navigation application, as identified in the adaptive framework
A navigation and object location device for the blind
Gemstone Team VisionTeam Vision's goal is to create a navigation system for the blind. To achieve
this, we took a multi-pronged approach. First, through surveys, we assessed the
needs of the blind community and developed a system around those needs. Then,
using recent technology, we combined a global positioning system (GPS), inertial
navigation unit (INU), computer vision algorithms, and audio and haptic
interfaces into one system. The GPS and INU work together to provide walking
directions from building to building when outdoors and the computer vision algorithms
identify and locate objects such as signs and landmarks, both indoors and
outdoors. The speech-based interface ties the GPS, INU, and computer vision
algorithms together into an interactive audio-based navigation device. Finally,
the haptic interface provides an alternative intuitive directional guidance system.
The resulting system guides users to speci ed buildings and to important objects
such as cellular telephones, wallets, or even restroom or exit signs
Recommended from our members
Review of substitutive assistive tools and technologies for people with visual impairments: recent advancements and prospects
YesThe development of many tools and technologies for people with visual impairment has become a major priority in the
field of assistive technology research. However, many of these technology advancements have limitations in terms of the
human aspects of the user experience (e.g., usability, learnability, and time to user adaptation) as well as difficulties in
translating research prototypes into production. Also, there was no clear distinction between the assistive aids of adults
and children, as well as between âpartial impairmentâ and âtotal blindnessâ. As a result of these limitations, the produced
aids have not gained much popularity and the intended users are still hesitant to utilise them. This paper presents a comprehensive review of substitutive interventions that aid in adapting to vision loss, centred on laboratory research studies
to assess user-system interaction and system validation. Depending on the primary cueing feedback signal offered to the
user, these technology aids are categorized as visual, haptics, or auditory-based aids. The context of use, cueing feedback
signals, and participation of visually impaired people in the evaluation are all considered while discussing these aids.
Based on the findings, a set of recommendations is suggested to assist the scientific community in addressing persisting
challenges and restrictions faced by both the totally blind and partially sighted people
Accessible Autonomy: Exploring Inclusive Autonomous Vehicle Design and Interaction for People who are Blind and Visually Impaired
Autonomous vehicles are poised to revolutionize independent travel for millions of people experiencing transportation-limiting visual impairments worldwide. However, the current trajectory of automotive technology is rife with roadblocks to accessible interaction and inclusion for this demographic. Inaccessible (visually dependent) interfaces and lack of information access throughout the trip are surmountable, yet nevertheless critical barriers to this potentially lifechanging technology. To address these challenges, the programmatic dissertation research presented here includes ten studies, three published papers, and three submitted papers in high impact outlets that together address accessibility across the complete trip of transportation. The first paper began with a thorough review of the fully autonomous vehicle (FAV) and blind and visually impaired (BVI) literature, as well as the underlying policy landscape. Results guided prejourney ridesharing needs among BVI users, which were addressed in paper two via a survey with (n=90) transit service drivers, interviews with (n=12) BVI users, and prototype design evaluations with (n=6) users, all contributing to the Autonomous Vehicle Assistant: an award-winning and accessible ridesharing app. A subsequent study with (n=12) users, presented in paper three, focused on prejourney mapping to provide critical information access in future FAVs. Accessible in-vehicle interactions were explored in the fourth paper through a survey with (n=187) BVI users. Results prioritized nonvisual information about the trip and indicated the importance of situational awareness. This effort informed the design and evaluation of an ultrasonic haptic HMI intended to promote situational awareness with (n=14) participants (paper five), leading to a novel gestural-audio interface with (n=23) users (paper six). Strong support from users across these studies suggested positive outcomes in pursuit of actionable situational awareness and control. Cumulative results from this dissertation research program represent, to our knowledge, the single most comprehensive approach to FAV BVI accessibility to date. By considering both pre-journey and in-vehicle accessibility, results pave the way for autonomous driving experiences that enable meaningful interaction for BVI users across the complete trip of transportation. This new mode of accessible travel is predicted to transform independent travel for millions of people with visual impairment, leading to increased independence, mobility, and quality of life
- âŠ