14,307 research outputs found
Weight Controlled Electric Skateboard
Technology and the way that humans interact is becoming more vital and omnipresent with every passing day. However, human interface device designers suffer from the increasingly popular “designed for me or people like me” syndrome. This design philosophy inherently limits accessibility and usability of technology to those like the designer. This places severe limits of usability to those who are not fully able as well as leaves non-traditional human interface devices unexplored. This project set out to explore a previously uncharted human interface device, on an electric skateboard, and compare it send user experience with industry leading human interface devices
A Framework for Mouse Emulation that Uses a Minimally Invasive Tongue Palate Control Device utilizing Resistopalatography
The ability to interface fluently with a robust Human Input Device is a major challenge facing patients with severe levels of disability. This paper describes a new method of computer interaction utilizing Force Sensitive Resistor Array Technology, embedded into an Intra-Oral device (Resistopalatography), to emulate a USB Human Interface Device using standard Drivers. The system is based around the patient using their tongue to manipulate these sensors in order to give a position and force measurement; these can then be analyzed to generate the necessary metrics to control a mouse for computer input
SmartState: A Protocol-driven Human Interface
Since the inception of human research studies, researchers must often
interact with participants on a set schedule to collect data. Researchers
manually perform many interactions, leading to considerable time and financial
expenses. Usually, user-provided data collection consists of surveys
administered via telephone or email. These methods are tedious for the survey
administrators, which could cause fatigue and potentially lead to collection
mistakes. This project leverages recent advancements in automatic speech
recognition, speech-to-text, natural language understanding (NLU), and
finite-state machines to automate research protocols. This generalized
application is fully customizable and irrespective of any research study. New
research protocols can be quickly created based on these parameters once
envisioned. Thus, we present SmartState, a fully-customizable, state-driven
protocol manager combined with supporting AI components to autonomously manage
user data and intelligently determine users' intentions through chat and
end-device interactions.Comment: 8 pages, 8 figure
- …