1,779 research outputs found
Adaptive physical human-robot interaction (PHRI) with a robotic nursing assistant.
Recently, more and more robots are being investigated for future applications in health-care. For instance, in nursing assistance, seamless Human-Robot Interaction (HRI) is very important for sharing workspaces and workloads between medical staff, patients, and robots. In this thesis we introduce a novel robot - the Adaptive Robot Nursing Assistant (ARNA) and its underlying components. ARNA has been designed specifically to assist nurses with day-to-day tasks such as walking patients, pick-and-place item retrieval, and routine patient health monitoring. An adaptive HRI in nursing applications creates a positive user experience, increase nurse productivity and task completion rates, as reported by experimentation with human subjects. ARNA has been designed to include interface devices such as tablets, force sensors, pressure-sensitive robot skins, LIDAR and RGBD camera. These interfaces are combined with adaptive controllers and estimators within a proposed framework that contains multiple innovations. A research study was conducted on methods of deploying an ideal HumanMachine Interface (HMI), in this case a tablet-based interface. Initial study points to the fact that a traded control level of autonomy is ideal for tele-operating ARNA by a patient. The proposed method of using the HMI devices makes the performance of a robot similar for both skilled and un-skilled workers. A neuro-adaptive controller (NAC), which contains several neural-networks to estimate and compensate for system non-linearities, was implemented on the ARNA robot. By linearizing the system, a cross-over usability condition is met through which humans find it more intuitive to learn to use the robot in any location of its workspace, A novel Base-Sensor Assisted Physical Interaction (BAPI) controller is introduced in this thesis, which utilizes a force-torque sensor at the base of the ARNA robot manipulator to detect full body collisions, and make interaction safer. Finally, a human-intent estimator (HIE) is proposed to estimate human intent while the robot and user are physically collaborating during certain tasks such as adaptive walking. A NAC with HIE module was validated on a PR2 robot through user studies. Its implementation on the ARNA robot platform can be easily accomplished as the controller is model-free and can learn robot dynamics online. A new framework, Directive Observer and Lead Assistant (DOLA), is proposed for ARNA which enables the user to interact with the robot in two modes: physically, by direct push-guiding, and remotely, through a tablet interface. In both cases, the human is being âobservedâ by the robot, then guided and/or advised during interaction. If the user has trouble completing the given tasks, the robot adapts their repertoire to lead users toward completing goals. The proposed framework incorporates interface devices as well as adaptive control systems in order to facilitate a higher performance interaction between the user and the robot than was previously possible. The ARNA robot was deployed and tested in a hospital environment at the School of Nursing of the University of Louisville. The user-experience tests were conducted with the help of healthcare professionals where several metrics including completion time, rate and level of user satisfaction were collected to shed light on the performance of various components of the proposed framework. The results indicate an overall positive response towards the use of such assistive robot in the healthcare environment. The analysis of these gathered data is included in this document. To summarize, this research study makes the following contributions: Conducting user experience studies with the ARNA robot in patient sitter and walker scenarios to evaluate both physical and non-physical human-machine interfaces. Evaluation and Validation of Human Intent Estimator (HIE) and Neuro-Adaptive Controller (NAC). Proposing the novel Base-Sensor Assisted Physical Interaction (BAPI) controller. Building simulation models for packaged tactile sensors and validating the models with experimental data. Description of Directive Observer and Lead Assistance (DOLA) framework for ARNA using adaptive interfaces
Soft manipulators and grippers: A review
Soft robotics is a growing area of research which utilizes the compliance and adaptability of soft structures to develop highly adaptive robotics for soft interactions. One area in which soft robotics has the ability to make significant impact is in the development of soft grippers and manipulators. With an increased requirement for automation, robotics systems are required to perform task in unstructured and not well defined environments; conditions which conventional rigid robotics are not best suited. This requires a paradigm shift in the methods and materials used to develop robots such that they can adapt to and work safely in human environments. One solution to this is soft robotics, which enables soft interactions with the surroundings while maintaining the ability to apply significant force. This review paper assesses the current materials and methods, actuation methods and sensors which are used in the development of soft manipulators. The achievements and shortcomings of recent technology in these key areas are evaluated, and this paper concludes with a discussion on the potential impacts of soft manipulators on industry and society
Survey of Visual and Force/Tactile Control of Robots for Physical Interaction in Spain
Sensors provide robotic systems with the information required to perceive the changes that happen in unstructured environments and modify their actions accordingly. The robotic controllers which process and analyze this sensory information are usually based on three types of sensors (visual, force/torque and tactile) which identify the most widespread robotic control strategies: visual servoing control, force control and tactile control. This paper presents a detailed review on the sensor architectures, algorithmic techniques and applications which have been developed by Spanish researchers in order to implement these mono-sensor and multi-sensor controllers which combine several sensors
Overcoming barriers and increasing independence: service robots for elderly and disabled people
This paper discusses the potential for service robots to overcome barriers and increase independence of
elderly and disabled people. It includes a brief overview of the existing uses of service robots by disabled and elderly
people and advances in technology which will make new uses possible and provides suggestions for some of these new
applications. The paper also considers the design and other conditions to be met for user acceptance. It also discusses
the complementarity of assistive service robots and personal assistance and considers the types of applications and
users for which service robots are and are not suitable
Challenges in flexible microsystem manufacturing : fabrication, robotic assembly, control, and packaging.
Microsystems have been investigated with renewed interest for the last three decades because of the emerging development of microelectromechanical system (MEMS) technology and the advancement of nanotechnology. The applications of microrobots and distributed sensors have the potential to revolutionize micro and nano manufacturing and have other important health applications for drug delivery and minimal invasive surgery. A class of microrobots studied in this thesis, such as the Solid Articulated Four Axis Microrobot (sAFAM) are driven by MEMS actuators, transmissions, and end-effectors realized by 3-Dimensional MEMS assembly. Another class of microrobots studied here, like those competing in the annual IEEE Mobile Microrobot Challenge event (MMC) are untethered and driven by external fields, such as magnetic fields generated by a focused permanent magnet. A third class of microsystems studied in this thesis includes distributed MEMS pressure sensors for robotic skin applications that are manufactured in the cleanroom and packaged in our lab.
In this thesis, we discuss typical challenges associated with the fabrication, robotic assembly and packaging of these microsystems. For sAFAM we discuss challenges arising from pick and place manipulation under microscopic closed-loop control, as well as bonding and attachment of silicon MEMS microparts. For MMC, we discuss challenges arising from cooperative manipulation of microparts that advance the capabilities of magnetic micro-agents. Custom microrobotic hardware configured and demonstrated during this research (such as the NeXus microassembly station) include micro-positioners, microscopes, and controllers driven via LabVIEW. Finally, we also discuss challenges arising in distributed sensor manufacturing. We describe sensor fabrication steps using clean-room techniques on Kapton flexible substrates, and present results of lamination, interconnection and testing of such sensors are presented
Toward Robots with Peripersonal Space Representation for Adaptive Behaviors
The abilities to adapt and act autonomously in an unstructured and
human-oriented environment are necessarily vital for the next generation of
robots, which aim to safely cooperate with humans. While this adaptability
is natural and feasible for humans, it is still very complex and challenging
for robots. Observations and findings from psychology and neuroscience in
respect to the development of the human sensorimotor system can inform
the development of novel approaches to adaptive robotics.
Among these is the formation of the representation of space closely surrounding
the body, the Peripersonal Space (PPS) , from multisensory sources
like vision, hearing, touch and proprioception, which helps to facilitate human
activities within their surroundings.
Taking inspiration from the virtual safety margin formed by the PPS representation
in humans, this thesis first constructs an equivalent model of the
safety zone for each body part of the iCub humanoid robot. This PPS layer
serves as a distributed collision predictor, which translates visually detected
objects approaching a robot\u2019s body parts (e.g., arm, hand) into the probabilities
of a collision between those objects and body parts. This leads to
adaptive avoidance behaviors in the robot via an optimization-based reactive
controller. Notably, this visual reactive control pipeline can also seamlessly
incorporate tactile input to guarantee safety in both pre- and post-collision
phases in physical Human-Robot Interaction (pHRI). Concurrently, the controller
is also able to take into account multiple targets (of manipulation reaching tasks) generated by a multiple Cartesian point planner. All components,
namely the PPS, the multi-target motion planner (for manipulation
reaching tasks), the reaching-with-avoidance controller and the humancentred
visual perception, are combined harmoniously to form a hybrid control
framework designed to provide safety for robots\u2019 interactions in a cluttered
environment shared with human partners.
Later, motivated by the development of manipulation skills in infants, in
which the multisensory integration is thought to play an important role, a
learning framework is proposed to allow a robot to learn the processes of
forming sensory representations, namely visuomotor and visuotactile, from
their own motor activities in the environment. Both multisensory integration
models are constructed with Deep Neural Networks (DNNs) in such a
way that their outputs are represented in motor space to facilitate the robot\u2019s
subsequent actions
Sensors for Robotic Hands: A Survey of State of the Art
Recent decades have seen significant progress in the field of artificial hands. Most of the
surveys, which try to capture the latest developments in this field, focused on actuation and control systems of these devices. In this paper, our goal is to provide a comprehensive survey of the sensors for artificial hands. In order to present the evolution of the field, we cover five year periods starting at the turn of the millennium. At each period, we present the robot hands with a focus on their sensor systems dividing them into categories, such as prosthetics, research devices, and industrial end-effectors.We also cover the sensors developed for robot hand usage in each era. Finally, the period between 2010 and 2015 introduces the reader to the state of the art and also hints to the future directions in the sensor development for artificial hands
The implications of embodiment for behavior and cognition: animal and robotic case studies
In this paper, we will argue that if we want to understand the function of
the brain (or the control in the case of robots), we must understand how the
brain is embedded into the physical system, and how the organism interacts with
the real world. While embodiment has often been used in its trivial meaning,
i.e. 'intelligence requires a body', the concept has deeper and more important
implications, concerned with the relation between physical and information
(neural, control) processes. A number of case studies are presented to
illustrate the concept. These involve animals and robots and are concentrated
around locomotion, grasping, and visual perception. A theoretical scheme that
can be used to embed the diverse case studies will be presented. Finally, we
will establish a link between the low-level sensory-motor processes and
cognition. We will present an embodied view on categorization, and propose the
concepts of 'body schema' and 'forward models' as a natural extension of the
embodied approach toward first representations.Comment: Book chapter in W. Tschacher & C. Bergomi, ed., 'The Implications of
Embodiment: Cognition and Communication', Exeter: Imprint Academic, pp. 31-5
- âŠ