17,255 research outputs found
Fast and Reliable Autonomous Surgical Debridement with Cable-Driven Robots Using a Two-Phase Calibration Procedure
Automating precision subtasks such as debridement (removing dead or diseased
tissue fragments) with Robotic Surgical Assistants (RSAs) such as the da Vinci
Research Kit (dVRK) is challenging due to inherent non-linearities in
cable-driven systems. We propose and evaluate a novel two-phase coarse-to-fine
calibration method. In Phase I (coarse), we place a red calibration marker on
the end effector and let it randomly move through a set of open-loop
trajectories to obtain a large sample set of camera pixels and internal robot
end-effector configurations. This coarse data is then used to train a Deep
Neural Network (DNN) to learn the coarse transformation bias. In Phase II
(fine), the bias from Phase I is applied to move the end-effector toward a
small set of specific target points on a printed sheet. For each target, a
human operator manually adjusts the end-effector position by direct contact
(not through teleoperation) and the residual compensation bias is recorded.
This fine data is then used to train a Random Forest (RF) to learn the fine
transformation bias. Subsequent experiments suggest that without calibration,
position errors average 4.55mm. Phase I can reduce average error to 2.14mm and
the combination of Phase I and Phase II can reduces average error to 1.08mm. We
apply these results to debridement of raisins and pumpkin seeds as fragment
phantoms. Using an endoscopic stereo camera with standard edge detection,
experiments with 120 trials achieved average success rates of 94.5%, exceeding
prior results with much larger fragments (89.4%) and achieving a speedup of
2.1x, decreasing time per fragment from 15.8 seconds to 7.3 seconds. Source
code, data, and videos are available at
https://sites.google.com/view/calib-icra/.Comment: Code, data, and videos are available at
https://sites.google.com/view/calib-icra/. Final version for ICRA 201
Automated Handling of Auxiliary Materials using a Multi-Kinematic Gripping System
Using a special, multi-kinematic gripping system, the vacuum bagging process in the manufacturing of carbon-fibre reinforced plastics (CFRP) can be automated. Using the example of a parabolic rear pressure bulkhead, the flexibility of the multi-kinematic system is used to handle largely different sized cut-pieces of auxiliary materials. Avoiding the need for special gripping systems for each part greatly reduces the cost for automation because it allows using a single system for a broad variety of different tasks. With a genetic algorithm for optimization, the high redundancy created by using several robots with each 6 or 7 degrees of freedom can be solved. The overall process is simulated using a 3D visualization environment and therefore can be programmed completely offline before being executed with real robot hardwar
Handle Anywhere: A Mobile Robot Arm for Providing Bodily Support to Elderly Persons
Age-related loss of mobility and increased risk of falling remain important
obstacles toward facilitating aging-in-place. Many elderly people lack the
coordination and strength necessary to perform common movements around their
home, such as getting out of bed or stepping into a bathtub. The traditional
solution has been to install grab bars on various surfaces; however, these are
often not placed in optimal locations due to feasibility constraints in room
layout. In this paper, we present a mobile robot that provides an older adult
with a handle anywhere in space - "handle anywhere". The robot consists of an
omnidirectional mobile base attached to a repositionable handle. We analyze the
postural changes in four activities of daily living and determine, in each, the
body pose that requires the maximal muscle effort. Using a simple model of the
human body, we develop a methodology to optimally place the handle to provide
the maximum support for the elderly person at the point of most effort. Our
model is validated with experimental trials. We discuss how the robotic device
could be used to enhance patient mobility and reduce the incidence of falls.Comment: 8 pages, 10 figure
Precision and power grip detection in egocentric hand-object Interaction using machine learning
This project, was carried out in Yverdon-les-Bains, Switzerland, between the University of Applied Sciences and Arts Western Switzerland (HEIG-VD / HES-SO) and the Centre Hospitalier Universitaire Vaudois (CHUV) in Lausanne, it focuses on the detection of grasp types from an egocentric point of view. The objective is to accurately determine the kind of grasp (power, precision and none) performed by a user based on images captured from their perspective. The successful implementation of this grasp detection system would greatly benefit the evaluation of patients undergoing upper limb rehabilitation. Various computer vision frameworks were utilized to detect hands, interacting objects, and depth information in the images. These extracted features were then fed into deep learning models for grasp prediction. Both custom recorded datasets and open-source datasets, such as EpicKitchen and the Yale dataset, were employed for training and evaluation. In conclusion, this project achieved satisfactory results in the detection of grasp types from an egocentric viewpoint, with a 0.76 F1-macro score in the final test set. The utilization of diverse videos, including custom recordings and publicly available datasets, facilitated comprehensive training and evaluation. A robust pipeline was developed through iterative refinement, enabling the extraction of crucial features from each frame to predict grasp types accurately. Furthermore, data mixtures were proposed to enhance dataset size and improve the generalization performance of the models, which played a crucial role in the project's final stages
Learning Singularity Avoidance
With the increase in complexity of robotic systems and the rise in non-expert
users, it can be assumed that task constraints are not explicitly known. In
tasks where avoiding singularity is critical to its success, this paper
provides an approach, especially for non-expert users, for the system to learn
the constraints contained in a set of demonstrations, such that they can be
used to optimise an autonomous controller to avoid singularity, without having
to explicitly know the task constraints. The proposed approach avoids
singularity, and thereby unpredictable behaviour when carrying out a task, by
maximising the learnt manipulability throughout the motion of the constrained
system, and is not limited to kinematic systems. Its benefits are demonstrated
through comparisons with other control policies which show that the constrained
manipulability of a system learnt through demonstration can be used to avoid
singularities in cases where these other policies would fail. In the absence of
the systems manipulability subject to a tasks constraints, the proposed
approach can be used instead to infer these with results showing errors less
than 10^-5 in 3DOF simulated systems as well as 10^-2 using a 7DOF real world
robotic system
- …