6 research outputs found
Autonomously Navigating a Surgical Tool Inside the Eye by Learning from Demonstration
A fundamental challenge in retinal surgery is safely navigating a surgical
tool to a desired goal position on the retinal surface while avoiding damage to
surrounding tissues, a procedure that typically requires tens-of-microns
accuracy. In practice, the surgeon relies on depth-estimation skills to
localize the tool-tip with respect to the retina in order to perform the
tool-navigation task, which can be prone to human error. To alleviate such
uncertainty, prior work has introduced ways to assist the surgeon by estimating
the tool-tip distance to the retina and providing haptic or auditory feedback.
However, automating the tool-navigation task itself remains unsolved and
largely unexplored. Such a capability, if reliably automated, could serve as a
building block to streamline complex procedures and reduce the chance for
tissue damage. Towards this end, we propose to automate the tool-navigation
task by learning to mimic expert demonstrations of the task. Specifically, a
deep network is trained to imitate expert trajectories toward various locations
on the retina based on recorded visual servoing to a given goal specified by
the user. The proposed autonomous navigation system is evaluated in simulation
and in physical experiments using a silicone eye phantom. We show that the
network can reliably navigate a needle surgical tool to various desired
locations within 137 microns accuracy in physical experiments and 94 microns in
simulation on average, and generalizes well to unseen situations such as in the
presence of auxiliary surgical tools, variable eye backgrounds, and brightness
conditions.Comment: Accepted to ICRA 202
Phantoms in medicine: the case of ophthalmology
Physical and in-silico phantoms have revealed extremely useful in the development of new surgical techniques and medical devices and for training purposes. The fabrication of eye phantoms requires knowledge of anatomy and physical principles beyond the eye physiology and medical instruments used in the clinical scenario. After a proper definition of phantoms and the discussion about their classification, the present work reviews the various phantoms developed in ophthalmology, illustrating the rationale of their design
Deep Learning Guided Autonomous Surgery: Guiding Small Needles into Sub-Millimeter Scale Blood Vessels
We propose a general strategy for autonomous guidance and insertion of a
needle into a retinal blood vessel. The main challenges underpinning this task
are the accurate placement of the needle-tip on the target vein and a careful
needle insertion maneuver to avoid double-puncturing the vein, while dealing
with challenging kinematic constraints and depth-estimation uncertainty.
Following how surgeons perform this task purely based on visual feedback, we
develop a system which relies solely on \emph{monocular} visual cues by
combining data-driven kinematic and contact estimation, visual-servoing, and
model-based optimal control. By relying on both known kinematic models, as well
as deep-learning based perception modules, the system can localize the surgical
needle tip and detect needle-tissue interactions and venipuncture events. The
outputs from these perception modules are then combined with a motion planning
framework that uses visual-servoing and optimal control to cannulate the target
vein, while respecting kinematic constraints that consider the safety of the
procedure. We demonstrate that we can reliably and consistently perform needle
insertion in the domain of retinal surgery, specifically in performing retinal
vein cannulation. Using cadaveric pig eyes, we demonstrate that our system can
navigate to target veins within 22 XY accuracy and perform the entire
procedure in less than 35 seconds on average, and all 24 trials performed on 4
pig eyes were successful. Preliminary comparison study against a human operator
show that our system is consistently more accurate and safer, especially during
safety-critical needle-tissue interactions. To the best of the authors'
knowledge, this work accomplishes a first demonstration of autonomous retinal
vein cannulation at a clinically-relevant setting using animal tissues
Autonomously Navigating a Surgical Tool Inside the Eye by Learning from Demonstration
A fundamental challenge in retinal surgery is safely navigating a surgical tool to a desired goal position on the retinal surface while avoiding damage to surrounding tissues, a procedure that typically requires tens-of-microns accuracy. In practice, the surgeon relies on depth-estimation skills to localize the tool-tip with respect to the retina in order to perform the tool-navigation task, which can be prone to human error. To alleviate such uncertainty, prior work has introduced ways to assist the surgeon by estimating the tool-tip distance to the retina and providing haptic or auditory feedback. However, automating the tool-navigation task itself remains unsolved and largely unexplored. Such a capability, if reliably automated, could serve as a building block to streamline complex procedures and reduce the chance for tissue damage. Towards this end, we propose to automate the tool-navigation task by learning to mimic expert demonstrations of the task. Specifically, a deep network is trained to imitate expert trajectories toward various locations on the retina based on recorded visual servoing to a given goal specified by the user. The proposed autonomous navigation system is evaluated in simulation and in physical experiments using a silicone eye phantom. We show that the network can reliably navigate a needle surgical tool to various desired locations within 137 micrometer accuracy in physical experiments and 94 micrometer in simulation on average, and generalizes well to unseen situations such as in the presence of auxiliary surgical tools, variable eye backgrounds, and brightness conditions
Visual Tracking of Instruments in Minimally Invasive Surgery
Reducing access trauma has been a focal point for modern surgery and tackling the challenges that arise from new operating techniques and instruments is an exciting and open area of research. Lack of awareness and control from indirect manipulation and visualization has created a need to augment the surgeon's understanding and perception of how their instruments interact with the patient's anatomy but current methods of achieving this are inaccurate and difficult to integrate into the surgical workflow. Visual methods have the potential to recover the position and orientation of the instruments directly in the reference frame of the observing camera without the need to introduce additional hardware to the operating room and perform complex calibration steps. This thesis explores how this problem can be solved with the fusion of coarse region and fine scale point features to enable the recovery of both the rigid and articulated degrees of freedom of laparoscopic and robotic instruments using only images provided by the surgical camera. Extensive experiments on different image features are used to determine suitable representations for reliable and robust pose estimation. Using this information a novel framework is presented which estimates 3D pose with a region matching scheme while using frame-to-frame optical flow to account for challenges due to symmetry in the instrument design. The kinematic structure of articulated robotic instruments is also used to track the movement of the head and claspers. The robustness of this method was evaluated on calibrated ex-vivo images and in-vivo sequences and comparative studies are performed with state-of-the-art kinematic assisted tracking methods
Vision-based Proximity Detection in Retinal Surgery
In retinal surgery, surgeons face difficulties such as indirect visualization of surgical targets, physiological tremor, and lack of tactile feedback, which increase the risk of retinal damage caused by incorrect surgical gestures. In this context, intraocular proximity sensing has the potential to overcome current technical limitations and increase surgical safety. In this paper, we present a system for detecting unintentional collisions between surgical tools and the retina using the visual feedback provided by the opthalmic stereo microscope. Using stereo images, proximity between surgical tools and the retinal surface can be detected when their relative stereo disparity is small. For this purpose, we developed a system comprised of two modules. The first is a module for tracking the surgical tool position on both stereo images. The second is a disparity tracking module for estimating a stereo disparity map of the retinal surface. Both modules were specially tailored for coping with the challenging visualization conditions in retinal surgery. The potential clinical value of the proposed method is demonstrated by extensive testing using a silicon phantom eye and recorded rabbit in vivo data