382 research outputs found

    Comparative evaluation of instrument segmentation and tracking methods in minimally invasive surgery

    Get PDF
    Intraoperative segmentation and tracking of minimally invasive instruments is a prerequisite for computer- and robotic-assisted surgery. Since additional hardware like tracking systems or the robot encoders are cumbersome and lack accuracy, surgical vision is evolving as promising techniques to segment and track the instruments using only the endoscopic images. However, what is missing so far are common image data sets for consistent evaluation and benchmarking of algorithms against each other. The paper presents a comparative validation study of different vision-based methods for instrument segmentation and tracking in the context of robotic as well as conventional laparoscopic surgery. The contribution of the paper is twofold: we introduce a comprehensive validation data set that was provided to the study participants and present the results of the comparative validation study. Based on the results of the validation study, we arrive at the conclusion that modern deep learning approaches outperform other methods in instrument segmentation tasks, but the results are still not perfect. Furthermore, we show that merging results from different methods actually significantly increases accuracy in comparison to the best stand-alone method. On the other hand, the results of the instrument tracking task show that this is still an open challenge, especially during challenging scenarios in conventional laparoscopic surgery

    A comprehensive survey on recent deep learning-based methods applied to surgical data

    Full text link
    Minimally invasive surgery is highly operator dependant with a lengthy procedural time causing fatigue to surgeon and risks to patients such as injury to organs, infection, bleeding, and complications of anesthesia. To mitigate such risks, real-time systems are desired to be developed that can provide intra-operative guidance to surgeons. For example, an automated system for tool localization, tool (or tissue) tracking, and depth estimation can enable a clear understanding of surgical scenes preventing miscalculations during surgical procedures. In this work, we present a systematic review of recent machine learning-based approaches including surgical tool localization, segmentation, tracking, and 3D scene perception. Furthermore, we provide a detailed overview of publicly available benchmark datasets widely used for surgical navigation tasks. While recent deep learning architectures have shown promising results, there are still several open research problems such as a lack of annotated datasets, the presence of artifacts in surgical scenes, and non-textured surfaces that hinder 3D reconstruction of the anatomical structures. Based on our comprehensive review, we present a discussion on current gaps and needed steps to improve the adaptation of technology in surgery.Comment: This paper is to be submitted to International journal of computer visio

    Artificial intelligence and automation in endoscopy and surgery

    Get PDF
    Modern endoscopy relies on digital technology, from high-resolution imaging sensors and displays to electronics connecting configurable illumination and actuation systems for robotic articulation. In addition to enabling more effective diagnostic and therapeutic interventions, the digitization of the procedural toolset enables video data capture of the internal human anatomy at unprecedented levels. Interventional video data encapsulate functional and structural information about a patient’s anatomy as well as events, activity and action logs about the surgical process. This detailed but difficult-to-interpret record from endoscopic procedures can be linked to preoperative and postoperative records or patient imaging information. Rapid advances in artificial intelligence, especially in supervised deep learning, can utilize data from endoscopic procedures to develop systems for assisting procedures leading to computer-assisted interventions that can enable better navigation during procedures, automation of image interpretation and robotically assisted tool manipulation. In this Perspective, we summarize state-of-the-art artificial intelligence for computer-assisted interventions in gastroenterology and surgery

    Artificial intelligence surgery: how do we get to autonomous actions in surgery?

    Get PDF
    Most surgeons are skeptical as to the feasibility of autonomous actions in surgery. Interestingly, many examples of autonomous actions already exist and have been around for years. Since the beginning of this millennium, the field of artificial intelligence (AI) has grown exponentially with the development of machine learning (ML), deep learning (DL), computer vision (CV) and natural language processing (NLP). All of these facets of AI will be fundamental to the development of more autonomous actions in surgery, unfortunately, only a limited number of surgeons have or seek expertise in this rapidly evolving field. As opposed to AI in medicine, AI surgery (AIS) involves autonomous movements. Fortuitously, as the field of robotics in surgery has improved, more surgeons are becoming interested in technology and the potential of autonomous actions in procedures such as interventional radiology, endoscopy and surgery. The lack of haptics, or the sensation of touch, has hindered the wider adoption of robotics by many surgeons; however, now that the true potential of robotics can be comprehended, the embracing of AI by the surgical community is more important than ever before. Although current complete surgical systems are mainly only examples of tele-manipulation, for surgeons to get to more autonomously functioning robots, haptics is perhaps not the most important aspect. If the goal is for robots to ultimately become more and more independent, perhaps research should not focus on the concept of haptics as it is perceived by humans, and the focus should be on haptics as it is perceived by robots/computers. This article will discuss aspects of ML, DL, CV and NLP as they pertain to the modern practice of surgery, with a focus on current AI issues and advances that will enable us to get to more autonomous actions in surgery. Ultimately, there may be a paradigm shift that needs to occur in the surgical community as more surgeons with expertise in AI may be needed to fully unlock the potential of AIS in a safe, efficacious and timely manner

    ToolNet: Holistically-Nested Real-Time Segmentation of Robotic Surgical Tools

    Get PDF
    Real-time tool segmentation from endoscopic videos is an essential part of many computer-assisted robotic surgical systems and of critical importance in robotic surgical data science. We propose two novel deep learning architectures for automatic segmentation of non-rigid surgical instruments. Both methods take advantage of automated deep-learning-based multi-scale feature extraction while trying to maintain an accurate segmentation quality at all resolutions. The two proposed methods encode the multi-scale constraint inside the network architecture. The first proposed architecture enforces it by cascaded aggregation of predictions and the second proposed network does it by means of a holistically-nested architecture where the loss at each scale is taken into account for the optimization process. As the proposed methods are for real-time semantic labeling, both present a reduced number of parameters. We propose the use of parametric rectified linear units for semantic labeling in these small architectures to increase the regularization ability of the design and maintain the segmentation accuracy without overfitting the training sets. We compare the proposed architectures against state-of-the-art fully convolutional networks. We validate our methods using existing benchmark datasets, including ex vivo cases with phantom tissue and different robotic surgical instruments present in the scene. Our results show a statistically significant improved Dice Similarity Coefficient over previous instrument segmentation methods. We analyze our design choices and discuss the key drivers for improving accuracy.Comment: Paper accepted at IROS 201

    μž„μƒμˆ κΈ° ν–₯상을 μœ„ν•œ λ”₯λŸ¬λ‹ 기법 연ꡬ: λŒ€μž₯λ‚΄μ‹œκ²½ 진단 및 λ‘œλ΄‡μˆ˜μˆ  술기 평가에 적용

    Get PDF
    ν•™μœ„λ…Όλ¬Έ (박사) -- μ„œμšΈλŒ€ν•™κ΅ λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ ν˜‘λ™κ³Όμ • μ˜μš©μƒμ²΄κ³΅ν•™μ „κ³΅, 2020. 8. 김희찬.This paper presents deep learning-based methods for improving performance of clinicians. Novel methods were applied to the following two clinical cases and the results were evaluated. In the first study, a deep learning-based polyp classification algorithm for improving clinical performance of endoscopist during colonoscopy diagnosis was developed. Colonoscopy is the main method for diagnosing adenomatous polyp, which can multiply into a colorectal cancer and hyperplastic polyps. The classification algorithm was developed using convolutional neural network (CNN), trained with colorectal polyp images taken by a narrow-band imaging colonoscopy. The proposed method is built around an automatic machine learning (AutoML) which searches for the optimal architecture of CNN for colorectal polyp image classification and trains the weights of the architecture. In addition, gradient-weighted class activation mapping technique was used to overlay the probabilistic basis of the prediction result on the polyp location to aid the endoscopists visually. To verify the improvement in diagnostic performance, the efficacy of endoscopists with varying proficiency levels were compared with or without the aid of the proposed polyp classification algorithm. The results confirmed that, on average, diagnostic accuracy was improved and diagnosis time was shortened in all proficiency groups significantly. In the second study, a surgical instruments tracking algorithm for robotic surgery video was developed, and a model for quantitatively evaluating the surgeons surgical skill based on the acquired motion information of the surgical instruments was proposed. The movement of surgical instruments is the main component of evaluation for surgical skill. Therefore, the focus of this study was develop an automatic surgical instruments tracking algorithm, and to overcome the limitations presented by previous methods. The instance segmentation framework was developed to solve the instrument occlusion issue, and a tracking framework composed of a tracker and a re-identification algorithm was developed to maintain the type of surgical instruments being tracked in the video. In addition, algorithms for detecting the tip position of instruments and arm-indicator were developed to acquire the movement of devices specialized for the robotic surgery video. The performance of the proposed method was evaluated by measuring the difference between the predicted tip position and the ground truth position of the instruments using root mean square error, area under the curve, and Pearsons correlation analysis. Furthermore, motion metrics were calculated from the movement of surgical instruments, and a machine learning-based robotic surgical skill evaluation model was developed based on these metrics. These models were used to evaluate clinicians, and results were similar in the developed evaluation models, the Objective Structured Assessment of Technical Skill (OSATS), and the Global Evaluative Assessment of Robotic Surgery (GEARS) evaluation methods. In this study, deep learning technology was applied to colorectal polyp images for a polyp classification, and to robotic surgery videos for surgical instruments tracking. The improvement in clinical performance with the aid of these methods were evaluated and verified.λ³Έ 논문은 μ˜λ£Œμ§„μ˜ μž„μƒμˆ κΈ° λŠ₯λ ₯을 ν–₯μƒμ‹œν‚€κΈ° μœ„ν•˜μ—¬ μƒˆλ‘œμš΄ λ”₯λŸ¬λ‹ 기법듀을 μ œμ•ˆν•˜κ³  λ‹€μŒ 두 가지 싀둀에 λŒ€ν•΄ μ μš©ν•˜μ—¬ κ·Έ κ²°κ³Όλ₯Ό ν‰κ°€ν•˜μ˜€λ‹€. 첫 번째 μ—°κ΅¬μ—μ„œλŠ” λŒ€μž₯λ‚΄μ‹œκ²½μœΌλ‘œ κ΄‘ν•™ 진단 μ‹œ, λ‚΄μ‹œκ²½ μ „λ¬Έμ˜μ˜ 진단 λŠ₯λ ₯을 ν–₯μƒμ‹œν‚€κΈ° μœ„ν•˜μ—¬ λ”₯λŸ¬λ‹ 기반의 μš©μ’… λΆ„λ₯˜ μ•Œκ³ λ¦¬μ¦˜μ„ κ°œλ°œν•˜κ³ , λ‚΄μ‹œκ²½ μ „λ¬Έμ˜μ˜ 진단 λŠ₯λ ₯ ν–₯상 μ—¬λΆ€λ₯Ό κ²€μ¦ν•˜κ³ μž ν•˜μ˜€λ‹€. λŒ€μž₯λ‚΄μ‹œκ²½ κ²€μ‚¬λ‘œ μ•”μ’…μœΌλ‘œ 증식할 수 μžˆλŠ” μ„ μ’…κ³Ό 과증식성 μš©μ’…μ„ μ§„λ‹¨ν•˜λŠ” 것은 μ€‘μš”ν•˜λ‹€. λ³Έ μ—°κ΅¬μ—μ„œλŠ” ν˜‘λŒ€μ—­ μ˜μƒ λ‚΄μ‹œκ²½μœΌλ‘œ μ΄¬μ˜ν•œ λŒ€μž₯ μš©μ’… μ˜μƒμœΌλ‘œ ν•©μ„±κ³± 신경망을 ν•™μŠ΅ν•˜μ—¬ λΆ„λ₯˜ μ•Œκ³ λ¦¬μ¦˜μ„ κ°œλ°œν•˜μ˜€λ‹€. μ œμ•ˆν•˜λŠ” μ•Œκ³ λ¦¬μ¦˜μ€ μžλ™ κΈ°κ³„ν•™μŠ΅ (AutoML) λ°©λ²•μœΌλ‘œ, λŒ€μž₯ μš©μ’… μ˜μƒμ— μ΅œμ ν™”λœ ν•©μ„±κ³± 신경망 ꡬ쑰λ₯Ό μ°Ύκ³  μ‹ κ²½λ§μ˜ κ°€μ€‘μΉ˜λ₯Ό ν•™μŠ΅ν•˜μ˜€λ‹€. λ˜ν•œ 기울기-κ°€μ€‘μΉ˜ 클래슀 ν™œμ„±ν™” 맡핑 기법을 μ΄μš©ν•˜μ—¬ κ°œλ°œν•œ ν•©μ„±κ³± 신경망 결과의 ν™•λ₯ μ  κ·Όκ±°λ₯Ό μš©μ’… μœ„μΉ˜μ— μ‹œκ°μ μœΌλ‘œ λ‚˜νƒ€λ‚˜λ„λ‘ ν•¨μœΌλ‘œ λ‚΄μ‹œκ²½ μ „λ¬Έμ˜μ˜ 진단을 돕도둝 ν•˜μ˜€λ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ, μˆ™λ ¨λ„ κ·Έλ£Ήλ³„λ‘œ λ‚΄μ‹œκ²½ μ „λ¬Έμ˜κ°€ μš©μ’… λΆ„λ₯˜ μ•Œκ³ λ¦¬μ¦˜μ˜ κ²°κ³Όλ₯Ό μ°Έκ³ ν•˜μ˜€μ„ λ•Œ 진단 λŠ₯λ ₯이 ν–₯μƒλ˜μ—ˆλŠ”μ§€ 비ꡐ μ‹€ν—˜μ„ μ§„ν–‰ν•˜μ˜€κ³ , λͺ¨λ“  κ·Έλ£Ήμ—μ„œ μœ μ˜λ―Έν•˜κ²Œ 진단 정확도가 ν–₯μƒλ˜κ³  진단 μ‹œκ°„μ΄ λ‹¨μΆ•λ˜μ—ˆμŒμ„ ν™•μΈν•˜μ˜€λ‹€. 두 번째 μ—°κ΅¬μ—μ„œλŠ” λ‘œλ΄‡μˆ˜μˆ  λ™μ˜μƒμ—μ„œ μˆ˜μˆ λ„κ΅¬ μœ„μΉ˜ 좔적 μ•Œκ³ λ¦¬μ¦˜μ„ κ°œλ°œν•˜κ³ , νšλ“ν•œ μˆ˜μˆ λ„κ΅¬μ˜ μ›€μ§μž„ 정보λ₯Ό λ°”νƒ•μœΌλ‘œ 수술자의 μˆ™λ ¨λ„λ₯Ό μ •λŸ‰μ μœΌλ‘œ ν‰κ°€ν•˜λŠ” λͺ¨λΈμ„ μ œμ•ˆν•˜μ˜€λ‹€. μˆ˜μˆ λ„κ΅¬μ˜ μ›€μ§μž„μ€ 수술자의 λ‘œλ΄‡μˆ˜μˆ  μˆ™λ ¨λ„λ₯Ό ν‰κ°€ν•˜κΈ° μœ„ν•œ μ£Όμš”ν•œ 정보이닀. λ”°λΌμ„œ λ³Έ μ—°κ΅¬λŠ” λ”₯λŸ¬λ‹ 기반의 μžλ™ μˆ˜μˆ λ„κ΅¬ 좔적 μ•Œκ³ λ¦¬μ¦˜μ„ κ°œλ°œν•˜μ˜€μœΌλ©°, λ‹€μŒ 두가지 μ„ ν–‰μ—°κ΅¬μ˜ ν•œκ³„μ μ„ κ·Ήλ³΅ν•˜μ˜€λ‹€. μΈμŠ€ν„΄μŠ€ λΆ„ν•  (Instance Segmentation) ν”„λ ˆμž„μ›μ„ κ°œλ°œν•˜μ—¬ 폐색 (Occlusion) 문제λ₯Ό ν•΄κ²°ν•˜μ˜€κ³ , 좔적기 (Tracker)와 μž¬μ‹λ³„ν™” (Re-Identification) μ•Œκ³ λ¦¬μ¦˜μœΌλ‘œ κ΅¬μ„±λœ 좔적 ν”„λ ˆμž„μ›μ„ κ°œλ°œν•˜μ—¬ λ™μ˜μƒμ—μ„œ μΆ”μ ν•˜λŠ” μˆ˜μˆ λ„κ΅¬μ˜ μ’…λ₯˜κ°€ μœ μ§€λ˜λ„λ‘ ν•˜μ˜€λ‹€. λ˜ν•œ λ‘œλ΄‡μˆ˜μˆ  λ™μ˜μƒμ˜ νŠΉμˆ˜μ„±μ„ κ³ λ €ν•˜μ—¬ μˆ˜μˆ λ„κ΅¬μ˜ μ›€μ§μž„μ„ νšλ“ν•˜κΈ°μœ„ν•΄ μˆ˜μˆ λ„κ΅¬ 끝 μœ„μΉ˜μ™€ λ‘œλ΄‡ νŒ”-인디케이터 (Arm-Indicator) 인식 μ•Œκ³ λ¦¬μ¦˜μ„ κ°œλ°œν•˜μ˜€λ‹€. μ œμ•ˆν•˜λŠ” μ•Œκ³ λ¦¬μ¦˜μ˜ μ„±λŠ₯은 μ˜ˆμΈ‘ν•œ μˆ˜μˆ λ„κ΅¬ 끝 μœ„μΉ˜μ™€ μ •λ‹΅ μœ„μΉ˜ κ°„μ˜ 평균 제곱근 였차, 곑선 μ•„λž˜ 면적, ν”Όμ–΄μŠ¨ μƒκ΄€λΆ„μ„μœΌλ‘œ ν‰κ°€ν•˜μ˜€λ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ, μˆ˜μˆ λ„κ΅¬μ˜ μ›€μ§μž„μœΌλ‘œλΆ€ν„° μ›€μ§μž„ μ§€ν‘œλ₯Ό κ³„μ‚°ν•˜κ³  이λ₯Ό λ°”νƒ•μœΌλ‘œ κΈ°κ³„ν•™μŠ΅ 기반의 λ‘œλ΄‡μˆ˜μˆ  μˆ™λ ¨λ„ 평가 λͺ¨λΈμ„ κ°œλ°œν•˜μ˜€λ‹€. κ°œλ°œν•œ 평가 λͺ¨λΈμ€ 기쑴의 Objective Structured Assessment of Technical Skill (OSATS), Global Evaluative Assessment of Robotic Surgery (GEARS) 평가 방법과 μœ μ‚¬ν•œ μ„±λŠ₯을 λ³΄μž„μ„ ν™•μΈν•˜μ˜€λ‹€. λ³Έ 논문은 μ˜λ£Œμ§„μ˜ μž„μƒμˆ κΈ° λŠ₯λ ₯을 ν–₯μƒμ‹œν‚€κΈ° μœ„ν•˜μ—¬ λŒ€μž₯ μš©μ’… μ˜μƒκ³Ό λ‘œλ΄‡μˆ˜μˆ  λ™μ˜μƒμ— λ”₯λŸ¬λ‹ κΈ°μˆ μ„ μ μš©ν•˜κ³  κ·Έ μœ νš¨μ„±μ„ ν™•μΈν•˜μ˜€μœΌλ©°, ν–₯후에 μ œμ•ˆν•˜λŠ” 방법이 μž„μƒμ—μ„œ μ‚¬μš©λ˜κ³  μžˆλŠ” 진단 및 평가 λ°©λ²•μ˜ λŒ€μ•ˆμ΄ 될 κ²ƒμœΌλ‘œ κΈ°λŒ€ν•œλ‹€.Chapter 1 General Introduction 1 1.1 Deep Learning for Medical Image Analysis 1 1.2 Deep Learning for Colonoscipic Diagnosis 2 1.3 Deep Learning for Robotic Surgical Skill Assessment 3 1.4 Thesis Objectives 5 Chapter 2 Optical Diagnosis of Colorectal Polyps using Deep Learning with Visual Explanations 7 2.1 Introduction 7 2.1.1 Background 7 2.1.2 Needs 8 2.1.3 Related Work 9 2.2 Methods 11 2.2.1 Study Design 11 2.2.2 Dataset 14 2.2.3 Preprocessing 17 2.2.4 Convolutional Neural Networks (CNN) 21 2.2.4.1 Standard CNN 21 2.2.4.2 Search for CNN Architecture 22 2.2.4.3 Searched CNN Training 23 2.2.4.4 Visual Explanation 24 2.2.5 Evaluation of CNN and Endoscopist Performances 25 2.3 Experiments and Results 27 2.3.1 CNN Performance 27 2.3.2 Results of Visual Explanation 31 2.3.3 Endoscopist with CNN Performance 33 2.4 Discussion 45 2.4.1 Research Significance 45 2.4.2 Limitations 47 2.5 Conclusion 49 Chapter 3 Surgical Skill Assessment during Robotic Surgery by Deep Learning-based Surgical Instrument Tracking 50 3.1 Introduction 50 3.1.1 Background 50 3.1.2 Needs 51 3.1.3 Related Work 52 3.2 Methods 56 3.2.1 Study Design 56 3.2.2 Dataset 59 3.2.3 Instance Segmentation Framework 63 3.2.4 Tracking Framework 66 3.2.4.1 Tracker 66 3.2.4.2 Re-identification 68 3.2.5 Surgical Instrument Tip Detection 69 3.2.6 Arm-Indicator Recognition 71 3.2.7 Surgical Skill Prediction Model 71 3.3 Experiments and Results 78 3.3.1 Performance of Instance Segmentation Framework 78 3.3.2 Performance of Tracking Framework 82 3.3.3 Evaluation of Surgical Instruments Trajectory 83 3.3.4 Evaluation of Surgical Skill Prediction Model 86 3.4 Discussion 90 3.4.1 Research Significance 90 3.4.2 Limitations 92 3.5 Conclusion 96 Chapter 4 Summary and Future Works 97 4.1 Thesis Summary 97 4.2 Limitations and Future Works 98 Bibliography 100 Abstract in Korean 116 Acknowledgement 119Docto
    • …
    corecore