15 research outputs found
A Multi-Robot Cooperation Framework for Sewing Personalized Stent Grafts
This paper presents a multi-robot system for manufacturing personalized
medical stent grafts. The proposed system adopts a modular design, which
includes: a (personalized) mandrel module, a bimanual sewing module, and a
vision module. The mandrel module incorporates the personalized geometry of
patients, while the bimanual sewing module adopts a learning-by-demonstration
approach to transfer human hand-sewing skills to the robots. The human
demonstrations were firstly observed by the vision module and then encoded
using a statistical model to generate the reference motion trajectories. During
autonomous robot sewing, the vision module plays the role of coordinating
multi-robot collaboration. Experiment results show that the robots can adapt to
generalized stent designs. The proposed system can also be used for other
manipulation tasks, especially for flexible production of customized products
and where bimanual or multi-robot cooperation is required.Comment: 10 pages, 12 figures, accepted by IEEE Transactions on Industrial
Informatics, Key words: modularity, medical device customization, multi-robot
system, robot learning, visual servoing, robot sewin
A Multi-Robot Cooperation Framework for Sewing Personalized Stent Grafts
This paper presents a multi-robot system for manufacturing personalized
medical stent grafts. The proposed system adopts a modular design, which
includes: a (personalized) mandrel module, a bimanual sewing module, and a
vision module. The mandrel module incorporates the personalized geometry of
patients, while the bimanual sewing module adopts a learning-by-demonstration
approach to transfer human hand-sewing skills to the robots. The human
demonstrations were firstly observed by the vision module and then encoded
using a statistical model to generate the reference motion trajectories. During
autonomous robot sewing, the vision module plays the role of coordinating
multi-robot collaboration. Experiment results show that the robots can adapt to
generalized stent designs. The proposed system can also be used for other
manipulation tasks, especially for flexible production of customized products
and where bimanual or multi-robot cooperation is required.Comment: 10 pages, 12 figures, accepted by IEEE Transactions on Industrial
Informatics, Key words: modularity, medical device customization, multi-robot
system, robot learning, visual servoing, robot sewin
A multi-robot cooperation framework for sewing personalized stent grafts
This paper presents a multi-robot system for manufacturing personalized medical stent grafts. The proposed system adopts a modular design, which includes: a (personalized) mandrel module, a bimanual sewing module, and a vision module. The mandrel module incorporates the personalized geometry of patients, while the bimanual sewing module adopts a learning-by-demonstration approach to transfer human hand-sewing skills to the robots. The human demonstrations were firstly observed by the vision module and then encoded using a statistical model to generate the reference motion trajectories. During autonomous robot sewing, the vision module plays the role of coordinating multi-robot collaboration. Experiment results show that the robots can adapt to generalized stent designs. The proposed system can also be used for other manipulation tasks, especially for flexible production of customized products and where bimanual or multi-robot cooperation is required
A Vision-guided Dual Arm Sewing System for Stent Graft Manufacturing
This paper presents an intelligent sewing system for personalized stent graft manufacturing, a challenging sewing task that is currently performed manually. Inspired by medical suturing robots, we have adopted a single-sided sewing technique using a curved needle to perform the task of sewing stents onto fabric. A motorized surgical needle driver was attached to a 7 d.o.f robot arm to manipulate the needle with a second robot controlling the position of the mandrel. A learningfrom-demonstration approach was used to program the robot to sew stents onto fabric. The demonstrated sewing skill was segmented to several phases, each of which was encoded with a Gaussian Mixture Model. Generalized sewing movements were then generated from these models and were used for task execution. During execution, a stereo vision system was adopted to guide the robots to adjust the learnt movements according to the needle pose. Two experiments are presented here with this system and the results show that our system can robustly perform the sewing task as well as adapt to various needle poses. The accuracy of the sewing system was within 2mm
Design, fabrication and testing a semi-automatic sewing device for personalized stent graft manufacturing
For the treatment of Abdominal Aortic Aneurysm (AAA), a personalised stent graft is used to ensure it fits tightly to the patients vessel geometry. A personalised stent graft is usually handmade which requires thousands of stitches and can take weeks or even months to complete. This delay may expose the patient to the risk of aneurysm rupture. This paper presents a robotic sewing device that can enhance the stent graft sewing speed by providing automated needle manipulation. It simplifies the sewing process and has the potential to achieve fully automated stent graft manufacturing via a vision-guided system. The device features a sewing probe that can switch a double pointed semi-circular needle between two movable jaws. This forgoes the need for manual needle handling including grasping, driving rotation, releasing and re-grasping, which requires a high level of manual dexterity and attention. This paper presents the design of the device, its mechanical synthesis and experimental validation. The focus of the paper is on the linkage parameter optimisation and needle locking mechanism design. The proposed device has been fabricated using 3D rapid prototyping techniques, and its performance has been compared with the conventional manual sewing method. The experimental results show that the device can achieve a 30% reduction of the completion time for a stitching task while achieving better consistency and quality of the stitches
λͺ¨μ ν리머ν°λΈλ₯Ό μ΄μ©ν 볡μ‘ν λ‘λ΄ μ무 νμ΅ λ° μΌλ°ν κΈ°λ²
νμλ
Όλ¬Έ (λ°μ¬) -- μμΈλνκ΅ λνμ : 곡과λν ν곡μ°μ£Όκ³΅νκ³Ό, 2020. 8. κΉνμ§.Learning from demonstrations (LfD) is a promising approach that enables robots to perform a specific movement. As robotic manipulations are substituting a variety of tasks, LfD algorithms are widely used and studied for specifying the robot configurations for the various types of movements.
This dissertation presents an approach based on parametric dynamic movement primitives (PDMP) as a motion representation algorithm which is one of relevant LfD techniques. Unlike existing motion representation algorithms, this work not only represents a prescribed motion but also computes the new behavior through a generalization of multiple demonstrations in the actual environment. The generalization process uses Gaussian process regression (GPR) by representing the nonlinear relationship between the PDMP parameters that determine motion and the corresponding environmental variables. The proposed algorithm shows that it serves as a powerful optimal and real-time motion planner among the existing planning algorithms when optimal demonstrations are provided as dataset.
In this dissertation, the safety of motion is also considered. Here, safety refers to keeping the system away from certain configurations that are unsafe. The safety criterion of the PDMP internal parameters are computed to check the safety. This safety criterion reflects the new behavior computed through the generalization process, as well as the individual motion safety of the demonstration set. The demonstrations causing unsafe movement are identified and removed. Also, the demolished demonstrations are replaced by proven demonstrations upon this criterion.
This work also presents an extension approach reducing the number of required demonstrations for the PDMP framework. This approach is effective where a single mission consists of multiple sub-tasks and requires numerous demonstrations in generalizing them. The whole trajectories in provided demonstrations are segmented into multiple sub-tasks representing unit motions. Then, multiple PDMPs are formed independently for correlated-segments. The phase-decision process determines which sub-task and associated PDMPs to be executed online, allowing multiple PDMPs to be autonomously configured within an integrated framework. GPR formulations are applied to obtain execution time and regional goal configuration for each sub-task.
Finally, the proposed approach and its extension are validated with the actual experiments of mobile manipulators. The first two scenarios regarding cooperative aerial transportation demonstrate the excellence of the proposed technique in terms of quick computation, generation of efficient movement, and safety assurance. The last scenario deals with two mobile manipulations using ground vehicles and shows the effectiveness of the proposed extension in executing complex missions.μμ° νμ΅ κΈ°λ²(Learning from demonstrations, LfD)μ λ‘λ΄μ΄ νΉμ λμμ μνν μ μλλ‘ νλ μ λ§ν λμ μμ± κΈ°λ²μ΄λ€. λ‘λ΄ μ‘°μκΈ°κ° μΈκ° μ¬νμμ λ€μν μ
무λ₯Ό λμ²΄ν΄ κ°μ λ°λΌ, λ€μν μ무λ₯Ό μννλ λ‘λ΄μ λμμ μμ±νκΈ° μν΄ LfD μκ³ λ¦¬μ¦λ€μ λ리 μ°κ΅¬λκ³ , μ¬μ©λκ³ μλ€.
λ³Έ λ
Όλ¬Έμ LfD κΈ°λ² μ€ λͺ¨μ
ν리머ν°λΈ κΈ°λ°μ λμ μ¬μμ± μκ³ λ¦¬μ¦μΈ Parametric dynamic movement primitives(PDMP)μ κΈ°μ΄ν μκ³ λ¦¬μ¦μ μ μνλ©°, μ΄λ₯Ό ν΅ν΄ λ€μν μ무λ₯Ό μννλ λͺ¨λ°μΌ μ‘°μκΈ°μ κΆ€μ μ μμ±νλ€. κΈ°μ‘΄μ λμ μ¬μμ± μκ³ λ¦¬μ¦κ³Ό λ¬λ¦¬, μ΄ μ°κ΅¬λ μ 곡λ μμ°μμ ννλ λμμ λ¨μν μ¬μμ±νλ κ²μ κ·ΈμΉμ§ μκ³ , μλ‘μ΄ νκ²½μ λ§κ² μΌλ°ν νλ κ³Όμ μ ν¬ν¨νλ€. μ΄ λ
Όλ¬Έμμ μ μνλ μΌλ°ν κ³Όμ μ PDMPsμ λ΄λΆ νλΌλ―Έν° κ°μΈ μ€νμΌ νλΌλ―Έν°μ νκ²½ λ³μ μ¬μ΄μ λΉμ ν κ΄κ³λ₯Ό κ°μ°μ€ νκ· κΈ°λ² (Gaussian process regression, GPR)μ μ΄μ©νμ¬ μμμ μΌλ‘ νννλ€. μ μλ κΈ°λ²μ λν μ΅μ μμ°λ₯Ό νμ΅νλ λ°©μμ ν΅ν΄ κ°λ ₯ν μ΅μ μ€μκ° κ²½λ‘ κ³ν κΈ°λ²μΌλ‘λ μμ©λ μ μλ€.
λ³Έ λ
Όλ¬Έμμλ λν λ‘λ΄μ ꡬλ μμ μ±λ κ³ λ €νλ€. κΈ°μ‘΄ μ°κ΅¬λ€μμ λ€λ£¨μ΄μ§ μμ° κ΄λ¦¬ κΈ°μ μ΄ λ‘λ΄μ ꡬλ ν¨μ¨μ±μ κ°μ νλ λ°©ν₯μΌλ‘ μ μλ κ²κ³Ό λ¬λ¦¬, μ΄ μ°κ΅¬λ κ°ν ꡬμ쑰건μΌλ‘ λ‘λ΄μ ꡬλ μμ μ±μ ν보νλ μμ° κ΄λ¦¬ κΈ°μ μ ν΅ν΄ μμ μ±μ κ³ λ €νλ μλ‘μ΄ λ°©μμ μ μνλ€. μ μλ λ°©μμ μ€νμΌ νλΌλ―Έν° κ° μμμ μμ μ± κΈ°μ€μ κ³μ°νλ©°, μ΄ μμ κΈ°μ€μ ν΅ν΄ μμ°μ μ κ±°νλ μΌλ ¨μ μμ
μ μννλ€. λν, μ κ±°λ μμλ₯Ό μμ κΈ°μ€μ λ°λΌ μ
μ¦λ μμλ‘ λ체νμ¬ μΌλ°ν μ±λ₯μ μ νμν€μ§ μλλ‘ μμλ₯Ό κ΄λ¦¬νλ€. μ΄λ₯Ό ν΅ν΄ λ€μμ μμ° κ°κ° κ°λ³ λμ μμ μ± λΏ μλλΌ μ¨λΌμΈ λμμ μμ μ±κΉμ§ κ³ λ €ν μ μμΌλ©°, μ€μκ° λ‘λ΄ μ‘°μκΈ° μ΄μ©μ μμ μ±μ΄ ν보λ μ μλ€. μ μλ μμ μ±μ κ³ λ €ν μμ° κ΄λ¦¬ κΈ°μ μ λν νκ²½μ μ μ μ€μ μ΄ λ³κ²½λμ΄ λͺ¨λ μμ°μ κ΅μ²΄ν΄μΌ ν μ μλ μν©μμ μ¬μ©ν μ μλ μμ°λ€μ νλ³νκ³ , ν¨μ¨μ μΌλ‘ μ¬μ¬μ©νλ λ° μμ©ν μ μλ€.
λν λ³Έ λ
Όλ¬Έμ 볡μ‘ν μ무μμ μ μ©λ μ μλ PDMPsμ νμ₯ κΈ°λ²μΈ seg-PDMPsλ₯Ό μ μνλ€. μ΄ μ κ·Όλ°©μμ 볡μ‘ν μλ¬΄κ° μΌλ°μ μΌλ‘ 볡μκ°μ κ°λ¨ν νμ μμ
μΌλ‘ ꡬμ±λλ€κ³ κ°μ νλ€. κΈ°μ‘΄ PDMPsμ λ¬λ¦¬ seg-PDMPsλ μ 체 κΆ€μ μ νμ μμ
μ λνλ΄λ μ¬λ¬ κ°μ λ¨μ λμμΌλ‘ λΆν νκ³ , κ° λ¨μλμμ λν΄ μ¬λ¬κ°μ PDMPsλ₯Ό ꡬμ±νλ€. κ° λ¨μ λμ λ³λ‘ μμ±λ PDMPsλ ν΅ν©λ νλ μμν¬λ΄μμ λ¨κ³ κ²°μ νλ‘μΈμ€λ₯Ό ν΅ν΄ μλμ μΌλ‘ νΈμΆλλ€. κ° λ¨κ³ λ³λ‘ λ¨μ λμμ μννκΈ° μν μκ° λ° νμ λͺ©νμ μ κ°μ°μ€ 곡μ νκ·(GPR)λ₯Ό μ΄μ©ν νκ²½λ³μμμμ κ΄κ³μμ ν΅ν΄ μ»λλ€. κ²°κ³Όμ μΌλ‘, μ΄ μ°κ΅¬λ μ 체μ μΌλ‘ μꡬλλ μμ°μ μλ₯Ό ν¨κ³Όμ μΌλ‘ μ€μΌ λΏ μλλΌ, κ° λ¨μλμμ νν μ±λ₯μ κ°μ νλ€.
μ μλ μκ³ λ¦¬μ¦μ νλ λͺ¨λ°μΌ λ‘λ΄ μ‘°μκΈ° μ€νμ ν΅νμ¬ κ²μ¦λλ€. μΈ κ°μ§μ μλ리μ€κ° λ³Έ λ
Όλ¬Έμμ λ€λ£¨μ΄μ§λ©°, ν곡 μ΄μ‘κ³Ό κ΄λ ¨λ 첫 λ κ°μ§ μλ리μ€λ PDMPs κΈ°λ²μ΄ λ‘λ΄ μ‘°μκΈ°μμ λΉ λ₯Έ μ μμ±, μ무 ν¨μ¨μ±κ³Ό μμ μ± λͺ¨λ λ§μ‘±νλ κ²μ μ
μ¦νλ€. λ§μ§λ§ μλ리μ€λ μ§μ μ°¨λμ μ΄μ©ν λ κ°μ λ‘λ΄ μ‘°μκΈ°μ λν μ€νμΌλ‘ 볡μ‘ν μ무 μνμ νκΈ° μν΄ νμ₯λ κΈ°λ²μΈ seg-PDMPsκ° ν¨κ³Όμ μΌλ‘ λ³ννλ νκ²½μμ μΌλ°νλ λμμ μμ±ν¨μ κ²μ¦νλ€.1 Introduction 1
1.1 Motivations 1
1.2 Literature Survey 3
1.2.1 Conventional Motion Planning in Mobile Manipulations 3
1.2.2 Motion Representation Algorithms 5
1.2.3 Safety-guaranteed Motion Representation Algorithms 7
1.3 Research Objectives and Contributions 7
1.3.1 Motion Generalization in Motion Representation Algorithm 9
1.3.2 Motion Generalization with Safety Guarantee 9
1.3.3 Motion Generalization for Complex Missions 10
1.4 Thesis Organization 11
2 Background 12
2.1 DMPs 12
2.2 Mobile Manipulation Systems 13
2.2.1 Single Mobile Manipulation 14
2.2.2 Cooperative Mobile Manipulations 14
2.3 Experimental Setup 17
2.3.1 Test-beds for Aerial Manipulators 17
2.3.2 Test-beds for Robot Manipulators with Ground Vehicles 17
3 Motion Generalization in Motion Representation Algorithm 22
3.1 Parametric Dynamic Movement Primitives 22
3.2 Generalization Process in PDMPs 26
3.2.1 Environmental Parameters 26
3.2.2 Mapping Function 26
3.3 Simulation Results 29
3.3.1 Two-dimensional Hurdling Motion 29
3.3.2 Cooperative Aerial Transportation 30
4 Motion Generalization with Safety Guarantee 36
4.1 Safety Criterion in Style Parameter 36
4.2 Demonstration Management 39
4.3 Simulation Validation 42
4.3.1 Two-dimensional Hurdling Motion 46
4.3.2 Cooperative Aerial Transportation 47
5 Motion Generalization for Complex Missions 51
5.1 Overall Structure of Seg-PDMPs 51
5.2 Motion Segments 53
5.3 Phase-decision Process 54
5.4 Seg-PDMPs for Single Phase 54
5.5 Simulation Results 55
5.5.1 Initial/terminal Offsets 56
5.5.2 Style Generalization 59
5.5.3 Recombination 61
6 Experimental Validation and Results 63
6.1 Cooperative Aerial Transportation 63
6.2 Cooperative Mobile Hang-dry Mission 70
6.2.1 Demonstrations 70
6.2.2 Simulation Validation 72
6.2.3 Experimental Results 78
7 Conclusions 82
Abstract (in Korean) 93Docto
Book of Abstracts 15th International Symposium on Computer Methods in Biomechanics and Biomedical Engineering and 3rd Conference on Imaging and Visualization
In this edition, the two events will run together as a single conference, highlighting the strong connection with the Taylor & Francis journals: Computer Methods in Biomechanics and Biomedical Engineering (John Middleton and Christopher Jacobs, Eds.) and Computer Methods in Biomechanics and Biomedical Engineering: Imaging and Visualization (JoΓ£oManuel R.S. Tavares, Ed.).
The conference has become a major international meeting on computational biomechanics, imaging andvisualization. In this edition, the main program includes 212 presentations. In addition, sixteen renowned researchers will give plenary keynotes, addressing current challenges in computational biomechanics and biomedical imaging.
In Lisbon, for the first time, a session dedicated to award the winner of the Best Paper in CMBBE Journal will take place.
We believe that CMBBE2018 will have a strong impact on the development of computational biomechanics and biomedical imaging and visualization, identifying emerging areas of research and promoting the collaboration and networking between participants. This impact is evidenced through the well-known research groups, commercial companies and scientific organizations, who continue to support and sponsor the CMBBE meeting
series. In fact, the conference is enriched with five workshops on specific scientific topics and commercial software.info:eu-repo/semantics/draf
Endoscopy
Endoscopy is a fast moving field, and new techniques are continuously emerging. In recent decades, endoscopy has evolved and branched out from a diagnostic modality to enhanced video and computer assisting imaging with impressive interventional capabilities. The modern endoscopy has seen advances not only in types of endoscopes available, but also in types of interventions amenable to the endoscopic approach. To date, there are a lot more developments that are being trialed. Modern endoscopic equipment provides physicians with the benefit of many technical advances. Endoscopy is an effective and safe procedure even in special populations including pediatric patients and renal transplant patients. It serves as the tool for diagnosis and therapeutic interventions of many organs including gastrointestinal tract, head and neck, urinary tract and others