4 research outputs found
Agent with Warm Start and Active Termination for Plane Localization in 3D Ultrasound
Standard plane localization is crucial for ultrasound (US) diagnosis. In
prenatal US, dozens of standard planes are manually acquired with a 2D probe.
It is time-consuming and operator-dependent. In comparison, 3D US containing
multiple standard planes in one shot has the inherent advantages of less
user-dependency and more efficiency. However, manual plane localization in US
volume is challenging due to the huge search space and large fetal posture
variation. In this study, we propose a novel reinforcement learning (RL)
framework to automatically localize fetal brain standard planes in 3D US. Our
contribution is two-fold. First, we equip the RL framework with a
landmark-aware alignment module to provide warm start and strong spatial bounds
for the agent actions, thus ensuring its effectiveness. Second, instead of
passively and empirically terminating the agent inference, we propose a
recurrent neural network based strategy for active termination of the agent's
interaction procedure. This improves both the accuracy and efficiency of the
localization system. Extensively validated on our in-house large dataset, our
approach achieves the accuracy of 3.4mm/9.6{\deg} and 2.7mm/9.1{\deg} for the
transcerebellar and transthalamic plane localization, respectively. Ourproposed
RL framework is general and has the potential to improve the efficiency and
standardization of US scanning.Comment: 9 pages, 5 figures, 1 table. Accepted by MICCAI 2019 (oral
Searching Collaborative Agents for Multi-plane Localization in 3D Ultrasound
3D ultrasound (US) is widely used due to its rich diagnostic information,
portability and low cost. Automated standard plane (SP) localization in US
volume not only improves efficiency and reduces user-dependence, but also
boosts 3D US interpretation. In this study, we propose a novel Multi-Agent
Reinforcement Learning (MARL) framework to localize multiple uterine SPs in 3D
US simultaneously. Our contribution is two-fold. First, we equip the MARL with
a one-shot neural architecture search (NAS) module to obtain the optimal agent
for each plane. Specifically, Gradient-based search using Differentiable
Architecture Sampler (GDAS) is employed to accelerate and stabilize the
training process. Second, we propose a novel collaborative strategy to
strengthen agents' communication. Our strategy uses recurrent neural network
(RNN) to learn the spatial relationship among SPs effectively. Extensively
validated on a large dataset, our approach achieves the accuracy of 7.05
degree/2.21mm, 8.62 degree/2.36mm and 5.93 degree/0.89mm for the mid-sagittal,
transverse and coronal plane localization, respectively. The proposed MARL
framework can significantly increase the plane localization accuracy and reduce
the computational cost and model size.Comment: Early accepted by MICCAI 202
Automatic Image Analysis and Recognition for Ultrasound Diagnosis and Treatment in Cardiac, Obstetrics and Radiology
Ultrasound image analysis and recognition techniques for improving workflow in diagnosis and treatment are introduced. Fully automatic techniques for applications of cardiac plane extraction, foetal weight measurement and ultrasound-CT image registration for liver surgery navigation are included. For standard plane extraction in 3D cardiac ultrasound, multiple cardiac landmarks defined in ultrasound cardiac examination guidelines are detected and localized by a Hough-forest-based detector, and by six standard cardiac planes, cardiac diagnosis is extracted following the guideline. For automatic foetal weight measurement, biparietal diameter (BPD), femur length (FL) and abdominal circumference (AC) are estimated by segmenting corresponding organs and regions from foetal ultrasound images. For ultrasound-CT liver image registration, initial alignment is obtained by localizing a corresponding portal vein branch from an intraoperative ultrasound and preoperative CT image pair. Then portal vein regions of the ultrasound-CT image pair are extracted by a machine learning method and are used for image registration