43 research outputs found
Recent Trends in Artificial Intelligence-Assisted Coronary Atherosclerotic Plaque Characterization
Coronary artery disease is a major cause of morbidity and mortality worldwide. Its underlying histopathology is the atherosclerotic plaque, which comprises lipid, fibrous and—when chronic—calcium components. Intravascular ultrasound (IVUS) and intravascular optical coherence tomography (IVOCT) performed during invasive coronary angiography are reference standards for characterizing the atherosclerotic plaque. Fine image spatial resolution attainable with contemporary coronary computed tomographic angiography (CCTA) has enabled noninvasive plaque assessment, including identifying features associated with vulnerable plaques known to presage acute coronary events. Manual interpretation of IVUS, IVOCT and CCTA images demands scarce physician expertise and high time cost. This has motivated recent research into and development of artificial intelligence (AI)-assisted methods for image processing, feature extraction, plaque identification and characterization. We performed parallel searches of the medical and technical literature from 1995 to 2021 focusing respectively on human plaque characterization using various imaging modalities and the use of AI-assisted computer aided diagnosis (CAD) to detect and classify atherosclerotic plaques, including their composition and the presence of high-risk features denoting vulnerable plaques. A total of 122 publications were selected for evaluation and the analysis was summarized in terms of data sources, methods—machine versus deep learning—and performance metrics. Trends in AI-assisted plaque characterization are detailed and prospective research challenges discussed. Future directions for the development of accurate and efficient CAD systems to characterize plaque noninvasively using CCTA are proposed.</jats:p
Recommended from our members
Detection, localization and quantification of non-calcified coronary plaques in contrast enhanced CT angiography
State-of-the-art imaging equipment has increased clinician's ability to make non-invasive diagnoses of coronary heart disease (CHD); however, high volumes of imaging data make manual abnormality detection cumbersome in practice. In addition, the interpretation of CTA heavily relies upon the previous knowledge of the clinician. These limitations have driven an intense research in the context of automated solutions for fast, reliable and accurate diagnosis. Accordingly, in this thesis, we present an automated framework for detection, localization and quantification of the non-calcified coronary plaques in cardiac computed tomography angiography (CTA).
The first contribution of the thesis is a coronary segmentation algorithm that is adaptive to the contrast agent and employs a hybrid energy incorporating local and global image statistics in a segmentation framework using partial differential equations (PDEs). Accordingly, we illustrated with the help of experimental evidence that a volume-specific intensity threshold leads to an improved segmentation in CTA. In the subsequent step, we employed a hybrid region-based energy for improved segmentation in CTA imagery. The hybrid energy couples an intensity-based local term with an efficient discontinuity-based global model of the image for optimal segmentation. The proposed method is less sensitive to the local optima problem and helps in reducing false positives, as well as it allows a certain degree of freedom for the initialization. Moreover, we employed an auto-correction feature for improved segmentation, as an auto-corrected mask captures the emerging peripheries of the coronary tree during the curve evolution. The effectiveness of the proposed model is demonstrated with the help of both qualitative and quantitative results, with a mean accuracy of 80% across the CTA dataset. The capability to address the variations in initial mask and localization radii simultaneously, makes our algorithm a feasible choice for coronary segmentation.
The second contribution of the thesis is an automatic approach to analyse the segmented coronary tree for the presence of non-calcified plaques. The specific focus of this work is detection of non-calcified plaques in CTA, as intensity overlap between blood, fat and non-calcified plaques make the detection challenging. Non-calcified plaques are identified based on mean radial profiles that average the image intensities in concentric rings around the vessel centreline. Subsequently, an SVM classifier is applied to differentiate the abnormal coronary segments from normal ones. A total of 32 CTA volumes have been analysed and a detection accuracy of 88.4% with respect to the manual expert has been achieved. For plaque-affected segments, we further proposed a derivative-based method to localize the position and length of the plaque inside the segment. The plaque localization accuracy has been around 83.2%. Moreover, the proposed model has been tested on three different CTA datasets and has produced consistent results, demonstrating its reproducibility for generic CTA data.
The final contribution of the thesis is a method to segment and quantify the non-calcified plaque. After evaluating the vessel wall thickness, posterior probability based voxel classification has been performed to quantify the lumen and plaque, respectively. Both qualitative and quantitative results demonstrate that the proposed model shows a good agreement with three independent experts. To optimize the processing time, we employed sparse field method in a level-set based active contour evolution
직접 볼륨 렌더링의 전이 함수 설계에 관한 연구
학위논문 (박사)-- 서울대학교 대학원 : 전기·컴퓨터공학부, 2017. 2. 신영길.Although direct volume rendering (DVR) has become a commodity, the design of transfer functions still a challenge. Transfer functions which map data values to optical properties (i.e., colors and opacities) highlight features of interests as well as hide unimportant regions, dramatically impacting on the quality of the visualization. Therefore, for the effective rendering of interesting features, the design of transfer functions is very important and challenging task. Furthermore, manipulation of these transfer functions is tedious and time-consuming task. In this paper, we propose a 3D spatial field for accurately identifying and visually distinguishing interesting features as well as a mechanism for data exploration using multi-dimensional transfer function.
First, we introduce a 3D spatial field for the effective visualization of constricted tubular structures, called as a stenosis map which stores the degree of constriction at each voxel. Constrictions within tubular structures are quantified by using newly proposed measures (i.e., line similarity measure and constriction measure) based on the localized structure analysis, and classified with a proposed transfer function mapping the degree of constriction to color and opacity. We show the application results of our method to the visualization of coronary artery stenoses. We present performance evaluations using twenty-eight clinical datasets, demonstrating high accuracy and efficacy of our proposed method.
Second, we propose a new multi-dimensional transfer function which incorporates texture features calculated from statistically homogeneous regions. This approach employs parallel coordinates to provide an intuitive interface for exploring a new multi-dimensional transfer function space. Three specific ways to use a new transfer function based on parallel coordinates enables the effective exploration of large and complex datasets. We present a mechanism for data exploration with a new transfer function space, demonstrating the practical efficacy of our proposed method.
Through a study on transfer function design for DVR, we propose two useful approaches. First method to saliently visualize the constrictions within tubular structures and interactively adjust the visual appearance of the constrictions delivers a substantial aid in radiologic practice. Furthermore, second method to classify objects with our intuitive interface utilizing parallel coordinates proves to be a powerful tool for complex data exploration.Chapter 1 Introduction 1
1.1 Background 1
1.1.1 Volume rendering 1
1.1.2 Computer-aided diagnosis 3
1.1.3 Parallel coordinates 5
1.2 Problem statement 8
1.3 Main contribution 12
1.4 Organization of dissertation 16
Chapter 2 Related Work 17
2.1 Transfer function 17
2.1.1 Transfer functions based on spatial characteristics 17
2.1.2 Opacity modulation techniques 20
2.1.3 Multi-dimensional transfer functions 22
2.1.4 Manipulation mechanism for transfer functions 25
2.2 Coronary artery stenosis 28
2.3 Parallel coordinates 32
Chapter 3 Volume Visualization of Constricted Tubular Structures 36
3.1 Overview 36
3.2 Localized structure analysis 37
3.3 Stenosis map 39
3.3.1 Overview 39
3.3.2 Detection of tubular structures 40
3.3.3 Stenosis map computation 49
3.4 Stenosis-based classification 52
3.4.1 Overview 52
3.4.2 Constriction-encoded volume rendering 52
3.4.3 Opacity modulation based on constriction 54
3.5 GPU implementation 57
3.6 Experimental results 59
3.6.1 Clinical data preparation 59
3.6.2 Qualitative evaluation 60
3.6.3 Quantitative evaluation 63
3.6.4 Comparison with previous methods 66
3.6.5 Parameter study 69
Chapter 4 Interactive Multi-Dimensional Transfer Function Using Adaptive Block Based Feature Analysis 73
4.1 Overview 73
4.2 Extraction of statistical features 74
4.3 Extraction of texture features 78
4.4 Multi-dimensional transfer function design using parallel coordinates 81
4.5 Experimental results 86
Chapter 5 Conclusion 90
Bibliography 92
초 록 107Docto
Automated Quantification of Atherosclerosis in CTA of Carotid Arteries
How is the human body built and how does it function? What are the causes of
disease, and where is disease located? Throughout the history of mankind these
questions were answered by the use of invasive methods that included the
“opening” of the human body, mainly cadavers. Thanks to these invasive
techniques the first precise and complete anatomy works started to appear in
the 16th century. The most influential works were published by Leonardo da
Vinci and the anatomist and physician Andreas Vesalius.
The discovery of X-rays in 1895, and their use for medical applications,
introduced a new era, in which non-invasive imaging of the functioning human
body became feasible. Nowadays, medical imaging includes many different
imaging modalities, such as X-ray, computed tomography (CT), magnetic
resonance imaging (MRI), ultrasound (US), nuclear and optical imaging, and
has become an indispensable diagnostic tool for a wide range of applications.
Initially, the application of medical imaging focused on the visualization of
anatomy and on the detection and localization of disease. However, with the
development of different modalities it has evolved into a much more versatile
tool providing important information on e.g. physiology and organ function,
biochemistry and metabolism using nuclear imaging (mainly positron emission
tomography (PET) imaging), molecular and processes on the molecular
and cellular level using molecular imaging techniques
Recommended from our members
Blood Vessel Segmentation and shape analysis for quantification of Coronary Artery Stenosis in CT Angiography
This thesis presents an automated framework for quantitative vascular shape analysis of the coronary arteries, which constitutes an important and fundamental component of an automated image-based diagnostic system. Firstly, an automated vessel segmentation algorithm is developed to extract the coronary arteries based on the framework of active contours. Both global and local intensity statistics are utilised in the energy functional calculation, which allows for dealing with non-uniform brightness conditions, while evolving the contour towards to the desired boundaries without being trapped in local minima. To suppress kissing vessel artifacts, a slice-by-slice correction scheme, based on multiple regions competition, is proposed to identify and track the kissing vessels throughout the transaxial images of the CTA data. Based on the resulting segmentation, we then present a dedicated algorithm to estimate the geometric parameters of the extracted arteries, with focus on vessel bifurcations. In particular, the centreline and associated reference surface of the coronary arteries, in the vicinity of arterial bifurcations, are determined by registering an elliptical cross sectional tube to the desired constituent branch. The registration problem is solved by a hybrid optimisation method, combining local greedy search and dynamic programming, which ensures the global optimality of the solution and permits the incorporation of any hard constraints posed to the tube model within a natural and direct framework. In contrast with conventional volume domain methods, this technique works directly on the mesh domain, thus alleviating the need for image upsampling. The performance of the proposed framework, in terms of efficiency and accuracy, is demonstrated on both synthetic and clinical image data. Experimental results have shown that our techniques are capable of extracting the major branches of the coronary arteries and estimating the related geometric parameters (i.e., the centreline and the reference surface) with a high degree of agreement to those obtained through manual delineation. Particularly, all of the major branches of coronary arteries are successfully detected by the proposed technique, with a voxel-wise error at 0.73 voxels to the manually delineated ground truth data. Through the application of the slice-by-slice correction scheme, the false positive metric, for those coronary segments affected by kissing vessel artifacts, reduces from 294% to 22.5%. In terms of the capability of the presented framework in defining the location of centrelines across vessel bifurcations, the mean square errors (MSE) of the resulting centreline, with respect to the ground truth data, is reduced by an average of 62.3%, when compared with initial estimation obtained using a topological thinning based algorithm