109 research outputs found
Learning to Generalize over Subpartitions for Heterogeneity-aware Domain Adaptive Nuclei Segmentation
Annotation scarcity and cross-modality/stain data distribution shifts are two
major obstacles hindering the application of deep learning models for nuclei
analysis, which holds a broad spectrum of potential applications in digital
pathology. Recently, unsupervised domain adaptation (UDA) methods have been
proposed to mitigate the distributional gap between different imaging
modalities for unsupervised nuclei segmentation in histopathology images.
However, existing UDA methods are built upon the assumption that data
distributions within each domain should be uniform. Based on the
over-simplified supposition, they propose to align the histopathology target
domain with the source domain integrally, neglecting severe intra-domain
discrepancy over subpartitions incurred by mixed cancer types and sampling
organs. In this paper, for the first time, we propose to explicitly consider
the heterogeneity within the histopathology domain and introduce open compound
domain adaptation (OCDA) to resolve the crux. In specific, a two-stage
disentanglement framework is proposed to acquire domain-invariant feature
representations at both image and instance levels. The holistic design
addresses the limitations of existing OCDA approaches which struggle to capture
instance-wise variations. Two regularization strategies are specifically
devised herein to leverage the rich subpartition-specific characteristics in
histopathology images and facilitate subdomain decomposition. Moreover, we
propose a dual-branch nucleus shape and structure preserving module to prevent
nucleus over-generation and deformation in the synthesized images. Experimental
results on both cross-modality and cross-stain scenarios over a broad range of
diverse datasets demonstrate the superiority of our method compared with
state-of-the-art UDA and OCDA methods
Hover-Net : simultaneous segmentation and classification of nuclei in multi-tissue histology images
Nuclear segmentation and classification within Haematoxylin & Eosin stained histology images is a fundamental prerequisite in the digital pathology work-flow. The development of automated methods for nuclear segmentation and classification enables the quantitative analysis of tens of thousands of nuclei within a whole-slide pathology image, opening up possibilities of further analysis of large-scale nuclear morphometry. However, automated nuclear segmentation and classification is faced with a major challenge in that there are several different types of nuclei, some of them exhibiting large intra-class variability such as the nuclei of tumour cells. Additionally, some of the nuclei are often clustered together. To address these challenges, we present a novel convolutional neural network for simultaneous nuclear segmentation and classification that leverages the instance-rich information encoded within the vertical and horizontal distances of nuclear pixels to their centres of mass. These distances are then utilised to separate clustered nuclei, resulting in an accurate segmentation, particularly in areas with overlapping instances. Then, for each segmented instance the network predicts the type of nucleus via a devoted up-sampling branch. We demonstrate state-of-the-art performance compared to other methods on multiple independent multi-tissue histology image datasets. As part of this work, we introduce a new dataset of Haematoxylin & Eosin stained colorectal adenocarcinoma image tiles, containing 24,319 exhaustively annotated nuclei with associated class labels
A Comprehensive Overview of Computational Nuclei Segmentation Methods in Digital Pathology
In the cancer diagnosis pipeline, digital pathology plays an instrumental
role in the identification, staging, and grading of malignant areas on biopsy
tissue specimens. High resolution histology images are subject to high variance
in appearance, sourcing either from the acquisition devices or the H\&E
staining process. Nuclei segmentation is an important task, as it detects the
nuclei cells over background tissue and gives rise to the topology, size, and
count of nuclei which are determinant factors for cancer detection. Yet, it is
a fairly time consuming task for pathologists, with reportedly high
subjectivity. Computer Aided Diagnosis (CAD) tools empowered by modern
Artificial Intelligence (AI) models enable the automation of nuclei
segmentation. This can reduce the subjectivity in analysis and reading time.
This paper provides an extensive review, beginning from earlier works use
traditional image processing techniques and reaching up to modern approaches
following the Deep Learning (DL) paradigm. Our review also focuses on the weak
supervision aspect of the problem, motivated by the fact that annotated data is
scarce. At the end, the advantages of different models and types of supervision
are thoroughly discussed. Furthermore, we try to extrapolate and envision how
future research lines will potentially be, so as to minimize the need for
labeled data while maintaining high performance. Future methods should
emphasize efficient and explainable models with a transparent underlying
process so that physicians can trust their output.Comment: 47 pages, 27 figures, 9 table
์ฌ์ธต์ ๊ฒฝ๋ง์ ์ด์ฉํ ์๋ํ๋ ์น๊ณผ ์๋ฃ์์ ๋ถ์
ํ์๋
ผ๋ฌธ(๋ฐ์ฌ) -- ์์ธ๋ํ๊ต๋ํ์ : ์น๊ณผ๋ํ ์น์๊ณผํ๊ณผ, 2021.8. ํ์ค์.๋ชฉ ์ : ์น๊ณผ ์์ญ์์๋ ์ฌ์ธต์ ๊ฒฝ๋ง(Deep Neural Network) ๋ชจ๋ธ์ ์ด์ฉํ ๋ฐฉ์ฌ์ ์ฌ์ง์์์ ์ํ๋ํธ ๋ถ๋ฅ, ๋ณ์ ์์น ํ์ง ๋ฑ์ ์ฐ๊ตฌ๋ค์ด ์งํ๋์์ผ๋, ์ต๊ทผ ๊ฐ๋ฐ๋ ํคํฌ์ธํธ ํ์ง(keypoint detection) ๋ชจ๋ธ ๋๋ ์ ์ฒด์ ๊ตฌํํ(panoptic segmentation) ๋ชจ๋ธ์ ์๋ฃ๋ถ์ผ์ ์ ์ฉํ ์ฐ๊ตฌ๋ ์์ง ๋ฏธ๋นํ๋ค. ๋ณธ ์ฐ๊ตฌ์ ๋ชฉ์ ์ ์น๊ทผ๋จ ๋ฐฉ์ฌ์ ์ฌ์ง์์ ํคํฌ์ธํธ ํ์ง๋ฅผ ์ด์ฉํด ์ํ๋ํธ ๊ณจ ์์ค ์ ๋๋ฅผ ํ์
ํ๋ ๋ชจ๋ธ๊ณผ panoptic segmentation์ ํ๋
ธ๋ผ๋ง์์์ ์ ์ฉํ์ฌ ๋ค์ํ ๊ตฌ์กฐ๋ฌผ๋ค์ ๊ตฌํํํ๋ ๋ชจ๋ธ์ ํ์ต์์ผ ์ง๋ฃ์ ๋ณด์กฐ์ ์ผ๋ก ํ์ฉ๋๋๋ก ๋ง๋ค์ด๋ณด๊ณ , ์ด ๋ชจ๋ธ๋ค์ ์ถ๋ก ๊ฒฐ๊ณผ๋ฅผ ํ๊ฐํด๋ณด๋ ๊ฒ์ด๋ค.
๋ฐฉ ๋ฒ: ๊ฐ์ฒด ํ์ง ๋ฐ ๊ตฌํํ์ ์์ด ๋๋ฆฌ ์ฐ๊ตฌ๋ ํฉ์ฑ๊ณฑ ์ ๊ฒฝ๋ง ๋ชจ๋ธ์ธ Mask-RCNN์ ํคํฌ์ธํธ ํ์ง๊ฐ ๊ฐ๋ฅํ ํํ๋ก ์ค๋นํ์ฌ ์น๊ทผ๋จ ๋ฐฉ์ฌ์ ์ฌ์ง์์ ์ํ๋ํธ์ top, apex, ๊ทธ๋ฆฌ๊ณ bone level ์ง์ ์ ์ข์ฐ๋ก ์ด 6์ง์ ํ์งํ๊ฒ๋ ํ์ต์ํจ ๋ค, ํ์ต์ ์ฌ์ฉ๋์ง ์์ ์ํ ๋ฐ์ดํฐ์
์ ๋์์ผ๋ก ํ์ง์ํจ๋ค. ํคํฌ์ธํธ ํ์ง ํ๊ฐ์ฉ ์งํ์ธ object keypoint similarity (OKS) ๋ฐ ์ด๋ฅผ ์ด์ฉํ average precision (AP) ๊ฐ์ ๊ณ์ฐํ๊ณ , ํ๊ท OKS๊ฐ์ ํตํด ๋ชจ๋ธ ๋ฐ ์น๊ณผ์์ฌ์ ๊ฒฐ๊ณผ๋ฅผ ๋น๊ตํ๋ค. ๋ํ, ํ์ง๋ ํคํฌ์ธํธ๋ฅผ ๋ฐํ์ผ๋ก ๋ฐฉ์ฌ์ ์ฌ์ง์์์์ ๊ณจ ์์ค ์ ๋๋ฅผ ์์นํํ๋ค.
Panoptic segmentation์ ์ํด์๋ ๊ธฐ์กด์ ๋ฒค์น๋งํฌ์์ ์ฐ์ํ ์ฑ์ ์ ๊ฑฐ๋ ์ ๊ฒฝ๋ง ๋ชจ๋ธ์ธ Panoptic DeepLab์ ํ๋
ธ๋ผ๋ง์์์์ ์ฃผ์ ๊ตฌ์กฐ๋ฌผ(์์
๋, ์์
๊ณจ, ํ์
๊ด, ํ์
๊ณจ, ์์ฐ์น, ์น๋ฃ๋ ์น์, ์ํ๋ํธ)์ ๊ตฌํํํ๋๋ก ํ์ต์ํจ ๋ค, ์ํ ๋ฐ์ดํฐ์
์์์ ๊ตฌํํ ๊ฒฐ๊ณผ์ panoptic / semantic / instance segmentation ๊ฐ๊ฐ์ ํ๊ฐ์งํ๋ค์ ์ ์ฉํ๊ณ , ํฝ์
๋ค์ ์ ๋ต(ground truth) ํด๋์ค์ ๋ชจ๋ธ์ด ์ถ๋ก ํ ํด๋์ค์ ๋ํ confusion matrix๋ฅผ ๊ณ์ฐํ๋ค.
๊ฒฐ ๊ณผ: OKS๊ฐ์ ๊ธฐ๋ฐ์ผ๋ก ๊ณ์ฐํ ํคํฌ์ธํธ ํ์ง AP๋, ๋ชจ๋ OKS threshold์ ๋ํ ํ๊ท ์ ๊ฒฝ์ฐ, ์์
์ํ๋ํธ์์๋ 0.761, ํ์
์ํ๋ํธ์์๋ 0.786์ด์๋ค. ํ๊ท OKS๋ ๋ชจ๋ธ์ด 0.8885, ์น๊ณผ์์ฌ๊ฐ 0.9012๋ก, ํต๊ณ์ ์ผ๋ก ์ ์๋ฏธํ ์ฐจ์ด๊ฐ ์์๋ค (p = 0.41). ๋ชจ๋ธ์ ํ๊ท OKS ๊ฐ์ ์ฌ๋์ ํคํฌ์ธํธ ์ด๋
ธํ
์ด์
์ ๊ท๋ถํฌ์์์ ์์ 66.92% ์์ค์ด์๋ค.
ํ๋
ธ๋ผ๋ง์์ ๊ตฌ์กฐ๋ฌผ ๊ตฌํํ์์๋, panoptic segmentation ํ๊ฐ์งํ์ธ panoptic quality ๊ฐ์ ๊ฒฝ์ฐ ๋ชจ๋ ํด๋์ค์ ํ๊ท ์ 80.47์ด์์ผ๋ฉฐ, ์น๋ฃ๋ ์น์๊ฐ 57.13์ผ๋ก ๊ฐ์ฅ ๋ฎ์๊ณ ํ์
๊ด์ด 65.97๋ก ๋๋ฒ์งธ๋ก ๋ฎ์ ๊ฐ์ ๋ณด์๋ค. Semantic segmentation ํ๊ฐ์งํ์ธ globalํ Intersection over Union (IoU) ๊ฐ์ ๋ชจ๋ ํด๋์ค ํ๊ท 0.795์์ผ๋ฉฐ, ํ์
๊ด์ด 0.639๋ก ๊ฐ์ฅ ๋ฎ์๊ณ ์น๋ฃ๋ ์น์๊ฐ 0.656์ผ๋ก ๋๋ฒ์งธ๋ก ๋ฎ์ ๊ฐ์ ๋ณด์๋ค. Confusion matrix ๊ณ์ฐ ๊ฒฐ๊ณผ, ground truth ํฝ์
๋ค ์ค ์ฌ๋ฐ๋ฅด๊ฒ ์ถ๋ก ๋ ํฝ์
๋ค์ ๋น์จ์ ํ์
๊ด์ด 0.802๋ก ๊ฐ์ฅ ๋ฎ์๋ค. ๊ฐ๋ณ ๊ฐ์ฒด์ ๋ํ IoU๋ฅผ ๊ธฐ๋ฐ์ผ๋ก ๊ณ์ฐํ Instance segmentation ํ๊ฐ์งํ์ธ AP๊ฐ์, ๋ชจ๋ IoU threshold์ ๋ํ ํ๊ท ์ ๊ฒฝ์ฐ, ์น๋ฃ๋ ์น์๊ฐ 0.316, ์ํ๋ํธ๊ฐ 0.414, ์์ฐ์น๊ฐ 0.520์ด์๋ค.
๊ฒฐ ๋ก : ํคํฌ์ธํธ ํ์ง ์ ๊ฒฝ๋ง ๋ชจ๋ธ์ ์ด์ฉํ์ฌ, ์น๊ทผ๋จ ๋ฐฉ์ฌ์ ์ฌ์ง์์ ์ํ๋ํธ์ ์ฃผ์ ์ง์ ์ ์ฌ๋๊ณผ ๋ค์ ์ ์ฌํ ์์ค์ผ๋ก ํ์งํ ์ ์์๋ค. ๋ํ, ํ์ง๋ ์ง์ ๋ค์ ๊ธฐ๋ฐ์ผ๋ก ๋ฐฉ์ฌ์ ์ฌ์ง์์์์ ์ํ๋ํธ ์ฃผ์ ๊ณจ ์์ค ๋น์จ ๊ณ์ฐ์ ์๋ํํ ์ ์๊ณ , ์ด ๊ฐ์ ์ํ๋ํธ ์ฃผ์์ผ์ ์ฌ๋ ๋ถ๋ฅ์ ์ฌ์ฉ๋ ์ ์๋ค. ํ๋
ธ๋ผ๋ง ์์์์๋ panoptic segmentation์ด ๊ฐ๋ฅํ ์ ๊ฒฝ๋ง ๋ชจ๋ธ์ ์ด์ฉํ์ฌ ์์
๋๊ณผ ํ์
๊ด์ ํฌํจํ ์ฃผ์ ๊ตฌ์กฐ๋ฌผ๋ค์ ๊ตฌํํํ ์ ์์๋ค. ๋ฐ๋ผ์, ์ด์ ๊ฐ์ด ๊ฐ ์์
์ ๋ง๋ ์ฌ์ธต์ ๊ฒฝ๋ง์ ์ ์ ํ ๋ฐ์ดํฐ๋ก ํ์ต์ํจ๋ค๋ฉด ์ง๋ฃ ๋ณด์กฐ ์๋จ์ผ๋ก ํ์ฉ๋ ์ ์๋ค.Purpose: In dentistry, deep neural network models have been applied in areas such as implant classification or lesion detection in radiographs. However, few studies have applied the recently developed keypoint detection model or panoptic segmentation model to medical or dental images. The purpose of this study is to train two neural network models to be used as aids in clinical practice and evaluate them: a model to determine the extent of implant bone loss using keypoint detection in periapical radiographs and a model that segments various structures on panoramic radiographs using panoptic segmentation.
Methods: Mask-RCNN, a widely studied convolutional neural network for object detection and instance segmentation, was constructed in a form that is capable of keypoint detection, and trained to detect six points of an implant in a periapical radiograph: left and right of the top, apex, and bone level. Next, a test dataset was used to evaluate the inference results. Object keypoint similarity (OKS), a metric to evaluate the keypoint detection task, and average precision (AP), based on the OKS values, were calculated. Furthermore, the results of the model and those arrived at by a dentist were compared using the mean OKS. Based on the detected keypoint, the peri-implant bone loss ratio was obtained from the radiograph.
For panoptic segmentation, Panoptic DeepLab, a neural network model ranked high in the previous benchmark, was trained to segment key structures in panoramic radiographs: maxillary sinus, maxilla, mandibular canal, mandible, natural tooth, treated tooth, and dental implant. Then, each evaluation metric of panoptic, semantic, and instance segmentation was applied to the inference results of the test dataset. Finally, the confusion matrix for the ground truth class of pixels and the class inferred by the model was obtained.
Results: The AP of keypoint detection for the average of all OKS thresholds was 0.761 for the upper implants and 0.786 for the lower implants. The mean OKS was 0.8885 for the model and 0.9012 for the dentist; thus, the difference was not statistically significant (p = 0.41). The mean OKS of the model was in the top 66.92% of the normal distribution of human keypoint annotations.
In panoramic radiograph segmentation, the average panoptic quality (PQ) of all classes was 80.47. The treated teeth showed the lowest PQ of 57.13, and the mandibular canal showed the second lowest PQ of 65.97. The Intersection over Union (IoU) was 0.795 on average for all classes, where the mandibular canal showed the lowest IoU of 0.639, and the treated tooth showed the second lowest IoU of 0.656. In the confusion matrix, the proportion of correctly inferred pixels among the ground truth pixels was the lowest in the mandibular canal at 0.802. The AP, averaged for all IoU thresholds, was 0.316 for the treated tooth, 0.414 for the dental implant, and 0.520 for the normal tooth.
Conclusion: Using the keypoint detection neural network model, it was possible to detect major landmarks around dental implants in periapical radiographs to a degree similar to that of human experts. In addition, it was possible to automate the calculation of the peri-implant bone loss ratio on periapical radiographs based on the detected keypoints, and this value could be used to classify the degree of peri-implantitis. In panoramic radiographs, the major structures including the maxillary sinus and the mandibular canal could be segmented using a neural network model capable of panoptic segmentation. Thus, if deep neural networks suitable for each task are trained using suitable datasets, the proposed approach can be used to assist dental clinicians.Chapter 1. Introduction 1
Chapter 2. Materials and methods 5
Chapter 3. Results 23
Chapter 4. Discussion 32
Chapter 5. Conclusions 45
Published papers related to this study 46
References 47
Abbreviations 52
Abstract in Korean 53
Acknowledgements 56๋ฐ
Cell Graph Transformer for Nuclei Classification
Nuclei classification is a critical step in computer-aided diagnosis with
histopathology images. In the past, various methods have employed graph neural
networks (GNN) to analyze cell graphs that model inter-cell relationships by
considering nuclei as vertices. However, they are limited by the GNN mechanism
that only passes messages among local nodes via fixed edges. To address the
issue, we develop a cell graph transformer (CGT) that treats nodes and edges as
input tokens to enable learnable adjacency and information exchange among all
nodes. Nevertheless, training the transformer with a cell graph presents
another challenge. Poorly initialized features can lead to noisy self-attention
scores and inferior convergence, particularly when processing the cell graphs
with numerous connections. Thus, we further propose a novel topology-aware
pretraining method that leverages a graph convolutional network (GCN) to learn
a feature extractor. The pre-trained features may suppress unreasonable
correlations and hence ease the finetuning of CGT. Experimental results suggest
that the proposed cell graph transformer with topology-aware pretraining
significantly improves the nuclei classification results, and achieves the
state-of-the-art performance. Code and models are available at
https://github.com/lhaof/CGTComment: AAAI 2024, Code and models are available at
https://github.com/lhaof/CG
Skeleton-Guided Instance Separation for Fine-Grained Segmentation in Microscopy
One of the fundamental challenges in microscopy (MS) image analysis is
instance segmentation (IS), particularly when segmenting cluster regions where
multiple objects of varying sizes and shapes may be connected or even
overlapped in arbitrary orientations. Existing IS methods usually fail in
handling such scenarios, as they rely on coarse instance representations such
as keypoints and horizontal bounding boxes (h-bboxes). In this paper, we
propose a novel one-stage framework named A2B-IS to address this challenge and
enhance the accuracy of IS in MS images. Our approach represents each instance
with a pixel-level mask map and a rotated bounding box (r-bbox). Unlike
two-stage methods that use box proposals for segmentations, our method
decouples mask and box predictions, enabling simultaneous processing to
streamline the model pipeline. Additionally, we introduce a Gaussian skeleton
map to aid the IS task in two key ways: (1) It guides anchor placement,
reducing computational costs while improving the model's capacity to learn
RoI-aware features by filtering out noise from background regions. (2) It
ensures accurate isolation of densely packed instances by rectifying erroneous
box predictions near instance boundaries. To further enhance the performance,
we integrate two modules into the framework: (1) An Atrous Attention Block
(A2B) designed to extract high-resolution feature maps with fine-grained
multiscale information, and (2) A Semi-Supervised Learning (SSL) strategy that
leverages both labeled and unlabeled images for model training. Our method has
been thoroughly validated on two large-scale MS datasets, demonstrating its
superiority over most state-of-the-art approaches
Point-supervised Single-cell Segmentation via Collaborative Knowledge Sharing
Despite their superior performance, deep-learning methods often suffer from
the disadvantage of needing large-scale well-annotated training data. In
response, recent literature has seen a proliferation of efforts aimed at
reducing the annotation burden. This paper focuses on a weakly-supervised
training setting for single-cell segmentation models, where the only available
training label is the rough locations of individual cells. The specific problem
is of practical interest due to the widely available nuclei counter-stain data
in biomedical literature, from which the cell locations can be derived
programmatically. Of more general interest is a proposed self-learning method
called collaborative knowledge sharing, which is related to but distinct from
the more well-known consistency learning methods. This strategy achieves
self-learning by sharing knowledge between a principal model and a very
light-weight collaborator model. Importantly, the two models are entirely
different in their architectures, capacities, and model outputs: In our case,
the principal model approaches the segmentation problem from an
object-detection perspective, whereas the collaborator model a sematic
segmentation perspective. We assessed the effectiveness of this strategy by
conducting experiments on LIVECell, a large single-cell segmentation dataset of
bright-field images, and on A431 dataset, a fluorescence image dataset in which
the location labels are generated automatically from nuclei counter-stain data.
Implementing code is available at https://github.com/jiyuuchc/lacss_ja
- โฆ