109 research outputs found

    Learning to Generalize over Subpartitions for Heterogeneity-aware Domain Adaptive Nuclei Segmentation

    Full text link
    Annotation scarcity and cross-modality/stain data distribution shifts are two major obstacles hindering the application of deep learning models for nuclei analysis, which holds a broad spectrum of potential applications in digital pathology. Recently, unsupervised domain adaptation (UDA) methods have been proposed to mitigate the distributional gap between different imaging modalities for unsupervised nuclei segmentation in histopathology images. However, existing UDA methods are built upon the assumption that data distributions within each domain should be uniform. Based on the over-simplified supposition, they propose to align the histopathology target domain with the source domain integrally, neglecting severe intra-domain discrepancy over subpartitions incurred by mixed cancer types and sampling organs. In this paper, for the first time, we propose to explicitly consider the heterogeneity within the histopathology domain and introduce open compound domain adaptation (OCDA) to resolve the crux. In specific, a two-stage disentanglement framework is proposed to acquire domain-invariant feature representations at both image and instance levels. The holistic design addresses the limitations of existing OCDA approaches which struggle to capture instance-wise variations. Two regularization strategies are specifically devised herein to leverage the rich subpartition-specific characteristics in histopathology images and facilitate subdomain decomposition. Moreover, we propose a dual-branch nucleus shape and structure preserving module to prevent nucleus over-generation and deformation in the synthesized images. Experimental results on both cross-modality and cross-stain scenarios over a broad range of diverse datasets demonstrate the superiority of our method compared with state-of-the-art UDA and OCDA methods

    Hover-Net : simultaneous segmentation and classification of nuclei in multi-tissue histology images

    Get PDF
    Nuclear segmentation and classification within Haematoxylin & Eosin stained histology images is a fundamental prerequisite in the digital pathology work-flow. The development of automated methods for nuclear segmentation and classification enables the quantitative analysis of tens of thousands of nuclei within a whole-slide pathology image, opening up possibilities of further analysis of large-scale nuclear morphometry. However, automated nuclear segmentation and classification is faced with a major challenge in that there are several different types of nuclei, some of them exhibiting large intra-class variability such as the nuclei of tumour cells. Additionally, some of the nuclei are often clustered together. To address these challenges, we present a novel convolutional neural network for simultaneous nuclear segmentation and classification that leverages the instance-rich information encoded within the vertical and horizontal distances of nuclear pixels to their centres of mass. These distances are then utilised to separate clustered nuclei, resulting in an accurate segmentation, particularly in areas with overlapping instances. Then, for each segmented instance the network predicts the type of nucleus via a devoted up-sampling branch. We demonstrate state-of-the-art performance compared to other methods on multiple independent multi-tissue histology image datasets. As part of this work, we introduce a new dataset of Haematoxylin & Eosin stained colorectal adenocarcinoma image tiles, containing 24,319 exhaustively annotated nuclei with associated class labels

    A Comprehensive Overview of Computational Nuclei Segmentation Methods in Digital Pathology

    Full text link
    In the cancer diagnosis pipeline, digital pathology plays an instrumental role in the identification, staging, and grading of malignant areas on biopsy tissue specimens. High resolution histology images are subject to high variance in appearance, sourcing either from the acquisition devices or the H\&E staining process. Nuclei segmentation is an important task, as it detects the nuclei cells over background tissue and gives rise to the topology, size, and count of nuclei which are determinant factors for cancer detection. Yet, it is a fairly time consuming task for pathologists, with reportedly high subjectivity. Computer Aided Diagnosis (CAD) tools empowered by modern Artificial Intelligence (AI) models enable the automation of nuclei segmentation. This can reduce the subjectivity in analysis and reading time. This paper provides an extensive review, beginning from earlier works use traditional image processing techniques and reaching up to modern approaches following the Deep Learning (DL) paradigm. Our review also focuses on the weak supervision aspect of the problem, motivated by the fact that annotated data is scarce. At the end, the advantages of different models and types of supervision are thoroughly discussed. Furthermore, we try to extrapolate and envision how future research lines will potentially be, so as to minimize the need for labeled data while maintaining high performance. Future methods should emphasize efficient and explainable models with a transparent underlying process so that physicians can trust their output.Comment: 47 pages, 27 figures, 9 table

    ์‹ฌ์ธต์‹ ๊ฒฝ๋ง์„ ์ด์šฉํ•œ ์ž๋™ํ™”๋œ ์น˜๊ณผ ์˜๋ฃŒ์˜์ƒ ๋ถ„์„

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ์น˜๊ณผ๋Œ€ํ•™ ์น˜์˜๊ณผํ•™๊ณผ, 2021.8. ํ•œ์ค‘์„.๋ชฉ ์ : ์น˜๊ณผ ์˜์—ญ์—์„œ๋„ ์‹ฌ์ธต์‹ ๊ฒฝ๋ง(Deep Neural Network) ๋ชจ๋ธ์„ ์ด์šฉํ•œ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์—์„œ์˜ ์ž„ํ”Œ๋ž€ํŠธ ๋ถ„๋ฅ˜, ๋ณ‘์†Œ ์œ„์น˜ ํƒ์ง€ ๋“ฑ์˜ ์—ฐ๊ตฌ๋“ค์ด ์ง„ํ–‰๋˜์—ˆ์œผ๋‚˜, ์ตœ๊ทผ ๊ฐœ๋ฐœ๋œ ํ‚คํฌ์ธํŠธ ํƒ์ง€(keypoint detection) ๋ชจ๋ธ ๋˜๋Š” ์ „์ฒด์  ๊ตฌํšํ™”(panoptic segmentation) ๋ชจ๋ธ์„ ์˜๋ฃŒ๋ถ„์•ผ์— ์ ์šฉํ•œ ์—ฐ๊ตฌ๋Š” ์•„์ง ๋ฏธ๋น„ํ•˜๋‹ค. ๋ณธ ์—ฐ๊ตฌ์˜ ๋ชฉ์ ์€ ์น˜๊ทผ๋‹จ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์—์„œ ํ‚คํฌ์ธํŠธ ํƒ์ง€๋ฅผ ์ด์šฉํ•ด ์ž„ํ”Œ๋ž€ํŠธ ๊ณจ ์†Œ์‹ค ์ •๋„๋ฅผ ํŒŒ์•…ํ•˜๋Š” ๋ชจ๋ธ๊ณผ panoptic segmentation์„ ํŒŒ๋…ธ๋ผ๋งˆ์˜์ƒ์— ์ ์šฉํ•˜์—ฌ ๋‹ค์–‘ํ•œ ๊ตฌ์กฐ๋ฌผ๋“ค์„ ๊ตฌํšํ™”ํ•˜๋Š” ๋ชจ๋ธ์„ ํ•™์Šต์‹œ์ผœ ์ง„๋ฃŒ์— ๋ณด์กฐ์ ์œผ๋กœ ํ™œ์šฉ๋˜๋„๋ก ๋งŒ๋“ค์–ด๋ณด๊ณ , ์ด ๋ชจ๋ธ๋“ค์˜ ์ถ”๋ก ๊ฒฐ๊ณผ๋ฅผ ํ‰๊ฐ€ํ•ด๋ณด๋Š” ๊ฒƒ์ด๋‹ค. ๋ฐฉ ๋ฒ•: ๊ฐ์ฒด ํƒ์ง€ ๋ฐ ๊ตฌํšํ™”์— ์žˆ์–ด ๋„๋ฆฌ ์—ฐ๊ตฌ๋œ ํ•ฉ์„ฑ๊ณฑ ์‹ ๊ฒฝ๋ง ๋ชจ๋ธ์ธ Mask-RCNN์„ ํ‚คํฌ์ธํŠธ ํƒ์ง€๊ฐ€ ๊ฐ€๋Šฅํ•œ ํ˜•ํƒœ๋กœ ์ค€๋น„ํ•˜์—ฌ ์น˜๊ทผ๋‹จ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์—์„œ ์ž„ํ”Œ๋ž€ํŠธ์˜ top, apex, ๊ทธ๋ฆฌ๊ณ  bone level ์ง€์ ์„ ์ขŒ์šฐ๋กœ ์ด 6์ง€์  ํƒ์ง€ํ•˜๊ฒŒ๋” ํ•™์Šต์‹œํ‚จ ๋’ค, ํ•™์Šต์— ์‚ฌ์šฉ๋˜์ง€ ์•Š์€ ์‹œํ—˜ ๋ฐ์ดํ„ฐ์…‹์„ ๋Œ€์ƒ์œผ๋กœ ํƒ์ง€์‹œํ‚จ๋‹ค. ํ‚คํฌ์ธํŠธ ํƒ์ง€ ํ‰๊ฐ€์šฉ ์ง€ํ‘œ์ธ object keypoint similarity (OKS) ๋ฐ ์ด๋ฅผ ์ด์šฉํ•œ average precision (AP) ๊ฐ’์„ ๊ณ„์‚ฐํ•˜๊ณ , ํ‰๊ท  OKS๊ฐ’์„ ํ†ตํ•ด ๋ชจ๋ธ ๋ฐ ์น˜๊ณผ์˜์‚ฌ์˜ ๊ฒฐ๊ณผ๋ฅผ ๋น„๊ตํ•œ๋‹ค. ๋˜ํ•œ, ํƒ์ง€๋œ ํ‚คํฌ์ธํŠธ๋ฅผ ๋ฐ”ํƒ•์œผ๋กœ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์ƒ์—์„œ์˜ ๊ณจ ์†Œ์‹ค ์ •๋„๋ฅผ ์ˆ˜์น˜ํ™”ํ•œ๋‹ค. Panoptic segmentation์„ ์œ„ํ•ด์„œ๋Š” ๊ธฐ์กด์˜ ๋ฒค์น˜๋งˆํฌ์—์„œ ์šฐ์ˆ˜ํ•œ ์„ฑ์ ์„ ๊ฑฐ๋‘” ์‹ ๊ฒฝ๋ง ๋ชจ๋ธ์ธ Panoptic DeepLab์„ ํŒŒ๋…ธ๋ผ๋งˆ์˜์ƒ์—์„œ ์ฃผ์š” ๊ตฌ์กฐ๋ฌผ(์ƒ์•…๋™, ์ƒ์•…๊ณจ, ํ•˜์•…๊ด€, ํ•˜์•…๊ณจ, ์ž์—ฐ์น˜, ์น˜๋ฃŒ๋œ ์น˜์•„, ์ž„ํ”Œ๋ž€ํŠธ)์„ ๊ตฌํšํ™”ํ•˜๋„๋ก ํ•™์Šต์‹œํ‚จ ๋’ค, ์‹œํ—˜ ๋ฐ์ดํ„ฐ์…‹์—์„œ์˜ ๊ตฌํšํ™” ๊ฒฐ๊ณผ์— panoptic / semantic / instance segmentation ๊ฐ๊ฐ์˜ ํ‰๊ฐ€์ง€ํ‘œ๋“ค์„ ์ ์šฉํ•˜๊ณ , ํ”ฝ์…€๋“ค์˜ ์ •๋‹ต(ground truth) ํด๋ž˜์Šค์™€ ๋ชจ๋ธ์ด ์ถ”๋ก ํ•œ ํด๋ž˜์Šค์— ๋Œ€ํ•œ confusion matrix๋ฅผ ๊ณ„์‚ฐํ•œ๋‹ค. ๊ฒฐ ๊ณผ: OKS๊ฐ’์„ ๊ธฐ๋ฐ˜์œผ๋กœ ๊ณ„์‚ฐํ•œ ํ‚คํฌ์ธํŠธ ํƒ์ง€ AP๋Š”, ๋ชจ๋“  OKS threshold์— ๋Œ€ํ•œ ํ‰๊ท ์˜ ๊ฒฝ์šฐ, ์ƒ์•… ์ž„ํ”Œ๋ž€ํŠธ์—์„œ๋Š” 0.761, ํ•˜์•… ์ž„ํ”Œ๋ž€ํŠธ์—์„œ๋Š” 0.786์ด์—ˆ๋‹ค. ํ‰๊ท  OKS๋Š” ๋ชจ๋ธ์ด 0.8885, ์น˜๊ณผ์˜์‚ฌ๊ฐ€ 0.9012๋กœ, ํ†ต๊ณ„์ ์œผ๋กœ ์œ ์˜๋ฏธํ•œ ์ฐจ์ด๊ฐ€ ์—†์—ˆ๋‹ค (p = 0.41). ๋ชจ๋ธ์˜ ํ‰๊ท  OKS ๊ฐ’์€ ์‚ฌ๋žŒ์˜ ํ‚คํฌ์ธํŠธ ์–ด๋…ธํ…Œ์ด์…˜ ์ •๊ทœ๋ถ„ํฌ์ƒ์—์„œ ์ƒ์œ„ 66.92% ์ˆ˜์ค€์ด์—ˆ๋‹ค. ํŒŒ๋…ธ๋ผ๋งˆ์˜์ƒ ๊ตฌ์กฐ๋ฌผ ๊ตฌํšํ™”์—์„œ๋Š”, panoptic segmentation ํ‰๊ฐ€์ง€ํ‘œ์ธ panoptic quality ๊ฐ’์˜ ๊ฒฝ์šฐ ๋ชจ๋“  ํด๋ž˜์Šค์˜ ํ‰๊ท ์€ 80.47์ด์—ˆ์œผ๋ฉฐ, ์น˜๋ฃŒ๋œ ์น˜์•„๊ฐ€ 57.13์œผ๋กœ ๊ฐ€์žฅ ๋‚ฎ์•˜๊ณ  ํ•˜์•…๊ด€์ด 65.97๋กœ ๋‘๋ฒˆ์งธ๋กœ ๋‚ฎ์€ ๊ฐ’์„ ๋ณด์˜€๋‹ค. Semantic segmentation ํ‰๊ฐ€์ง€ํ‘œ์ธ globalํ•œ Intersection over Union (IoU) ๊ฐ’์€ ๋ชจ๋“  ํด๋ž˜์Šค ํ‰๊ท  0.795์˜€์œผ๋ฉฐ, ํ•˜์•…๊ด€์ด 0.639๋กœ ๊ฐ€์žฅ ๋‚ฎ์•˜๊ณ  ์น˜๋ฃŒ๋œ ์น˜์•„๊ฐ€ 0.656์œผ๋กœ ๋‘๋ฒˆ์งธ๋กœ ๋‚ฎ์€ ๊ฐ’์„ ๋ณด์˜€๋‹ค. Confusion matrix ๊ณ„์‚ฐ ๊ฒฐ๊ณผ, ground truth ํ”ฝ์…€๋“ค ์ค‘ ์˜ฌ๋ฐ”๋ฅด๊ฒŒ ์ถ”๋ก ๋œ ํ”ฝ์…€๋“ค์˜ ๋น„์œจ์€ ํ•˜์•…๊ด€์ด 0.802๋กœ ๊ฐ€์žฅ ๋‚ฎ์•˜๋‹ค. ๊ฐœ๋ณ„ ๊ฐ์ฒด์— ๋Œ€ํ•œ IoU๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ๊ณ„์‚ฐํ•œ Instance segmentation ํ‰๊ฐ€์ง€ํ‘œ์ธ AP๊ฐ’์€, ๋ชจ๋“  IoU threshold์— ๋Œ€ํ•œ ํ‰๊ท ์˜ ๊ฒฝ์šฐ, ์น˜๋ฃŒ๋œ ์น˜์•„๊ฐ€ 0.316, ์ž„ํ”Œ๋ž€ํŠธ๊ฐ€ 0.414, ์ž์—ฐ์น˜๊ฐ€ 0.520์ด์—ˆ๋‹ค. ๊ฒฐ ๋ก : ํ‚คํฌ์ธํŠธ ํƒ์ง€ ์‹ ๊ฒฝ๋ง ๋ชจ๋ธ์„ ์ด์šฉํ•˜์—ฌ, ์น˜๊ทผ๋‹จ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์—์„œ ์ž„ํ”Œ๋ž€ํŠธ์˜ ์ฃผ์š” ์ง€์ ์„ ์‚ฌ๋žŒ๊ณผ ๋‹ค์†Œ ์œ ์‚ฌํ•œ ์ˆ˜์ค€์œผ๋กœ ํƒ์ง€ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ๋˜ํ•œ, ํƒ์ง€๋œ ์ง€์ ๋“ค์„ ๊ธฐ๋ฐ˜์œผ๋กœ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์ƒ์—์„œ์˜ ์ž„ํ”Œ๋ž€ํŠธ ์ฃผ์œ„ ๊ณจ ์†Œ์‹ค ๋น„์œจ ๊ณ„์‚ฐ์„ ์ž๋™ํ™”ํ•  ์ˆ˜ ์žˆ๊ณ , ์ด ๊ฐ’์€ ์ž„ํ”Œ๋ž€ํŠธ ์ฃผ์œ„์—ผ์˜ ์‹ฌ๋„ ๋ถ„๋ฅ˜์— ์‚ฌ์šฉ๋  ์ˆ˜ ์žˆ๋‹ค. ํŒŒ๋…ธ๋ผ๋งˆ ์˜์ƒ์—์„œ๋Š” panoptic segmentation์ด ๊ฐ€๋Šฅํ•œ ์‹ ๊ฒฝ๋ง ๋ชจ๋ธ์„ ์ด์šฉํ•˜์—ฌ ์ƒ์•…๋™๊ณผ ํ•˜์•…๊ด€์„ ํฌํ•จํ•œ ์ฃผ์š” ๊ตฌ์กฐ๋ฌผ๋“ค์„ ๊ตฌํšํ™”ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ๋”ฐ๋ผ์„œ, ์ด์™€ ๊ฐ™์ด ๊ฐ ์ž‘์—…์— ๋งž๋Š” ์‹ฌ์ธต์‹ ๊ฒฝ๋ง์„ ์ ์ ˆํ•œ ๋ฐ์ดํ„ฐ๋กœ ํ•™์Šต์‹œํ‚จ๋‹ค๋ฉด ์ง„๋ฃŒ ๋ณด์กฐ ์ˆ˜๋‹จ์œผ๋กœ ํ™œ์šฉ๋  ์ˆ˜ ์žˆ๋‹ค.Purpose: In dentistry, deep neural network models have been applied in areas such as implant classification or lesion detection in radiographs. However, few studies have applied the recently developed keypoint detection model or panoptic segmentation model to medical or dental images. The purpose of this study is to train two neural network models to be used as aids in clinical practice and evaluate them: a model to determine the extent of implant bone loss using keypoint detection in periapical radiographs and a model that segments various structures on panoramic radiographs using panoptic segmentation. Methods: Mask-RCNN, a widely studied convolutional neural network for object detection and instance segmentation, was constructed in a form that is capable of keypoint detection, and trained to detect six points of an implant in a periapical radiograph: left and right of the top, apex, and bone level. Next, a test dataset was used to evaluate the inference results. Object keypoint similarity (OKS), a metric to evaluate the keypoint detection task, and average precision (AP), based on the OKS values, were calculated. Furthermore, the results of the model and those arrived at by a dentist were compared using the mean OKS. Based on the detected keypoint, the peri-implant bone loss ratio was obtained from the radiograph. For panoptic segmentation, Panoptic DeepLab, a neural network model ranked high in the previous benchmark, was trained to segment key structures in panoramic radiographs: maxillary sinus, maxilla, mandibular canal, mandible, natural tooth, treated tooth, and dental implant. Then, each evaluation metric of panoptic, semantic, and instance segmentation was applied to the inference results of the test dataset. Finally, the confusion matrix for the ground truth class of pixels and the class inferred by the model was obtained. Results: The AP of keypoint detection for the average of all OKS thresholds was 0.761 for the upper implants and 0.786 for the lower implants. The mean OKS was 0.8885 for the model and 0.9012 for the dentist; thus, the difference was not statistically significant (p = 0.41). The mean OKS of the model was in the top 66.92% of the normal distribution of human keypoint annotations. In panoramic radiograph segmentation, the average panoptic quality (PQ) of all classes was 80.47. The treated teeth showed the lowest PQ of 57.13, and the mandibular canal showed the second lowest PQ of 65.97. The Intersection over Union (IoU) was 0.795 on average for all classes, where the mandibular canal showed the lowest IoU of 0.639, and the treated tooth showed the second lowest IoU of 0.656. In the confusion matrix, the proportion of correctly inferred pixels among the ground truth pixels was the lowest in the mandibular canal at 0.802. The AP, averaged for all IoU thresholds, was 0.316 for the treated tooth, 0.414 for the dental implant, and 0.520 for the normal tooth. Conclusion: Using the keypoint detection neural network model, it was possible to detect major landmarks around dental implants in periapical radiographs to a degree similar to that of human experts. In addition, it was possible to automate the calculation of the peri-implant bone loss ratio on periapical radiographs based on the detected keypoints, and this value could be used to classify the degree of peri-implantitis. In panoramic radiographs, the major structures including the maxillary sinus and the mandibular canal could be segmented using a neural network model capable of panoptic segmentation. Thus, if deep neural networks suitable for each task are trained using suitable datasets, the proposed approach can be used to assist dental clinicians.Chapter 1. Introduction 1 Chapter 2. Materials and methods 5 Chapter 3. Results 23 Chapter 4. Discussion 32 Chapter 5. Conclusions 45 Published papers related to this study 46 References 47 Abbreviations 52 Abstract in Korean 53 Acknowledgements 56๋ฐ•

    Cell Graph Transformer for Nuclei Classification

    Full text link
    Nuclei classification is a critical step in computer-aided diagnosis with histopathology images. In the past, various methods have employed graph neural networks (GNN) to analyze cell graphs that model inter-cell relationships by considering nuclei as vertices. However, they are limited by the GNN mechanism that only passes messages among local nodes via fixed edges. To address the issue, we develop a cell graph transformer (CGT) that treats nodes and edges as input tokens to enable learnable adjacency and information exchange among all nodes. Nevertheless, training the transformer with a cell graph presents another challenge. Poorly initialized features can lead to noisy self-attention scores and inferior convergence, particularly when processing the cell graphs with numerous connections. Thus, we further propose a novel topology-aware pretraining method that leverages a graph convolutional network (GCN) to learn a feature extractor. The pre-trained features may suppress unreasonable correlations and hence ease the finetuning of CGT. Experimental results suggest that the proposed cell graph transformer with topology-aware pretraining significantly improves the nuclei classification results, and achieves the state-of-the-art performance. Code and models are available at https://github.com/lhaof/CGTComment: AAAI 2024, Code and models are available at https://github.com/lhaof/CG

    Skeleton-Guided Instance Separation for Fine-Grained Segmentation in Microscopy

    Full text link
    One of the fundamental challenges in microscopy (MS) image analysis is instance segmentation (IS), particularly when segmenting cluster regions where multiple objects of varying sizes and shapes may be connected or even overlapped in arbitrary orientations. Existing IS methods usually fail in handling such scenarios, as they rely on coarse instance representations such as keypoints and horizontal bounding boxes (h-bboxes). In this paper, we propose a novel one-stage framework named A2B-IS to address this challenge and enhance the accuracy of IS in MS images. Our approach represents each instance with a pixel-level mask map and a rotated bounding box (r-bbox). Unlike two-stage methods that use box proposals for segmentations, our method decouples mask and box predictions, enabling simultaneous processing to streamline the model pipeline. Additionally, we introduce a Gaussian skeleton map to aid the IS task in two key ways: (1) It guides anchor placement, reducing computational costs while improving the model's capacity to learn RoI-aware features by filtering out noise from background regions. (2) It ensures accurate isolation of densely packed instances by rectifying erroneous box predictions near instance boundaries. To further enhance the performance, we integrate two modules into the framework: (1) An Atrous Attention Block (A2B) designed to extract high-resolution feature maps with fine-grained multiscale information, and (2) A Semi-Supervised Learning (SSL) strategy that leverages both labeled and unlabeled images for model training. Our method has been thoroughly validated on two large-scale MS datasets, demonstrating its superiority over most state-of-the-art approaches

    Point-supervised Single-cell Segmentation via Collaborative Knowledge Sharing

    Full text link
    Despite their superior performance, deep-learning methods often suffer from the disadvantage of needing large-scale well-annotated training data. In response, recent literature has seen a proliferation of efforts aimed at reducing the annotation burden. This paper focuses on a weakly-supervised training setting for single-cell segmentation models, where the only available training label is the rough locations of individual cells. The specific problem is of practical interest due to the widely available nuclei counter-stain data in biomedical literature, from which the cell locations can be derived programmatically. Of more general interest is a proposed self-learning method called collaborative knowledge sharing, which is related to but distinct from the more well-known consistency learning methods. This strategy achieves self-learning by sharing knowledge between a principal model and a very light-weight collaborator model. Importantly, the two models are entirely different in their architectures, capacities, and model outputs: In our case, the principal model approaches the segmentation problem from an object-detection perspective, whereas the collaborator model a sematic segmentation perspective. We assessed the effectiveness of this strategy by conducting experiments on LIVECell, a large single-cell segmentation dataset of bright-field images, and on A431 dataset, a fluorescence image dataset in which the location labels are generated automatically from nuclei counter-stain data. Implementing code is available at https://github.com/jiyuuchc/lacss_ja
    • โ€ฆ
    corecore