31 research outputs found

    Dual-attention Focused Module for Weakly Supervised Object Localization

    Get PDF
    The research on recognizing the most discriminative regions provides referential information for weakly supervised object localization with only image-level annotations. However, the most discriminative regions usually conceal the other parts of the object, thereby impeding entire object recognition and localization. To tackle this problem, the Dual-attention Focused Module (DFM) is proposed to enhance object localization performance. Specifically, we present a dual attention module for information fusion, consisting of a position branch and a channel one. In each branch, the input feature map is deduced into an enhancement map and a mask map, thereby highlighting the most discriminative parts or hiding them. For the position mask map, we introduce a focused matrix to enhance it, which utilizes the principle that the pixels of an object are continuous. Between these two branches, the enhancement map is integrated with the mask map, aiming at partially compensating the lost information and diversifies the features. With the dual-attention module and focused matrix, the entire object region could be precisely recognized with implicit information. We demonstrate outperforming results of DFM in experiments. In particular, DFM achieves state-of-the-art performance in localization accuracy in ILSVRC 2016 and CUB-200-2011.Comment: 8 pages, 6 figures and 4 table

    Optic Cup Segmentation Using Large Pixel Patch Based CNNs

    Full text link
    Optic cup(OC) segmentation on color fundus image is essential for the calculation of cup-to-disk ratio and fundus morphological analysis, which are very important references in the diagnosis of glaucoma. In this paper we proposed an OC segmentation method using convolutional neural networks(CNNs) to learn from big size patch belong to each pixel. The segmentation result is achieved by classification of each pixel patch and postprocessing. With large pixel patch, the network could learn more global information around each pixel and make a better judgement during classification. We tested this method on public dataset Drishti-GS and achieved average F-Score of 93.73% and average overlapping error of 12.25%, which is better than state-of-the-art algorithms. This method could be used for fundus morphological analysis, and could also be employed to other medical image segmentation works which the boundary of the target area is fuzzy

    Weakly-Supervised Simultaneous Evidence Identification and Segmentation for Automated Glaucoma Diagnosis

    No full text
    Evidence identification, optic disc segmentation and automated glaucoma diagnosis are the most clinically significant tasks for clinicians to assess fundus images. However, delivering the three tasks simultaneously is extremely challenging due to the high variability of fundus structure and lack of datasets with complete annotations. In this paper, we propose an innovative Weakly-Supervised Multi-Task Learning method (WSMTL) for accurate evidence identification, optic disc segmentation and automated glaucoma diagnosis. The WSMTL method only uses weak-label data with binary diagnostic labels (normal/glaucoma) for training, while obtains pixel-level segmentation mask and diagnosis for testing. The WSMTL is constituted by a skip and densely connected CNN to capture multi-scale discriminative representation of fundus structure; a well-designed pyramid integration structure to generate high-resolution evidence map for evidence identification, in which the pixels with higher value represent higher confidence to highlight the abnormalities; a constrained clustering branch for optic disc segmentation; and a fully-connected discriminator for automated glaucoma diagnosis. Experimental results show that our proposed WSMTL effectively and simultaneously delivers evidence identification, optic disc segmentation (89.6% TP Dice), and accurate glaucoma diagnosis (92.4% AUC). This endows our WSMTL a great potential for the effective clinical assessment of glaucoma

    Hierarchical Contour Closure-Based Holistic Salient Object Detection

    No full text

    Saliency-based segmentation of optic disc in retinal images

    No full text
    Abstract Accurate segmentation of optic disc (OD) is significant for the automation of retinal analysis and retinal diseases screening. This paper proposes a novel optic disc segmentation method based on the saliency. It includes two stages:optic disc location and saliency-based segmentation. In the location stage, the OD is detected using a matched template and the density of the vessels. In the segmentation stage, we treat the OD as the salient object and formulate it as a saliency detection problem. To measure the saliency of a region, the boundary prior and the connectivity prior are exploited. Then geodesic distance to the window boundary is computed to measure the cost the region spends to reach the window boundary. After a threshold and ellipse fitting, we obtain the OD. Experimental results on two public databases for OD segmentation show that the proposed method achieves the- state-of-the-art performance

    Hierarchical contour closure-based holistic salient object detection

    No full text
    Abstract Most existing salient object detection methods compute the saliency for pixels, patches, or superpixels by contrast. Such fine-grained contrast-based salient object detection methods are stuck with saliency attenuation of the salient object and saliency overestimation of the background when the image is complicated. To better compute the saliency for complicated images, we propose a hierarchical contour closure-based holistic salient object detection method, in which two saliency cues, i.e., closure completeness and closure reliability, are thoroughly exploited. The former pops out the holistic homogeneous regions bounded by completely closed outer contours, and the latter highlights the holistic homogeneous regions bounded by averagely highly reliable outer contours. Accordingly, we propose two computational schemes to compute the corresponding saliency maps in a hierarchical segmentation space. Finally, we propose a framework to combine the two saliency maps, obtaining the final saliency map. Experimental results on three publicly available datasets show that even each single saliency map is able to reach the state-of-the-art performance. Furthermore, our framework, which combines two saliency maps, outperforms the state of the arts. Additionally, we show that the proposed framework can be easily used to extend existing methods and further improve their performances substantially

    A spatial-aware joint optic disc and cup segmentation method

    No full text
    Abstract When dealing with the optic disc and cup in the optical nerve head images, their joint segmentation confronts two critical problems. One is that the spatial layout of the vessels in the optic nerve head images is variant. The other is that the landmarks for the optic cup boundaries are spatially sparse and at small spatial scale. To solve these two problems, we propose a spatial-aware joint segmentation method by explicitly considering the spatial locations of the pixels and learning the multi-scale spatially dense features. We formulate the joint segmentation task from a probabilistic perspective, and derive a spatial-aware maximum conditional probability framework and the corresponding error function. Accordingly, we provide an end-to-end solution by designing a spatial-aware neural network. It consists of three modules: the atrous CNN module to extract the spatially dense features, the pyramid filtering module to produce the spatial-aware multi-scale features, and the spatial-aware segmentation module to predict the labels of pixels. We validate the state-of-the-art performances of our spatial-aware segmentation method on two public datasets, i.e., ORIGA and DRISHTI. Based on the segmentation masks, we quantify the cup-to-disk values and apply them to the glaucoma screening. High correlation between the cup-to-disk values and the risks of the glaucoma is validated on the dataset ORIGA
    corecore