22 research outputs found

    3D Automatic Segmentation Method for Retinal Optical Coherence Tomography Volume Data Using Boundary Surface Enhancement

    Full text link
    With the introduction of spectral-domain optical coherence tomography (SDOCT), much larger image datasets are routinely acquired compared to what was possible using the previous generation of time-domain OCT. Thus, there is a critical need for the development of 3D segmentation methods for processing these data. We present here a novel 3D automatic segmentation method for retinal OCT volume data. Briefly, to segment a boundary surface, two OCT volume datasets are obtained by using a 3D smoothing filter and a 3D differential filter. Their linear combination is then calculated to generate new volume data with an enhanced boundary surface, where pixel intensity, boundary position information, and intensity changes on both sides of the boundary surface are used simultaneously. Next, preliminary discrete boundary points are detected from the A-Scans of the volume data. Finally, surface smoothness constraints and a dynamic threshold are applied to obtain a smoothed boundary surface by correcting a small number of error points. Our method can extract retinal layer boundary surfaces sequentially with a decreasing search region of volume data. We performed automatic segmentation on eight human OCT volume datasets acquired from a commercial Spectralis OCT system, where each volume of data consisted of 97 OCT images with a resolution of 496 512; experimental results show that this method can accurately segment seven layer boundary surfaces in normal as well as some abnormal eyes.Comment: 27 pages, 19 figure

    ViLTA: Enhancing Vision-Language Pre-training through Textual Augmentation

    Full text link
    Vision-language pre-training (VLP) methods are blossoming recently, and its crucial goal is to jointly learn visual and textual features via a transformer-based architecture, demonstrating promising improvements on a variety of vision-language tasks. Prior arts usually focus on how to align visual and textual features, but strategies for improving the robustness of model and speeding up model convergence are left insufficiently explored. In this paper, we propose a novel method ViLTA, comprising of two components to further facilitate the model to learn fine-grained representations among image-text pairs. For Masked Language Modeling (MLM), we propose a cross-distillation method to generate soft labels to enhance the robustness of model, which alleviates the problem of treating synonyms of masked words as negative samples in one-hot labels. For Image-Text Matching (ITM), we leverage the current language encoder to synthesize hard negatives based on the context of language input, encouraging the model to learn high-quality representations by increasing the difficulty of the ITM task. By leveraging the above techniques, our ViLTA can achieve better performance on various vision-language tasks. Extensive experiments on benchmark datasets demonstrate that the effectiveness of ViLTA and its promising potential for vision-language pre-training.Comment: 15 pages, 5 figure

    Comparison of the clinical characteristics and prognosis between clear cell carcinomas and high-grade serous ovarian carcinomas

    Get PDF
    Objectives: To compare the clinical characteristics and prognosis of women with clear cell versus high-grade serous ovarian carcinoma. Material and methods: Retrospective analysis of the clinical data of 50 cases patients with ovarian clear cell carcinoma (OCCC) and 103 cases with high-grade serous ovarian carcinoma (HGSOC), who were initially treated and completed standardized therapy in Affiliated Hospital of Qingdao University from January 2013 to December 2017. Results: There were significant differences in age, gravidity (G > 1), chief complaint, with ovarian endometriosis, tumor diameter, unilateral or bilateral, cystic and solid tumor, CA125, HE4, CA199, lactate dehydrogenase (LDH), and FIGO stage between the two groups. The differences in the prognosis between OCCC patients and HGSOC patients with early stage (FIGO I–II) were not statistically significant. The 5-year overall survival and progression-free survival of OCCC patients were significantly worse than those of HGSOC patients with advanced stage (FIGO III–IV) (p < 0.05). FIGO stage and non-R0 resection were independent risk factors affecting the prognosis of patients with ovarian clear cell carcinoma, screening by Cox regression analysis. FIGO stage, the lowest value of CA125, and non-R0 resection were independent risk factors affecting the prognosis of patients with high-grade serous ovarian cancer. Conclusions: The clinical characteristics and prognosis of OCCC are different from those of HGSOC. Ovarian clear cell carcinoma (OCCC) patients have a significantly worse prognosis than those with HGSOC in the advanced stage (FIGO Ⅲ–Ⅳ). Satisfactory tumor resection is an essential factor related to the prognosis of patients with OCCC and HGSOC

    MDAN-UNet: Multi-Scale and Dual Attention Enhanced Nested U-Net Architecture for Segmentation of Optical Coherence Tomography Images

    No full text
    Optical coherence tomography (OCT) is an optical high-resolution imaging technique for ophthalmic diagnosis. In this paper, we take advantages of multi-scale input, multi-scale side output and dual attention mechanism and present an enhanced nested U-Net architecture (MDAN-UNet), a new powerful fully convolutional network for automatic end-to-end segmentation of OCT images. We have evaluated two versions of MDAN-UNet (MDAN-UNet-16 and MDAN-UNet-32) on two publicly available benchmark datasets which are the Duke Diabetic Macular Edema (DME) dataset and the RETOUCH dataset, in comparison with other state-of-the-art segmentation methods. Our experiment demonstrates that MDAN-UNet-32 achieved the best performance, followed by MDAN-UNet-16 with smaller parameter, for multi-layer segmentation and multi-fluid segmentation respectively

    Intelligent Diagnosis Method for Rotating Machinery Using Dictionary Learning and Singular Value Decomposition

    No full text
    Rotating machinery is widely used in industrial applications. With the trend towards more precise and more critical operating conditions, mechanical failures may easily occur. Condition monitoring and fault diagnosis (CMFD) technology is an effective tool to enhance the reliability and security of rotating machinery. In this paper, an intelligent fault diagnosis method based on dictionary learning and singular value decomposition (SVD) is proposed. First, the dictionary learning scheme is capable of generating an adaptive dictionary whose atoms reveal the underlying structure of raw signals. Essentially, dictionary learning is employed as an adaptive feature extraction method regardless of any prior knowledge. Second, the singular value sequence of learned dictionary matrix is served to extract feature vector. Generally, since the vector is of high dimensionality, a simple and practical principal component analysis (PCA) is applied to reduce dimensionality. Finally, the K-nearest neighbor (KNN) algorithm is adopted for identification and classification of fault patterns automatically. Two experimental case studies are investigated to corroborate the effectiveness of the proposed method in intelligent diagnosis of rotating machinery faults. The comparison analysis validates that the dictionary learning-based matrix construction approach outperforms the mode decomposition-based methods in terms of capacity and adaptability for feature extraction

    Efficient Deep Learning-Based Automated Pathology Identification in Retinal Optical Coherence Tomography Images

    No full text
    We present an automatic method based on transfer learning for the identification of dry age-related macular degeneration (AMD) and diabetic macular edema (DME) from retinal optical coherence tomography (OCT) images. The algorithm aims to improve the classification performance of retinal OCT images and shorten the training time. Firstly, we remove the last several layers from the pre-trained Inception V3 model and regard the remaining part as a fixed feature extractor. Then, the features are used as input of a convolutional neural network (CNN) designed to learn the feature space shifts. The experimental results on two different retinal OCT images datasets demonstrate the effectiveness of the proposed method
    corecore