3,781 research outputs found

    A New Automatic Watercolour Painting Algorithm Based on Dual Stream Image Segmentation Model with Colour Space Estimation

    Get PDF
    Image processing plays a crucial role in automatic watercolor painting by manipulating the digital image to achieve the desired watercolor effect. segmentation in automatic watercolor painting algorithms is essential for region-based processing, color mixing and blending, capturing brushwork and texture, and providing artistic control over the final result. It allows for more realistic and expressive watercolor-like paintings by processing different image regions individually and applying appropriate effects to each segment. Hence, this paper proposed an effective Dual Stream Exception Maximization (DSEM) for automatic image segmentation. DSEM combines both color and texture information to segment an image into meaningful regions. This approach begins by converting the image from the RGB color space to a perceptually-based color space, such as CIELAB, to account for variations in lighting conditions and human perception of color.  With the color space conversion, DSEM extracts relevant features from the image. Color features are computed based on the values of the color channels in the chosen color space, capturing the nuances of color distribution within the image. Simultaneously, texture features are derived by computing statistical measures such as local variance or co-occurrence matrices, capturing the textural characteristics of the image. Finally, the model is applied over the deep learning model for the classification of the color space in the painting. Simulation analysis is performed compared with conventional segmentation techniques such a CNN and RNN. The comparative analysis states that the proposed DSEM exhibits superior performance compared to conventional techniques in terms of color space estimation, texture analysis and region merging. The performance of classification with DSEM is ~12% higher than the conventional techniques

    Automated Low-Cost Malaria Detection System in Thin Blood Slide Images Using Mobile Phones

    Get PDF
    Malaria, a deadly disease which according to the World Health Organisation (WHO) is responsible for the fatal illness in 200 million people around the world in 2010, is diagnosed using peripheral blood examination. The work undertaken in this research programme aims to develop an automated malaria parasite-detection system, using microscopic-image processing, that can be incorporated onto mobile phones. In this research study, the main objective is to achieve the performance equal to or better than the manual microscopy, which is the gold standard in malaria diagnosis, in order to produce a reliable automated diagnostic platform without expert intervention, for the effective treatment and eradication of the deadly disease. The work contributed to the field of mathematical morphology by proposing a novel method called the Annular Ring Ratio transform for blood component identification. It has also proposed an automated White Blood Cell and Red Blood Cell differentiation algorithm, which when combined with ARR transform method, has wide applications not only for malaria diagnosis but also for many blood related analysis involving microscopic examination. The research has undertaken investigations on infected cell identification which aids in the calculation of parasitemia, the measure of infection. In addition, an automated diagnostic tool to detect the sexual stage (gametocytes) of the species P.falciparum for post-treatment malaria diagnosis was developed. Furthermore, a parallel investigation was carried out on automated malaria diagnosis on fluorescent thin blood films and a WBC and infected cell differentiation algorithm was proposed. Finally, a mobile phone application based on the morphological image processing algorithms proposed in this thesis was developed. A complete malaria diagnostic unit using the mobile phones attached to a portable microscope was set up which has enormous potential not only for malaria diagnosis but also for the blood parasitological field where advancement in medical diagnostics using cellular smart phone technology is widely acknowledged

    PWD-3DNet: A deep learning-based fully-automated segmentation of multiple structures on temporal bone CT scans

    Get PDF
    The temporal bone is a part of the lateral skull surface that contains organs responsible for hearing and balance. Mastering surgery of the temporal bone is challenging because of this complex and microscopic three-dimensional anatomy. Segmentation of intra-temporal anatomy based on computed tomography (CT) images is necessary for applications such as surgical training and rehearsal, amongst others. However, temporal bone segmentation is challenging due to the similar intensities and complicated anatomical relationships among crit- ical structures, undetectable small structures on standard clinical CT, and the amount of time required for manual segmentation. This paper describes a single multi-class deep learning-based pipeline as the first fully automated algorithm for segmenting multiple temporal bone structures from CT volumes, including the sigmoid sinus, facial nerve, inner ear, malleus, incus, stapes, internal carotid artery and internal auditory canal. The proposed fully convolutional network, PWD-3DNet, is a patch-wise densely connected (PWD) three-dimensional (3D) network. The accuracy and speed of the proposed algorithm was shown to surpass current manual and semi-automated segmentation techniques. The experimental results yielded significantly high Dice similar- ity scores and low Hausdorff distances for all temporal bone structures with an average of 86% and 0.755 millimeter (mm), respectively. We illustrated that overlapping in the inference sub-volumes improves the segmentation performance. Moreover, we proposed augmentation layers by using samples with various transformations and image artefacts to increase the robustness of PWD-3DNet against image acquisition protocols, such as smoothing caused by soft tissue scanner settings and larger voxel sizes used for radiation reduction. The proposed algorithm was tested on low-resolution CTs acquired by another center with different scanner parameters than the ones used to create the algorithm and shows potential for application beyond the particular training data used in the study

    PWD-3DNet: A Deep Learning-Based Fully-Automated Segmentation of Multiple Structures on Temporal Bone CT Scans

    Get PDF
    The temporal bone is a part of the lateral skull surface that contains organs responsible for hearing and balance. Mastering surgery of the temporal bone is challenging because of this complex and microscopic three-dimensional anatomy. Segmentation of intra-temporal anatomy based on computed tomography (CT) images is necessary for applications such as surgical training and rehearsal, amongst others. However, temporal bone segmentation is challenging due to the similar intensities and complicated anatomical relationships among critical structures, undetectable small structures on standard clinical CT, and the amount of time required for manual segmentation. This paper describes a single multi-class deep learning-based pipeline as the first fully automated algorithm for segmenting multiple temporal bone structures from CT volumes, including the sigmoid sinus, facial nerve, inner ear, malleus, incus, stapes, internal carotid artery and internal auditory canal. The proposed fully convolutional network, PWD-3DNet, is a patch-wise densely connected (PWD) three-dimensional (3D) network. The accuracy and speed of the proposed algorithm was shown to surpass current manual and semi-automated segmentation techniques. The experimental results yielded significantly high Dice similarity scores and low Hausdorff distances for all temporal bone structures with an average of 86% and 0.755 millimeter (mm), respectively. We illustrated that overlapping in the inference sub-volumes improves the segmentation performance. Moreover, we proposed augmentation layers by using samples with various transformations and image artefacts to increase the robustness of PWD-3DNet against image acquisition protocols, such as smoothing caused by soft tissue scanner settings and larger voxel sizes used for radiation reduction. The proposed algorithm was tested on low-resolution CTs acquired by another center with different scanner parameters than the ones used to create the algorithm and shows potential for application beyond the particular training data used in the study

    Computationally Efficient Algorithm for Detecting Moving Objects with Moving Background

    Get PDF
    Abstract— The area of moving object detection has been a constant topic of research in more than a decade, where a research community have witnessed various significant contribution in the past that mitigates the problem of real-time and moving object detection. In our prior studies, we have addressed such issues using various sophisticated technique yielding superior results. But, it is felt that some light weight algorithm is required for the purpose of performing moving object detection with complete retention of object detection accuracy. This paper have presented a very simple algorithm that uses visual descriptor for extracting the dynamic features during fast transition of frames. The proposed algorithm is tested with one of the most significant work done recently on same purpose with respect to precision and recall rate along with analysis of processing time of proposed algorithm

    ShadowSense: Unsupervised Domain Adaptation and Feature Fusion for Shadow-Agnostic Tree Crown Detection from RGB-Thermal Drone Imagery

    Full text link
    Accurate detection of individual tree crowns from remote sensing data poses a significant challenge due to the dense nature of forest canopy and the presence of diverse environmental variations, e.g., overlapping canopies, occlusions, and varying lighting conditions. Additionally, the lack of data for training robust models adds another limitation in effectively studying complex forest conditions. This paper presents a novel method for detecting shadowed tree crowns and provides a challenging dataset comprising roughly 50k paired RGB-thermal images to facilitate future research for illumination-invariant detection. The proposed method (ShadowSense) is entirely self-supervised, leveraging domain adversarial training without source domain annotations for feature extraction and foreground feature alignment for feature pyramid networks to adapt domain-invariant representations by focusing on visible foreground regions, respectively. It then fuses complementary information of both modalities to effectively improve upon the predictions of an RGB-trained detector and boost the overall accuracy. Extensive experiments demonstrate the superiority of the proposed method over both the baseline RGB-trained detector and state-of-the-art techniques that rely on unsupervised domain adaptation or early image fusion. Our code and data are available: https://github.com/rudrakshkapil/ShadowSenseComment: Accepted in IEEE/CVF Winter Applications of Computer Vision (WACV) 2024 main conference! 8 pages (11 with bibliography), 5 figures, 3 table

    Variational and deep learning segmentation of very-low-contrast X-ray computed tomography images of carbon/epoxy woven composites

    Get PDF
    The purpose of this work is to find an effective image segmentation method for lab-based micro-tomography (mu-CT) data of carbon fiber reinforced polymers (CFRP) with insufficient contrast-to-noise ratio. The segmentation is the first step in creating a realistic geometry (based on mu-CT) for finite element modelling of textile composites on meso-scale. Noise in X-ray imaging data of carbon/polymer composites forms a challenge for this segmentation due to the very low X-ray contrast between fiber and polymer and unclear fiber gradients. To the best of our knowledge, segmentation of mu-CT images of carbon/polymer textile composites with low resolution data (voxel size close to the fiber diameter) remains poorly documented. In this paper, we propose and evaluate different approaches for solving the segmentation problem: variational on the one hand and deep-learning-based on the other. In the author's view, both strategies present a novel and reliable ground for the segmentation of mu-CT data of CFRP woven composites. The predictions of both approaches were evaluated against a manual segmentation of the volume, constituting our "ground truth", which provides quantitative data on the segmentation accuracy. The highest segmentation accuracy (about 4.7% in terms of voxel-wise Dice similarity) was achieved using the deep learning approach with U-Net neural network
    corecore