23 research outputs found

    Extraction of Text from Optic Nerve Optical Coherence Tomography Reports

    Full text link
    Purpose: The purpose of this study was to develop and evaluate rule-based algorithms to enhance the extraction of text data, including retinal nerve fiber layer (RNFL) values and other ganglion cell count (GCC) data, from Zeiss Cirrus optical coherence tomography (OCT) scan reports. Methods: DICOM files that contained encapsulated PDF reports with RNFL or Ganglion Cell in their document titles were identified from a clinical imaging repository at a single academic ophthalmic center. PDF reports were then converted into image files and processed using the PaddleOCR Python package for optical character recognition. Rule-based algorithms were designed and iteratively optimized for improved performance in extracting RNFL and GCC data. Evaluation of the algorithms was conducted through manual review of a set of RNFL and GCC reports. Results: The developed algorithms demonstrated high precision in extracting data from both RNFL and GCC scans. Precision was slightly better for the right eye in RNFL extraction (OD: 0.9803 vs. OS: 0.9046), and for the left eye in GCC extraction (OD: 0.9567 vs. OS: 0.9677). Some values presented more challenges in extraction, particularly clock hours 5 and 6 for RNFL thickness, and signal strength for GCC. Conclusions: A customized optical character recognition algorithm can identify numeric results from optical coherence scan reports with high precision. Automated processing of PDF reports can greatly reduce the time to extract OCT results on a large scale

    Nitrogen/sulfur dual-doping of reduced graphene oxide harvesting hollow ZnSnS3 nano-microcubes with superior sodium storage

    Get PDF
    Bimetallic sulfides have exhibited promising applications in advanced sodium-ion batteries (SIBs) due to their relatively high electronic conductivity and electrochemical activity. In this study, for the first time, the N/S dual-doped reduced graphene oxide (rGO) encapsulating hollow ZnSnS3 nano-microcubes (N/S-rGO@ZnSnS3) is designed to improve the sluggish reaction kinetics, poor cycling stability and unsatisfactory rate capability of metal sulfides. To examine this design, the cycling stability and rate capability of the desired anode material is studied in detail. It is found that N/S-rGO@ZnSnS3 hybrid delivers a high discharge capacity of 501.7 mAh g−1 after 100 cycles at 0.1 A g−1, and a reversible capacity of 290.7 mAh g−1 after 500 cycles at 1.0 A g−1 with a capacity fading of 0.06% per cycle. The cycling stability as well as rate capability of N/S-rGO@ZnSnS3 are superior to those of the pristine hollow ZnSnS3 cubes/un-doped rGO composite. It is convinced that the electrode performance is strongly rooted in its structural conformation. Furthermore, the structural evolutions of ZnSnS3 reactions with sodium are revealed by in situ X-ray diffraction combined with ex situ X-ray photoelectron spectroscopy, which provides a valuable revelation for the understanding of reaction mechanism toward bimetallic sulfides and beyond

    Laser-Visible Face Image Translation and Recognition Based on CycleGAN and Spectral Normalization

    No full text
    The range-gated laser imaging instrument can capture face images in a dark environment, which provides a new idea for long-distance face recognition at night. However, the laser image has low contrast, low SNR and no color information, which affects observation and recognition. Therefore, it becomes important to convert laser images into visible images and then identify them. For image translation, we propose a laser-visible face image translation model combined with spectral normalization (SN-CycleGAN). We add spectral normalization layers to the discriminator to solve the problem of low image translation quality caused by the difficulty of training the generative adversarial network. The content reconstruction loss function based on the Y channel is added to reduce the error mapping. The face generated by the improved model on the self-built laser-visible face image dataset has better visual quality, which reduces the error mapping and basically retains the structural features of the target compared with other models. The FID value of evaluation index is 36.845, which is 16.902, 13.781, 10.056, 57.722, 62.598 and 0.761 lower than the CycleGAN, Pix2Pix, UNIT, UGATIT, StarGAN and DCLGAN models, respectively. For the face recognition of translated images, we propose a laser-visible face recognition model based on feature retention. The shallow feature maps with identity information are directly connected to the decoder to solve the problem of identity information loss in network transmission. The domain loss function based on triplet loss is added to constrain the style between domains. We use pre-trained FaceNet to recognize generated visible face images and obtain the recognition accuracy of Rank-1. The recognition accuracy of the images generated by the improved model reaches 76.9%, which is greatly improved compared with the above models and 19.2% higher than that of laser face recognition

    A Lightweight Object Detection Algorithm for Remote Sensing Images Based on Attention Mechanism and YOLOv5s

    No full text
    The specific characteristics of remote sensing images, such as large directional variations, large target sizes, and dense target distributions, make target detection a challenging task. To improve the detection performance of models while ensuring real-time detection, this paper proposes a lightweight object detection algorithm based on an attention mechanism and YOLOv5s. Firstly, a depthwise-decoupled head (DD-head) module and spatial pyramid pooling cross-stage partial GSConv (SPPCSPG) module were constructed to replace the coupled head and the spatial pyramid pooling-fast (SPPF) module of YOLOv5s. A shuffle attention (SA) mechanism was introduced in the head structure to enhance spatial attention and reconstruct channel attention. A content-aware reassembly of features (CARAFE) module was introduced in the up-sampling operation to reassemble feature points with similar semantic information. In the neck structure, a GSConv module was introduced to maintain detection accuracy while reducing the number of parameters. Experimental results on remote sensing datasets, RSOD and DIOR, showed an improvement of 1.4% and 1.2% in mean average precision accuracy compared with the original YOLOv5s algorithm. Moreover, the algorithm was also tested on conventional object detection datasets, PASCAL VOC and MS COCO, which showed an improvement of 1.4% and 3.1% in mean average precision accuracy. Therefore, the experiments showed that the constructed algorithm not only outperformed the original network on remote sensing images but also performed better than the original network on conventional object detection images

    Real-world agreement of same-visit Tono-Pen vs Goldmann applanation intraocular pressure measurements using electronic health records

    No full text
    Purpose: To compare intraocular pressure (IOP) obtained with Tono-Pen (TP) and Goldmann applanation (GAT) using large-scale electronic health records (EHR). Design: Retrospective cohort study. Methods: A single pair of eligible TP/GAT IOP readings was randomly selected from the EHR for each ophthalmology patient at an academic ophthalmology center (2013–2022), yielding 4550 eligible measurements. We used Bland-Altman analysis to describe agreement between TP/GAT IOP differences and mean IOP measurements. We also used multivariable logistic regression to identify factors associated with different IOP readings in the same eye, including demographics, glaucoma diagnosis, and central corneal thickness (CCT). Primary outcome metrics were discrepant measurements between TP and GAT as defined by two methods: Outcome A (normal TP despite elevated GAT measurements), and Outcome B (TP and GAT IOP differences ≄6 mmHg). Result: The mean TP/GAT IOP difference was 0.15 mmHg ( ± 5.49 mmHg 95% CI). There was high correlation between the measurements (r = 0.790, p 16.5 mmHg (Fig. 4). Discrepant measurements accounted for 2.6% (N = 116) and 5.2% (N = 238) for outcomes A and B respectively. Patients with thinner CCT had higher odds of discrepant IOP (OR 0.88 per 25 Όm increase, CI [0.84–0.92], p < 0.0001; OR 0.88 per 25 Όm increase, CI [0.84–0.92], p < 0.0001 for outcomes A and B respectively). Conclusion: In a real-world academic practice setting, TP and GAT IOP measurements demonstrated close agreement, although 2.6% of measurements showed elevated GAT IOP despite normal TP measurements, and 5.2% of measurements were ≄6 mmHg apart

    Strawberry Maturity Recognition Algorithm Combining Dark Channel Enhancement and YOLOv5

    No full text
    Aiming at the problems of low accuracy of strawberry fruit picking and large rate of mispicking or missed picking, YOLOv5 combined with dark channel enhancement is proposed. In “Fengxiang” strawberry, the criterion of “bad fruit” is added to the conventional three criteria of ripeness, near-ripeness, and immaturity, because some of the bad fruits are close to the color of ripe fruits, but the fruits are small and dry. The training accuracy of the four kinds of strawberries with different ripeness is above 85%, and the testing accuracy is above 90%. Then, to meet the demand of all-day picking and address the problem of low illumination of images collected at night, an enhancement algorithm is proposed to enhance the images, which are recognized. We compare the actual detection results of the five enhancement algorithms, i.e., histogram equalization, Laplace transform, gamma transform, logarithmic variation, and dark channel enhancement processing under the different numbers of fruits, periods, and video tests. The results show that combined with dark channel enhancement, YOLOv5 has the highest recognition rate. Finally, the experimental results demonstrate that YOLOv5 is better than SSD, DSSD, and EfficientDet in terms of recognition accuracy, and the correct rate can reach more than 90%. Meanwhile, the method has good robustness in complex environments such as partial occlusion and multiple fruits

    Image Processing for Laser Imaging Using Adaptive Homomorphic Filtering and Total Variation

    No full text
    Laser active imaging technology has important practical value and broad application prospects in military fields such as target detection, radar reconnaissance, and precise guidance. However, factors such as uneven laser illuminance, atmospheric backscatter, and the imaging system itself will introduce noise, which will affect the quality of the laser active imaging image, resulting in image contrast decline and blurring image edges and details. Therefore, an image denoising algorithm based on homomorphic filtering and total variation cascade is proposed in this paper, which strives to reduce the noise while retaining the edge features of the image to the maximum extent. Firstly, the image type is determined according to the characteristics of the laser image, and then the speckle noise in the low-frequency region is suppressed by adaptive homomorphic filtering. Finally, the image denoising method of minimizing the total variation is adopted for the impulse noise and Gaussian noise. Experimental results show that compared with separate homomorphic filtering, total variation filtering, and median filtering, the proposed algorithm significantly improves the contrast, retains edge details, achieves the expected effect. It can better adjust the image brightness and is beneficial for subsequent processing
    corecore