32 research outputs found

    Accelerating magnetic induction tomography‐based imaging through heterogeneous parallel computing

    Get PDF
    Magnetic Induction Tomography (MIT) is a non‐invasive imaging technique, which has applications in both industrial and clinical settings. In essence, it is capable of reconstructing the electromagnetic parameters of an object from measurements made on its surface. With the exploitation of parallelism, it is possible to achieve high quality inexpensive MIT images for biomedical applications on clinically relevant time scales. In this paper we investigate the performance of different parallel implementations of the forward eddy current problem, which is the main computational component of the inverse problem through which measured voltages are converted into images. We show that a heterogeneous parallel method that exploits multiple CPUs and GPUs can provide a high level of parallel scaling, leading to considerably improved runtimes. We also show how multiple GPUs can be used in conjunction with deal.II, a widely‐used open source finite element library

    Energy Efficiency Improvements in Dry Drilling with Optimised Diamond-Like Carbon Coating

    Get PDF
    We demonstrate enhancements of performance and energy efficiency of cutting tools by deposition of diamond-like carbon (DLC) coatings on machine parts. DLC was deposited on steel drill bits, using plasma enhanced chemical vapour deposition (PECVD) with the acetylene precursor diluted with argon, to produce a surface with low friction and low wear rate. Drill bit performance in dry drilling of aluminium was quantified by analysis of power consumption and swarf flow. Optimised deposition conditions produced drill bits with greatly enhanced performance over uncoated drill bits, showing a 25% reduction in swarf clogging, a 36% reduction in power consumption and a greater than five-fold increase in lifetime. Surface analysis with scanning electron microscopy shows that DLC coated drills exhibit much lower aluminium build up on the trailing shank of the drill, enhancing the anti-adhering properties of the drill and reducing heat generation during operation, resulting in the observed improvements in efficiency. Variation of drilling efficiency with argon dilution of precursor is related to changes in the microstructure of the DLC coating

    A critical review of digital technology innovations for early warning of water-related disease outbreaks associated with climatic hazards

    Get PDF
    Water-related climatic disasters pose a significant threat to human health due to the potential of disease outbreaks, which are exacerbated by climate change. Therefore, it is crucial to predict their occurrence with sufficient lead time to allow for contingency plans to reduce risks to the population. Opportunities to address this challenge can be found in the rapid evolution of digital technologies. This study conducted a critical analysis of recent publications investigating advanced technologies and digital innovations for forecasting, alerting, and responding to water-related extreme events, particularly flooding, which is often linked to disaster-related disease outbreaks. The results indicate that certain digital innovations, such as portable and local sensors integrated with web-based platforms are new era for predicting events, developing control strategies and establishing early warning systems. Other technologies, such as augmented reality, virtual reality, and social media, can be more effective for monitoring flood spread, disseminating before/during the event information, and issuing warnings or directing emergency responses. The study also identified that the collection and translation of reliable data into information can be a major challenge for effective early warning systems and the adoption of digital innovations in disaster management. Augmented reality, and digital twin technologies should be further explored as valuable tools for better providing of communicating complex information on disaster development and response strategies to a wider range of audiences, particularly non-experts. This can help to increase community engagement in designing and operating effective early warning systems that can reduce the health impact of climatic disasters

    Automated Assessment of Image Quality in 2D Echocardiography Using Deep Learning

    Get PDF
    Echocardiography is the most used modality for assessing cardiac functions. The reliability of the echocardiographic measurements, however, depends on the quality of the images. Currently, the method of image quality assessment is a subjective process, where an echocardiography specialist visually inspects the images. An automated image quality assessment system is thus required. Here, we have reported on the feasibility of using deep learning for developing such automated quality scoring systems. A scoring system was proposed to include specific quality attributes for on-axis, contrast/gain and left ventricular (LV) foreshortening of the apical view. We prepared and used 1,039 echocardiographic patient datasets for model development and testing. Average accuracy of at least 86% was obtained with computation speed at 0.013ms per frame which indicated the feasibility for real-time deployment

    Doppler assessment of aortic stenosis: a 25-operator study demonstrating why reading the peak velocity is superior to velocity time integral

    Get PDF
    Aims Measurements with superior reproducibility are useful clinically and research purposes. Previous reproducibility studies of Doppler assessment of aortic stenosis (AS) have compared only a pair of observers and have not explored the mechanism by which disagreement between operators occurs. Using custom-designed software which stored operators’ traces, we investigated the reproducibility of peak and velocity time integral (VTI) measurements across a much larger group of operators and explored the mechanisms by which disagreement arose. Methods and results Twenty-five observers reviewed continuous wave (CW) aortic valve (AV) and pulsed wave (PW) left ventricular outflow tract (LVOT) Doppler traces from 20 sequential cases of AS in random order. Each operator unknowingly measured each peak velocity and VTI twice. VTI tracings were stored for comparison. Measuring the peak is much more reproducible than VTI for both PW (coefficient of variation 10.1 vs. 18.0%; P < 0.001) and CW traces (coefficient of variation 4.0 vs. 10.2%; P < 0.001). VTI is inferior because the steep early and late parts of the envelope are difficult to trace reproducibly. Dimensionless index improves reproducibility because operators tended to consistently over-read or under-read on LVOT and AV traces from the same patient (coefficient of variation 9.3 vs. 17.1%; P < 0.001). Conclusion It is far more reproducible to measure the peak of a Doppler trace than the VTI, a strategy that reduces measurement variance by approximately six-fold. Peak measurements are superior to VTI because tracing the steep slopes in the early and late part of the VTI envelope is difficult to achieve reproducibly

    Automated speckle tracking algorithm to aid on-axis imaging in echocardiography

    Get PDF
    Obtaining a “correct” view in echocardiography is a subjective process in which an operator attempts to obtain images conforming to consensus standard views. Real-time objective quantification of image alignment may assist less experienced operators, but no reliable index yet exists. We present a fully automated algorithm for detecting incorrect medial/lateral translation of an ultrasound probe by image analysis. The ability of the algorithm to distinguish optimal from sub-optimal four-chamber images was compared to that of specialists—the current “gold-standard.” The orientation assessments produced by the automated algorithm correlated well with consensus visual assessments of the specialists (r=0.87r=0.87) and compared favourably with the correlation between individual specialists and the consensus, 0.82±0.09. Each individual specialist’s assessments were within the consensus of other specialists, 75±14% of the time, and the algorithm’s assessments were within the consensus of specialists 85% of the time. The mean discrepancy in probe translation values between individual specialists and their consensus was 0.97±0.87  cm, and between the automated algorithm and specialists’ consensus was 0.92±0.70  cm. This technology could be incorporated into hardware to provide real-time guidance for image optimisation—a potentially valuable tool both for training and quality control

    Frame rate required for speckle tracking echocardiography: A quantitative clinical study with open-source, vendor-independent software

    Get PDF
    Background Assessing left ventricular function with speckle tracking is useful in patient diagnosis but requires a temporal resolution that can follow myocardial motion. In this study we investigated the effect of different frame rates on the accuracy of speckle tracking results, highlighting the temporal resolution where reliable results can be obtained. Material and methods 27 patients were scanned at two different frame rates at their resting heart rate. From all acquired loops, lower temporal resolution image sequences were generated by dropping frames, decreasing the frame rate by up to 10-fold. Results Tissue velocities were estimated by automated speckle tracking. Above 40 frames/s the peak velocity was reliably measured. When frame rate was lower, the inter-frame interval containing the instant of highest velocity also contained lower velocities, and therefore the average velocity in that interval was an underestimate of the clinically desired instantaneous maximum velocity. Conclusions The higher the frame rate, the more accurately maximum velocities are identified by speckle tracking, until the frame rate drops below 40 frames/s, beyond which there is little increase in peak velocity. We provide in an online supplement the vendor-independent software we used for automatic speckle-tracked velocity assessment to help others working in this field

    Automated aortic Doppler flow tracing for reproducible research and clinical measurements

    Get PDF
    In clinical practice, echocardiographers are often unkeen to make the significant time investment to make additional multiple measurements of Doppler velocity. Main hurdle to obtaining multiple measurements is the time required to manually trace a series of Doppler traces. To make it easier to analyze more beats, we present the description of an application system for automated aortic Doppler envelope quantification, compatible with a range of hardware platforms. It analyses long Doppler strips, spanning many heartbeats, and does not require electrocardiogram to separate individual beats. We tested its measurement of velocity-time-integral and peak-velocity against the reference standard defined as the average of three experts who each made three separate measurements. The automated measurements of velocity-time-integral showed strong correspondence (R2 = 0.94) and good Bland-Altman agreement (SD = 1.39 cm) with the reference consensus expert values, and indeed performed as well as the individual experts ( R2 = 0.90 to 0.96, SD = 1.05 to 1.53 cm). The same performance was observed for peak-velocities; ( R2 = 0.98, SD = 3.07 cm/s) and ( R2 = 0.93 to 0.98, SD = 2.96 to 5.18 cm/s). This automated technology allows > 10 times as many beats to be analyzed compared to the conventional manual approach. This would make clinical and research protocols more precise for the same operator effort

    Open-source, vendor-independent, automated multi-beat tissue Doppler echocardiography analysis

    Get PDF
    Current guidelines for measuring cardiac function by tissue Doppler recommend using multiple beats, but this has a time cost for human operators. We present an open-source, vendor-independent, drag-and-drop software capable of automating the measurement process. A database of ~8000 tissue Doppler beats (48 patients) from the septal and lateral annuli were analyzed by three expert echocardiographers. We developed an intensity- and gradient-based automated algorithm to measure tissue Doppler velocities. We tested its performance against manual measurements from the expert human operators. Our algorithm showed strong agreement with expert human operators. Performance was indistinguishable from a human operator: for algorithm, mean difference and SDD from the mean of human operators’ estimates 0.48 ± 1.12 cm/s (R2= 0.82); for the humans individually this was 0.43 ± 1.11 cm/s (R2= 0.84), −0.88 ± 1.12 cm/s (R2= 0.84) and 0.41 ± 1.30 cm/s (R2= 0.78). Agreement between operators and the automated algorithm was preserved when measuring at either the edge or middle of the trace. The algorithm was 10-fold quicker than manual measurements (p < 0.001). This open-source, vendor-independent, drag-and-drop software can make peak velocity measurements from pulsed wave tissue Doppler traces as accurately as human experts. This automation permits rapid, bias-resistant multi-beat analysis from spectral tissue Doppler images.European Research Council and British Heart Foundatio
    corecore