24 research outputs found

    Real-Time Virtual Pathology Using Signal Analysis and Synthesis

    Get PDF
    This dissertation discusses the modeling and simulation (M& S) research in the area of real-time virtual pathology using signal analysis and synthesis. The goal of this research is to contribute to the research in the M&S area of generating simulated outputs of medical diagnostics tools to supplement training of medical students with human patient role players. To become clinically competent physicians, medical students must become skilled in the areas of doctor-patient communication, eliciting the patient\u27s history, and performing the physical exam. The use of Standardized Patients (SPs), individuals trained to realistically portray patients, has become common practice. SPs provide the medical student with a means to learn in a safe, realistic setting, while providing a way to reliably test students\u27 clinical skills. The range of clinical problems an SP can portray, however, is limited. SPs are usually healthy individuals with few or no abnormal physical findings. Some SPs have been trained to simulate physical abnormalities, such as breathing through one lung, voluntarily and increasing blood pressure. But, there are many abnormalities that SPs cannot simulate. The research encompassed developing methods and algorithms to be incorporated into the previous work of McKenzie, el al. [1]–[3] for simulating abnormal heart sounds in a Standardized Patient (SP), which may be utilized in a modified electronic stethoscope. The methods and algorithms are specific to the real-time modeling of human body sounds through modifying the sounds from a real person with various abnormalities. The main focus of the research involved applying methods from tempo and beat analysis of acoustic musical signals for heart signal analysis, specifically in detecting the heart rate and heartbeat locations. In addition, the research included an investigation and selection of an adaptive noise cancellation filtering method to separate heart sounds from lung sounds. A model was developed to use a heart/lung sound signal as input to efficiently and accurately separate heart sound and lung sound signals, characterize the heart sound signal when appropriate, replace the heart or lung sound signal with a reference pathology signal containing an abnormality such as a crackle or murmur, and then recombine the original heart or lung sound signal with the modified pathology signal for presentation to the student. After completion of the development of the model, the model was validated. The validation included both a qualitative assessment and a quantitative assessment. The qualitative assessment drew on the visual and auditory analysis of SMEs, and the quantitative assessment utilized simulated data to verify key portions of the model

    Effects of antiplatelet therapy on stroke risk by brain imaging features of intracerebral haemorrhage and cerebral small vessel diseases: subgroup analyses of the RESTART randomised, open-label trial

    Get PDF
    Background Findings from the RESTART trial suggest that starting antiplatelet therapy might reduce the risk of recurrent symptomatic intracerebral haemorrhage compared with avoiding antiplatelet therapy. Brain imaging features of intracerebral haemorrhage and cerebral small vessel diseases (such as cerebral microbleeds) are associated with greater risks of recurrent intracerebral haemorrhage. We did subgroup analyses of the RESTART trial to explore whether these brain imaging features modify the effects of antiplatelet therapy

    Circulating microRNAs in sera correlate with soluble biomarkers of immune activation but do not predict mortality in ART treated individuals with HIV-1 infection: A case control study

    Get PDF
    Introduction: The use of anti-retroviral therapy (ART) has dramatically reduced HIV-1 associated morbidity and mortality. However, HIV-1 infected individuals have increased rates of morbidity and mortality compared to the non-HIV-1 infected population and this appears to be related to end-organ diseases collectively referred to as Serious Non-AIDS Events (SNAEs). Circulating miRNAs are reported as promising biomarkers for a number of human disease conditions including those that constitute SNAEs. Our study sought to investigate the potential of selected miRNAs in predicting mortality in HIV-1 infected ART treated individuals. Materials and Methods: A set of miRNAs was chosen based on published associations with human disease conditions that constitute SNAEs. This case: control study compared 126 cases (individuals who died whilst on therapy), and 247 matched controls (individuals who remained alive). Cases and controls were ART treated participants of two pivotal HIV-1 trials. The relative abundance of each miRNA in serum was measured, by RTqPCR. Associations with mortality (all-cause, cardiovascular and malignancy) were assessed by logistic regression analysis. Correlations between miRNAs and CD4+ T cell count, hs-CRP, IL-6 and D-dimer were also assessed. Results: None of the selected miRNAs was associated with all-cause, cardiovascular or malignancy mortality. The levels of three miRNAs (miRs -21, -122 and -200a) correlated with IL-6 while miR-21 also correlated with D-dimer. Additionally, the abundance of miRs -31, -150 and -223, correlated with baseline CD4+ T cell count while the same three miRNAs plus miR- 145 correlated with nadir CD4+ T cell count. Discussion: No associations with mortality were found with any circulating miRNA studied. These results cast doubt onto the effectiveness of circulating miRNA as early predictors of mortality or the major underlying diseases that contribute to mortality in participants treated for HIV-1 infection

    Development and Validation of a Risk Score for Chronic Kidney Disease in HIV Infection Using Prospective Cohort Data from the D:A:D Study

    Get PDF
    Ristola M. on työryhmien DAD Study Grp ; Royal Free Hosp Clin Cohort ; INSIGHT Study Grp ; SMART Study Grp ; ESPRIT Study Grp jäsen.Background Chronic kidney disease (CKD) is a major health issue for HIV-positive individuals, associated with increased morbidity and mortality. Development and implementation of a risk score model for CKD would allow comparison of the risks and benefits of adding potentially nephrotoxic antiretrovirals to a treatment regimen and would identify those at greatest risk of CKD. The aims of this study were to develop a simple, externally validated, and widely applicable long-term risk score model for CKD in HIV-positive individuals that can guide decision making in clinical practice. Methods and Findings A total of 17,954 HIV-positive individuals from the Data Collection on Adverse Events of Anti-HIV Drugs (D:A:D) study with >= 3 estimated glomerular filtration rate (eGFR) values after 1 January 2004 were included. Baseline was defined as the first eGFR > 60 ml/min/1.73 m2 after 1 January 2004; individuals with exposure to tenofovir, atazanavir, atazanavir/ritonavir, lopinavir/ritonavir, other boosted protease inhibitors before baseline were excluded. CKD was defined as confirmed (>3 mo apart) eGFR In the D:A:D study, 641 individuals developed CKD during 103,185 person-years of follow-up (PYFU; incidence 6.2/1,000 PYFU, 95% CI 5.7-6.7; median follow-up 6.1 y, range 0.3-9.1 y). Older age, intravenous drug use, hepatitis C coinfection, lower baseline eGFR, female gender, lower CD4 count nadir, hypertension, diabetes, and cardiovascular disease (CVD) predicted CKD. The adjusted incidence rate ratios of these nine categorical variables were scaled and summed to create the risk score. The median risk score at baseline was -2 (interquartile range -4 to 2). There was a 1: 393 chance of developing CKD in the next 5 y in the low risk group (risk score = 5, 505 events), respectively. Number needed to harm (NNTH) at 5 y when starting unboosted atazanavir or lopinavir/ritonavir among those with a low risk score was 1,702 (95% CI 1,166-3,367); NNTH was 202 (95% CI 159-278) and 21 (95% CI 19-23), respectively, for those with a medium and high risk score. NNTH was 739 (95% CI 506-1462), 88 (95% CI 69-121), and 9 (95% CI 8-10) for those with a low, medium, and high risk score, respectively, starting tenofovir, atazanavir/ritonavir, or another boosted protease inhibitor. The Royal Free Hospital Clinic Cohort included 2,548 individuals, of whom 94 individuals developed CKD (3.7%) during 18,376 PYFU (median follow-up 7.4 y, range 0.3-12.7 y). Of 2,013 individuals included from the SMART/ESPRIT control arms, 32 individuals developed CKD (1.6%) during 8,452 PYFU (median follow-up 4.1 y, range 0.6-8.1 y). External validation showed that the risk score predicted well in these cohorts. Limitations of this study included limited data on race and no information on proteinuria. Conclusions Both traditional and HIV-related risk factors were predictive of CKD. These factors were used to develop a risk score for CKD in HIV infection, externally validated, that has direct clinical relevance for patients and clinicians to weigh the benefits of certain antiretrovirals against the risk of CKD and to identify those at greatest risk of CKD.Peer reviewe

    Proceedings of the Virtual 3rd UK Implementation Science Research Conference : Virtual conference. 16 and 17 July 2020.

    Get PDF

    Effect of angiotensin-converting enzyme inhibitor and angiotensin receptor blocker initiation on organ support-free days in patients hospitalized with COVID-19

    Get PDF
    IMPORTANCE Overactivation of the renin-angiotensin system (RAS) may contribute to poor clinical outcomes in patients with COVID-19. Objective To determine whether angiotensin-converting enzyme (ACE) inhibitor or angiotensin receptor blocker (ARB) initiation improves outcomes in patients hospitalized for COVID-19. DESIGN, SETTING, AND PARTICIPANTS In an ongoing, adaptive platform randomized clinical trial, 721 critically ill and 58 non–critically ill hospitalized adults were randomized to receive an RAS inhibitor or control between March 16, 2021, and February 25, 2022, at 69 sites in 7 countries (final follow-up on June 1, 2022). INTERVENTIONS Patients were randomized to receive open-label initiation of an ACE inhibitor (n = 257), ARB (n = 248), ARB in combination with DMX-200 (a chemokine receptor-2 inhibitor; n = 10), or no RAS inhibitor (control; n = 264) for up to 10 days. MAIN OUTCOMES AND MEASURES The primary outcome was organ support–free days, a composite of hospital survival and days alive without cardiovascular or respiratory organ support through 21 days. The primary analysis was a bayesian cumulative logistic model. Odds ratios (ORs) greater than 1 represent improved outcomes. RESULTS On February 25, 2022, enrollment was discontinued due to safety concerns. Among 679 critically ill patients with available primary outcome data, the median age was 56 years and 239 participants (35.2%) were women. Median (IQR) organ support–free days among critically ill patients was 10 (–1 to 16) in the ACE inhibitor group (n = 231), 8 (–1 to 17) in the ARB group (n = 217), and 12 (0 to 17) in the control group (n = 231) (median adjusted odds ratios of 0.77 [95% bayesian credible interval, 0.58-1.06] for improvement for ACE inhibitor and 0.76 [95% credible interval, 0.56-1.05] for ARB compared with control). The posterior probabilities that ACE inhibitors and ARBs worsened organ support–free days compared with control were 94.9% and 95.4%, respectively. Hospital survival occurred in 166 of 231 critically ill participants (71.9%) in the ACE inhibitor group, 152 of 217 (70.0%) in the ARB group, and 182 of 231 (78.8%) in the control group (posterior probabilities that ACE inhibitor and ARB worsened hospital survival compared with control were 95.3% and 98.1%, respectively). CONCLUSIONS AND RELEVANCE In this trial, among critically ill adults with COVID-19, initiation of an ACE inhibitor or ARB did not improve, and likely worsened, clinical outcomes. TRIAL REGISTRATION ClinicalTrials.gov Identifier: NCT0273570

    Contemporary Profile of Seizures in Neonates: A Prospective Cohort Study

    No full text
    OBJECTIVE: To determine the contemporary etiology, burden, and short-term outcomes of seizures in neonates monitored with continuous video-electroencephalogram (cEEG). STUDY DESIGN: We prospectively collected data from 426 consecutive neonates (56% male, 88% term) ≤44 weeks postmenstrual age with clinically suspected seizures and/or electrographic seizures. Subjects were assessed between January 2013 and April 2015 at seven U.S. tertiary care pediatric centers following American Clinical Neurophysiology Society (ACNS) guidelines for cEEG for at risk neonates. Seizure etiology, burden, management and outcome were determined by chart review using a case report form designed at study onset. RESULTS: The most common seizure etiologies were hypoxic-ischemic encephalopathy (38%), ischemic stroke (18%), and intracranial hemorrhage (11%). Seizure burden was high, with 59% having ≥7 electrographic seizures and 16% having status epilepticus; 52% received ≥2 anti-seizure medications. During the neonatal admission, 17% died; 49% of survivors had abnormal neurological examination at hospital discharge. In an adjusted analysis, high seizure burden was a significant risk factor for mortality, length of hospital stay, and abnormal neurological examination at discharge. CONCLUSIONS: In this large contemporary profile of consecutively enrolled newborns with seizures treated at centers using cEEG per ACNS guidelines, about half had high seizure burden, received ≥2 anti-seizure medications, and/or died or had abnormal examination at discharge. Higher seizure burden was associated with increased morbidity and mortality. These findings underscore the importance of accurate determination of neonatal seizure frequency and etiology, and a potential for improved outcome if seizure burden is reduced
    corecore