36 research outputs found

    Personalised approaches to antithrombotic therapies: insights from linked electronic health records

    Get PDF
    Antithrombotic drugs are increasingly used for the prevention of atherothrombotic events in cardiovascular diseases and represent a paradigm for the study of personalised medicine because of the need to balance potential benefits with the substantial risks of bleeding harms. To be effective, personalised medicine needs validated prognostic risk models, rich phenotypes, and patient monitoring over time. The opportunity to use linked electronic health records has potential advantages; we have rich longitudinal data spanning patients’ entire journey through the healthcare system including primary care visits, clinical biomarkers, hospital admissions, hospital procedures and prescribed medication. Challenges include structuring the data into research-ready format and accurately defining clinical endpoints and handling missing data. The data used in this thesis was from the CALIBER platform: linked routinely-collected electronic health records from general practices, hospitals admissions, myocardial infarction registry and death registry for 2 million patients in England from 1997 to 2010. In this thesis I (1) developed comprehensive bleeding phenotypes in linked electronic health records, (2) assessed the incidence and prognosis of bleeding in atrial fibrillation and coronary disease patients in England, (3) developed and validated prognostic models for atherothrombotic and bleeding events in stable myocardial infarction survivors pertaining to the benefits and harms of prolonged dual antiplatelet therapy, (4) assessed the predictors and outcomes associated with time in therapeutic range for patients treated with oral anticoagulants (5) assessed the predictive value of novel measures of international normalised ratio control in patients treated with oral anticoagulants for atherothrombotic and bleeding outcomes. Taken together these findings offer researchers scalable methodological approaches, that may be applied to other diseases and treatments with crucial benefits and harms considerations, and demonstrates how records used in clinical practice maybe harnessed to improve treatment decisions, monitoring and overall care of a cardiovascular disease population treated with a class of drugs

    Risk factors, outcomes and healthcare utilisation in individuals with multimorbidity including heart failure, chronic kidney disease and type 2 diabetes mellitus: a national electronic health record study

    Get PDF
    BACKGROUND: Heart failure (HF), type 2 diabetes (T2D) and chronic kidney disease (CKD) commonly coexist. We studied characteristics, prognosis and healthcare utilisation of individuals with two of these conditions. METHODS: We performed a retrospective, population-based linked electronic health records study from 1998 to 2020 in England to identify individuals diagnosed with two of: HF, T2D or CKD. We described cohort characteristics at time of second diagnosis and estimated risk of developing the third condition and mortality using Kaplan-Meier and Cox regression models. We also estimated rates of healthcare utilisation in primary care and hospital settings in follow-up. FINDINGS: We identified cohorts of 64 226 with CKD and HF, 82 431 with CKD and T2D, and 13 872 with HF and T2D. Compared with CKD and T2D, those with CKD and HF and HF and T2D had more severe risk factor profile. At 5 years, incidence of the third condition and all-cause mortality occurred in 37% (95% CI: 35.9%, 38.1%%) and 31.3% (30.4%, 32.3%) in HF+T2D, 8.7% (8.4%, 9.0%) and 51.6% (51.1%, 52.1%) in HF+CKD, and 6.8% (6.6%, 7.0%) and 17.9% (17.6%, 18.2%) in CKD+T2D, respectively. In each of the three multimorbid groups, the order of the first two diagnoses was also associated with prognosis. In multivariable analyses, we identified risk factors for developing the third condition and mortality, such as age, sex, medical history and the order of disease diagnosis. Inpatient and outpatient healthcare utilisation rates were highest in CKD and HF, and lowest in CKD and T2D. INTERPRETATION: HF, CKD and T2D carry significant mortality and healthcare burden in combination. Compared with other disease pairs, individuals with CKD and HF had the most severe risk factor profile, prognosis and healthcare utilisation. Service planning, policy and prevention must take into account and monitor data across conditions

    Assessing the efficacy of oral immunotherapy for the desensitisation of peanut allergy in children (STOP II): a phase 2 randomised controlled trial

    Get PDF
    SummaryBackgroundSmall studies suggest peanut oral immunotherapy (OIT) might be effective in the treatment of peanut allergy. We aimed to establish the efficacy of OIT for the desensitisation of children with allergy to peanuts.MethodsWe did a randomised controlled crossover trial to compare the efficacy of active OIT (using characterised peanut flour; protein doses of 2–800 mg/day) with control (peanut avoidance, the present standard of care) at the NIHR/Wellcome Trust Cambridge Clinical Research Facility (Cambridge, UK). Randomisation (1:1) was by use of an audited online system; group allocation was not masked. Eligible participants were aged 7–16 years with an immediate hypersensitivity reaction after peanut ingestion, positive skin prick test to peanuts, and positive by double-blind placebo-controlled food challenge (DBPCFC). We excluded participants if they had a major chronic illness, if the care provider or a present household member had suspected or diagnosed allergy to peanuts, or if there was an unwillingness or inability to comply with study procedures. Our primary outcome was desensitisation, defined as negative peanut challenge (1400 mg protein in DBPCFC) at 6 months (first phase). Control participants underwent OIT during the second phase, with subsequent DBPCFC. Immunological parameters and disease-specific quality-of-life scores were measured. Analysis was by intention to treat. Fisher's exact test was used to compare the proportion of those with desensitisation to peanut after 6 months between the active and control group at the end of the first phase. This trial is registered with Current Controlled Trials, number ISRCTN62416244.FindingsThe primary outcome, desensitisation, was recorded for 62% (24 of 39 participants; 95% CI 45–78) in the active group and none of the control group after the first phase (0 of 46; 95% CI 0–9; p<0·001). 84% (95% CI 70–93) of the active group tolerated daily ingestion of 800 mg protein (equivalent to roughly five peanuts). Median increase in peanut threshold after OIT was 1345 mg (range 45–1400; p<0·001) or 25·5 times (range 1·82–280; p<0·001). After the second phase, 54% (95% CI 35–72) tolerated 1400 mg challenge (equivalent to roughly ten peanuts) and 91% (79–98) tolerated daily ingestion of 800 mg protein. Quality-of-life scores improved (decreased) after OIT (median change −1·61; p<0·001). Side-effects were mild in most participants. Gastrointestinal symptoms were, collectively, most common (31 participants with nausea, 31 with vomiting, and one with diarrhoea), then oral pruritus after 6·3% of doses (76 participants) and wheeze after 0·41% of doses (21 participants). Intramuscular adrenaline was used after 0·01% of doses (one participant).InterpretationOIT successfully induced desensitisation in most children within the study population with peanut allergy of any severity, with a clinically meaningful increase in peanut threshold. Quality of life improved after intervention and there was a good safety profile. Immunological changes corresponded with clinical desensitisation. Further studies in wider populations are recommended; peanut OIT should not be done in non-specialist settings, but it is effective and well tolerated in the studied age group.FundingMRC-NIHR partnership

    Elevated plasma triglyceride concentration and risk of adverse clinical outcomes in 1.5 million people: a CALIBER linked electronic health record study

    Get PDF
    BACKGROUND: Assessing the spectrum of disease risk associated with hypertriglyceridemia is needed to inform potential benefits from emerging triglyceride lowering treatments. We sought to examine the associations between a full range of plasma triglyceride concentration with five clinical outcomes. METHODS: We used linked data from primary and secondary care for 15 M people, to explore the association between triglyceride concentration and risk of acute pancreatitis, chronic pancreatitis, new onset diabetes, myocardial infarction and all-cause mortality, over a median of 6-7 years follow up. RESULTS: Triglyceride concentration was available for 1,530,411 individuals (mean age 56·6 ± 15·6 years, 51·4% female), with a median of 1·3 mmol/L (IQR: 0.9.to 1.9). Severe hypertriglyceridemia, defined as > 10 mmol/L, was identified in 3289 (0·21%) individuals including 620 with > 20 mmol/L. In multivariable analyses, a triglyceride concentration > 20 mmol/L was associated with very high risk for acute pancreatitis (Hazard ratio (HR) 13·55 (95% CI 9·15-20·06)); chronic pancreatitis (HR 25·19 (14·91-42·55)); and high risk for diabetes (HR 5·28 (4·51-6·18)) and all-cause mortality (HR 3·62 (2·82-4·65)) when compared to the reference category of ≤ 1·7 mmol/L. An association with myocardial infarction, however, was only observed for more moderate hypertriglyceridaemia between 1.7 and 10 mmol/L. We found a risk interaction with age, with higher risks for all outcomes including mortality among those ≤ 40 years compared to > 40 years. CONCLUSIONS: We highlight an exponential association between severe hypertriglyceridaemia and risk of incident acute and chronic pancreatitis, new diabetes, and mortality, especially at younger ages, but not for myocardial infarction for which only moderate hypertriglyceridemia conferred risk

    Identifying subtypes of chronic kidney disease with machine learning: development, internal validation and prognostic validation using linked electronic health records in 350,067 individuals

    Get PDF
    BACKGROUND: Although chronic kidney disease (CKD) is associated with high multimorbidity, polypharmacy, morbidity and mortality, existing classification systems (mild to severe, usually based on estimated glomerular filtration rate, proteinuria or urine albumin-creatinine ratio) and risk prediction models largely ignore the complexity of CKD, its risk factors and its outcomes. Improved subtype definition could improve prediction of outcomes and inform effective interventions. METHODS: We analysed individuals ≥18 years with incident and prevalent CKD (n = 350,067 and 195,422 respectively) from a population-based electronic health record resource (2006-2020; Clinical Practice Research Datalink, CPRD). We included factors (n = 264 with 2670 derived variables), e.g. demography, history, examination, blood laboratory values and medications. Using a published framework, we identified subtypes through seven unsupervised machine learning (ML) methods (K-means, Diana, HC, Fanny, PAM, Clara, Model-based) with 66 (of 2670) variables in each dataset. We evaluated subtypes for: (i) internal validity (within dataset, across methods); (ii) prognostic validity (predictive accuracy for 5-year all-cause mortality and admissions); and (iii) medications (new and existing by British National Formulary chapter). FINDINGS: After identifying five clusters across seven approaches, we labelled CKD subtypes: 1. Early-onset, 2. Late-onset, 3. Cancer, 4. Metabolic, and 5. Cardiometabolic. Internal validity: We trained a high performing model (using XGBoost) that could predict disease subtypes with 95% accuracy for incident and prevalent CKD (Sensitivity: 0.81-0.98, F1 score:0.84-0.97). Prognostic validity: 5-year all-cause mortality, hospital admissions, and incidence of new chronic diseases differed across CKD subtypes. The 5-year risk of mortality and admissions in the overall incident CKD population were highest in cardiometabolic subtype: 43.3% (42.3-42.8%) and 29.5% (29.1-30.0%), respectively, and lowest in the early-onset subtype: 5.7% (5.5-5.9%) and 18.7% (18.4-19.1%). MEDICATIONS: Across CKD subtypes, the distribution of prescription medication classes at baseline varied, with highest medication burden in cardiometabolic and metabolic subtypes, and higher burden in prevalent than incident CKD. INTERPRETATION: In the largest CKD study using ML, to-date, we identified five distinct subtypes in individuals with incident and prevalent CKD. These subtypes have relevance to study of aetiology, therapeutics and risk prediction. FUNDING: AstraZeneca UK Ltd, Health Data Research UK

    Estimating excess 1-year mortality associated with the COVID-19 pandemic according to underlying conditions and age: a population-based cohort study.

    Get PDF
    BACKGROUND: The medical, societal, and economic impact of the coronavirus disease 2019 (COVID-19) pandemic has unknown effects on overall population mortality. Previous models of population mortality are based on death over days among infected people, nearly all of whom thus far have underlying conditions. Models have not incorporated information on high-risk conditions or their longer-term baseline (pre-COVID-19) mortality. We estimated the excess number of deaths over 1 year under different COVID-19 incidence scenarios based on varying levels of transmission suppression and differing mortality impacts based on different relative risks for the disease. METHODS: In this population-based cohort study, we used linked primary and secondary care electronic health records from England (Health Data Research UK-CALIBER). We report prevalence of underlying conditions defined by Public Health England guidelines (from March 16, 2020) in individuals aged 30 years or older registered with a practice between 1997 and 2017, using validated, openly available phenotypes for each condition. We estimated 1-year mortality in each condition, developing simple models (and a tool for calculation) of excess COVID-19-related deaths, assuming relative impact (as relative risks [RRs]) of the COVID-19 pandemic (compared with background mortality) of 1·5, 2·0, and 3·0 at differing infection rate scenarios, including full suppression (0·001%), partial suppression (1%), mitigation (10%), and do nothing (80%). We also developed an online, public, prototype risk calculator for excess death estimation. FINDINGS: We included 3 862 012 individuals (1 957 935 [50·7%] women and 1 904 077 [49·3%] men). We estimated that more than 20% of the study population are in the high-risk category, of whom 13·7% were older than 70 years and 6·3% were aged 70 years or younger with at least one underlying condition. 1-year mortality in the high-risk population was estimated to be 4·46% (95% CI 4·41-4·51). Age and underlying conditions combined to influence background risk, varying markedly across conditions. In a full suppression scenario in the UK population, we estimated that there would be two excess deaths (vs baseline deaths) with an RR of 1·5, four with an RR of 2·0, and seven with an RR of 3·0. In a mitigation scenario, we estimated 18 374 excess deaths with an RR of 1·5, 36 749 with an RR of 2·0, and 73 498 with an RR of 3·0. In a do nothing scenario, we estimated 146 996 excess deaths with an RR of 1·5, 293 991 with an RR of 2·0, and 587 982 with an RR of 3·0. INTERPRETATION: We provide policy makers, researchers, and the public a simple model and an online tool for understanding excess mortality over 1 year from the COVID-19 pandemic, based on age, sex, and underlying condition-specific estimates. These results signal the need for sustained stringent suppression measures as well as sustained efforts to target those at highest risk because of underlying conditions with a range of preventive interventions. Countries should assess the overall (direct and indirect) effects of the pandemic on excess mortality. FUNDING: National Institute for Health Research University College London Hospitals Biomedical Research Centre, Health Data Research UK

    Identifying subtypes of heart failure from three electronic health record sources with machine learning: an external, prognostic, and genetic validation study

    Get PDF
    BACKGROUND: Machine learning has been used to analyse heart failure subtypes, but not across large, distinct, population-based datasets, across the whole spectrum of causes and presentations, or with clinical and non-clinical validation by different machine learning methods. Using our published framework, we aimed to discover heart failure subtypes and validate them upon population representative data. METHODS: In this external, prognostic, and genetic validation study we analysed individuals aged 30 years or older with incident heart failure from two population-based databases in the UK (Clinical Practice Research Datalink [CPRD] and The Health Improvement Network [THIN]) from 1998 to 2018. Pre-heart failure and post-heart failure factors (n=645) included demographic information, history, examination, blood laboratory values, and medications. We identified subtypes using four unsupervised machine learning methods (K-means, hierarchical, K-Medoids, and mixture model clustering) with 87 of 645 factors in each dataset. We evaluated subtypes for (1) external validity (across datasets); (2) prognostic validity (predictive accuracy for 1-year mortality); and (3) genetic validity (UK Biobank), association with polygenic risk score (PRS) for heart failure-related traits (n=11), and single nucleotide polymorphisms (n=12). FINDINGS: We included 188 800, 124 262, and 9573 individuals with incident heart failure from CPRD, THIN, and UK Biobank, respectively, between Jan 1, 1998, and Jan 1, 2018. After identifying five clusters, we labelled heart failure subtypes as (1) early onset, (2) late onset, (3) atrial fibrillation related, (4) metabolic, and (5) cardiometabolic. In the external validity analysis, subtypes were similar across datasets (c-statistics: THIN model in CPRD ranged from 0·79 [subtype 3] to 0·94 [subtype 1], and CPRD model in THIN ranged from 0·79 [subtype 1] to 0·92 [subtypes 2 and 5]). In the prognostic validity analysis, 1-year all-cause mortality after heart failure diagnosis (subtype 1 0·20 [95% CI 0·14-0·25], subtype 2 0·46 [0·43-0·49], subtype 3 0·61 [0·57-0·64], subtype 4 0·11 [0·07-0·16], and subtype 5 0·37 [0·32-0·41]) differed across subtypes in CPRD and THIN data, as did risk of non-fatal cardiovascular diseases and all-cause hospitalisation. In the genetic validity analysis the atrial fibrillation-related subtype showed associations with the related PRS. Late onset and cardiometabolic subtypes were the most similar and strongly associated with PRS for hypertension, myocardial infarction, and obesity (p<0·0009). We developed a prototype app for routine clinical use, which could enable evaluation of effectiveness and cost-effectiveness. INTERPRETATION: Across four methods and three datasets, including genetic data, in the largest study of incident heart failure to date, we identified five machine learning-informed subtypes, which might inform aetiological research, clinical risk prediction, and the design of heart failure trials. FUNDING: European Union Innovative Medicines Initiative-2

    Differences in Treatment Patterns and Outcomes of Acute Myocardial Infarction for Low- and High-Income Patients in 6 Countries

    Get PDF
    IMPORTANCE: Differences in the organization and financing of health systems may produce more or less equitable outcomes for advantaged vs disadvantaged populations. We compared treatments and outcomes of older high- and low-income patients across 6 countries. OBJECTIVE: To determine whether treatment patterns and outcomes for patients presenting with acute myocardial infarction differ for low- vs high-income individuals across 6 countries. DESIGN, SETTING, AND PARTICIPANTS: Serial cross-sectional cohort study of all adults aged 66 years or older hospitalized with acute myocardial infarction from 2013 through 2018 in the US, Canada, England, the Netherlands, Taiwan, and Israel using population-representative administrative data. EXPOSURES: Being in the top and bottom quintile of income within and across countries. MAIN OUTCOMES AND MEASURES: Thirty-day and 1-year mortality; secondary outcomes included rates of cardiac catheterization and revascularization, length of stay, and readmission rates. RESULTS: We studied 289 376 patients hospitalized with ST-segment elevation myocardial infarction (STEMI) and 843 046 hospitalized with non-STEMI (NSTEMI). Adjusted 30-day mortality generally was 1 to 3 percentage points lower for high-income patients. For instance, 30-day mortality among patients admitted with STEMI in the Netherlands was 10.2% for those with high income vs 13.1% for those with low income (difference, -2.8 percentage points [95% CI, -4.1 to -1.5]). One-year mortality differences for STEMI were even larger than 30-day mortality, with the highest difference in Israel (16.2% vs 25.3%; difference, -9.1 percentage points [95% CI, -16.7 to -1.6]). In all countries, rates of cardiac catheterization and percutaneous coronary intervention were higher among high- vs low-income populations, with absolute differences ranging from 1 to 6 percentage points (eg, 73.6% vs 67.4%; difference, 6.1 percentage points [95% CI, 1.2 to 11.0] for percutaneous intervention in England for STEMI). Rates of coronary artery bypass graft surgery for patients with STEMI in low- vs high-income strata were similar but for NSTEMI were generally 1 to 2 percentage points higher among high-income patients (eg, 12.5% vs 11.0% in the US; difference, 1.5 percentage points [95% CI, 1.3 to 1.8 ]). Thirty-day readmission rates generally also were 1 to 3 percentage points lower and hospital length of stay generally was 0.2 to 0.5 days shorter for high-income patients. CONCLUSIONS AND RELEVANCE: High-income individuals had substantially better survival and were more likely to receive lifesaving revascularization and had shorter hospital lengths of stay and fewer readmissions across almost all countries. Our results suggest that income-based disparities were present even in countries with universal health insurance and robust social safety net systems

    UK phenomics platform for developing and validating electronic health record phenotypes: CALIBER

    Get PDF
    Objective: Electronic health records (EHRs) are a rich source of information on human diseases, but the information is variably structured, fragmented, curated using different coding systems, and collected for purposes other than medical research. We describe an approach for developing, validating, and sharing reproducible phenotypes from national structured EHR in the United Kingdom with applications for translational research. Materials and Methods: We implemented a rule-based phenotyping framework, with up to 6 approaches of validation. We applied our framework to a sample of 15 million individuals in a national EHR data source (population-based primary care, all ages) linked to hospitalization and death records in England. Data comprised continuous measurements (for example, blood pressure; medication information; coded diagnoses, symptoms, procedures, and referrals), recorded using 5 controlled clinical terminologies: (1) read (primary care, subset of SNOMED-CT [Systematized Nomenclature of Medicine Clinical Terms]), (2) International Classification of Diseases–Ninth Revision and Tenth Revision (secondary care diagnoses and cause of mortality), (3) Office of Population Censuses and Surveys Classification of Surgical Operations and Procedures, Fourth Revision (hospital surgical procedures), and (4) DMþD prescription codes. Results: Using the CALIBER phenotyping framework, we created algorithms for 51 diseases, syndromes, biomarkers, and lifestyle risk factors and provide up to 6 validation approaches. The EHR phenotypes are curated in the open-access CALIBER Portal (https://www.caliberresearch.org/portal) and have been used by 40 national and international research groups in 60 peer-reviewed publications. Conclusions: We describe a UK EHR phenomics approach within the CALIBER EHR data platform with initial evidence of validity and use, as an important step toward international use of UK EHR data for health research

    Identifying adults at high-risk for change in weight and BMI in England: a longitudinal, large-scale, population-based cohort study using electronic health records.

    Get PDF
    Funder: Department of HealthFunder: Medical Research CouncilBackgroundTargeted obesity prevention policies would benefit from the identification of population groups with the highest risk of weight gain. The relative importance of adult age, sex, ethnicity, geographical region, and degree of social deprivation on weight gain is not known. We aimed to identify high-risk groups for changes in weight and BMI using electronic health records (EHR).MethodsIn this longitudinal, population-based cohort study we used linked EHR data from 400 primary care practices (via the Clinical Practice Research Datalink) in England, accessed via the CALIBER programme. Eligible participants were aged 18-74 years, were registered at a general practice clinic, and had BMI and weight measurements recorded between Jan 1, 1998, and June 30, 2016, during the period when they had eligible linked data with at least 1 year of follow-up time. We calculated longitudinal changes in BMI over 1, 5, and 10 years, and investigated the absolute risk and odds ratios (ORs) of transitioning between BMI categories (underweight, normal weight, overweight, obesity class 1 and 2, and severe obesity [class 3]), as defined by WHO. The associations of demographic factors with BMI transitions were estimated by use of logistic regression analysis, adjusting for baseline BMI, family history of cardiovascular disease, use of diuretics, and prevalent chronic conditions.FindingsWe included 2 092 260 eligible individuals with more than 9 million BMI measurements in our study. Young adult age was the strongest risk factor for weight gain at 1, 5, and 10 years of follow-up. Compared with the oldest age group (65-74 years), adults in the youngest age group (18-24 years) had the highest OR (4·22 [95% CI 3·86-4·62]) and greatest absolute risk (37% vs 24%) of transitioning from normal weight to overweight or obesity at 10 years. Likewise, adults in the youngest age group with overweight or obesity at baseline were also at highest risk to transition to a higher BMI category; OR 4·60 (4·06-5·22) and absolute risk (42% vs 18%) of transitioning from overweight to class 1 and 2 obesity, and OR 5·87 (5·23-6·59) and absolute risk (22% vs 5%) of transitioning from class 1 and 2 obesity to class 3 obesity. Other demographic factors were consistently less strongly associated with these transitions; for example, the OR of transitioning from normal weight to overweight or obesity in people living in the most socially deprived versus least deprived areas was 1·23 (1·18-1·27), for men versus women was 1·12 (1·08-1·16), and for Black individuals versus White individuals was 1·13 (1·04-1·24). We provide an open access online risk calculator, and present high-resolution obesity risk charts over a 1-year, 5-year, and 10-year follow-up period.InterpretationA radical shift in policy is required to focus on individuals at the highest risk of weight gain (ie, young adults aged 18-24 years) for individual-level and population-level prevention of obesity and its long-term consequences for health and health care.FundingThe British Hearth Foundation, Health Data Research UK, the UK Medical Research Council, and the National Institute for Health Research
    corecore