9 research outputs found

    Payments by Industry to Residency Program Directors in the United States: A Cross-Sectional Study

    No full text
    Purpose To assess the proportion, nature, and extent of financial payments from industry to residency program directors in the United States. Method This cross-sectional study used opensource data from Doximity and the Centers for Medicare and Medicaid (CMS) open payments database. Profiles of 4,686 residency program directors from 28 different specialties were identified using Doximity and matched to records in the CMS database. All payments received per residency program director over the years 2014 to 2018 were extracted, including amount in U.S. dollars, payment year, and nature of payment (research versus general payments). Total payments (research plus general payments) received over the 5 years were added up per residency program director. Only personal payments were included. Results Overall, 74% (3,465/4,686) of all residency program directors received 1 or more personal payments, totaling 77,058,139,withamedianof77,058,139, with a median of 216 (interquartile range, 00-2,150) and a mean of 16,444(standarddeviation,16,444 (standard deviation, 183,061) per residency program director over the 5 years. Ninety-five percent of total payment value were general payments, and 5% were personal research payments. About 11% (536/4,686) of residency program directors received more than 10,000,while310,000, while 3% (133/4,686) received more than 100,000 in the study years. There was a substantial difference in the proportion (P < .001), nature (P < .001), and amount (P < .001) of payments of residency program directors between specialties. Almost all residency program directors of interventional radiology (96% [74/77]), vascular surgery (96% [53/55]), and orthopedic surgery (92% [184/201]) received payments, while only one-third to one-half of those in preventive medicine (29% [18/62]), pediatrics (43% [90/211]), and pathology (51% [73/143]) received payments. Conclusions Industry payments to residency program directors are common, although large variation exists between specialties. The majority of direct payments to residency program directors are for non-researchrelated activities

    Analysis of Online Reviews of Orthopaedic Surgeons and Orthopaedic Practices Using Natural Language Processing

    No full text
    BACKGROUND: There is growing interest in measuring and improving patient experience. Machine learning-based natural language processing techniques may help identify instructive themes in online comments written by patients about their healthcare provider. Separating individual surgeon and orthopaedic office reviews, we analyzed themes that are discussed based on the rating category, the association with review length, the number of people posting more than one review for a surgeon or office, the mean number of reviews per rating category, and the difference in review tones. METHODS: On Yelp.com, we collected 11,614 free-text reviews-together with a one- to five-star rating-of orthopaedic surgeons. Using natural language processing, we identified the most frequently occurring word combinations among rating categories. Themes were derived by categorizing word combinations. Dominant tones (emotional and language styles) were assessed by the IBM Watson Tone Analyzer. We calculated chi-square tests for linear trend and Spearman's rank correlation coefficients to assess differences among rating category. RESULTS: For individual surgeons and orthopaedic offices, themes such as logistics, care and compassion, trust, recommendation, and customer service varied among rating categories. More positive reviews are shorter for individual surgeons and orthopaedic offices, while rating category was comparable among people posting more than one review for both groups. Tones of joy and confidence were associated with higher ratings. Sadness and tentative tones were associated with lower ratings. DISCUSSION: For individual orthopaedic surgeons and orthopaedic offices, patient experience may be influenced mostly by the patient-clinician relationship. Training in more effective communication strategies may help improve self-reported patient experience

    Diagnosis of dorsal screw penetration after volar plating of a distal radial fracture: Intraoperative dorsal tangential views versus 3D fluoroscopy

    No full text
    Aims The aim of this study was to investigate whether intraoperative 3D fluoroscopic imaging outperforms dorsal tangential views in the detection of dorsal cortex screw penetration after volar plating of an intra-articular distal radial fracture, as identified on postoperative CT imaging. Methods A total of 165 prospectively enrolled patients who underwent volar plating for an intra-articular distal radial fracture were retrospectively evaluated to study three intraoperative imaging protocols: 1) standard 2D fluoroscopic imaging with anteroposterior (AP) and elevated lateral images (n = 55); 2) 2D fluoroscopic imaging with AP, lateral, and dorsal tangential views images (n = 50); and 3) 3D fluoroscopy (n = 60). Multiplanar reconstructions of postoperative CT scans served as the reference standard. Results In order to detect dorsal screw penetration, the sensitivity of dorsal tangential views was 39% with a negative predictive value (NPV) of 91% and an accuracy of 91%; compared with a sensitivity of 25% for 3D fluoroscopy with a NPV of 93% and an accuracy of 93%. On the postoperative CT scans, we found penetrating screws in: 1) 40% of patients in the 2D fluoroscopy group; 2) in 32% of those in the 2D fluoroscopy group with AP, lateral, and dorsal tangential views; and 3) in 25% of patients in the 3D fluoroscopy group. In all three groups, the second compartment was prone to penetration, while the postoperative incidence decreased when more advanced imaging was used. There were no penetrating screws in the third compartment (extensor pollicis longus groove) in the 3D fluoroscopy groups, and one in the dorsal tangential views group. Conclusion Advanced intraoperative imaging helps to identify screws which have penetrated the dorsal compartments of the wrist. However, based on diagnostic performance characteristics, one cannot conclude that 3D fluoroscopy outperforms dorsal tangential views when used for this purpose. Dorsal tangential views are sufficiently accurate to detect dorsal screw penetration, and arguably more efficacious than 3D fluoroscopy

    What are the best diagnostic tests for diagnosing bacterial arthritis of a native joint?: a systematic review of 27 studies

    No full text
    AIMS: This study aimed to answer two questions: what are the best diagnostic methods for diagnosing bacterial arthritis of a native joint?; and what are the most commonly used definitions for bacterial arthritis of a native joint? METHODS: We performed a search of PubMed, Embase, and Cochrane libraries for relevant studies published between January 1980 and April 2020. Of 3,209 identified studies, we included 27 after full screening. Sensitivity, specificity, area under the curve, and Youden index of diagnostic tests were extracted from included studies. We grouped test characteristics per diagnostic modality. We extracted the definitions used to establish a definitive diagnosis of bacterial arthritis of a native joint per study. RESULTS: Overall, 28 unique diagnostic tests for diagnosing bacterial arthritis of a native joint were identified. The following five tests were deemed most useful: serum ESR (sensitivity: 34% to 100%, specificity: 23% to 93%), serum CRP (sensitivity: 58% to 100%, specificity: 0% to 96%), serum procalcitonin (sensitivity: 0% to 100%, specificity: 68% to 100%), the proportion of synovial polymorphonuclear cells (sensitivity: 42% to 100%, specificity: 54% to 94%), and the gram stain of synovial fluid (sensitivity: 27% to 81%, specificity: 99% to 100%). CONCLUSION: Diagnostic methods with relatively high sensitivities, such as serum CRP, ESR, and synovial polymorphonuclear cells, are useful for screening. Diagnostic methods with a relatively high specificity, such as serum procalcitonin and synovial fluid gram stain, are useful for establishing a diagnosis of bacterial arthritis. This review helps to interpret the value of various diagnostic tests for diagnosing bacterial arthritis of a native joint in clinical practice. Cite this article: Bone Joint J 2021;103-B(12):1745-1753

    Antihypertensive drugs demonstrate varying levels of hip fracture risk: A systematic review and meta-analysis

    No full text
    Objective: By aggregating the literature, we evaluated the association between use of specific antihypertensive drugs and the risk of hip fractures compared with nonuse. Study design and setting: We systematically searched the Pubmed, Embase, and Cochrane databases from inception of each database until July 30, 2020 to identify articles including patients 18 years of age or older reporting on the association between antihypertensive drugs and the risk of hip fracture. Antihypertensive drugs were restricted to thiazides; beta-blockers; calcium-channel blockers; angiotensin-converting enzyme (ACE) inhibitors; and angiotensin receptor blockers. Nonusers encompass all patients that are not using the specific antihypertensive drug that has been reported. Unadjusted odds ratios with 95% confidence intervals (CIs) of the association between antihypertensive drug use and hip fractures were reported. Meta-analysis was performed when a minimum of five studies were identified for each antihypertensive drug class. Quality assessment was done using ROBINS-I tool. The GRADE approach was used to evaluate the certainty of the evidence. Results: Of 962 citations, 22 observational studies were included; 9 studies had a cohort design and 13 studies were case-control studies. No randomized controlled trials were identified. We found very low certainty of evidence that both thiazides (pooled odds ratio: 0.85, 95% CI 0.73 to 0.99, p = 0.04) as well as beta-blockers (pooled odds ratio: 0.88, 95% CI 0.79 to 0.98, p = 0.02) were associated with a reduced hip fracture risk as compared to specific nonuse. One study, reporting on angiotensin receptor blockers, also suggested a protective effect for hip fractures, whereas we found conflicting findings in four studies for calcium-channel blockers and in two studies for ACE inhibitors. Conclusion: Among 22 observational studies, we found very low certainty of evidence that, compared to specific nonuse of antihypertensive drugs, use of thiazides, beta-blockers, and angiotensin receptor blockers were associated with a reduced protective hip fracture risk, while conflicting findings for calcium-channel blockers and ACE inhibitors were found. Given the low quality of included studies, further research –randomized controlled trials– are needed to definitively assess the causal relationship between specific antihypertensive drug classes and (relatively infrequent) hip fractures

    Is Deep Learning On Par with Human Observers for Detection of Radiographically Visible and Occult Fractures of the Scaphoid?

    No full text
    BACKGROUND: Preliminary experience suggests that deep learning algorithms are nearly as good as humans in detecting common, displaced, and relatively obvious fractures (such as, distal radius or hip fractures). However, it is not known whether this also is true for subtle or relatively nondisplaced fractures that are often difficult to see on radiographs, such as scaphoid fractures. QUESTIONS/PURPOSES: (1) What is the diagnostic accuracy, sensitivity, and specificity of a deep learning algorithm in detecting radiographically visible and occult scaphoid fractures using four radiographic imaging views? (2) Does adding patient demographic (age and sex) information improve the diagnostic performance of the deep learning algorithm? (3) Are orthopaedic surgeons better at diagnostic accuracy, sensitivity, and specificity compared with deep learning? (4) What is the interobserver reliability among five human observers and between human consensus and deep learning algorithm? METHODS: We retrospectively searched the picture archiving and communication system (PACS) to identify 300 patients with a radiographic scaphoid series, until we had 150 fractures (127 visible on radiographs and 23 only visible on MRI) and 150 non-fractures with a corresponding CT or MRI as the reference standard for fracture diagnosis. At our institution, MRIs are usually ordered for patients with scaphoid tenderness and normal radiographs, and a CT with radiographically visible scaphoid fracture. We used a deep learning algorithm (a convolutional neural network [CNN]) for automated fracture detection on radiographs. Deep learning, an advanced subset of artificial intelligence, combines artificial neuronal layers to resemble a neuron cell. CNNs-essentially deep learning algorithms resembling interconnected neurons in the human brain-are most commonly used for image analysis. Area under the receiver operating characteristic curve (AUC) was used to evaluate the algorithm's diagnostic performance. An AUC of 1.0 would indicate perfect prediction, whereas 0.5 would indicate that a prediction is no better than a flip of a coin. The probability of a scaphoid fracture generated by the CNN, sex, and age were included in a multivariable logistic regression to determine whether this would improve the algorithm's diagnostic performance. Diagnostic performance characteristics (accuracy, sensitivity, and specificity) and reliability (kappa statistic) were calculated for the CNN and for the five orthopaedic surgeon observers in our study. RESULTS: The algorithm had an AUC of 0.77 (95% CI 0.66 to 0.85), 72% accuracy (95% CI 60% to 84%), 84% sensitivity (95% CI 0.74 to 0.94), and 60% specificity (95% CI 0.46 to 0.74). Adding age and sex did not improve diagnostic performance (AUC 0.81 [95% CI 0.73 to 0.89]). Orthopaedic surgeons had better specificity (0.93 [95% CI 0.93 to 0.99]; p < 0.01), while accuracy (84% [95% CI 81% to 88%]) and sensitivity (0.76 [95% CI 0.70 to 0.82]; p = 0.29) did not differ between the algorithm and human observers. Although the CNN was less specific in diagnosing relatively obvious fractures, it detected five of six occult scaphoid fractures that were missed by all human observers. The interobserver reliability among the five surgeons was substantial (Fleiss' kappa = 0.74 [95% CI 0.66 to 0.83]), but the reliability between the algorithm and human observers was only fair (Cohen's kappa = 0.34 [95% CI 0.17 to 0.50]). CONCLUSIONS: Initial experience with our deep learning algorithm suggests that it has trouble identifying scaphoid fractures that are obvious to human observers. Thirteen false positive suggestions were made by the CNN, which were correctly detected by the five surgeons. Research with larger datasets-preferably also including information from physical examination-or further algorithm refinement is merited. LEVEL OF EVIDENCE: Level III, diagnostic study

    Do 3-D Printed Handheld Models Improve Surgeon Reliability for Recognition of Intraarticular Distal Radius Fracture Characteristics?

    No full text
    BACKGROUND: For fracture care, radiographs and two-dimensional (2-D) and three-dimensional (3-D) CT are primarily used for preoperative planning and postoperative evaluation. Intraarticular distal radius fractures are technically challenging to treat, and meticulous preoperative planning is paramount to improve the patient's outcome. Three-dimensionally printed handheld models might improve the surgeon's interpretation of specific fracture characteristics and patterns preoperatively and could therefore be clinically valuable; however, the additional value of 3-D printed handheld models for fractures of the distal radius, a high-volume and commonly complex fracture due to its intraarticular configuration, has yet to be determined. QUESTIONS/PURPOSES: (1) Does the reliability of assessing specific fracture characteristics that guide surgical decision-making for distal radius fractures improve with 3-D printed handheld models? (2) Does surgeon agreement on the overall fracture classification improve with 3-D printed handheld models? (3) Does the surgeon's confidence improve when assessing the overall fracture configuration with an additional 3-D model? METHODS: We consecutively included 20 intraarticular distal radius fractures treated at a Level 1 trauma center between May 2018 and November 2018. Ten surgeons evaluated the presence or absence of specific fracture characteristics (volar rim fracture, die punch, volar lunate facet, dorsal comminution, step-off > 2 mm, and gap > 2 mm), fracture classification according to the AO/Orthopaedic Trauma Association (OTA) classification scheme, and their confidence in assessing the overall fracture according to the classification scheme, rated on a scale from 0 to 10 (0 = not at all confident to 10 = very confident). Of 10 participants regularly treating distal radius fractures, seven were orthopaedic trauma surgeons and three upper limb surgeons with experience levels ranging from 1 to 25 years after completion of residency training. Fractures were assessed twice, with 1 month between each assessment. Initially, fractures were assessed using radiographs and 2-D and 3-D CT images (conventional assessment); the second time, the evaluation was based on radiographs and 2-D and 3-D CT images with an additional 3-D handheld model (3-D printed handheld model assessment). On both occasions, fracture characteristics were evaluated upon a surgeon's own interpretation, without specific instruction before assessment. We provided a sheet demonstrating the AO/OTA classification scheme before evaluation on each session. Multi-rater Fleiss's kappa was used to determine intersurgeon reliability for assessing fracture characteristics and classification. Confidence regarding assessment of the overall fracture classification was assessed using a paired t-test. RESULTS: We found that 3-D printed models of intraarticular distal radius fractures led to no change in kappa values for the reliability of all characteristics: volar rim (conventional kappa 0.19 [95% CI 0.06 to 0.32], kappa for 3-D handheld model 0.23 [95% CI 0.11 to 0.36], difference of kappas 0.04 [95% CI -0.14 to 0.22]; p = 0.66), die punch (conventional kappa 0.38 [95% CI 0.15 to 0.61], kappa for 3-D handheld model 0.50 [95% CI 0.23 to 0.78], difference of kappas 0.12 [95% CI -0.23 to 0.47]; p = 0.52), volar lunate facet (conventional kappa 0.31 [95% CI 0.14 to 0.49], kappa for 3-D handheld model 0.48 [95% CI 0.23 to 0.72], difference of kappas 0.17 [95% CI -0.12 to 0.46]; p = 0.26), dorsal comminution (conventional kappa 0.36 [95% CI 0.13 to 0.58], kappa for 3-D handheld model 0.31 [95% CI 0.11 to 0.51], difference of kappas -0.05 [95% CI -0.34 to 0.24]; p = 0.74), step-off > 2 mm (conventional kappa 0.55 [95% CI 0.29 to 0.82], kappa for 3-D handheld model 0.58 [95% CI 0.31 to 0.85], difference of kappas 0.03 [95% CI -0.34 to 0.40]; p = 0.87), gap > 2 mm (conventional kappa 0.59 [95% CI 0.39 to 0.79], kappa for 3-D handheld model 0.69 [95% CI 0.50 to 0.89], difference of kappas 0.10 [95% CI -0.17 to 0.37]; p = 0.48). Although there appeared to be categorical improvement in kappa values for some fracture characteristics, overlapping CIs indicated no change. Fracture classification did not improve (conventional diagnostics: kappa 0.27 [95% CI 0.14 to 0.39], conventional diagnostics with an additional 3-D handheld model: kappa 0.25 [95% CI 0.15 to 0.35], difference of kappas: -0.02 [95% CI -0.18 to 0.14]; p = 0.81). There was no improvement in self-assessed confidence in terms of assessment of overall fracture configuration when a 3-D model was added to the evaluation process (conventional diagnostics 7.8 [SD 0.79 {95% CI 7.2 to 8.3}], 3-D handheld model 8.5 [SD 0.71 {95% CI 8.0 to 9.0}], difference of score: 0.7 [95% CI -1.69 to 0.16], p = 0.09). CONCLUSIONS: Intersurgeon reliability for evaluating the characteristics of and classifying intraarticular distal radius fractures did not improve with an additional 3-D model. Further studies should evaluate the added value of 3-D printed handheld models for teaching surgical residents and medical trainees to define the future role of 3-D printing in caring for fractures of the distal radius. LEVEL OF EVIDENCE: Level II, diagnostic study

    What Are the Applications and Limitations of Artificial Intelligence for Fracture Detection and Classification in Orthopaedic Trauma Imaging?: A Systematic Review

    No full text
    BACKGROUND: Artificial-intelligence algorithms derive rules and patterns from large amounts of data to calculate the probabilities of various outcomes using new sets of similar data. In medicine, artificial intelligence (AI) has been applied primarily to image-recognition diagnostic tasks and evaluating the probabilities of particular outcomes after treatment. However, the performance and limitations of AI in the automated detection and classification of fractures has not been examined comprehensively. QUESTION/PURPOSES: In this systematic review, we asked (1) What is the proportion of correctly detected or classified fractures and the area under the receiving operating characteristic (AUC) curve of AI fracture detection and classification models? (2) What is the performance of AI in this setting compared with the performance of human examiners? METHODS: The PubMed, Embase, and Cochrane databases were systematically searched from the start of each respective database until September 6, 2018, using terms related to "fracture", "artificial intelligence", and "detection, prediction, or evaluation." Of 1221 identified studies, we retained 10 studies: eight studies involved fracture detection (ankle, hand, hip, spine, wrist, and ulna), one addressed fracture classification (diaphyseal femur), and one addressed both fracture detection and classification (proximal humerus). We registered the review before data collection (PROSPERO: CRD42018110167) and used the Preferred Reporting Items for Systematic Reviews and Meta-analyses (PRISMA). We reported the range of the accuracy and AUC for the performance of the predicted fracture detection and/or classification task. An AUC of 1.0 would indicate perfect prediction, whereas 0.5 would indicate a prediction is no better than a flip-of-a-coin. We conducted quality assessment using a seven-item checklist based on a modified methodologic index for nonrandomized studies instrument (MINORS). RESULTS: For fracture detection, the AUC in five studies reflected near perfect prediction (range, 0.95-1.0), and the accuracy in seven studies ranged from 83% to 98%. For fracture classification, the AUC was 0.94 in one study, and the accuracy in two studies ranged from 77% to 90%. In two studies AI outperformed human examiners for detecting and classifying hip and proximal humerus fractures, and one study showed equivalent performance for detecting wrist, hand and ankle fractures. CONCLUSIONS: Preliminary experience with fracture detection and classification using AI shows promising performance. AI may enhance processing and communicating probabilistic tasks in medicine, including orthopaedic surgery. At present, inadequate reference standard assignments to train and test AI is the biggest hurdle before integration into clinical workflow. The next step will be to apply AI to more challenging diagnostic and therapeutic scenarios when there is absence of certitude. Future studies should also seek to address legal regulation and better determine feasibility of implementation in clinical practice. LEVEL OF EVIDENCE: Level II, diagnostic study

    A Machine Learning Algorithm to Predict the Probability of (Occult) Posterior Malleolar Fractures Associated With Tibial Shaft Fractures to Guide Malleolus First Fixation.

    No full text
    OBJECTIVES: To develop an accurate machine learning (ML) predictive model incorporating patient, fracture, and trauma characteristics to identify individual patients at risk of an (occult) PMF. METHODS: Databases of 2 studies including patients with TSFs from 2 Level 1 trauma centers were combined for analysis. Using ten-fold cross-validation, 4 supervised ML algorithms were trained in recognizing patterns associated with PMFs: (1) Bayes point machine; (2) support vector machine; (3) neural network; and (4) boosted decision tree. Performance of each ML algorithm was evaluated and compared based on (1) C-statistic; (2) calibration slope and intercept; and (3) Brier score. The best-performing ML algorithm was incorporated into an online open-access prediction tool. RESULTS: Total data set included 263 patients, of which 28% had a PMF. Training of the Bayes point machine resulted in the best-performing prediction model reflected by good C-statistic, calibration slope, calibration intercept, and Brier score of 0.89, 1.02, -0.06, and 0.106, respectively. This prediction model was deployed as an open-access online prediction tool. CONCLUSION: A ML-based prediction model accurately predicted the probability of a (occult) PMF in patients with a TSF based on patient- and fracture-specific characteristics. This prediction model can guide surgeons in their diagnostic workup and preoperative planning. Further research is required to externally validate the model before implementation in clinical practice. LEVEL OF EVIDENCE: Prognostic Level III. See Instructions for Authors for a complete description of levels of evidence
    corecore