247 research outputs found

    Query Generation based on Generative Adversarial Networks

    Full text link
    Many problems in database systems, such as cardinality estimation, database testing and optimizer tuning, require a large query load as data. However, it is often difficult to obtain a large number of real queries from users due to user privacy restrictions or low frequency of database access. Query generation is one of the approaches to solve this problem. Existing query generation methods, such as random generation and template-based generation, do not consider the relationship between the generated queries and existing queries, or even generate semantically incorrect queries. In this paper, we propose a query generation framework based on generative adversarial networks (GAN) to generate query load that is similar to the given query load. In our framework, we use a syntax parser to transform the query into a parse tree and traverse the tree to obtain the sequence of production rules corresponding to the query. The generator of GAN takes a fixed distribution prior as input and outputs the query sequence, and the discriminator takes the real query and the fake query generated by the generator as input and outputs a gradient to guide the generator learning. In addition, we add context-free grammar and semantic rules to the generation process, which ensures that the generated queries are syntactically and semantically correct. We conduct experiments to evaluate our approach on real-world dataset, which show that our approach can generate new query loads with a similar distribution to a given query load, and that the generated queries are syntactically correct with no semantic errors. The generated query loads are used in downstream task, and the results show a significant improvement in the models trained with the expanded query loads using our approach

    The Irreversible Pollution Game

    Get PDF
    We study a 2-country differential game with irreversible pollution. Irresability is of a hard type: above a certain threshold level of pollution, the self-cleaning capacity of Nature drops to zero. Accordingly, the game includes a non-concave feature, and we characterize both the cooperative and non-cooperative versions with this general non-LQ property. We deliver full analytical results for the existence of Markov Perfect Equilibria. We first demonstrate that when pollution costs are equal across players (symmetry), irreversible pollution regimes are more frequently reached than under cooperation. Second, we study the implications of asymmetry in the pollution cost. We find far nontrivial results on the reachability of the irreversible regime. However, we unambiguously prove that, for the same total cost of pollution, provided the irreversible regime is reached in both the symmetric and asymmetric cases, long-term pollution is larger in the symmetric case, reflecting more intensive free-riding under symmetry

    Optimal behavior under pollution irreversibility risk and distance to the irreversibility thresholds: A global approach

    Get PDF
    We study optimal behavior under irreversible pollution risk. Irreversibility comes from the decay rate of pollution sharply dropping (possibly to zero) above a threshold pollution level. In addition, the economy can instantaneously move from a reversible to an irreversible pollution mode, following a Poisson process, the irreversible mode being an absorbing state. The resulting non-convex optimal pollution control is therefore piecewise deterministic. First, we are able to characterize analytically and globally the optimal emission policy using dynamic programming. Second, we prove that for any value of the Poisson probability, the optimal emission policy leads to more pollution with the irreversibility risk than without in a neighborhood of the pollution irreversibility threshold. Third, we find that this local result does not necessarily hold if actual pollution is far enough from the irreversibility threshold. Our results enhance the importance of the avoidability of the latter threshold in the optimal economic behavior under the irreversibility risk

    A Mild Dyssynchronous Contraction Pattern Detected by SPECT Myocardial Perfusion Imaging Predicts Super-Response to Cardiac Resynchronization Therapy

    Get PDF
    Background: Using single photon emission computed tomography myocardial perfusion imaging (SPECT MPI) with phase analysis (PA), we aimed to identify the predictive value of a new contraction pattern in cardiac resynchronization therapy (CRT) response. Methods: Left ventricular mechanical dyssynchrony (LVMD) was evaluated using SPECT MPI with PA in non-ischemic dilated cardiomyopathy (DCM) patients with left bundle branch block (LBBB) indicated for CRT. CRT super-response was defined as LV ejection fraction (EF) ≥50% or an absolute increase of LVEF \u3e15%. The LV contraction was categorized as the mild dyssynchronous pattern when the phase standard deviation (PSD) ≤ 40.3° and phase histogram bandwidth (PBW) ≤ 111.9°, otherwise it was defined as severe dyssynchronous pattern which was further characterized as U-shaped, heterogeneous or homogenous pattern. Results: The final cohort comprised 74 patients, including 32 (43.2%) in mild dyssynchronous group, 17 (23%) in U-shaped group, 19 (25.7%) in heterogeneous group, and 6 (8.1%) in homogenous group. The mild dyssynchronous group had lower PSD and PBW than U-shaped, heterogeneous, and homogenous groups ( \u3c 0.0001). Compared to patients with the heterogeneous pattern, the odds ratios (ORs) with 95% confidence intervals (CIs) for CRT super-response were 10.182(2.43-42.663), 12.8(2.545-64.372), and 2.667(0.327-21.773) for patients with mild dyssynchronous, U-shaped, and homogenous pattern, respectively. After multivariable adjustment, mild dyssynchronous group remained associated with increased CRT super-response (adjusted OR 5.709, 95% CI 1.152-28.293). Kaplan-Meier curves showed that mild dyssynchronous group demonstrated a better long-term prognosis. Conclusions: The mild dyssynchronous pattern in patients with DCM is associated with an increased CRT super-response and better long-term prognosis

    A new method using deep transfer learning on ECG to predict the response to cardiac resynchronization therapy

    Full text link
    Background: Cardiac resynchronization therapy (CRT) has emerged as an effective treatment for heart failure patients with electrical dyssynchrony. However, accurately predicting which patients will respond to CRT remains a challenge. This study explores the application of deep transfer learning techniques to train a predictive model for CRT response. Methods: In this study, the short-time Fourier transform (STFT) technique was employed to transform ECG signals into two-dimensional images. A transfer learning approach was then applied on the MIT-BIT ECG database to pre-train a convolutional neural network (CNN) model. The model was fine-tuned to extract relevant features from the ECG images, and then tested on our dataset of CRT patients to predict their response. Results: Seventy-one CRT patients were enrolled in this study. The transfer learning model achieved an accuracy of 72% in distinguishing responders from non-responders in the local dataset. Furthermore, the model showed good sensitivity (0.78) and specificity (0.79) in identifying CRT responders. The performance of our model outperformed clinic guidelines and traditional machine learning approaches. Conclusion: The utilization of ECG images as input and leveraging the power of transfer learning allows for improved accuracy in identifying CRT responders. This approach offers potential for enhancing patient selection and improving outcomes of CRT

    Collaborative simulation method with spatiotemporal synchronization process control

    Get PDF
    When designing a complex mechatronics system, such as high speed trains, it is relatively difficult to effectively simulate the entire system’s dynamic behaviors because it involves multi-disciplinary subsystems. Currently, a most practical approach for multi-disciplinary simulation is interface based coupling simulation method, but it faces a twofold challenge: spatial and time unsynchronizations among multi-directional coupling simulation of subsystems. A new collaborative simulation method with spatiotemporal synchronization process control is proposed for coupling simulating a given complex mechatronics system across multiple subsystems on different platforms. The method consists of 1) a coupler-based coupling mechanisms to define the interfacing and interaction mechanisms among subsystems, and 2) a simulation process control algorithm to realize the coupling simulation in a spatiotemporal synchronized manner. The test results from a case study show that the proposed method 1) can certainly be used to simulate the sub-systems interactions under different simulation conditions in an engineering system, and 2) effectively supports multi-directional coupling simulation among multi-disciplinary subsystems. This method has been successfully applied in China high speed train design and development processes, demonstrating that it can be applied in a wide range of engineering systems design and simulation with improved efficiency and effectiveness

    A Robust and Powerful Set-Valued Approach to Rare Variant Association Analyses of Secondary Traits in Case-Control Sequencing Studies

    Get PDF
    In many case-control designs of genome-wide association (GWAS) or next generation sequencing (NGS) studies, extensive data on secondary traits that may correlate and share the common genetic variants with the primary disease are available. Investigating these secondary traits can provide critical insights into the disease etiology or pathology, and enhance the GWAS or NGS results. Methods based on logistic regression (LG) were developed for this purpose. However, for the identification of rare variants (RVs), certain inadequacies in the LG models and algorithmic instability can cause severely inflated type I error, and significant loss of power, when the two traits are correlated and the RV is associated with the disease, especially at stringent significance levels. To address this issue, we propose a novel set-valued (SV) method that models a binary trait by dichotomization of an underlying continuous variable, and incorporate this into the genetic association model as a critical component. Extensive simulations and an analysis of seven secondary traits in a GWAS of benign ethnic neutropenia show that the SV method consistently controls type I error well at stringent significance levels, has larger power than the LG-based methods, and is robust in performance to effect pattern of the genetic variant (risk or protective), rare or common variants, rare or common diseases, and trait distributions. Because of the SV method’s striking and profound advantage, we strongly recommend the SV method be employed instead of the LG-based methods for secondary traits analyses in case-control sequencing studies

    Machine Learning for Prediction of Sudden Cardiac Death in Heart Failure Patients With Low Left Ventricular Ejection Fraction: Study Protocol for a Retrospective Multicentre Registry in China

    Get PDF
    Introduction: Left ventricular ejection fraction (LVEF) ≤35%, as current significant implantable cardioverter-defibrillator (ICD) indication for primary prevention of sudden cardiac death (SCD) in heart failure (HF) patients, has been widely recognised to be inefficient. Improvement of patient selection for low LVEF (≤35%) is needed to optimise deployment of ICD. Most of the existing prediction models are not appropriate to identify ICD candidates at high risk of SCD in HF patients with low LVEF. Compared with traditional statistical analysis, machine learning (ML) can employ computer algorithms to identify patterns in large datasets, analyse rules automatically and build both linear and non-linear models in order to make data-driven predictions. This study is aimed to develop and validate new models using ML to improve the prediction of SCD in HF patients with low LVEF. Methods and analysis: We will conduct a retroprospective, multicentre, observational registry of Chinese HF patients with low LVEF. The HF patients with LVEF ≤35% after optimised medication at least 3 months will be enrolled in this study. The primary endpoints are all-cause death and SCD. The secondary endpoints are malignant arrhythmia, sudden cardiac arrest, cardiopulmonary resuscitation and rehospitalisation due to HF. The baseline demographic, clinical, biological, electrophysiological, social and psychological variables will be collected. Both ML and traditional multivariable Cox proportional hazards regression models will be developed and compared in the prediction of SCD. Moreover, the ML model will be validated in a prospective study. Ethics and dissemination: The study protocol has been approved by the Ethics Committee of the First Affiliated Hospital of Nanjing Medical University (2017-SR-06). All results of this study will be published in international peer-reviewed journals and presented at relevant conferences
    • …
    corecore