5,270 research outputs found

    Uncertainty Quantification of Future Design Rainfall Depths in Korea

    Get PDF
    One of the most common ways to investigate changes in future rainfall extremes is to use future rainfall data simulated by climate models with climate change scenarios. However, the projected future design rainfall intensity varies greatly depending on which climate model is applied. In this study, future rainfall Intensity???Duration???Frequency (IDF) curves are projected using various combinations of climate models. Future Ensemble Average (FEA) is calculated using a total of 16 design rainfall intensity ensembles, and uncertainty of FEA is quantified using the coefficient of variation of ensembles. The FEA and its uncertainty vary widely depending on how the climate model combination is constructed, and the uncertainty of the FEA depends heavily on the inclusion of specific climate model combinations at each site. In other words, we found that unconditionally using many ensemble members did not help to reduce the uncertainty of future IDF curves. Finally, a method for constructing ensemble members that reduces the uncertainty of future IDF curves is proposed, which will contribute to minimizing confusion among policy makers in developing climate change adaptation policies

    Non-uniform Feature Sampling for Decision Tree Ensembles

    Full text link
    We study the effectiveness of non-uniform randomized feature selection in decision tree classification. We experimentally evaluate two feature selection methodologies, based on information extracted from the provided dataset: (i)(i) \emph{leverage scores-based} and (ii)(ii) \emph{norm-based} feature selection. Experimental evaluation of the proposed feature selection techniques indicate that such approaches might be more effective compared to naive uniform feature selection and moreover having comparable performance to the random forest algorithm [3]Comment: 7 pages, 7 figures, 1 tabl

    HoloDetect: Few-Shot Learning for Error Detection

    Full text link
    We introduce a few-shot learning framework for error detection. We show that data augmentation (a form of weak supervision) is key to training high-quality, ML-based error detection models that require minimal human involvement. Our framework consists of two parts: (1) an expressive model to learn rich representations that capture the inherent syntactic and semantic heterogeneity of errors; and (2) a data augmentation model that, given a small seed of clean records, uses dataset-specific transformations to automatically generate additional training data. Our key insight is to learn data augmentation policies from the noisy input dataset in a weakly supervised manner. We show that our framework detects errors with an average precision of ~94% and an average recall of ~93% across a diverse array of datasets that exhibit different types and amounts of errors. We compare our approach to a comprehensive collection of error detection methods, ranging from traditional rule-based methods to ensemble-based and active learning approaches. We show that data augmentation yields an average improvement of 20 F1 points while it requires access to 3x fewer labeled examples compared to other ML approaches.Comment: 18 pages

    Ensemble Reinforcement Learning: A Survey

    Full text link
    Reinforcement Learning (RL) has emerged as a highly effective technique for addressing various scientific and applied problems. Despite its success, certain complex tasks remain challenging to be addressed solely with a single model and algorithm. In response, ensemble reinforcement learning (ERL), a promising approach that combines the benefits of both RL and ensemble learning (EL), has gained widespread popularity. ERL leverages multiple models or training algorithms to comprehensively explore the problem space and possesses strong generalization capabilities. In this study, we present a comprehensive survey on ERL to provide readers with an overview of recent advances and challenges in the field. First, we introduce the background and motivation for ERL. Second, we analyze in detail the strategies that have been successfully applied in ERL, including model averaging, model selection, and model combination. Subsequently, we summarize the datasets and analyze algorithms used in relevant studies. Finally, we outline several open questions and discuss future research directions of ERL. By providing a guide for future scientific research and engineering applications, this survey contributes to the advancement of ERL.Comment: 42 page
    corecore