194,352 research outputs found

    Bagged ensemble of Fuzzy C-Means classifiers for nuclear transient identification

    Get PDF
    This paper presents an ensemble-based scheme for nuclear transient identification. The approach adopted to construct the ensemble of classifiers is bagging; the novelty consists in using supervised fuzzy C-means (FCM) classifiers as base classifiers of the ensemble. The performance of the proposed classification scheme has been verified by comparison with a single supervised, evolutionary-optimized FCM classifier with respect of the task of classifying artificial datasets. The results obtained indicate that in the cases of datasets of large or very small sizes and/or complex decision boundaries, the bagging ensembles can improve classification accuracy. Then, the approach has been applied to the identification of simulated transients in the feedwater system of a boiling water reactor (BWR)

    Bagged ensemble of Fuzzy C-Means classifiers for nuclear transient identification

    Get PDF
    This paper presents an ensemble-based scheme for nuclear transient identification. The approach adopted to construct the ensemble of classifiers is bagging; the novelty consists in using supervised fuzzy C-means (FCM) classifiers as base classifiers of the ensemble. The performance of the proposed classification scheme has been verified by comparison with a single supervised, evolutionary-optimized FCM classifier with respect of the task of classifying artificial datasets. The results obtained indicate that in the cases of datasets of large or very small sizes and/or complex decision boundaries, the bagging ensembles can improve classification accuracy. Then, the approach has been applied to the identification of simulated transients in the feedwater system of a boiling water reactor (BWR)

    Classification of multiwavelength transients with machine learning

    Get PDF
    With the advent of powerful telescopes such as the Square Kilometre Array (SKA), its precursor MeerKAT and the Large Synoptic Survey Telescope (LSST), we are entering a golden era of multiwavelength transient astronomy. The large MeerKAT science project ThunderKAT may dramatically increase the detected number of radio transients. Currently radio transient datasets are still very small, allowing spectroscopic classification of all objects of interest. As the event rate increases, follow-up resources must be prioritised by making use of early classification of the radio data. Machine learning algorithms have proven themselves invaluable in the context of optical astronomy, however it has yet to be applied to radio transients. In the burgeoning era of multimessenger astronomy, incorporating data from different telescopes such as MeerLICHT, Fermi, LSST and the gravitational wave observatory LIGO could significantly improve classification of events. Here we present MALT (Machine Learning for Transients): a general machine learning pipeline for multiwavelength transient classification. In order to make use of most machine learning algorithms, "features" must be extracted from complex and often high dimensional datasets. In our approach, we first interpolate the data onto a uniform grid using Gaussian processes, we then perform a wavelet decomposition and finally reduce the dimensionality using principal component analysis. We then classify the light curves with the popular machine learning algorithm random forests. For the first time, we apply machine learning to the classification of radio transients. Unfortunately publicly available radio transient data is scarce and our dataset consists of just 87 light curves, with several classes only consisting of a single example. However machine learning is often applied to such small datasets by making use of data augmentation. We develop a novel data augmentation technique based on Gaussian processes, able to generate new data statistically consistent with the original. As the dataset is currently small, three studies were done on the effect of the training set. The classifier was trained on a non-representative training set, achieving an overall accuracy of 77.8% over all 11 classes with the known 87 lightcurves with just eight hours of observations. The expected increase in performance, as more training data are acquired, is shown by training the classifier on a simulated representative training set, achieving an average accuracy of 95.8% across all 11 classes. Finally, the effectiveness of including multiwavelength data for general transient classification is demonstrated. First the classifier is trained on wavelet features and a contextual feature, achieving an average accuracy of 72.9%. The classifier was then trained on wavelet features and a contextual feature, together with a single optical flux feature. This addition improves the overall accuracy to 94.7%. This work provides a general approach for multiwavelength transient classification and shows that machine learning can be highly effective at classifying the influx of radio transients anticipated with MeerKAT and other radio telescopes

    Data Dropout: Optimizing Training Data for Convolutional Neural Networks

    Full text link
    Deep learning models learn to fit training data while they are highly expected to generalize well to testing data. Most works aim at finding such models by creatively designing architectures and fine-tuning parameters. To adapt to particular tasks, hand-crafted information such as image prior has also been incorporated into end-to-end learning. However, very little progress has been made on investigating how an individual training sample will influence the generalization ability of a model. In other words, to achieve high generalization accuracy, do we really need all the samples in a training dataset? In this paper, we demonstrate that deep learning models such as convolutional neural networks may not favor all training samples, and generalization accuracy can be further improved by dropping those unfavorable samples. Specifically, the influence of removing a training sample is quantifiable, and we propose a Two-Round Training approach, aiming to achieve higher generalization accuracy. We locate unfavorable samples after the first round of training, and then retrain the model from scratch with the reduced training dataset in the second round. Since our approach is essentially different from fine-tuning or further training, the computational cost should not be a concern. Our extensive experimental results indicate that, with identical settings, the proposed approach can boost performance of the well-known networks on both high-level computer vision problems such as image classification, and low-level vision problems such as image denoising
    corecore