345 research outputs found

    A randomized neural network for data streams

    Get PDF
    © 2017 IEEE. Randomized neural network (RNN) is a highly feasible solution in the era of big data because it offers a simple and fast working principle in processing dynamic and evolving data streams. This paper proposes a novel RNN, namely recurrent type-2 random vector functional link network (RT2McRVFLN), which provides a highly scalable solution for data streams in a strictly online and integrated framework. It is built upon the psychologically inspired concept of metacognitive learning, which covers three basic components of human learning: what-to-learn, how-to-learn, and when-to-learn. The what-to-learn selects important samples on the fly with the use of online active learning scenario, which renders our algorithm an online semi-supervised algorithm. The how-to-learn process combines an open structure of evolving concept and a randomized learning algorithm of random vector functional link network (RVFLN). The efficacy of the RT2McRVFLN has been numerically validated through two real-world case studies and comparisons with its counterparts, which arrive at a conclusive finding that our algorithm delivers a tradeoff between accuracy and simplicity

    An Optimized Recursive General Regression Neural Network Oracle for the Prediction and Diagnosis of Diabetes

    Get PDF
    Diabetes is a serious, chronic disease that has been seeing a rise in the number of cases and prevalence over the past few decades. It can lead to serious complications and can increase the overall risk of dying prematurely. Data-oriented prediction models have become effective tools that help medical decision-making and diagnoses in which the use of machine learning in medicine has increased substantially. This research introduces the Recursive General Regression Neural Network Oracle (R-GRNN Oracle) and is applied on the Pima Indians Diabetes dataset for the prediction and diagnosis of diabetes. The R-GRNN Oracle (Bani-Hani, 2017) is an enhancement to the GRNN Oracle developed by Masters et al. in 1998, in which the recursive model is created of two oracles: one within the other. Several classifiers, along with the R-GRNN Oracle and the GRNN Oracle, are applied to the dataset, they are: Support Vector Machine (SVM), Multilayer Perceptron (MLP), Probabilistic Neural Network (PNN), Gaussian NaEF;ve Bayes (GNB), K-Nearest Neighbor (KNN), and Random Forest (RF). Genetic Algorithm (GA) was used for feature selection as well as the hyperparameter optimization of SVM and MLP, and Grid Search (GS) was used to optimize the hyperparameters of KNN and RF. The performance metrics accuracy, AUC, sensitivity, and specificity were recorded for each classifier. The R-GRNN Oracle was able to achieve the highest accuracy, AUC, and sensitivity (81.14%, 86.03%, and 63.80%, respectively), while the optimized MLP had the highest specificity (89.71%)

    Autoregressive time series prediction by means of fuzzy inference systems using nonparametric residual variance estimation

    Get PDF
    We propose an automatic methodology framework for short- and long-term prediction of time series by means of fuzzy inference systems. In this methodology, fuzzy techniques and statistical techniques for nonparametric residual variance estimation are combined in order to build autoregressive predictive models implemented as fuzzy inference systems. Nonparametric residual variance estimation plays a key role in driving the identification and learning procedures. Concrete criteria and procedures within the proposed methodology framework are applied to a number of time series prediction problems. The learn from examples method introduced by Wang and Mendel (W&M) is used for identification. The Levenberg–Marquardt (L–M) optimization method is then applied for tuning. The W&M method produces compact and potentially accurate inference systems when applied after a proper variable selection stage. The L–M method yields the best compromise between accuracy and interpretability of results, among a set of alternatives. Delta test based residual variance estimations are used in order to select the best subset of inputs to the fuzzy inference systems as well as the number of linguistic labels for the inputs. Experiments on a diverse set of time series prediction benchmarks are compared against least-squares support vector machines (LS-SVM), optimally pruned extreme learning machine (OP-ELM), and k-NN based autoregressors. The advantages of the proposed methodology are shown in terms of linguistic interpretability, generalization capability and computational cost. Furthermore, fuzzy models are shown to be consistently more accurate for prediction in the case of time series coming from real-world applications.Ministerio de Ciencia e Innovación TEC2008-04920Junta de Andalucía P08-TIC-03674, IAC07-I-0205:33080, IAC08-II-3347:5626

    LARSEN-ELM: Selective Ensemble of Extreme Learning Machines using LARS for Blended Data

    Get PDF
    Extreme learning machine (ELM) as a neural network algorithm has shown its good performance, such as fast speed, simple structure etc, but also, weak robustness is an unavoidable defect in original ELM for blended data. We present a new machine learning framework called LARSEN-ELM for overcoming this problem. In our paper, we would like to show two key steps in LARSEN-ELM. In the first step, preprocessing, we select the input variables highly related to the output using least angle regression (LARS). In the second step, training, we employ Genetic Algorithm (GA) based selective ensemble and original ELM. In the experiments, we apply a sum of two sines and four datasets from UCI repository to verify the robustness of our approach. The experimental results show that compared with original ELM and other methods such as OP-ELM, GASEN-ELM and LSBoost, LARSEN-ELM significantly improve robustness performance while keeping a relatively high speed.Comment: Accepted for publication in Neurocomputing, 01/19/201

    Quantifying the reliability of fault classifiers

    No full text
    International audienceFault diagnostics problems can be formulated as classification tasks. Due to limited data and to uncertainty, classification algorithms are not perfectly accurate in practical applications. Maintenance decisions based on erroneous fault classifications result in inefficient resource allocations and/or operational disturbances. Thus, knowing the accuracy of classifiers is important to give confidence in the maintenance decisions. The average accuracy of a classifier on a test set of data patterns is often used as a measure of confidence in the performance of a specific classifier. However, the performance of a classifier can vary in different regions of the input data space. Several techniques have been proposed to quantify the reliability of a classifier at the level of individual classifications. Many of the proposed techniques are only applicable to specific classifiers, such as ensemble techniques and support vector machines. In this paper, we propose a meta approach based on the typicalness framework (Kolmogorov's concept of randomness), which is independent of the applied classifier. We apply the approach to a case of fault diagnosis in railway turnout systems and compare the results obtained with both extreme learning machines and echo state networks

    Load forecast on a Micro Grid level through Machine Learning algorithms

    Get PDF
    As Micro Redes constituem um sector em crescimento da indústria energética, representando uma mudança de paradigma, desde as remotas centrais de geração até à produção mais localizada e distribuída. A capacidade de isolamento das principais redes elétricas e atuar de forma independente tornam as Micro Redes em sistemas resilientes, capazes de conduzir operações flexíveis em paralelo com a prestação de serviços que tornam a rede mais competitiva. Como tal, as Micro Redes fornecem energia limpa eficiente de baixo custo, aprimoram a coordenação dos ativos e melhoram a operação e estabilidade da rede regional de eletricidade, através da capacidade de resposta dinâmica aos recursos energéticos. Para isso, necessitam de uma coordenação de gestão inteligente que equilibre todas as tecnologias ao seu dispor. Daqui surge a necessidade de recorrer a modelos de previsão de carga e de produção robustos e de confiança, que interligam a alocação dos recursos da rede perante as necessidades emergentes. Sendo assim, foi desenvolvida a metodologia HALOFMI, que tem como principal objetivo a criação de um modelo de previsão de carga para 24 horas. A metodologia desenvolvida é constituída, numa primeira fase, por uma abordagem híbrida de multinível para a criação e escolha de atributos, que alimenta uma rede neuronal (Multi-Layer Perceptron) sujeita a um ajuste de híper-parâmetros. Posto isto, numa segunda fase são testados dois modos de aplicação e gestão de dados para a Micro Rede. A metodologia desenvolvida é aplicada em dois casos de estudo: o primeiro é composto por perfis de carga agregados correspondentes a dados de clientes em Baixa Tensão Normal e de Unidades de Produção e Autoconsumo (UPAC). Este caso de estudo apresenta-se como um perfil de carga elétrica regular e com contornos muito suaves. O segundo caso de estudo diz respeito a uma ilha turística e representa um perfil irregular de carga, com variações bruscas e difíceis de prever e apresenta um desafio maior em termos de previsão a 24-horas A partir dos resultados obtidos, é avaliado o impacto da integração de uma seleção recursiva inteligente de atributos, seguido por uma viabilização do processo de redução da dimensão de dados para o operador da Micro Rede, e por fim uma comparação de estimadores usados no modelo de previsão, através de medidores de erros na performance do algoritmo.Micro Grids constitute a growing sector of the energetic industry, representing a paradigm shift from the central power generation plans to a more distributed generation. The capacity to work isolated from the main electric grid make the MG resilient system, capable of conducting flexible operations while providing services that make the network more competitive. Additionally, Micro Grids supply clean and efficient low-cost energy, enhance the flexible assets coordination and improve the operation and stability of the of the local electric grid, through the capability of providing a dynamic response to the energetic resources. For that, it is required an intelligent coordination which balances all the available technologies. With this, rises the need to integrate accurate and robust load and production forecasting models into the MG management platform, thus allowing a more precise coordination of the flexible resource according to the emerging demand needs. For these reasons, the HALOFMI methodology was developed, which focus on the creation of a precise 24-hour load forecast model. This methodology includes firstly, a hybrid multi-level approach for the creation and selection of features. Then, these inputs are fed to a Neural Network (Multi-Layer Perceptron) with hyper-parameters tuning. In a second phase, two ways of data operation are compared and assessed, which results in the viability of the network operating with a reduced number of training days without compromising the model's performance. Such process is attained through a sliding window application. Furthermore, the developed methodology is applied in two case studies, both with 15-minute timesteps: the first one is composed by aggregated load profiles of Standard Low Voltage clients, including production and self-consumption units. This case study presents regular and very smooth load profile curves. The second case study concerns a touristic island and represents an irregular load curve with high granularity with abrupt variations. From the attained results, it is evaluated the impact of integrating a recursive intelligent feature selection routine, followed by an assessment on the sliding window application and at last, a comparison on the errors coming from different estimators for the model, through several well-defined performance metrics

    Explainable online ensemble of deep neural network pruning for time series forecasting

    Get PDF
    Both the complex and evolving nature of time series data make forecasting among one of the most challenging tasks in machine learning. Typical methods for forecasting are designed to model time-evolving dependencies between data observations. However, it is generally accepted that none of them are universally valid for every application. Therefore, methods for learning heterogeneous ensembles by combining a diverse set of forecasters together appears as a promising solution to tackle this task. While several approaches in the context of time series forecasting have focused on how to combine individual models in an ensemble, ranging from simple and enhanced averaging tactics to applying meta-learning methods, few works have tackled the task of ensemble pruning, i.e. individual model selection to take part in the ensemble. In addition, in classical ML literature, ensemble pruning techniques are mostly restricted to operate in a static manner. To deal with changes in the relative performance of models as well as changes in the data distribution, we employ gradient-based saliency maps for online ensemble pruning of deep neural networks. This method consists of generating individual models’ performance saliency maps that are subsequently used to prune the ensemble by taking into account both aspects of accuracy and diversity. In addition, the saliency maps can be exploited to provide suitable explanations for the reason behind selecting specific models to construct an ensemble that plays the role of a forecaster at a certain time interval or instant. An extensive empirical study on many real-world datasets demonstrates that our method achieves excellent or on par results in comparison to the state-of-the-art approaches as well as several baselines. Our code is available on Github (https://github.com/MatthiasJakobs/os-pgsm/tree/ecml_journal_2022)

    Development of 2D Curve-Fitting Genetic/Gene-Expression Programming Technique for Efficient Time-series Financial Forecasting

    Get PDF
    Stock market prediction is of immense interest to trading companies and buyers due to high profit margins. Therefore, precise prediction of the measure of increase or decrease of stock prices also plays an important role in buying/selling activities. This research presents a specialised extension to the genetic algorithms (GA) known as the genetic programming (GP) and gene expression programming (GEP) to explore and investigate the outcome of the GEP criteria on the stock market price prediction. The research presented in this paper aims at the modelling and prediction of short-to-medium term stock value fluctuations in the market via genetically tuned stock market parameters. The technique uses hierarchically defined GP and GEP techniques to tune algebraic functions representing the fittest equation for stock market activities. The proposed methodology is evaluated against five well-known stock market companies with each having its own trading circumstances during the past 20+ years. The proposed GEP/GP methodologies were evaluated based on variable window/population sizes, selection methods, and Elitism, Rank and Roulette selection methods. The Elitism-based approach showed promising results with a low error-rate in the resultant pattern matching with an overall accuracy of 93.46% for short term 5-day and 92.105 for medium-term 56-day tradin
    corecore