31,277 research outputs found

    Handling Concept Drift for Predictions in Business Process Mining

    Get PDF
    Predictive services nowadays play an important role across all business sectors. However, deployed machine learning models are challenged by changing data streams over time which is described as concept drift. Prediction quality of models can be largely influenced by this phenomenon. Therefore, concept drift is usually handled by retraining of the model. However, current research lacks a recommendation which data should be selected for the retraining of the machine learning model. Therefore, we systematically analyze different data selection strategies in this work. Subsequently, we instantiate our findings on a use case in process mining which is strongly affected by concept drift. We can show that we can improve accuracy from 0.5400 to 0.7010 with concept drift handling. Furthermore, we depict the effects of the different data selection strategies

    Comparative analysis of clustering-based remaining-time predictive process monitoring approaches

    Get PDF
    Predictive process monitoring aims to accurately predict a variable of interest (e.g. remaining time) or the future state of the process instance (e.g. outcome or next step). Various studies have been explored to develop models with greater predictive power. However, comparing the various studies is difficult as different datasets, parameters and evaluation measures have been used. This paper seeks to address this problem with a focus on studies that adopt a clustering-based approach to predict the remaining time to the end of the process instance. A systematic literature review is undertaken to identify existing studies that adopt a clustering-based remaining-time predictive process monitoring approach and performs a comparative analysis to compare and benchmark the output of the identified studies using five real-life event logs

    Business Process Models in the Context of Predictive Process Monitoring

    Get PDF
    Predictive process monitoring is a subject of growing interest in academic research. As a result, an increased number of papers on this topic have been published. Due to the high complexity in this research area a wide range of different experimental setups and methods have been applied which makes it very difficult to reliably compare research results. This paper's objective is to investigate how business process models and their characteristics are used during experimental setups and how they can contribute to academic research. First, a literature review is conducted to analyze and discuss the awareness of business process models in experimental setups. Secondly, the paper discusses identified research problems and proposes the concept of a web-based business process model metric suite and the idea of ranked metrics. Through a metric suite researchers and practitioners can automatically evaluate business process model characteristics in their future work. Further, a contextualization of metrics by introducing a ranking of characteristics can potentially indicate how the outcome of experimental setups will be. Hence, the paper's work demonstrates the importance of business process models and their characteristics in the context of predictive process monitoring and proposes the concept of a tool approach and ranking to reliably evaluate business process models characteristics

    Stream Learning in Energy IoT Systems: A Case Study in Combined Cycle Power Plants

    Get PDF
    The prediction of electrical power produced in combined cycle power plants is a key challenge in the electrical power and energy systems field. This power production can vary depending on environmental variables, such as temperature, pressure, and humidity. Thus, the business problem is how to predict the power production as a function of these environmental conditions, in order to maximize the profit. The research community has solved this problem by applying Machine Learning techniques, and has managed to reduce the computational and time costs in comparison with the traditional thermodynamical analysis. Until now, this challenge has been tackled from a batch learning perspective, in which data is assumed to be at rest, and where models do not continuously integrate new information into already constructed models. We present an approach closer to the Big Data and Internet of Things paradigms, in which data are continuously arriving and where models learn incrementally, achieving significant enhancements in terms of data processing (time, memory and computational costs), and obtaining competitive performances. This work compares and examines the hourly electrical power prediction of several streaming regressors, and discusses about the best technique in terms of time processing and predictive performance to be applied on this streaming scenario.This work has been partially supported by the EU project iDev40. This project has received funding from the ECSEL Joint Undertaking (JU) under grant agreement No 783163. The JU receives support from the European Union’s Horizon 2020 research and innovation programme and Austria, Germany, Belgium, Italy, Spain, Romania. It has also been supported by the Basque Government (Spain) through the project VIRTUAL (KK-2018/00096), and by Ministerio de Economía y Competitividad of Spain (Grant Ref. TIN2017-85887-C2-2-P)

    Do Google Searches Help in Nowcasting Private Consumption?: A Real-Time Evidence for the US

    Get PDF
    In this paper, we investigate whether the Google search activity can help in nowcasting the year-on-year growth rates of monthly US private consumption using a real-time data set. The Google-based forecasts are compared to those based on a benchmark AR(1) model and the models including the consumer surveys and financial indicators. According to the Diebold-Mariano test of equal predictive ability, the null hypothesis can be rejected suggesting that Google-based forecasts are significantly more accurate than those of the benchmark model. At the same time, the corresponding null hypothesis cannot be rejected for models with consumer surveys and financial variables. Moreover, when we apply the test of superior predictive ability (Hansen, 2005) that controls for possible data-snooping biases, we are able to reject the null hypothesis that the benchmark model is not inferior to any alternative model forecasts. Furthermore, the results of the model confidence set (MCS) procedure (Hansen et al., 2005) suggest that the autoregressive benchmark is not selected into a set of the best forecasting models. Apart from several Google-based models, the MCS contains also some models including survey-based indicators and financial variables. We conclude that Google searches do help improving the nowcasts of the private consumption in US.Google indicators, real-time nowcasting, principal components, US private consumption

    Exploring Interpretability for Predictive Process Analytics

    Full text link
    Modern predictive analytics underpinned by machine learning techniques has become a key enabler to the automation of data-driven decision making. In the context of business process management, predictive analytics has been applied to making predictions about the future state of an ongoing business process instance, for example, when will the process instance complete and what will be the outcome upon completion. Machine learning models can be trained on event log data recording historical process execution to build the underlying predictive models. Multiple techniques have been proposed so far which encode the information available in an event log and construct input features required to train a predictive model. While accuracy has been a dominant criterion in the choice of various techniques, they are often applied as a black-box in building predictive models. In this paper, we derive explanations using interpretable machine learning techniques to compare and contrast the suitability of multiple predictive models of high accuracy. The explanations allow us to gain an understanding of the underlying reasons for a prediction and highlight scenarios where accuracy alone may not be sufficient in assessing the suitability of techniques used to encode event log data to features used by a predictive model. Findings from this study motivate the need and importance to incorporate interpretability in predictive process analytics.Comment: 15 pages, 7 figure

    Modeling Financial Time Series with Artificial Neural Networks

    Full text link
    Financial time series convey the decisions and actions of a population of human actors over time. Econometric and regressive models have been developed in the past decades for analyzing these time series. More recently, biologically inspired artificial neural network models have been shown to overcome some of the main challenges of traditional techniques by better exploiting the non-linear, non-stationary, and oscillatory nature of noisy, chaotic human interactions. This review paper explores the options, benefits, and weaknesses of the various forms of artificial neural networks as compared with regression techniques in the field of financial time series analysis.CELEST, a National Science Foundation Science of Learning Center (SBE-0354378); SyNAPSE program of the Defense Advanced Research Project Agency (HR001109-03-0001
    corecore