4,323 research outputs found

    Comparative Evaluation of Log-Based Process Performance Analysis Techniques

    Get PDF
    Käesolev magistritöö võrdleb erinevaid protsessikaeve uurimustöid ning liigitab neid järgnevate näitajate põhjal: aeg, kvaliteet ja ressursikasutus. Magistritöö põhjendab protsessikaeve meetodite kasutamist ja nende pakutavat lisandväärtust. Pakume ühiseid mõõtühikuid ja parameetreid, mida saab kasutada protsessi tulemuslikkuse analüüsimeetodite hindamiseks. Lisaks eelnevale kirjeldab lõputöö kirjanduses esinevaid tarkvaralahendusi ja algoritme.This paper gives a comparative overview of process mining performance studies and clusters them based on proposed metrics: time, quality and resources. This thesis provides an explanation of reasons for using process mining performance techniques and shows what value they can bring. We provide common metrics and unit of measurement that can be used to evaluate process performance analysis methods. Also, the paper describes tools and algorithms that have been implemented in the literature

    Limit-order completion time in the London stock market

    Get PDF
    This study develops an econometric model of limit-order completion time using survival analysis. Time-to-completion for both buy and sell limit orders is estimated using tick-by-tick UK order data. The study investigates the explanatory power of variables that measure order characteristics and market conditions, such as the limitorder price, limit-order size, best bid-offer spread, and market volatility. The generic results show that limit-order completion time depends on some variables more than on others. This study also provides an investigation of how the dynamics of the market are incorporated into models of limit-order completion. The empirical results show that time-varying variables capture the state of an order book in a better way than static ones. Moreover, this study provides an examination of the prediction accuracy of the proposed models. In addition, this study provides an investigation of the intra-day pattern of order submission and time-of-day effects on limit-order completion time in the UK market. It provides evidence showing that limit orders placed in the afternoon period are expected to have the shortest completion times while orders placed in the mid-day period are expected to have the longest completion times, and the sensitivities of limit-order completion time to the explanatory variables vary over the trading day

    Experimental analysis of computer system dependability

    Get PDF
    This paper reviews an area which has evolved over the past 15 years: experimental analysis of computer system dependability. Methodologies and advances are discussed for three basic approaches used in the area: simulated fault injection, physical fault injection, and measurement-based analysis. The three approaches are suited, respectively, to dependability evaluation in the three phases of a system's life: design phase, prototype phase, and operational phase. Before the discussion of these phases, several statistical techniques used in the area are introduced. For each phase, a classification of research methods or study topics is outlined, followed by discussion of these methods or topics as well as representative studies. The statistical techniques introduced include the estimation of parameters and confidence intervals, probability distribution characterization, and several multivariate analysis methods. Importance sampling, a statistical technique used to accelerate Monte Carlo simulation, is also introduced. The discussion of simulated fault injection covers electrical-level, logic-level, and function-level fault injection methods as well as representative simulation environments such as FOCUS and DEPEND. The discussion of physical fault injection covers hardware, software, and radiation fault injection methods as well as several software and hybrid tools including FIAT, FERARI, HYBRID, and FINE. The discussion of measurement-based analysis covers measurement and data processing techniques, basic error characterization, dependency analysis, Markov reward modeling, software-dependability, and fault diagnosis. The discussion involves several important issues studies in the area, including fault models, fast simulation techniques, workload/failure dependency, correlated failures, and software fault tolerance

    On the Inability of Markov Models to Capture Criticality in Human Mobility

    Get PDF
    We examine the non-Markovian nature of human mobility by exposing the inability of Markov models to capture criticality in human mobility. In particular, the assumed Markovian nature of mobility was used to establish a theoretical upper bound on the predictability of human mobility (expressed as a minimum error probability limit), based on temporally correlated entropy. Since its inception, this bound has been widely used and empirically validated using Markov chains. We show that recurrent-neural architectures can achieve significantly higher predictability, surpassing this widely used upper bound. In order to explain this anomaly, we shed light on several underlying assumptions in previous research works that has resulted in this bias. By evaluating the mobility predictability on real-world datasets, we show that human mobility exhibits scale-invariant long-range correlations, bearing similarity to a power-law decay. This is in contrast to the initial assumption that human mobility follows an exponential decay. This assumption of exponential decay coupled with Lempel-Ziv compression in computing Fano's inequality has led to an inaccurate estimation of the predictability upper bound. We show that this approach inflates the entropy, consequently lowering the upper bound on human mobility predictability. We finally highlight that this approach tends to overlook long-range correlations in human mobility. This explains why recurrent-neural architectures that are designed to handle long-range structural correlations surpass the previously computed upper bound on mobility predictability

    Malware in the Future? Forecasting of Analyst Detection of Cyber Events

    Full text link
    There have been extensive efforts in government, academia, and industry to anticipate, forecast, and mitigate cyber attacks. A common approach is time-series forecasting of cyber attacks based on data from network telescopes, honeypots, and automated intrusion detection/prevention systems. This research has uncovered key insights such as systematicity in cyber attacks. Here, we propose an alternate perspective of this problem by performing forecasting of attacks that are analyst-detected and -verified occurrences of malware. We call these instances of malware cyber event data. Specifically, our dataset was analyst-detected incidents from a large operational Computer Security Service Provider (CSSP) for the U.S. Department of Defense, which rarely relies only on automated systems. Our data set consists of weekly counts of cyber events over approximately seven years. Since all cyber events were validated by analysts, our dataset is unlikely to have false positives which are often endemic in other sources of data. Further, the higher-quality data could be used for a number for resource allocation, estimation of security resources, and the development of effective risk-management strategies. We used a Bayesian State Space Model for forecasting and found that events one week ahead could be predicted. To quantify bursts, we used a Markov model. Our findings of systematicity in analyst-detected cyber attacks are consistent with previous work using other sources. The advanced information provided by a forecast may help with threat awareness by providing a probable value and range for future cyber events one week ahead. Other potential applications for cyber event forecasting include proactive allocation of resources and capabilities for cyber defense (e.g., analyst staffing and sensor configuration) in CSSPs. Enhanced threat awareness may improve cybersecurity.Comment: Revised version resubmitted to journa

    Modeling Customer Experience in a Contact Center through Process Log Mining

    Get PDF
    The use of data mining and modeling methods in service industry is a promising avenue for optimizing current processes in a targeted manner, ultimately reducing costs and improving customer experience. However, the introduction of such tools in already established pipelines often must adapt to the way data is sampled and to its content. In this study, we tackle the challenge of characterizing and predicting customer experience having available only process log data with time-stamp information, without any ground truth feedback from the customers. As a case study, we consider the context of a contact center managed by TeleWare and analyze phone call logs relative to a two months span. We develop an approach to interpret the phone call process events registered in the logs and infer concrete points of improvement in the service management. Our approach is based on latent tree modeling and multi-class Naïve Bayes classification, which jointly allow us to infer a spectrum of customer experiences and test their predictability based on the current data sampling strategy. Moreover, such approach can overcome limitations in customer feedback collection and sharing across organizations, thus having wide applicability and being complementary to tools relying on more heavily constrained data

    Aligning observed and modeled behavior

    Get PDF

    Complexity in daily life – a 3D-visualization showing activity patterns in their contexts

    Get PDF
    This article attacks the difficulties to make well informed empirically grounded descriptions and analyses of everyday life activity patterns. At a first glance, everyday life seems to be very simple and everybody has experiences from it, but when we try to investigate it from a scientific perspective, its complexity is overwhelming. There are enormous variations in interests and activity patterns among individuals, between households and socio-economic groups in the population. Therefore, and in spite of good intentions, traditional methods and means to visualize and analyze often lead to over-simplifications. The aim of this article is to present a visualization method that might inspire social scientists to tackle the complexity of everyday life from a new angle, starting with a visual overview of the individual's time use in her daily life, subsequently aggregating to time use in her household, further at group and population levels without leaving the individual out of sight. Thereby variations and complexity might be treated as assets in the interpretation rather than obstacles. To exemplify the method we show how activities in a daily life project are distributed among household members and between men and women in a population.household division of labour, time-geography, 3D method, visualization, diaries, everyday life, activity patterns. Complexity in daily life – a 3D-visualization showing activity patterns in their contexts
    corecore