6,271 research outputs found
LSTM Networks for Data-Aware Remaining Time Prediction of Business Process Instances
Predicting the completion time of business process instances would be a very
helpful aid when managing processes under service level agreement constraints.
The ability to know in advance the trend of running process instances would
allow business managers to react in time, in order to prevent delays or
undesirable situations. However, making such accurate forecasts is not easy:
many factors may influence the required time to complete a process instance. In
this paper, we propose an approach based on deep Recurrent Neural Networks
(specifically LSTMs) that is able to exploit arbitrary information associated
to single events, in order to produce an as-accurate-as-possible prediction of
the completion time of running instances. Experiments on real-world datasets
confirm the quality of our proposal.Comment: Article accepted for publication in 2017 IEEE Symposium on Deep
Learning (IEEE DL'17) @ SSC
A deep learning integrated Lee-Carter model
In the field of mortality, the Lee–Carter based approach can be considered the milestone
to forecast mortality rates among stochastic models. We could define a “Lee–Carter model family”
that embraces all developments of this model, including its first formulation (1992) that remains the
benchmark for comparing the performance of future models. In the Lee–Carter model, the kt parameter,
describing the mortality trend over time, plays an important role about the future mortality behavior.
The traditional ARIMA process usually used to model kt shows evident limitations to describe the future
mortality shape. Concerning forecasting phase, academics should approach a more plausible way in
order to think a nonlinear shape of the projected mortality rates. Therefore, we propose an alternative
approach the ARIMA processes based on a deep learning technique. More precisely, in order to catch
the pattern of kt series over time more accurately, we apply a Recurrent Neural Network with a Long
Short-Term Memory architecture and integrate the Lee–Carter model to improve its predictive capacity.
The proposed approach provides significant performance in terms of predictive accuracy and also allow
for avoiding the time-chunks’ a priori selection. Indeed, it is a common practice among academics to
delete the time in which the noise is overflowing or the data quality is insufficient. The strength of
the Long Short-Term Memory network lies in its ability to treat this noise and adequately reproduce it
into the forecasted trend, due to its own architecture enabling to take into account significant long-term
patterns
- …