185,061 research outputs found
Usage of Network Simulators in Machine-Learning-Assisted 5G/6G Networks
Without any doubt, Machine Learning (ML) will be an important driver of
future communications due to its foreseen performance when applied to complex
problems. However, the application of ML to networking systems raises concerns
among network operators and other stakeholders, especially regarding
trustworthiness and reliability. In this paper, we devise the role of network
simulators for bridging the gap between ML and communications systems. In
particular, we present an architectural integration of simulators in ML-aware
networks for training, testing, and validating ML models before being applied
to the operative network. Moreover, we provide insights on the main challenges
resulting from this integration, and then give hints discussing how they can be
overcome. Finally, we illustrate the integration of network simulators into
ML-assisted communications through a proof-of-concept testbed implementation of
a residential Wi-Fi network
Holistic Measures for Evaluating Prediction Models in Smart Grids
The performance of prediction models is often based on "abstract metrics"
that estimate the model's ability to limit residual errors between the observed
and predicted values. However, meaningful evaluation and selection of
prediction models for end-user domains requires holistic and
application-sensitive performance measures. Inspired by energy consumption
prediction models used in the emerging "big data" domain of Smart Power Grids,
we propose a suite of performance measures to rationally compare models along
the dimensions of scale independence, reliability, volatility and cost. We
include both application independent and dependent measures, the latter
parameterized to allow customization by domain experts to fit their scenario.
While our measures are generalizable to other domains, we offer an empirical
analysis using real energy use data for three Smart Grid applications:
planning, customer education and demand response, which are relevant for energy
sustainability. Our results underscore the value of the proposed measures to
offer a deeper insight into models' behavior and their impact on real
applications, which benefit both data mining researchers and practitioners.Comment: 14 Pages, 8 figures, Accepted and to appear in IEEE Transactions on
Knowledge and Data Engineering, 2014. Authors' final version. Copyright
transferred to IEE
On A Simpler and Faster Derivation of Single Use Reliability Mean and Variance for Model-Based Statistical Testing
Markov chain usage-based statistical testing has proved sound and effective in providing audit trails of evidence in certifying software-intensive systems. The system end-toend reliability is derived analytically in closed form, following an arc-based Bayesian model. System reliability is represented by an important statistic called single use reliability, and defined as the probability of a randomly selected use being successful. This paper continues our earlier work on a simpler and faster derivation of the single use reliability mean, and proposes a new derivation of the single use reliability variance by applying a well-known theorem and eliminating the need to compute the second moments of arc
failure probabilities. Our new results complete a new analysis that could be shown to be simpler, faster, and more direct while also rendering a more intuitive explanation. Our new
theory is illustrated with three simple Markov chain usage models with manual derivations and experimental results
The Art of Fault Injection
Classical greek philosopher considered the foremost virtues to be temperance, justice, courage, and prudence. In this paper we relate these cardinal virtues to the correct methodological approaches that researchers should follow when setting up a fault injection experiment. With this work we try to understand where the "straightforward pathway" lies, in order to highlight those common methodological errors that deeply influence the coherency and the meaningfulness of fault injection experiments. Fault injection is like an art, where the success of the experiments depends on a very delicate balance between modeling, creativity, statistics, and patience
Evaluation of A Resilience Embedded System Using Probabilistic Model-Checking
If a Micro Processor Unit (MPU) receives an external electric signal as
noise, the system function will freeze or malfunction easily. A new resilience
strategy is implemented in order to reset the MPU automatically and stop the
MPU from freezing or malfunctioning. The technique is useful for embedded
systems which work in non-human environments. However, evaluating resilience
strategies is difficult because their effectiveness depends on numerous,
complex, interacting factors.
In this paper, we use probabilistic model checking to evaluate the embedded
systems installed with the above mentioned new resilience strategy. Qualitative
evaluations are implemented with 6 PCTL formulas, and quantitative evaluations
use two kinds of evaluation. One is system failure reduction, and the other is
ADT (Average Down Time), the industry standard. Our work demonstrates the
benefits brought by the resilience strategy. Experimental results indicate that
our evaluation is cost-effective and reliable.Comment: In Proceedings ESSS 2014, arXiv:1405.055
A Quality Model for Actionable Analytics in Rapid Software Development
Background: Accessing relevant data on the product, process, and usage
perspectives of software as well as integrating and analyzing such data is
crucial for getting reliable and timely actionable insights aimed at
continuously managing software quality in Rapid Software Development (RSD). In
this context, several software analytics tools have been developed in recent
years. However, there is a lack of explainable software analytics that software
practitioners trust. Aims: We aimed at creating a quality model (called
Q-Rapids quality model) for actionable analytics in RSD, implementing it, and
evaluating its understandability and relevance. Method: We performed workshops
at four companies in order to determine relevant metrics as well as product and
process factors. We also elicited how these metrics and factors are used and
interpreted by practitioners when making decisions in RSD. We specified the
Q-Rapids quality model by comparing and integrating the results of the four
workshops. Then we implemented the Q-Rapids tool to support the usage of the
Q-Rapids quality model as well as the gathering, integration, and analysis of
the required data. Afterwards we installed the Q-Rapids tool in the four
companies and performed semi-structured interviews with eight product owners to
evaluate the understandability and relevance of the Q-Rapids quality model.
Results: The participants of the evaluation perceived the metrics as well as
the product and process factors of the Q-Rapids quality model as
understandable. Also, they considered the Q-Rapids quality model relevant for
identifying product and process deficiencies (e.g., blocking code situations).
Conclusions: By means of heterogeneous data sources, the Q-Rapids quality model
enables detecting problems that take more time to find manually and adds
transparency among the perspectives of system, process, and usage.Comment: This is an Author's Accepted Manuscript of a paper to be published by
IEEE in the 44th Euromicro Conference on Software Engineering and Advanced
Applications (SEAA) 2018. The final authenticated version will be available
onlin
Models and metrics for software management and engineering
This paper attempts to characterize and present a state of the art view of several quantitative models and metrics of the software life cycle. These models and metrics can be used to aid in managing and engineering software projects. They deal with various aspects of the software process and product, including resources allocation and estimation, changes and errors, size, complexity and reliability. Some indication is given of the extent to which the various models have been used and the success they have achieved
- …