61,438 research outputs found
Early Warning Analysis for Social Diffusion Events
There is considerable interest in developing predictive capabilities for
social diffusion processes, for instance to permit early identification of
emerging contentious situations, rapid detection of disease outbreaks, or
accurate forecasting of the ultimate reach of potentially viral ideas or
behaviors. This paper proposes a new approach to this predictive analytics
problem, in which analysis of meso-scale network dynamics is leveraged to
generate useful predictions for complex social phenomena. We begin by deriving
a stochastic hybrid dynamical systems (S-HDS) model for diffusion processes
taking place over social networks with realistic topologies; this modeling
approach is inspired by recent work in biology demonstrating that S-HDS offer a
useful mathematical formalism with which to represent complex, multi-scale
biological network dynamics. We then perform formal stochastic reachability
analysis with this S-HDS model and conclude that the outcomes of social
diffusion processes may depend crucially upon the way the early dynamics of the
process interacts with the underlying network's community structure and
core-periphery structure. This theoretical finding provides the foundations for
developing a machine learning algorithm that enables accurate early warning
analysis for social diffusion events. The utility of the warning algorithm, and
the power of network-based predictive metrics, are demonstrated through an
empirical investigation of the propagation of political memes over social media
networks. Additionally, we illustrate the potential of the approach for
security informatics applications through case studies involving early warning
analysis of large-scale protests events and politically-motivated cyber
attacks
Analysis of Daily Streamflow Complexity by Kolmogorov Measures and Lyapunov Exponent
Analysis of daily streamflow variability in space and time is important for
water resources planning, development, and management. The natural variability
of streamflow is being complicated by anthropogenic influences and climate
change, which may introduce additional complexity into the phenomenological
records. To address this question for daily discharge data recorded during the
period 1989-2016 at twelve gauging stations on Brazos River in Texas (USA), we
use a set of novel quantitative tools: Kolmogorov complexity (KC) with its
derivative associated measures to assess complexity, and Lyapunov time (LT) to
assess predictability. We find that all daily discharge series exhibit long
memory with an increasing downflow tendency, while the randomness of the series
at individual sites cannot be definitively concluded. All Kolmogorov complexity
measures have relatively small values with the exception of the USGS (United
States Geological Survey) 08088610 station at Graford, Texas, which exhibits
the highest values of these complexity measures. This finding may be attributed
to the elevated effect of human activities at Graford, and proportionally
lesser effect at other stations. In addition, complexity tends to decrease
downflow, meaning that larger catchments are generally less influenced by
anthropogenic activity. The correction on randomness of Lyapunov time
(quantifying predictability) is found to be inversely proportional to the
Kolmogorov complexity, which strengthens our conclusion regarding the effect of
anthropogenic activities, considering that KC and LT are distinct measures,
based on rather different techniques
Formal Executable Models for Automatic Detection of Timing Anomalies
A timing anomaly is a counterintuitive timing behavior in the sense that a local fast execution slows down an overall global execution. The presence of such behaviors is inconvenient for the WCET analysis which requires, via abstractions, a certain monotony property to compute safe bounds. In this paper we explore how to systematically execute a previously proposed formal definition of timing anomalies. We ground our work on formal designs of architecture models upon which we employ guided model checking techniques. Our goal is towards the automatic detection of timing anomalies in given computer architecture designs
- ā¦