219 research outputs found

    Mean-Payoff Optimization in Continuous-Time Markov Chains with Parametric Alarms

    Full text link
    Continuous-time Markov chains with alarms (ACTMCs) allow for alarm events that can be non-exponentially distributed. Within parametric ACTMCs, the parameters of alarm-event distributions are not given explicitly and can be subject of parameter synthesis. An algorithm solving the ε\varepsilon-optimal parameter synthesis problem for parametric ACTMCs with long-run average optimization objectives is presented. Our approach is based on reduction of the problem to finding long-run average optimal strategies in semi-Markov decision processes (semi-MDPs) and sufficient discretization of parameter (i.e., action) space. Since the set of actions in the discretized semi-MDP can be very large, a straightforward approach based on explicit action-space construction fails to solve even simple instances of the problem. The presented algorithm uses an enhanced policy iteration on symbolic representations of the action space. The soundness of the algorithm is established for parametric ACTMCs with alarm-event distributions satisfying four mild assumptions that are shown to hold for uniform, Dirac and Weibull distributions in particular, but are satisfied for many other distributions as well. An experimental implementation shows that the symbolic technique substantially improves the efficiency of the synthesis algorithm and allows to solve instances of realistic size.Comment: This article is a full version of a paper accepted to the Conference on Quantitative Evaluation of SysTems (QEST) 201

    Chip firing and all-terminal network reliability bounds

    Get PDF
    AbstractThe (all-terminal) reliability of a graph G is the probability that all vertices are in the same connected component, given that vertices are always operational but edges fail independently each with probability p. Computing reliability is #P-complete, and hence is expected to be intractable. Consequently techniques for efficiently (and effectively) bounding reliability have been the major thrust of research in the area. We utilize a deep connection between reliability and chip firings on graphs to improve previous bounds for reliability

    Universal Reliability Bounds for Sparse Networks

    Get PDF
    Consider a graph with perfect nodes and edges subject to independent random failures with identical probability.The all-terminal reliability (ATR) is the probability that the resulting subgraph is connected. First, we fully characterize uniformly least reliable graphs (ULRG) whose co-rank is not greater than four. Universal reliability bounds are here introduced for those graphs. It is formally proved that ULRG are invariant under bridge-contractions, and maximize the number of bridges among all connected simple graphs with a prescribed number of nodes and edges. A closed-form for the maximum number of bridges is also given, which has an intrinsic interest from a graphtheoretic point of view. Finally, the cost-reliability trade-off is discussed, comparing the number of edges required to reduce the reliability gaps between the least and most reliable graphs. A remarkable conclusion is that the network design is critical under rare event failures, where the reliability-gap between least and most-reliable networks is monotonically increasing with the number of terminalsAgencia Nacional de Investigación e Innovació

    Recent Advances in Fully Dynamic Graph Algorithms

    Full text link
    In recent years, significant advances have been made in the design and analysis of fully dynamic algorithms. However, these theoretical results have received very little attention from the practical perspective. Few of the algorithms are implemented and tested on real datasets, and their practical potential is far from understood. Here, we present a quick reference guide to recent engineering and theory results in the area of fully dynamic graph algorithms

    Network Robustness: Diffusing Information Despite Adversaries

    Get PDF
    In this thesis, we consider the problem of diffusing information resiliently in networks that contain misbehaving nodes. Previous strategies to achieve resilient information diffusion typically require the normal nodes to hold some global information, such as the topology of the network and the identities of non-neighboring nodes. However, these assumptions are not suitable for large-scale networks and this necessitates our study of resilient algorithms based on only local information. We propose a consensus algorithm where, at each time-step, each normal node removes the extreme values in its neighborhood and updates its value as a weighted average of its own value and the remaining values. We show that traditional topological metrics (such as connectivity of the network) fail to capture such dynamics. Thus, we introduce a topological property termed as network robustness and show that this concept, together with its variants, is the key property to characterize the behavior of a class of resilient algorithms that use purely local information. We then investigate the robustness properties of complex networks. Specifically, we consider common random graph models for complex networks, including the preferential attachment model, the Erdos-Renyi model, and the geometric random graph model, and compare the metrics of connectivity and robustness in these models. While connectivity and robustness are greatly different in general (i.e., there exist graphs which are highly connected but with poor robustness), we show that the notions of robustness and connectivity are equivalent in the preferential attachment model, cannot be very different in the geometric random graph model, and share the same threshold functions in the Erdos-Renyi model, which gives us more insight about the structure of complex networks. Finally, we provide a construction method for robust graphs

    Overlapping Community Detection in Networks: the State of the Art and Comparative Study

    Full text link
    This paper reviews the state of the art in overlapping community detection algorithms, quality measures, and benchmarks. A thorough comparison of different algorithms (a total of fourteen) is provided. In addition to community level evaluation, we propose a framework for evaluating algorithms' ability to detect overlapping nodes, which helps to assess over-detection and under-detection. After considering community level detection performance measured by Normalized Mutual Information, the Omega index, and node level detection performance measured by F-score, we reached the following conclusions. For low overlapping density networks, SLPA, OSLOM, Game and COPRA offer better performance than the other tested algorithms. For networks with high overlapping density and high overlapping diversity, both SLPA and Game provide relatively stable performance. However, test results also suggest that the detection in such networks is still not yet fully resolved. A common feature observed by various algorithms in real-world networks is the relatively small fraction of overlapping nodes (typically less than 30%), each of which belongs to only 2 or 3 communities.Comment: This paper (final version) is accepted in 2012. ACM Computing Surveys, vol. 45, no. 4, 2013 (In press) Contact: [email protected]
    corecore