588 research outputs found
Dynamics and stability of small social networks
The choices and behaviours of individuals in social systems combine in unpredictable ways to create complex, often surprising, social outcomes. The structure of these behaviours, or interactions between individuals, can be represented as a social network. These networks are not static but vary over time as connections are made and broken or change in intensity. Generally these changes are gradual, but in some cases individuals disagree and as a result "fall out" with each other, i.e. , actively end their relationship by ceasing all contact. These "fallouts" have been shown to be capable of fragmenting the social network into disconnected parts. Fragmentation can impair the functioning of social networks and it is thus important to better understand the social processes that have such consequences. In this thesis we investigate the question of how networks fragment: what mechanism drives the changes that ultimately result in fragmentation? To do so, we also aim to understand the necessary conditions for fragmentation to be possible and identify the connections that are most important for the cohesion of the network. To answer these questions, we need a model of social network dynamics that is stable enough such that fragmentation does not occur spontaneously, but is simultaneously dynamic enough to allow the system to react to perturbations (i.e. , disagreements). We present such a model and show that it is able to grow and maintain networks exhibiting the characteristic properties of social networks, and does so using local behavioural rules inspired by sociological theory. We then provide a detailed investigation of fragmentation and confirm basic intuitions on the importance of bridges for network cohesion. Furthermore, we show that this topological feature alone does not explain which points of the network are most vulnerable to fragmentation. Rather, we find that dependencies between edges are crucial for understanding subtle differences between stable and vulnerable bridges. This understandingof the vulnerability of different network components is likely to be valuable for preventing fragmentation and limiting the impact of social fallou
Assortativity Effects on Diffusion-like Processes in Scale-free Networks
We study the variation in epidemic thresholds in complex networks with different assortativity properties. We determine the thresholds by applying spectral analysis to the matrices associated to the graphs. In order to produce graphs with a specific assortativity we introduce a procedure to sample the space of all the possible networks with a given degree sequence. Our analysis shows that while disassortative networks have an higher epidemiological threshold, assortative networks have a slower diffusion time for diseases. We also used these networks for evaluating the effects of assortativity in a specific dynamic model of sandpile. We show that immunization procedures give different results according to the assortativity of the network considered
Synchronization in complex networks
Synchronization processes in populations of locally interacting elements are
in the focus of intense research in physical, biological, chemical,
technological and social systems. The many efforts devoted to understand
synchronization phenomena in natural systems take now advantage of the recent
theory of complex networks. In this review, we report the advances in the
comprehension of synchronization phenomena when oscillating elements are
constrained to interact in a complex network topology. We also overview the new
emergent features coming out from the interplay between the structure and the
function of the underlying pattern of connections. Extensive numerical work as
well as analytical approaches to the problem are presented. Finally, we review
several applications of synchronization in complex networks to different
disciplines: biological systems and neuroscience, engineering and computer
science, and economy and social sciences.Comment: Final version published in Physics Reports. More information
available at http://synchronets.googlepages.com
Challenges in Complex Systems Science
FuturICT foundations are social science, complex systems science, and ICT.
The main concerns and challenges in the science of complex systems in the
context of FuturICT are laid out in this paper with special emphasis on the
Complex Systems route to Social Sciences. This include complex systems having:
many heterogeneous interacting parts; multiple scales; complicated transition
laws; unexpected or unpredicted emergence; sensitive dependence on initial
conditions; path-dependent dynamics; networked hierarchical connectivities;
interaction of autonomous agents; self-organisation; non-equilibrium dynamics;
combinatorial explosion; adaptivity to changing environments; co-evolving
subsystems; ill-defined boundaries; and multilevel dynamics. In this context,
science is seen as the process of abstracting the dynamics of systems from
data. This presents many challenges including: data gathering by large-scale
experiment, participatory sensing and social computation, managing huge
distributed dynamic and heterogeneous databases; moving from data to dynamical
models, going beyond correlations to cause-effect relationships, understanding
the relationship between simple and comprehensive models with appropriate
choices of variables, ensemble modeling and data assimilation, modeling systems
of systems of systems with many levels between micro and macro; and formulating
new approaches to prediction, forecasting, and risk, especially in systems that
can reflect on and change their behaviour in response to predictions, and
systems whose apparently predictable behaviour is disrupted by apparently
unpredictable rare or extreme events. These challenges are part of the FuturICT
agenda
A wireless sensor network system for border security and crossing detection
The protection of long stretches of countries’ borders has posed a number of challenges. Effective and continuous monitoring of a border requires the implementation of multi-surveillance technologies, such as Wireless Sensor Networks (WSN), that work as an integrated unit to meet the desired goals. The research presented in this thesis investigates the application of topologically Linear WSN (LWSNs) to international border monitoring and surveillance. The main research questions studied here are: What is the best form of node deployment and hierarchy? What is the minimum number of sensor nodes to achieve k− barrier coverage in a given belt region? iven an appropriate network density, how do we determine if a region is indeed k−barrier covered? What are the factors that affect barrier coverage? How to organise nodes into logical segments to perform in-network processing of data? How to transfer information from the networks to the end users while maintaining critical QoS measures such as timeliness and accuracy. To address these questions, we propose an architecture that specifies a mechanism to assign nodes to various network levels depending on their location. These levels are used by a cross-layer communication protocol to achieve data delivery at the lowest possible cost and minimal delivery delay. Building on this levelled architecture, we study the formation of weak and strong barriers and how they determine border crossing detection probability. We propose new method to calculate the required node density to provide higher intruder detection rate. Then, we study the effect of people movement models on the border crossing detection probability. At the data link layer, new energy balancing along with shifted MAC protocol are introduced to further increase the network lifetime and delivery speed. In addition, at network layer, a routing protocol called Level Division raph (LD ) is developed. LD utilises a complex link cost measurement to insure best QoS data delivery to the sink node at the lowest possible cost. The proposed system has the ability to work independently or cooperatively with other monitoring technologies, such as drowns and mobile monitoring stations. The performance of the proposed work is extensively evaluated analytically and in simulation using real-life conditions and parameters. The simulation results show significant performance gains when comparing LD to its best rivals in the literature Dynamic Source Routing. Compared to DSR, LD achieves higher performance in terms of average end-to-end delays by up to 95%, packet delivery ratio by up to 20%, and throughput by up to 60%, while maintaining similar performance in terms of normalised routing load and energy consumption
High-Performance Modelling and Simulation for Big Data Applications
This open access book was prepared as a Final Publication of the COST Action IC1406 “High-Performance Modelling and Simulation for Big Data Applications (cHiPSet)“ project. Long considered important pillars of the scientific method, Modelling and Simulation have evolved from traditional discrete numerical methods to complex data-intensive continuous analytical optimisations. Resolution, scale, and accuracy have become essential to predict and analyse natural and complex systems in science and engineering. When their level of abstraction raises to have a better discernment of the domain at hand, their representation gets increasingly demanding for computational and data resources. On the other hand, High Performance Computing typically entails the effective use of parallel and distributed processing units coupled with efficient storage, communication and visualisation systems to underpin complex data-intensive applications in distinct scientific and technical domains. It is then arguably required to have a seamless interaction of High Performance Computing with Modelling and Simulation in order to store, compute, analyse, and visualise large data sets in science and engineering. Funded by the European Commission, cHiPSet has provided a dynamic trans-European forum for their members and distinguished guests to openly discuss novel perspectives and topics of interests for these two communities. This cHiPSet compendium presents a set of selected case studies related to healthcare, biological data, computational advertising, multimedia, finance, bioinformatics, and telecommunications
TOWER: Topology Optimization for netWork Enhanced Resilience
7th International Conference on Data Communication Networking - DCNET 2016 , 26/07/2016-28/07/2016, Lisboa, PortugalNowadays society is more and more dependent on critical infrastructures. Critical network infrastructures (CNI) are communication networks whose disruption can create a severe impact on other systems including critical infrastructures. In this work, we propose TOWER, a framework for the provision of adequate strategies to optimize service provision and system resilience in CNIs. The goal of TOWER is being able to compute new network topologies for CNIs under the event of malicious attacks. For doing this, TOWER takes into account a risk analysis of the CNI, the results from a cyber-physical IDS and a multilayer model
of the network, for taking into account all the existing dependences. TOWER analyses the network structure in order to determine the best strategy for obtaining a network topology, taking into account the existing dependences and the potential conflicting interests when not all requirements can be met. Finally, we present some lines for further development of TOWER.European Commissio
High-Performance Modelling and Simulation for Big Data Applications
This open access book was prepared as a Final Publication of the COST Action IC1406 “High-Performance Modelling and Simulation for Big Data Applications (cHiPSet)“ project. Long considered important pillars of the scientific method, Modelling and Simulation have evolved from traditional discrete numerical methods to complex data-intensive continuous analytical optimisations. Resolution, scale, and accuracy have become essential to predict and analyse natural and complex systems in science and engineering. When their level of abstraction raises to have a better discernment of the domain at hand, their representation gets increasingly demanding for computational and data resources. On the other hand, High Performance Computing typically entails the effective use of parallel and distributed processing units coupled with efficient storage, communication and visualisation systems to underpin complex data-intensive applications in distinct scientific and technical domains. It is then arguably required to have a seamless interaction of High Performance Computing with Modelling and Simulation in order to store, compute, analyse, and visualise large data sets in science and engineering. Funded by the European Commission, cHiPSet has provided a dynamic trans-European forum for their members and distinguished guests to openly discuss novel perspectives and topics of interests for these two communities. This cHiPSet compendium presents a set of selected case studies related to healthcare, biological data, computational advertising, multimedia, finance, bioinformatics, and telecommunications
- …