902 research outputs found

    Survey on replication techniques for distributed system

    Get PDF
    Distributed systems mainly provide access to a large amount of data and computational resources through a wide range of interfaces. Besides its dynamic nature, which means that resources may enter and leave the environment at any time, many distributed systems applications will be running in an environment where faults are more likely to occur due to their ever-increasing scales and the complexity. Due to diverse faults and failures conditions, fault tolerance has become a critical element for distributed computing in order for the system to perform its function correctly even in the present of faults. Replication techniques primarily concentrate on the two fault tolerance manners precisely masking the failures as well as reconfigure the system in response. This paper presents a brief survey on different replication techniques such as Read One Write All (ROWA), Quorum Consensus (QC), Tree Quorum (TQ) Protocol, Grid Configuration (GC) Protocol, Two-Replica Distribution Techniques (TRDT), Neighbour Replica Triangular Grid (NRTG) and Neighbour Replication Distributed Techniques (NRDT). These techniques have its own redeeming features and shortcoming which forms the subject matter of this survey

    Novelty circular neighboring technique using reactive fault tolerance method

    Get PDF
    The availability of the data in a distributed system can be increase by implementing fault tolerance mechanism in the system. Reactive method in fault tolerance mechanism deals with restarting the failed services, placing redundant copies of data in multiple nodes across network, in other words data replication and migrating the data for recovery. Even if the idea of data replication is solid, the challenge is to choose the right replication technique that able to provide better data availability as well as consistency that involves read and write operations on the redundant copies. Circular Neighboring Replication (CNR) technique exploits neighboring policy in replicating the data items in the system performs well with regards to lower copies needed to maintain the system availability at the highest. In a performance analysis with existing techniques, results show that CNR improves system availability by average 37% by offering only two replicas needed to maintain data availability and consistency. The study demonstrates the possibility of the proposed technique and the potential of deploying in larger and complex environment

    Binary vote assignment on grid quorum replication technique with association rule

    Get PDF
    One of the biggest challenges that data grids users have to face today relates to the improvement of the data management. Organizations need to provide current data to users who may be geographically remote and to handle a volume of requests of data distributed around multiple sites in distributed environment. Therefore, the storage, availability, and consistency are important issues to be addressed to allow efficient and safe data access from many different sites. One way to effectively cope with these challenges is to rely on the replication technique. Replication is a useful technique for distributed database systems. Through this technique, a data can be accessed from multiple locations. Thus, replication increases data availability and accessibility to users. When one site fails, user still can access the same data at another site. Techniques such as Read-One-Write-All (ROWA), Hierarchical Replication Scheme (HRS) and Branch Replication Scheme (BRS) are the popular techniques being used for replication and data management. However, these techniques have its weaknesses in terms of communication costs that is the total replication servers needed to replicate the data. Furthermore, these techniques also do not consider the correlation between data during the fragmentation process. The knowledge about data correlation can be extracted from historical data using techniques of the data mining field. Without proper strategies, replication increases job execution time. In this research, the some-data-to-some-sites scheme called Binary Vote Assignment on Grid Quorum with Association (BV AGQAR) is proposed to manage replication for meaningful fragmented data in distributed database environment with low communication cost and processing time for a transaction. The main feature of BV AGQ-AR is that the technique integrates replication and data mining technique allowing meaningful extraction of knowledge from large data sets. Performance of the BVAGQ-AR technique comprised the following steps. First step is mining the data by using Apriori algorithm from Association Rules. It is used to discover the correlation between data. For the second step, the database is fragmented based on the data mining analysis results. This technique is executed to make sure data replication can be effectively done while saving cost. Then, the databases that are resulted after the fragmentation process are allocated at their assigned sites. Finally, after allocation process, each site has a database file and ready for any transaction and replication process. Finally, the result of the experiments shows that BV AGQ-AR can preserve the data consistency with the lowest communication cost and processing time for a transaction as compared to BCSA, PRA, ROW A, HRS and BRS

    Scale and abstraction : the sensitivity of fire regime simulation to nuisance parameters

    Get PDF
    Fire plays a key role in ecosystem dynamics and its impact on environmental, social and economic assets is increasingly a critical area of research. Fire regime simulation models are one of many approaches that provide insights into the relative importance of factors driving the dynamics of fire-vegetation systems. Fire propagates as a contagious process and simulation is an approach that captures this behaviour explicitly, integrating spatial and temporal data to produce auto-correlated patterns of fire regimes. However, when formulating these models, time and many aspects of space must be made discrete. These parameters are 'nuisance parameters': parameters necessary for the model formulation but not otherwise of interest. Fire growth simulations are therefore discrete approximations of continuous non-linear systems, and it might be expected that the values chosen for these nuisance parameters will be important. While it is well known that discrete geometries have consequences for the shape and area of simulated fires, no research has investigated the consequence this may have for estimates of the relative importance of the various drivers of fire regimes. I argue that nuisance parameters can be demonstrated to be unimportant for this class of model. I use the idea of 'importance' to underline the need for context with such an assertion. With sufficient replication, any parameter can be found statistically significant. A parameter is important, on the other hand, if different values produce qualitatively different outcomes. Models are commonly either re-parameterised to account for changes in resolution or scaling-up methods applied if such exist. I will further argue that such differences as there are in model outputs due to spatial resolution, cannot be accounted for by either re-parameterising or using a common approach that allows resolution to vary over the spatial extent. A set of experiments were devised using a published fire regime simulation model, modified, verified and validated, to isolate just those aspects of the model's sensitivity to resolution and discrete geometries that are unavoidable or intrinsic to these choices. This new model was used to test the above hypotheses, using peer-reviewed treatments that stand as yardsticks by which formal estimates of the importance of nuisance parameters can be made. As estimated by the model, neither spatio-temporal resolution nor any of the various choices available for discrete geometries, altered the model predictions. As expected, it is spatial resolution that has the greatest impact on running times for the model but this study finds that neither calibration, nor taking an approach that allows resolution to vary over the spatial extent, can account for differences in model outputs that arise from running simulations at coarser resolutions. All models are abstractions and a good model should ideally hold over levels of abstraction. This is rarely the case, but this study shows that results obtained through simulation in estimating the drivers of fire frequency in large landscapes, are robust with regard to these aspects of abstraction. This adds considerable confidence to a significant body of work that has used this approach over the last two decades

    Development Bvag Replication Prototype In Distributed Database Environment

    Get PDF
    Replication is a useful technique for distributed database systems. Through this technique, a data can be accessed from multiple locations. Thus, it increases data availability and accessibility to users. When one site fails, user still can access the same data at another site. Techniques such as Read-One-Write-All (ROWA), Hierarchical Replication Scheme (HRS) and Branch Replication Scheme (BRS) are the popular techniques being used for replication and data management. However, these techniques have its weaknesses in terms of communication costs. Consequently, ROWA, HRS and BRS take long executing time for a transaction since these techniques have to replicate its data to all servers. In this research, the some-data-to-some-sites scheme called Binary Vote Assignment on Grid (BVAG) is proposed. It works by considering neighbors binary vote assignment to its logical grid structure on fragmented data copies in order to manage transactions in the systems. For simplicity, the neighbours are assigned with vote one or zero. The assignment provides minimum communication cost due to the minimum number of quorum size required. In addition, it minimizes the storage capacity needed since we store database that has been fragmented. The development of prototype for the BVAG replication techniques were carried out using HTML, CSS, JavaScript and PHP. The prototype was developed in order to produce a web-based application that utilize BVAG replication techniques. From the development, the web based BVAG were developed and tested with basic data input to replicate the data. From the results, it shows the web based BVAG prototype is working and able to replicate data to the neighbour’s site

    The hippocampal formation from a machine learning perspective

    Get PDF
    Nos dias de hoje, existem diversos tipos de sensores que conseguem captar uma grande quantidade de dados em curtos espaços de tempo. Em muitas situações, as informações obtidas pelos diferentes sensores traduzem fenómenos específicos, através de dados obtidos em diferentes formatos. Nesses casos, torna-se difícil saber quais as relações entre os dados e/ou identificar se os diferentes dados traduzem uma certa condição. Neste contexto, torna-se relevante desenvolver sistemas que tenham capacidade de analisar grandes quantidades de dados num menor tempo possível, produzindo informação válida a partir da informação recolhida. O cérebro dos animais é um órgão biológico capaz de fazer algo semelhante com a informação obtida pelos sentidos, que traduzem fenómenos específicos. Dentro do cérebro, existe um elemento chamado Hipocampo, que se encontra situado na área do lóbulo temporal. A sua função principal consiste em analisar os elementos previamente codificados pelo Entorhinal Cortex, dando origem à formação de novas memórias. Sendo o Hipocampo um órgão que foi sofrendo evoluções ao longo do tempos, é importante perceber qual é o seu funcionamento e, se possível, tentar encontrar modelos computacionais que traduzam o seu mecanismo. Desde a remoção do Hipocampo num paciente que sofria de convulsões, ficou claro que, sem esse elemento, não seria possível memorizar lugares ou eventos ocorridos num determinado espaço de tempo. Essa funcionalidade é obtida através de um conjunto específico de células chamadas de Grid Cells, que estão situadas na área do Entorhinal Cortex, mas também das Place Cells, Head Direction Cells e Boundary Vector Cells. Neste âmbito, o principal objetivo desta Dissertação consiste em descrever os principais mecanismos biológicos localizados no Hipocampo e definir modelos computacionais que consigam simular as funções mais críticas de ambos os Hipocampos e da área do Entorhinal Cortex.Nowadays, sensor devices are able to generate huge amounts of data in short periods of time. In many situations, that data, collected by many different sensors, translates a specific phenomenon, but is presented in very different types and formats. In these cases, it is hard to determine how these distinct types of data are related to each other or translate a certain condition. In this context, it would be of great importance to develop a system capable of analysing such data in the smallest amount time to produce valid information. The brain is a biological organ capable of such decisions. Inside the brain, there is an element called Hippocampus, that is situated in the Temporal Lobe area. Its main function is to analyse the sensorial data encoded by the Entorhinal Cortex to create new memories. Since the Hippocampus has evolved for thousands of years to perform these tasks, it is of high importance to try to understand its functioning and to model it, i.e. to define a set of computer algorithms that approximates it. Since the removal of the Hippocampus from a patient suffering from seizures, the scientific community believes that the Hippocampus is crucial for memory formation and for spatial navigation. Without it, it wouldn’t be possible to memorize places and events that happened in a specific time or place. Such functionality is achieved with the help of set of cells called Grid Cells, present in the Entorhinal Cortex area, but also with Place Cells, Head Direction Cells and Boundary Vector Cells. The combined information analysed by those cells allows the unique identification of places or events. The main objective of the work developed in this Thesis consists in describing the biological mechanisms present in the Hippocampus area and to define potential computer models that allow the simulation of all or the most critical functions of both the Hippocampus and the Entorhinal Cortex areas

    A Grid-Enabled Infrastructure for Resource Sharing, E-Learning, Searching and Distributed Repository Among Universities

    Get PDF
    In the recent years, service-based approaches for sharing of data among repositories and online learning are rising to prominence because of their potential to meet the requirements in the area of high performance computing. Developing education based grid services and assuring high availability reliability and scalability are demanding in web service architectures. On the other hand, grid computing provides flexibility towards aggregating distributed CPU, memory, storage, data and supports large number of distributed resource sharing to provide the full potential for education like applications to share the knowledge that can be attainable on any single system. However, the literature shows that the potential of grid resources for educational purposes is not being utilized yet. In this paper, an education based grid framework architecture that provides promising platform to support sharing of geographically dispersed learning content among universities is developed. It allows students, faculty and researchers to share and gain knowledge in their area of interest by using e-learning, searching and distributed repository services among universities from anywhere, anytime. Globus toolkit 5.2.5 (GTK) software is used as grid middleware that provides resource access, discovery and management, data movement, security, and so forth. Furthermore, this work uses the OGSA-DAI that provides database access and operations. The resulting infrastructure enables users to discover education services and interact with them using the grid portal

    Monitoring and Optimization of ATLAS Tier 2 Center GoeGrid

    Get PDF
    The demand on computational and storage resources is growing along with the amount of infor- mation that needs to be processed and preserved. In order to ease the provisioning of the digital services to the growing number of consumers, more and more distributed computing systems and platforms are actively developed and employed. The building block of the distributed computing infrastructure are single computing centers, similar to the Worldwide LHC Computing Grid, Tier 2 centre GoeGrid. The main motivation of this thesis was the optimization of GoeGrid perfor- mance by efficient monitoring. The goal has been achieved by means of the GoeGrid monitoring information analysis. The data analysis approach was based on the adaptive-network-based fuzzy inference system (ANFIS) and machine learning algorithm such as Linear Support Vector Machine (SVM). The main object of the research was the digital service, since availability, reliability and ser- viceability of the computing platform can be measured according to the constant and stable provisioning of the services. Due to the widely used concept of the service oriented architecture (SOA) for large computing facilities, in advance knowing of the service state as well as the quick and accurate detection of its disability allows to perform the proactive management of the com- puting facility. The proactive management is considered as a core component of the computing facility management automation concept, such as Autonomic Computing. Thus in time as well as in advance and accurate identification of the provided service status can be considered as a contribution to the computing facility management automation, which is directly related to the provisioning of the stable and reliable computing resources. Based on the case studies, performed using the GoeGrid monitoring data, consideration of the approaches as generalized methods for the accurate and fast identification and prediction of the service status is reasonable. Simplicity and low consumption of the computing resources allow to consider the methods in the scope of the Autonomic Computing component

    Stories from different worlds in the universe of complex systems: A journey through microstructural dynamics and emergent behaviours in the human heart and financial markets

    Get PDF
    A physical system is said to be complex if it exhibits unpredictable structures, patterns or regularities emerging from microstructural dynamics involving a large number of components. The study of complex systems, known as complexity science, is maturing into an independent and multidisciplinary area of research seeking to understand microscopic interactions and macroscopic emergence across a broad spectrum systems, such as the human brain and the economy, by combining specific modelling techniques, data analytics, statistics and computer simulations. In this dissertation we examine two different complex systems, the human heart and financial markets, and present various research projects addressing specific problems in these areas. Cardiac fibrillation is a diffuse pathology in which the periodic planar electrical conduction across the cardiac tissue is disrupted and replaced by fast and disorganised electrical waves. In spite of a century-long history of research, numerous debates and disputes on the mechanisms of cardiac fibrillation are still unresolved while the outcomes of clinical treatments remain far from satisfactory. In this dissertation we use cellular automata and mean-field models to qualitatively replicate the onset and maintenance of cardiac fibrillation from the interactions among neighboring cells and the underlying topology of the cardiac tissue. We use these models to study the transition from paroxysmal to persistent atrial fibrillation, the mechanisms through which the gap-junction enhancer drug Rotigaptide terminates cardiac fibrillation and how focal and circuital drivers of fibrillation may co-exist as projections of transmural electrical activities. Financial markets are hubs in which heterogeneous participants, such as humans and algorithms, adopt different strategic behaviors to exchange financial assets. In recent decades the widespread adoption of algorithmic trading, the electronification of financial transactions, the increased competition among trading venues and the use of sophisticated financial instruments drove the transformation of financial markets into a global and interconnected complex system. In this thesis we introduce agent-based and state-space models to describe specific microstructural dynamics in the stock and foreign exchange markets. We use these models to replicate the emergence of cross-currency correlations from the interactions between heterogeneous participants in the currency market and to disentangle the relationships between price fluctuations, market liquidity and demand/supply imbalances in the stock market.Open Acces
    corecore