689 research outputs found

    An Evaluation Framework for Reputation Management Systems

    Get PDF
    Reputation management (RM) is employed in distributed and peer-to-peer networks to help users compute a measure of trust in other users based on initial belief, observed behavior, and run-time feedback. These trust values influence how, or with whom, a user will interact. Existing literature on RM focuses primarily on algorithm development, not comparative analysis. To remedy this, we propose an evaluation framework based on the trace-simulator paradigm. Trace file generation emulates a variety of network configurations, and particular attention is given to modeling malicious user behavior. Simulation is trace-based and incremental trust calculation techniques are developed to allow experimentation with networks of substantial size. The described framework is available as open source so that researchers can evaluate the effectiveness of other reputation management techniques and/or extend functionality. This chapter reports on our framework’s design decisions. Our goal being to build a general-purpose simulator, we have the opportunity to characterize the breadth of existing RM systems. Further, we demonstrate our tool using two reputation algorithms (EigenTrust and a modified TNA-SL) under varied network conditions. Our analysis permits us to make claims about the algorithms’ comparative merits. We conclude that such systems, assuming their distribution is secure, are highly effective at managing trust, even against adversarial collectives

    SDSF : social-networking trust based distributed data storage and co-operative information fusion.

    Get PDF
    As of 2014, about 2.5 quintillion bytes of data are created each day, and 90% of the data in the world was created in the last two years alone. The storage of this data can be on external hard drives, on unused space in peer-to-peer (P2P) networks or using the more currently popular approach of storing in the Cloud. When the users store their data in the Cloud, the entire data is exposed to the administrators of the services who can view and possibly misuse the data. With the growing popularity and usage of Cloud storage services like Google Drive, Dropbox etc., the concerns of privacy and security are increasing. Searching for content or documents, from this distributed stored data, given the rate of data generation, is a big challenge. Information fusion is used to extract information based on the query of the user, and combine the data and learn useful information. This problem is challenging if the data sources are distributed and heterogeneous in nature where the trustworthiness of the documents may be varied. This thesis proposes two innovative solutions to resolve both of these problems. Firstly, to remedy the situation of security and privacy of stored data, we propose an innovative Social-based Distributed Data Storage and Trust based co-operative Information Fusion Framework (SDSF). The main objective is to create a framework that assists in providing a secure storage system while not overloading a single system using a P2P like approach. This framework allows the users to share storage resources among friends and acquaintances without compromising the security or privacy and enjoying all the benefits that the Cloud storage offers. The system fragments the data and encodes it to securely store it on the unused storage capacity of the data owner\u27s friends\u27 resources. The system thus gives a centralized control to the user over the selection of peers to store the data. Secondly, to retrieve the stored distributed data, the proposed system performs the fusion also from distributed sources. The technique uses several algorithms to ensure the correctness of the query that is used to retrieve and combine the data to improve the information fusion accuracy and efficiency for combining the heterogeneous, distributed and massive data on the Cloud for time critical operations. We demonstrate that the retrieved documents are genuine when the trust scores are also used while retrieving the data sources. The thesis makes several research contributions. First, we implement Social Storage using erasure coding. Erasure coding fragments the data, encodes it, and through introduction of redundancy resolves issues resulting from devices failures. Second, we exploit the inherent concept of trust that is embedded in social networks to determine the nodes and build a secure net-work where the fragmented data should be stored since the social network consists of a network of friends, family and acquaintances. The trust between the friends, and availability of the devices allows the user to make an informed choice about where the information should be stored using `k\u27 optimal paths. Thirdly, for the purpose of retrieval of this distributed stored data, we propose information fusion on distributed data using a combination of Enhanced N-grams (to ensure correctness of the query), Semantic Machine Learning (to extract the documents based on the context and not just bag of words and also considering the trust score) and Map Reduce (NSM) Algorithms. Lastly we evaluate the performance of distributed storage of SDSF using era- sure coding and identify the social storage providers based on trust and evaluate their trustworthiness. We also evaluate the performance of our information fusion algorithms in distributed storage systems. Thus, the system using SDSF framework, implements the beneficial features of P2P networks and Cloud storage while avoiding the pitfalls of these systems. The multi-layered encrypting ensures that all other users, including the system administrators cannot decode the stored data. The application of NSM algorithm improves the effectiveness of fusion since large number of genuine documents are retrieved for fusion

    Optimisation techniques for data distribution in Volunteer Computing

    Get PDF
    Volunteer Computing is a new paradigm of distributed computing where the ordinary computer owners volunteer their computing power and storage capability to scientific projects. The increasing number of internet connected PCs allows Volunteer Computing to provide more computing power and storage capacity than what can be achieved with supercomputers, clusters and grids. However, volunteer computing projects rely on a centralized infrastructure for distributing data. This can affect the scalability of data intensive projects and when the projects participants increases. In this thesis, a new approach is proposed to incorporate P2P techniques into volunteer computing projects and apply trust management to optimize the use of P2P techniques in these projects. This approach adopted a P2P technique to form a decentralized data centres layer based on the resources of participants of volunteer computing projects. VASCODE framework is based on Attic File System to enable building the decentralized data centres and makes use of trust framework to provide the necessary data to users to select the optimum data centres for downloading data. Empirical evaluation demonstrated that the proposed approaches can achieve better scalability and performance as compared to the central server approach used in BOINC projects. In addition, it shows that clients with the support of trust framework have reliable and consistent download times because using trust allows them select the optimum data centres and avoid the malicious behaviour of data centres

    Student Digital Piracy In The Florida State University System:an Exploratory Study On Its Infrastructural Effects

    Get PDF
    Digital piracy is a problem that may never disappear from society. Through readily available resources such as those found in a university, students will always have access to illegal goods. While piracy is a global phenomenon, an institution\u27s resources combined with the typical college student\u27s lack of funds makes it more lucrative. Students use a number of methods to justify their actions ranging from previewing media to bringing justice to a corrupt company. While trying to understand the mindset of pirates is one route to deal with piracy, corporations attempted to alleviate the situation using added software encoding. These messages are not always effective, and in some cases caused further damage to consumer morale. Furthermore, students such as Joel Tenenbaum, who continued to pirate music despite warnings from his parents and the recording industry, exemplify the type of person that is unfazed by legal threats, leading to a question of ethics. Students may not feel that downloading is stealing despite numerous warnings from the Digital Millennium Copyright Act and other major media organizations. The predominant solution used by universities involves monitoring the students\u27 network connection to detect Peer-to-Peer (P2P) connections or other connections that involve the transferring of copyrighted goods. Unfortunately, the current tools contain flaws that a crafty student may easily circumvent, undermining any attempts a university\u27s IT department may use to deter piracy. This study explored the nature of piracy prevention tools used by IT departments in the Florida State University System in order to determine their relative effectiveness. The study also looked into the opinions of the Information Security Officer in terms of alternative piracy prevention techniques that do not involve legal action and monitoring. It was found that most institutions do not use a formal piece of software that monitors for infringing data. They also stated that while their current techniques can do its required task, it was not perfected to a point where it could run autonomously. Furthermore, institutions agreed that students lack proper ethics and concern over the matter of copyright, but were not fully convinced that other preventions methods would be effective. The study ultimately considered monitoring techniques a short-term solution and that more research should be put into finding long-term solutions. It also implied that IT departments should be better funded in order to keep up with the technological gap

    Measuring Infringement of Intellectual Property Rights

    Get PDF
    © Crown Copyright 2014. You may re-use this information (excluding logos) free of charge in any format or medium, under the terms of the Open Government Licence. To view this licence, visit http://www.nationalarchives.gov. uk/doc/open-government-licence/ Where we have identified any third party copyright information you will need to obtain permission from the copyright holders concernedThe review is wide-ranging in scope and overall our findings evidence a lack of appreciation among those producing research for the high-level principles of measurement and assessment of scale. To date, the approaches adopted by industry seem more designed for internal consumption and are usually contingent on particular technologies and/or sector perspectives. Typically, there is a lack of transparency in the methodologies and data used to form the basis of claims, making much of this an unreliable basis for policy formulation. The research approaches we found are characterised by a number of features that can be summarised as a preference for reactive approaches that look to establish snapshots of an important issue at the time of investigation. Most studies are ad hoc in nature and on the whole we found a lack of sustained longitudinal approaches that would develop the appreciation of change. Typically the studies are designed to address specific hypotheses that might serve to support the position of the particular commissioning body. To help bring some structure to this area, we propose a framework for the assessment of the volume of infringement in each different area. The underlying aim is to draw out a common approach wherever possible in each area, rather than being drawn initially to the differences in each field. We advocate on-going survey tracking of the attitudes, perceptions and, where practical, behaviours of both perpetrators and claimants in IP infringement. Clearly, the nature of perpetrators, claimants and enforcement differs within each IPR but in our view the assessment for each IPR should include all of these elements. It is important to clarify that the key element of the survey structure is the adoption of a survey sampling methodology and smaller volumes of representative participation. Once selection is given the appropriate priority, a traditional offline survey will have a part to play, but as the opportunity arises, new technological methodologies, particularly for the voluntary monitoring of online behaviour, can add additional detail to the overall assessment of the scale of activity. This framework can be applied within each of the IP right sectors: copyright, trademarks,patents, and design rights. It may well be that the costs involved with this common approach could be mitigated by a syndicated approach to the survey elements. Indeed, a syndicated approach has a number of advantages in addition to cost. It could be designed to reduce any tendency either to hide inappropriate/illegal activity or alternatively exaggerate its volume to fit with the theme of the survey. It also has the scope to allow for monthly assessments of attitudes rather than being vulnerable to unmeasured seasonal impacts

    Doctor of Philosophy

    Get PDF
    dissertationWe propose a collective approach for harnessing the idle resources (cpu, storage, and bandwidth) of nodes (e.g., home desktops) distributed across the Internet. Instead of a purely peer-to-peer (P2P) approach, we organize participating nodes to act collectively using collective managers (CMs). Participating nodes provide idle resources to CMs, which unify these resources to run meaningful distributed services for external clients. We do not assume altruistic users or employ a barter-based incentive model; instead, participating nodes provide resources to CMs for long durations and are compensated in proportion to their contribution. In this dissertation we discuss the challenges faced by collective systems, present a design that addresses these challenges, and study the effect of selfish nodes. We believe that the collective service model is a useful alternative to the dominant pure P2P and centralized work queue models. It provides more effective utilization of idle resources, has a more meaningful economic model, and is better suited for building legal and commercial distributed services. We demonstrate the value of our work by building two distributed services using the collective approach. These services are a collective content distribution service and a collective data backup service

    A Mini Review of Peer-to-Peer (P2P) for Vehicular Communication

    Get PDF
    In recent times, peer-to-peer (P2P) has evolved, where it leverages the capability to scale compared to server-based networks. Consequently, P2P has appeared to be the future distributed systems in emerging several applications. P2P is actually a disruptive technology for setting up applications that scale to numerous concurrent individuals. Thus, in a P2P distributed system, individuals become themselves as peers through contributing, sharing, and managing the resources in a network. In this paper, P2P for vehicular communication is explored. A comprehensive of the functioning concept of both P2P along with vehicular communication is examined. In addition, the advantages are furthermore conversed for a far better understanding on the implementation

    A Balanced Trust-Based Method to Counter Sybil and Spartacus Attacks in Chord

    Get PDF
    A Sybil attack is one of the main challenges to be addressed when securing peer-to-peer networks, especially those based on Distributed Hash Tables (DHTs). Tampering routing tables by means of multiple fake identities can make routing, storing, and retrieving operations significantly more difficult and time-consuming. Countermeasures based on trust and reputation have already proven to be effective in some contexts, but one variant of the Sybil attack, the Spartacus attack, is emerging as a new threat and its effects are even riskier and more difficult to stymie. In this paper, we first improve a well-known and deployed DHT (Chord) through a solution mixing trust with standard operations, for facing a Sybil attack affecting either routing or storage and retrieval operations. This is done by maintaining the least possible overhead for peers. Moreover, we extend the solution we propose in order for it to be resilient also against a Spartacus attack, both for an iterative and for a recursive lookup procedure. Finally, we validate our findings by showing that the proposed techniques outperform other trust-based solutions already known in the literature as well
    • …
    corecore