20 research outputs found

    A Robust Fault-Tolerant and Scalable Cluster-wide Deduplication for Shared-Nothing Storage Systems

    Full text link
    Deduplication has been largely employed in distributed storage systems to improve space efficiency. Traditional deduplication research ignores the design specifications of shared-nothing distributed storage systems such as no central metadata bottleneck, scalability, and storage rebalancing. Further, deduplication introduces transactional changes, which are prone to errors in the event of a system failure, resulting in inconsistencies in data and deduplication metadata. In this paper, we propose a robust, fault-tolerant and scalable cluster-wide deduplication that can eliminate duplicate copies across the cluster. We design a distributed deduplication metadata shard which guarantees performance scalability while preserving the design constraints of shared- nothing storage systems. The placement of chunks and deduplication metadata is made cluster-wide based on the content fingerprint of chunks. To ensure transactional consistency and garbage identification, we employ a flag-based asynchronous consistency mechanism. We implement the proposed deduplication on Ceph. The evaluation shows high disk-space savings with minimal performance degradation as well as high robustness in the event of sudden server failure.Comment: 6 Pages including reference

    Improving k-nn search and subspace clustering based on local intrinsic dimensionality

    Get PDF
    In several novel applications such as multimedia and recommender systems, data is often represented as object feature vectors in high-dimensional spaces. The high-dimensional data is always a challenge for state-of-the-art algorithms, because of the so-called curse of dimensionality . As the dimensionality increases, the discriminative ability of similarity measures diminishes to the point where many data analysis algorithms, such as similarity search and clustering, that depend on them lose their effectiveness. One way to handle this challenge is by selecting the most important features, which is essential for providing compact object representations as well as improving the overall search and clustering performance. Having compact feature vectors can further reduce the storage space and the computational complexity of search and learning tasks. Support-Weighted Intrinsic Dimensionality (support-weighted ID) is a new promising feature selection criterion that estimates the contribution of each feature to the overall intrinsic dimensionality. Support-weighted ID identifies relevant features locally for each object, and penalizes those features that have locally lower discriminative power as well as higher density. In fact, support-weighted ID measures the ability of each feature to locally discriminate between objects in the dataset. Based on support-weighted ID, this dissertation introduces three main research contributions: First, this dissertation proposes NNWID-Descent, a similarity graph construction method that utilizes the support-weighted ID criterion to identify and retain relevant features locally for each object and enhance the overall graph quality. Second, with the aim to improve the accuracy and performance of cluster analysis, this dissertation introduces k-LIDoids, a subspace clustering algorithm that extends the utility of support-weighted ID within a clustering framework in order to gradually select the subset of informative and important features per cluster. k-LIDoids is able to construct clusters together with finding a low dimensional subspace for each cluster. Finally, using the compact object and cluster representations from NNWID-Descent and k-LIDoids, this dissertation defines LID-Fingerprint, a new binary fingerprinting and multi-level indexing framework for the high-dimensional data. LID-Fingerprint can be used for hiding the information as a way of preventing passive adversaries as well as providing an efficient and secure similarity search and retrieval for the data stored on the cloud. When compared to other state-of-the-art algorithms, the good practical performance provides an evidence for the effectiveness of the proposed algorithms for the data in high-dimensional spaces

    Survey of existing fingerprint countermeasures

    No full text
    A number of fingerprint countermeasures have claimed that their countermeasures can prevent the user away from fingerprint tracking. The research attempts to prove their claims and requires to study the effectiveness of their fingerprint prevention and observe the results of side effects after defence in order to assist the attentive user to know the limitation of current approaches. Under investigation, all countermeasures will be installed on the web browser and visit the developed hybrid fingerprint website in order to know the efficiency of the fingerprint resistance of all types. The research shows that all fingerprint countermeasures nowadays are unable to obstruct all kinds of the fingerprint tracking and countermeasures that use the blocking technique appear more side effects to the web browser than other techniques. Also, the increasing number of fingerprint attributes are increasing cause of unusual combination inside the Internet browser

    Web-based relay management with biometric authentication

    Get PDF
    This thesis proposes a web-based system for managing digital relay settings. These relays are deployed in the power system to protect sensitive and expensive equipment from physical damage during system faults and overload conditions. Providing this capability exposes these devices to the same cyber security threats that corporations have faced for many years.;This thesis investigates the risks and requirements for deploying the proposed system. A breakdown in the protection that these relays provide would cause power outages. The cost of outages can be significant. Therefore cyber security is critical in the system design. Cyber security requirements for the power industry identify access control as an important aspect for the protection of its infrastructure. If properly implemented, biometrics can be used to strengthen access control to computer systems.;The web-based relay management system uses fingerprint authentication along with a username and password to provide access control. Website users are given access to functionality based on user roles. Only high level users may attempt relay setting modification. The relay management system interacts with a database that stores the current relay settings, relay setting restrictions, and a queue of relay updates. A process is implemented to verify attempted setting changes against these setting restrictions. This provides an extra security layer if users attempt harmful changes to protection schemes. Valid setting changes are added to the queue and a separate relay update program communicates these changes to the relay. The database and relay update program protect the relays from direct modification. These features combined with biometric authentication provide a strong layered scheme for protecting relays, while supplying an easy to use interface for remotely using their capabilities

    Indoor localisation by using wireless sensor nodes

    Get PDF
    This study is devoted to investigating and developing WSN based localisation approaches with high position accuracies indoors. The study initially summarises the design and implementation of localisation systems and WSN architecture together with the characteristics of LQI and RSSI values. A fingerprint localisation approach is utilised for indoor positioning applications. A k-nearest neighbourhood algorithm (k-NN) is deployed, using Euclidean distances between the fingerprint database and the object fingerprints, to estimate unknown object positions. Weighted LQI and RSSI values are calculated and the k-NN algorithm with different weights is utilised to improve the position detection accuracy. Different weight functions are investigated with the fingerprint localisation technique. A novel weight function which produced the maximum position accuracy is determined and employed in calculations. The study covered designing and developing the centroid localisation (CL) and weighted centroid localisation (WCL) approaches by using LQI values. A reference node localisation approach is proposed. A star topology of reference nodes are to be utilized and a 3-NN algorithm is employed to determine the nearest reference nodes to the object location. The closest reference nodes are employed to each nearest reference nodes and the object locations are calculated by using the differences between the closest and nearest reference nodes. A neighbourhood weighted localisation approach is proposed between the nearest reference nodes in star topology. Weights between nearest reference nodes are calculated by using Euclidean and physical distances. The physical distances between the object and the nearest reference nodes are calculated and the trigonometric techniques are employed to derive the object coordinates. An environmentally adaptive centroid localisation approach is proposed.Weighted standard deviation (STD) techniques are employed adaptively to estimate the unknown object positions. WSNs with minimum RSSI mean values are considered as reference nodes across the sensing area. The object localisation is carried out in two phases with respect to these reference nodes. Calculated object coordinates are later translated into the universal coordinate system to determine the actual object coordinates. Virtual fingerprint localisation technique is introduced to determine the object locations by using virtual fingerprint database. A physical fingerprint database is organised in the form of virtual database by using LQI distribution functions. Virtual database elements are generated among the physical database elements with linear and exponential distribution functions between the fingerprint points. Localisation procedures are repeated with virtual database and localisation accuracies are improved compared to the basic fingerprint approach. In order to reduce the computation time and effort, segmentation of the sensing area is introduced. Static and dynamic segmentation techniques are deployed. Segments are defined by RSS ranges and the unknown object is localised in one of these segments. Fingerprint techniques are applied only in the relevant segment to find the object location. Finally, graphical user interfaces (GUI) are utilised with application program interfaces (API), in all calculations to visualise unknown object locations indoors

    Fitness value based evolution algorithm approach for text steganalysis model

    Get PDF
    In this paper, we present a new alternative method for text steganalysis based on an evolution algorithm, implemented using the Java Evolution Algorithms Package (JEAP).The main objective of this paper is to detect the existence of hidden messages ased on fitness values of a text description.It is found that the detection performance has been influenced by two groups of fitness values which are good fitness value and bad fitness value. This paper provides a valuable insight into the development and enhancement of the text steganalysis domain

    An Analysis of Cholesteric Spherical Reflector Identifiers for Object Authenticity Verification

    Get PDF
    Arrays of Cholesteric Spherical Reflectors (CSRs), microscopic cholesteric liquid crystals in a spherical shape, have been argued to become a game-changing technology in anti-counterfeiting. Used to build identifiable tags or coating, called CSR IDs, they can supply objects with unclonable fingerprint-like characteristics, making it possible to authenticate objects. In a previous study, we have shown how to extract minutiæ from CSR IDs. In this journal version, we build on that previous research, consolidate the methodology, and test it over CSR IDs obtained by different production processes. We measure the robustness and reliability of our procedure on large and variegate sets of CSR IDs’ images taken with a professional microscope (Laboratory Data set) and with a microscope that could be used in a realistic scenario (Realistic Data set). We measure intra-distance and interdistance, proving that we can distinguish images coming from the same CSR ID from images of different CSR IDs. However, without surprise, images in Laboratory Data set have an intra-distance that on average is less, and with less variance, than the intra-distance between responses from Realistic Data set. With this evidence, we discuss a few requirements for an anti-counterfeiting technology based on CSRs
    corecore