246 research outputs found

    Impact of Fuzzy Logic in Object-Oriented Database Through Blockchain

    Get PDF
    In this article, we show that applying fuzzy reasoning to an object-arranged data set produces noticeably better results than applying it to a social data set by applying it to both social and object-situated data sets. A Relational Data Base Management System (RDBMS) product structure offers a practical and efficient way to locate, store, and retrieve accurate data included inside a data collection. In any case, clients typically have to make vague, ambiguous, or fanciful requests. Our work allows clients the freedom to utilise FRDB to examine the database in everyday language, enabling us to provide a range of solutions that would benefit clients in a variety of ways. Given that the degree of attributes in a fuzzy knowledge base goes from 0 to 1, the term "fuzzy" was coined. This is due to the base's fictitious formalization's reliance on fuzzy reasoning. In order to lessen the fuzziness of the fuzzy social data set as a result of the abundance of uncertainty and vulnerabilities in clinical medical services information, a fuzzy article located information base is designed here for the Health-Care space. In order to validate the presentation and sufficiency of the fuzzy logic on both data sets, certain fuzzy questions are thus posed of the fuzzy social data set and the fuzzy item-situated information base.

    Temporal RDF(S) Data Storage and Query with HBase

    Get PDF
    Resource Description Framework (RDF) is a metadata model recommended by World Wide Web Consortium (W3C) for describing the Web resources. With the arrival of the era of Big Data, very large amounts of RDF data are continuously being created and need to be stored for management. The traditional centralized RDF storage models cannot meet the need of largescale RDF data storage. Meanwhile, the importance of temporal information management and processing has been acknowledged by academia and industry. In this paper, we propose a storage model to store temporal RDF based on HBase. The proposed storage model applies the built-in time mechanism of HBase. Our experiments on LUBM dataset with temporal information added show that our storage model can store large temporal RDF data and obtain good query efficiency

    Heterogeneous data source integration for smart grid ecosystems based on metadata mining

    Get PDF
    The arrival of new technologies related to smart grids and the resulting ecosystem of applications andmanagement systems pose many new problems. The databases of the traditional grid and the variousinitiatives related to new technologies have given rise to many different management systems with several formats and different architectures. A heterogeneous data source integration system is necessary toupdate these systems for the new smart grid reality. Additionally, it is necessary to take advantage of theinformation smart grids provide. In this paper, the authors propose a heterogeneous data source integration based on IEC standards and metadata mining. Additionally, an automatic data mining framework isapplied to model the integrated information.Ministerio de EconomĂ­a y Competitividad TEC2013-40767-

    Low-latency, query-driven analytics over voluminous multidimensional, spatiotemporal datasets

    Get PDF
    2017 Summer.Includes bibliographical references.Ubiquitous data collection from sources such as remote sensing equipment, networked observational devices, location-based services, and sales tracking has led to the accumulation of voluminous datasets; IDC projects that by 2020 we will generate 40 zettabytes of data per year, while Gartner and ABI estimate 20-35 billion new devices will be connected to the Internet in the same time frame. The storage and processing requirements of these datasets far exceed the capabilities of modern computing hardware, which has led to the development of distributed storage frameworks that can scale out by assimilating more computing resources as necessary. While challenging in its own right, storing and managing voluminous datasets is only the precursor to a broader field of study: extracting knowledge, insights, and relationships from the underlying datasets. The basic building block of this knowledge discovery process is analytic queries, encompassing both query instrumentation and evaluation. This dissertation is centered around query-driven exploratory and predictive analytics over voluminous, multidimensional datasets. Both of these types of analysis represent a higher-level abstraction over classical query models; rather than indexing every discrete value for subsequent retrieval, our framework autonomously learns the relationships and interactions between dimensions in the dataset (including time series and geospatial aspects), and makes the information readily available to users. This functionality includes statistical synopses, correlation analysis, hypothesis testing, probabilistic structures, and predictive models that not only enable the discovery of nuanced relationships between dimensions, but also allow future events and trends to be predicted. This requires specialized data structures and partitioning algorithms, along with adaptive reductions in the search space and management of the inherent trade-off between timeliness and accuracy. The algorithms presented in this dissertation were evaluated empirically on real-world geospatial time-series datasets in a production environment, and are broadly applicable across other storage frameworks

    Optimization of Columnar NoSQL Data Warehouse Model with Clarans Clustering Algorithm

    Get PDF
    In order to perfectly meet the needs of business leaders, decision-makers have resorted to the integration of external sources (such as Linked Open Data) in the decision-making system in order to enrich their existing data warehouses with new concepts contributing to bring added value to their organizations, enhance its productivity and retain its customers. However, the traditional data warehouse environment is not suitable to support external Big Data. To deal with this new challenge, several researches are oriented towards the direct conversion of classical relational data warehouse to a columnar NoSQL data warehouse, whereas the existing advanced works based on clustering algorithms are very limited and have several shortcomings. In this context, our paper proposes a new solution that conceives an optimized columnar data warehouse based on CLARANS clustering algorithm that has proven its effectiveness in generating optimal column families. Experimental results improve the validity of our system by performing a detailed comparative study between the existing advanced approaches and our proposed optimized method

    A Crowdsourcing Based Framework for Sentiment Analysis: A Product Reputation

    Get PDF
    As social networking has spread, people started sharing their personal opinions and thoughts widely via these online platforms. The resulting vast valuable data represent a rich source for companies to deduct their products’ reputation from both social media and crowds’ judgments. To exploit this wealth of data, a framework was proposed to collect opinions and rating scores respectively from social media and crowdsourcing platform to perform sentiment analysis, provide insights about a product and give consumers’ tendencies. During the analysis process, a consumer category (strict) is excluded from the process of reaching a majority consensus. To overcome this, a fuzzy clustering is used to compute consumers’ credibility. The key novelty of our approach is the new layer of validity check using a crowdsourcing component that ensures that the results obtained from social media are supported by opinions extracted directly from real-life consumers. Finally, experiments are carried out to validate this model (Twitter and Facebook were used as data sources). The obtained results show that this approach is more efficient and accurate than existing solutions thanks to our two-layer validity check design
    • …
    corecore