57 research outputs found

    A GPFCSP-Based Fuzzy XQuery Interpreter

    Get PDF
    Nowadays XQuery has become the strongest standard for querying XML data. However, most of the real world information is in the form of imprecise, vague, ambiguous, uncertain and incomplete values. That is why there is a need for a flexible query language in which users can formulate queries that arise from their own criteria. In this paper, we propose an implementation of the Fuzzy XQuery - an extension of the XQuery query language based on the fuzzy set theory. In particular, we provide priority, threshold and fuzzy expressions for handling flexible queries. In addition, we have implemented an interpreter for this language by using the GPFCSP concept in Java and eXist-db environment

    Razvoj proširenja xquery interpretera baziran na fazi logici sa prioritetima

    Get PDF
    In many real-world applications, information is often imprecise and uncertain. With the popularity of web-based applications, huge amounts of data are available on the web, and XML (eXtensible Markup Language) has become the de facto standard for data exchange over the internet. The XQuery is the language for querying XML data. However, XML and XQuery suffer from incapability of representing and manipulating imprecise and uncertain data. Consequently, this work represents fuzzy data in XML documents and extends XQuery language as providing a more flexible XQuery language by using the fuzzy set theory. In this thesis, an extension of the XQuery query, called Fuzzy XQuery is described. It allows users to define priority, threshold and fuzzy expressions in their queries. Users also can predefine linguistic terms to use them in querying. An algorithm for calculating the global constraint satisfaction degree using the Generalized Prioritized Fuzzy Constraint Satisfaction Problem (GPFCSP) is introduced. Furthermore, Fuzzy XQuery Interpreter (FXI) is implemented allowing execution of fuzzy XQuery queries based on open source technologies and native XML open- source database. Additionally, innovative methods for computing fuzzy set compatibility and introducing order over fuzzy sets have been implemented, which give serious improvements in computational performance compared to previous implementations

    The development a dedicated system for the application fuzzy logic in systems for manage of XML documents

    Get PDF
    U stvarnom životu većina informacija dolazi kao neprecizne ili nepotpune vrednosti.  XML tehnologija je napravila veliki napredak u oblasti skladištenja i prenosa podataka.  Doktorska disertacija definiše fazi XML sintaksu koja kombinuje neodređenosti u vrednostima XML-a i neodređenosti u strukturi XML dokumenata. Takođe, rad proširuje standardnu XQuery upitnu sintaksu fazi elementima i uvodi prioritete i pragove zadovoljenja pomoću GPFCSP-a. Za razliku od drugih radova koji se fokusiraju na postavljanje teorija i definisanje sintaksi, rad posebnu pažnju posvećuje praktičnoj upotrebi definisanih sintaksi. Za potrebe istraživanja razvijen je softverski paket koji omogućava rad kako sa standardnim, tako i sa fazi XML, XSD i DTD dokumentima, kao i postavljanje i izvršavanje prioritizovanih fazi XQuery upita. Alat je testiran nad primerima iz prakse.In real life, as opposed to virtual, most information comes in the form of imprecise or incomplete values. XML  technology has made great progress in the field of storage and data transfer. This doctoral dissertation  developed XML extension which combines  indefiniteness in the values of XML and indefiniteness in the structure of XML into a single fuzzy XML extension. Furthermore,  dissertation  expands XQuery syntax with fuzzy values and  includes priorities and thresholds in fuzzy XQuery extension using  GPFCSP for the first time. Unlike other papers that focus on setting up theories and defining syntax, in this  dissertation, special attention  is turned to their practical use. A tool for working with XML, XSD and DTD documents and prioritized fuzzy XQuery extension queries has been developed. The tool has been tested on practical examples

    The development a dedicated system for the application fuzzy logic in systems for manage of XML documents

    Get PDF
    U stvarnom životu većina informacija dolazi kao neprecizne ili nepotpune vrednosti.  XML tehnologija je napravila veliki napredak u oblasti skladištenja i prenosa podataka.  Doktorska disertacija definiše fazi XML sintaksu koja kombinuje neodređenosti u vrednostima XML-a i neodređenosti u strukturi XML dokumenata. Takođe, rad proširuje standardnu XQuery upitnu sintaksu fazi elementima i uvodi prioritete i pragove zadovoljenja pomoću GPFCSP-a. Za razliku od drugih radova koji se fokusiraju na postavljanje teorija i definisanje sintaksi, rad posebnu pažnju posvećuje praktičnoj upotrebi definisanih sintaksi. Za potrebe istraživanja razvijen je softverski paket koji omogućava rad kako sa standardnim, tako i sa fazi XML, XSD i DTD dokumentima, kao i postavljanje i izvršavanje prioritizovanih fazi XQuery upita. Alat je testiran nad primerima iz prakse.In real life, as opposed to virtual, most information comes in the form of imprecise or incomplete values. XML  technology has made great progress in the field of storage and data transfer. This doctoral dissertation  developed XML extension which combines  indefiniteness in the values of XML and indefiniteness in the structure of XML into a single fuzzy XML extension. Furthermore,  dissertation  expands XQuery syntax with fuzzy values and  includes priorities and thresholds in fuzzy XQuery extension using  GPFCSP for the first time. Unlike other papers that focus on setting up theories and defining syntax, in this  dissertation, special attention  is turned to their practical use. A tool for working with XML, XSD and DTD documents and prioritized fuzzy XQuery extension queries has been developed. The tool has been tested on practical examples

    Impliance: A Next Generation Information Management Appliance

    Full text link
    ably successful in building a large market and adapting to the changes of the last three decades, its impact on the broader market of information management is surprisingly limited. If we were to design an information management system from scratch, based upon today's requirements and hardware capabilities, would it look anything like today's database systems?" In this paper, we introduce Impliance, a next-generation information management system consisting of hardware and software components integrated to form an easy-to-administer appliance that can store, retrieve, and analyze all types of structured, semi-structured, and unstructured information. We first summarize the trends that will shape information management for the foreseeable future. Those trends imply three major requirements for Impliance: (1) to be able to store, manage, and uniformly query all data, not just structured records; (2) to be able to scale out as the volume of this data grows; and (3) to be simple and robust in operation. We then describe four key ideas that are uniquely combined in Impliance to address these requirements, namely the ideas of: (a) integrating software and off-the-shelf hardware into a generic information appliance; (b) automatically discovering, organizing, and managing all data - unstructured as well as structured - in a uniform way; (c) achieving scale-out by exploiting simple, massive parallel processing, and (d) virtualizing compute and storage resources to unify, simplify, and streamline the management of Impliance. Impliance is an ambitious, long-term effort to define simpler, more robust, and more scalable information systems for tomorrow's enterprises.Comment: This article is published under a Creative Commons License Agreement (http://creativecommons.org/licenses/by/2.5/.) You may copy, distribute, display, and perform the work, make derivative works and make commercial use of the work, but, you must attribute the work to the author and CIDR 2007. 3rd Biennial Conference on Innovative Data Systems Research (CIDR) January 710, 2007, Asilomar, California, US

    Vereinheitlichte Anfrageverarbeitung in heterogenen und verteilten Multimediadatenbanken

    Get PDF
    Multimedia retrieval is an essential part of today's world. This situation is observable in industrial domains, e.g., medical imaging, as well as in the private sector, visible by activities in manifold Social Media platforms. This trend led to the creation of a huge environment of multimedia information retrieval services offering multimedia resources for almost any user requests. Indeed, the encompassed data is in general retrievable by (proprietary) APIs and query languages, but unfortunately a unified access is not given due to arising interoperability issues between those services. In this regard, this thesis focuses on two application scenarios, namely a medical retrieval system supporting a radiologist's workflow, as well as an interoperable image retrieval service interconnecting diverse data silos. The scientific contribution of this dissertation is split in three different parts: the first part of this thesis improves the metadata interoperability issue. Here, major contributions to a community-driven, international standardization have been proposed leading to the specification of an API and ontology to enable a unified annotation and retrieval of media resources. The second part issues a metasearch engine especially designed for unified retrieval in distributed and heterogeneous multimedia retrieval environments. This metasearch engine is capable of being operated in a federated as well as autonomous manner inside the aforementioned application scenarios. The remaining third part ensures an efficient retrieval due to the integration of optimization techniques for multimedia retrieval in the overall query execution process of the metasearch engine.Egal ob im industriellen Bereich oder auch im Social Media - multimediale Daten nehmen eine immer zentralere Rolle ein. Aus diesem fortlaufendem Entwicklungsprozess entwickelten sich umfangreiche Informationssysteme, die Daten für zahlreiche Bedürfnisse anbieten. Allerdings ist ein einheitlicher Zugriff auf jene verteilte und heterogene Landschaft von Informationssystemen in der Praxis nicht gewährleistet. Und dies, obwohl die Datenbestände meist über Schnittstellen abrufbar sind. Im Detail widmet sich diese Arbeit mit der Bearbeitung zweier Anwendungsszenarien. Erstens, einem medizinischen System zur Diagnoseunterstützung und zweitens einer interoperablen, verteilten Bildersuche. Der wissenschaftliche Teil der vorliegenden Dissertation gliedert sich in drei Teile: Teil eins befasst sich mit dem Problem der Interoperabilität zwischen verschiedenen Metadatenformaten. In diesem Bereich wurden maßgebliche Beiträge für ein internationales Standardisierungsverfahren entwickelt. Ziel war es, einer Ontologie, sowie einer Programmierschnittstelle einen vereinheitlichten Zugriff auf multimediale Informationen zu ermöglichen. In Teil zwei wird eine externe Metasuchmaschine vorgestellt, die eine einheitliche Anfrageverarbeitung in heterogenen und verteilten Multimediadatenbanken ermöglicht. In den Anwendungsszenarien wird zum einen auf eine föderative, als auch autonome Anfrageverarbeitung eingegangen. Abschließend werden in Teil drei Techniken zur Optimierung von verteilten multimedialen Anfragen präsentiert

    Migrating 120,000 Legacy Publications from Several Systems into a Current Research Information System Using Advanced Data Wrangling Techniques

    Get PDF
    This article describes a complex CRIS (current research information system) implementation project involving the migration of around 120,000 legacy publication records from three different systems. The project, undertaken by Tampere University, encountered several challenges in data diversity, data quality, and resource allocation. To handle the extensive and heterogenous dataset, innovative approaches such as machine learning techniques and various data wrangling tools were used to process data, correct errors, and merge information from different sources. Despite significant delays and unforeseen obstacles, the project was ultimately successful in achieving its goals. The project served as a valuable learning experience, highlighting the importance of data quality and standardized practices, and the need for dedicated resources in handling complex data migration projects in research organizations. This study stands out for its comprehensive documentation of the data wrangling and migration process, which has been less explored in the context of CRIS literature.Peer reviewe

    Parsing for agile modeling

    Get PDF
    Agile modeling refers to a set of methods that allow for a quick initial development of an importer and its further refinement. These requirements are not met simultaneously by the current parsing technology. Problems with parsing became a bottleneck in our research of agile modeling. In this thesis we introduce a novel approach to specify and build parsers. Our approach allows for expressive, tolerant and composable parsers without sacrificing performance. The approach is based on a context-sensitive extension of parsing expression grammars that allows a grammar engineer to specify complex language restrictions. To insure high parsing performance we automatically analyze a grammar definition and choose different parsing strategies for different parts of the grammar. We show that context-sensitive parsing expression grammars allow for highly composable, tolerant and variable-grained parsers that can be easily refined. Different parsing strategies significantly insure high-performance of parsers without sacrificing expressiveness of the underlying grammars

    Resource-aware business process management : analysis and support

    Get PDF
    corecore