19,902 research outputs found

    Term-Specific Eigenvector-Centrality in Multi-Relation Networks

    Get PDF
    Fuzzy matching and ranking are two information retrieval techniques widely used in web search. Their application to structured data, however, remains an open problem. This article investigates how eigenvector-centrality can be used for approximate matching in multi-relation graphs, that is, graphs where connections of many different types may exist. Based on an extension of the PageRank matrix, eigenvectors representing the distribution of a term after propagating term weights between related data items are computed. The result is an index which takes the document structure into account and can be used with standard document retrieval techniques. As the scheme takes the shape of an index transformation, all necessary calculations are performed during index tim

    Search Engines Giving You Garbage? Put A Corc In It, Implementing The Cooperative Online Resource Catalog

    Full text link
    This paper presents an implementation strategy for adding Internet resources to a library online catalog using OCLC\u27s Cooperative Online Resource Catalog (CORC). Areas of consideration include deciding which electronic resources to include in the online catalog and how to select them. The value and importance of pathfinders in creating electronic bibliographies and the role of library staff in updating them is introduced. Using an electronic suggestion form as a means of Internet resource collection development is another innovative method of enriching library collections. Education and training for cataloging staff on Dublin Core elements is also needed. Attention should be paid to the needs of distance learners in providing access to Internet resources. The significance of evaluating the appropriateness of Internet resources for library collections is emphasized

    The Christian Reformed Church Periodical Index: A Local Solution to Indexing Periodicals

    Get PDF
    This article describes the creation of a web-based database that indexes less well-known periodical titles of importance to scholars in the Christian Reformed Church, and generally not covered by other indexing services. The author explains how the data from the index, originally stored in a card catalog, was moved online to a text-based system, and eventually into its present form in a web-based system. Highlighting some of the challenges that were overcome in creating this resource, brief details are provided on how the data is stored and retrieved in the web environment, on how the data are searched and presented to the researcher, and on the methods used to keep the database current

    Semantic Support for Log Analysis of Safety-Critical Embedded Systems

    Full text link
    Testing is a relevant activity for the development life-cycle of Safety Critical Embedded systems. In particular, much effort is spent for analysis and classification of test logs from SCADA subsystems, especially when failures occur. The human expertise is needful to understand the reasons of failures, for tracing back the errors, as well as to understand which requirements are affected by errors and which ones will be affected by eventual changes in the system design. Semantic techniques and full text search are used to support human experts for the analysis and classification of test logs, in order to speedup and improve the diagnosis phase. Moreover, retrieval of tests and requirements, which can be related to the current failure, is supported in order to allow the discovery of available alternatives and solutions for a better and faster investigation of the problem.Comment: EDCC-2014, BIG4CIP-2014, Embedded systems, testing, semantic discovery, ontology, big dat

    Reconciliation through Description: Using Metadata to Realize the Vision of the National Research Centre for Truth and Reconciliation

    Get PDF
    PostprintThis articlewill discuss the history and context surrounding the document collection and statement gathering mandates of the Truth and Reconciliation Commission of Canada and the challenges the newly established National Research Centre for Truth and Reconciliation will face in applying the Commission’s metadata set in the realization of its vision. By working respectfully with Indigenous people through the implementation of Indigenous knowledge best practices and the application of contrasting traditional/nontraditional, archival/user-generated, and institutional/Indigenous descriptive elements, the Centre will attempt to create a “living archive” and facilitate Indigenous participation, collaboration, and ultimately, the process of reconciliation.https://www-tandfonline-com.uwinnipeg.idm.oclc.org/doi/full/10.1080/01639374.2015.100871

    Impliance: A Next Generation Information Management Appliance

    Full text link
    ably successful in building a large market and adapting to the changes of the last three decades, its impact on the broader market of information management is surprisingly limited. If we were to design an information management system from scratch, based upon today's requirements and hardware capabilities, would it look anything like today's database systems?" In this paper, we introduce Impliance, a next-generation information management system consisting of hardware and software components integrated to form an easy-to-administer appliance that can store, retrieve, and analyze all types of structured, semi-structured, and unstructured information. We first summarize the trends that will shape information management for the foreseeable future. Those trends imply three major requirements for Impliance: (1) to be able to store, manage, and uniformly query all data, not just structured records; (2) to be able to scale out as the volume of this data grows; and (3) to be simple and robust in operation. We then describe four key ideas that are uniquely combined in Impliance to address these requirements, namely the ideas of: (a) integrating software and off-the-shelf hardware into a generic information appliance; (b) automatically discovering, organizing, and managing all data - unstructured as well as structured - in a uniform way; (c) achieving scale-out by exploiting simple, massive parallel processing, and (d) virtualizing compute and storage resources to unify, simplify, and streamline the management of Impliance. Impliance is an ambitious, long-term effort to define simpler, more robust, and more scalable information systems for tomorrow's enterprises.Comment: This article is published under a Creative Commons License Agreement (http://creativecommons.org/licenses/by/2.5/.) You may copy, distribute, display, and perform the work, make derivative works and make commercial use of the work, but, you must attribute the work to the author and CIDR 2007. 3rd Biennial Conference on Innovative Data Systems Research (CIDR) January 710, 2007, Asilomar, California, US
    corecore