25,747 research outputs found

    Data quality: Some comments on the NASA software defect datasets

    Get PDF
    Background-Self-evidently empirical analyses rely upon the quality of their data. Likewise, replications rely upon accurate reporting and using the same rather than similar versions of datasets. In recent years, there has been much interest in using machine learners to classify software modules into defect-prone and not defect-prone categories. The publicly available NASA datasets have been extensively used as part of this research. Objective-This short note investigates the extent to which published analyses based on the NASA defect datasets are meaningful and comparable. Method-We analyze the five studies published in the IEEE Transactions on Software Engineering since 2007 that have utilized these datasets and compare the two versions of the datasets currently in use. Results-We find important differences between the two versions of the datasets, implausible values in one dataset and generally insufficient detail documented on dataset preprocessing. Conclusions-It is recommended that researchers 1) indicate the provenance of the datasets they use, 2) report any preprocessing in sufficient detail to enable meaningful replication, and 3) invest effort in understanding the data prior to applying machine learners

    Design and implementation of a filter engine for semantic web documents

    Get PDF
    This report describes our project that addresses the challenge of changes in the semantic web. Some studies have already been done for the so-called adaptive semantic web, such as applying inferring rules. In this study, we apply the technology of Event Notification System (ENS). Treating changes as events, we developed a notification system for such events

    Answer Sets for Consistent Query Answering in Inconsistent Databases

    Full text link
    A relational database is inconsistent if it does not satisfy a given set of integrity constraints. Nevertheless, it is likely that most of the data in it is consistent with the constraints. In this paper we apply logic programming based on answer sets to the problem of retrieving consistent information from a possibly inconsistent database. Since consistent information persists from the original database to every of its minimal repairs, the approach is based on a specification of database repairs using disjunctive logic programs with exceptions, whose answer set semantics can be represented and computed by systems that implement stable model semantics. These programs allow us to declare persistence by defaults and repairing changes by exceptions. We concentrate mainly on logic programs for binary integrity constraints, among which we find most of the integrity constraints found in practice.Comment: 34 page

    A Framework for Reference Management in the Semantic Web

    No full text
    Much of the semantic web relies upon open and unhindered interoperability between diverse systems. The successful convergence of multiple ontologies and referencing schemes is key. This is hampered by a lack of any means for managing and communicating co-references. We have therefore developed an ontology and framework for the exploration and resolution of potential co-references, in the semantic web at large, that allow the user to a) discover and record uniquely identifying attributes b) interface candidates with and create pipelines of other systems for reference management c) record identified duplicates in a usable and retrievable manner, and d) provide a consistent reference service for accessing them. This paper describes this ontology and a framework of web services designed to support and utilise it

    Final report on the farmer's aid in plant disease diagnoses

    Get PDF
    This report is the final report on the FAD project. The FAD project was initiated in september 1985 to test the expert system shell Babylon by developing a prototype crop disease diagnosis system in it. A short overview of the history of the project and the main problems encountered is given in chapter 1. Chapter 2 describes the result of an attempt to integrate JSD with modelling techniques like generalisation and aggregation and chapter 3 concentrates on the method we used to elicit phytopathological knowledge from specialists. Chapter 4 gives the result of knowledge acquisition for the 10 wheat diseases most commonly occurring in the Netherlands. The user interface is described briefly in chapter 5 and chapter 6 gives an overview of the additions to the implementation we made to the version of FAD reported in our second report. Chapter 7, finally, summarises the conclusions of the project and gives recommendations for follow-up projects

    Self unbound: ego dissolution in psychedelic experience

    Get PDF
    Users of psychedelic drugs often report that their sense of being a self or ‘I’ distinct from the rest of the world has diminished or altogether dissolved. Neuroscientific study of such ‘ego dissolution’ experiences offers a window onto the nature of self-awareness. We argue that ego dissolution is best explained by an account that explains self-awareness as resulting from the integrated functioning of hierarchical predictive models which posit the existence of a stable and unchanging entity to which representations are bound. Combining recent work on the ‘integrative self' and the phenomenon of self-binding with predictive processing principles yields an explanation of ego dissolution according to which self-representation is a useful Cartesian fiction: an ultimately false representation of a simple and enduring substance to which attributes are bound which serves to integrate and unify cognitive processing across levels and domains. The self-model is not a mere narrative posit, as some have suggested; it has a more robust and ubiquitous cognitive function than that. But this does not mean, as others have claimed, that the self-model has the right attributes to qualify as a self. It performs some of the right kinds of functions, but it is not the right kind of entity. Ego dissolution experiences reveal that the self-model plays an important binding function in cognitive processing, but the self does not exist

    Open issues in semantic query optimization in relational DBMS

    Get PDF
    After two decades of research into Semantic Query Optimization (SQO) there is clear agreement as to the efficacy of SQO. However, although there are some experimental implementations there are still no commercial implementations. We first present a thorough analysis of research into SQO. We identify three problems which inhibit the effective use of SQO in Relational Database Management Systems(RDBMS). We then propose solutions to these problems and describe first steps towards the implementation of an effective semantic query optimizer for relational databases

    Bringing self assessment home: repository profiling and key lines of enquiry within DRAMBORA

    Get PDF
    Digital repositories are a manifestation of complex organizational, financial, legal, technological, procedural, and political interrelationships. Accompanying each of these are innate uncertainties, exacerbated by the relative immaturity of understanding prevalent within the digital preservation domain. Recent efforts have sought to identify core characteristics that must be demonstrable by successful digital repositories, expressed in the form of check-list documents, intended to support the processes of repository accreditation and certification. In isolation though, the available guidelines lack practical applicability; confusion over evidential requirements and difficulties associated with the diversity that exists among repositories (in terms of mandate, available resources, supported content and legal context) are particularly problematic. A gap exists between the available criteria and the ways and extent to which conformity can be demonstrated. The Digital Repository Audit Method Based on Risk Assessment (DRAMBORA) is a methodology for undertaking repository self assessment, developed jointly by the Digital Curation Centre (DCC) and DigitalPreservationEurope (DPE). DRAMBORA requires repositories to expose their organization, policies and infrastructures to rigorous scrutiny through a series of highly structured exercises, enabling them to build a comprehensive registry of their most pertinent risks, arranged into a structure that facilitates effective management. It draws on experiences accumulated throughout 18 evaluative pilot assessments undertaken in an internationally diverse selection of repositories, digital libraries and data centres (including institutions and services such as the UK National Digital Archive of Datasets, the National Archives of Scotland, Gallica at the National Library of France and the CERN Document Server). Other organizations, such as the British Library, have been using sections of DRAMBORA within their own risk assessment procedures. Despite the attractive benefits of a bottom up approach, there are implicit challenges posed by neglecting a more objective perspective. Following a sustained period of pilot audits undertaken by DPE, DCC and the DELOS Digital Preservation Cluster aimed at evaluating DRAMBORA, it was stated that had respective project members not been present to facilitate each assessment, and contribute their objective, external perspectives, the results may have been less useful. Consequently, DRAMBORA has developed in a number of ways, to enable knowledge transfer from the responses of comparable repositories, and incorporate more opportunities for structured question sets, or key lines of enquiry, that provoke more comprehensive awareness of the applicability of particular threats and opportunities

    Some Varieties of Superparadox. The implications of dynamic contradiction, the characteristic form of breakdown of breakdown of sense to which self-reference is prone

    Get PDF
    The Problem of the Paradoxes came to the fore in philosophy and mathematics with the discovery of Russell's Paradox in 1901. It is the "forgotten" intellectual-scientific problem of the Twentieth Century, because for more than sixty years a pretence was maintained, by a consensus of logicians, that the problem had been "solved"

    Guaranteeing no interaction between functional dependencies and tree-like inclusion dependencies

    Get PDF
    Functional dependencies (FDs) and inclusion dependencies (INDs) are the most fundamental integrity constraints that arise in practice in relational databases. A given set of FDs does not interact with a given set of INDs if logical implication of any FD can be determined solely by the given set of FDs, and logical implication of any IND can be determined solely by the given set of INDs. The set of tree-like INDs constitutes a useful subclass of INDs whose implication problem is polynomial time decidable. We exhibit a necessary and sufficient condition for a set of FDs and tree-like INDs not to interact; this condition can be tested in polynomial time
    corecore