1,869 research outputs found
Mistakes in medical ontologies: Where do they come from and how can they be detected?
We present the details of a methodology for quality assurance in large medical terminologies and describe three algorithms that can help terminology developers and users to identify potential mistakes. The methodology is based in part on linguistic criteria and in part on logical and ontological principles governing sound classifications. We conclude by outlining the results of applying the methodology in the form of a taxonomy different types of errors and potential errors detected in SNOMED-CT
"More Deterministic" vs. "Smaller" Buechi Automata for Efficient LTL Model Checking
The standard technique for LTL model checking (M\models\neg\vi) consists on translating the negation of the LTL specification, \vi, into a B\"uchi automaton A_\vi, and then on checking if the product M \times A_\vi has an empty language. The efforts to maximize the efficiency of this process have so far concentrated on developing translation algorithms producing B\"uchi automata which are ``{\em as small as possible}'', under the implicit conjecture that this fact should make the final product smaller. In this paper we build on a different conjecture and present an alternative approach in which we generate instead B\"uchi automata which are ``{\em as deterministic as possible}'', in the sense that we try to reduce as much as we are able to the presence of non-deterministic decision states in A_\vi. We motivate our choice and present some empirical tests to support this approach
Abstract Canonical Inference
An abstract framework of canonical inference is used to explore how different
proof orderings induce different variants of saturation and completeness.
Notions like completion, paramodulation, saturation, redundancy elimination,
and rewrite-system reduction are connected to proof orderings. Fairness of
deductive mechanisms is defined in terms of proof orderings, distinguishing
between (ordinary) "fairness," which yields completeness, and "uniform
fairness," which yields saturation.Comment: 28 pages, no figures, to appear in ACM Trans. on Computational Logi
A new methodology for automatic fault tree construction based on component and mark libraries
During the design stage of the development of a new system, automated fault tree construction would produce results a lot sooner than the manual process and hence be highly beneficial in order to modify the system design based on identified weakest areas. Although much work has been performed in this area, the construction of fault trees is still generally done manually. In this paper, a new methodology of constructing fault trees from a system description is proposed. Multi-state input/output tables are introduced, which have the capability to capture output deviations during the normal operation of a component as well as under the influence of abnormality or failure. Two libraries, namely, a component library and a mark library, are introduced. The former stores component models and the latter stores a range of marks. The main purpose of a mark is to identify a certain feature of the system, such as a feedback loop or multiple redundancies. These two libraries are used to redraw the system in a graphical environment where the designer can witness the system come together and also input the necessary failure data for each component. An algorithm has been developed, that uses input/output tables and marks, to automatically construct fault trees for failure modes of interest. In order to demonstrate this methodology, it is applied to an automotive emission control system, and a fault tree is generated using the algorithm developed in this work
All the World's a (Hyper)Graph: A Data Drama
We introduce Hyperbard, a dataset of diverse relational data representationsderived from Shakespeare's plays. Our representations range from simple graphscapturing character co-occurrence in single scenes to hypergraphs encodingcomplex communication settings and character contributions as hyperedges withedge-specific node weights. By making multiple intuitive representationsreadily available for experimentation, we facilitate rigorous representationrobustness checks in graph learning, graph mining, and network analysis,highlighting the advantages and drawbacks of specific representations.Leveraging the data released in Hyperbard, we demonstrate that many solutionsto popular graph mining problems are highly dependent on the representationchoice, thus calling current graph curation practices into question. As anhomage to our data source, and asserting that science can also be art, wepresent all our points in the form of a play.<br
Semantic optimisation in datalog programs
Bibliography: leaves 138-142.Datalog is the fusion of Prolog and Database technologies aimed at producing an efficient, logic-based, declarative language for databases. This fusion takes the best of logic programming for the syntax of Datalog, and the best of database systems for the operational part of Datalog. As is the case with all declarative languages, optimisation is necessary to improve the efficiency of programs. Semantic optimisation uses meta-knowledge describing the data in the database to optimise queries and rules, aiming to reduce the resources required to answer queries. In this thesis, I analyse prior work that has been done on semantic optimisation and then propose an optimisation system for Datalog that includes optimisation of recursive programs and a semantic knowledge management module. A language, DatalogiC, which is an extension of Datalog that allows semantic knowledge to be expressed, has also been devised as an implementation vehicle. Finally, empirical results concerning the benefits of semantic optimisation are reported
Recommended from our members
Detecting Personal Life Events from Social Media
Social media has become a dominating force over the past 15 years, with the rise of sites such as Facebook, Instagram, and Twitter. Some of us have been with these sites since the start, posting all about our personal lives and building up a digital identify of ourselves.
But within this myriad of posts, what actually matters to us, and what do our digital identities tell people about ourselves? One way that we can start to filter through this data, is to build classifiers that can identify posts about our personal life events, allowing us to start to self reflect on what we share online.
The advantages of this type of technology also have direct merits within marketing, allowing companies to target customers with better products. We also suggest that the techniques and methodologies built throughout this thesis also have opportunities to support research within other areas such as cyber bullying, and radicalisation detection.
The aim of this thesis is to build upon the under researched area of life event detection, specifically targeting Twitter, and Instagram. Our goal is to develop classifiers that identify a list of life events inspired by cognitive psychology, where we target a total of seven within this thesis.
To achieve this we look to answer three research questions covered in each of our empirical chapters. In our first empirical chapter, we ask; What features would improve the classification of important life events. To answer this, we look at first extracting a new dataset from Twitter targeting the following events: Getting Married, Having Children, Starting School, Falling in Love, and Death of a Parent. We look at three new feature sets: interactions, content, and semantic features, and compare against a current state of the art technique.
In our second empirical chapter, we draw inspiration from cheminformatics, and frequent sub-graph mining to ask; Could the inclusion of semantic and syntactic patterns improve performance in our life event classifier. Here we look at expanding our tweets into semantic networks, as well as consider two forms of syntactic relationships between tokens. We then mine for frequent sub-graphs amongst our tweet graphs, and use these as features in our classifier. Our results produce F1 scores of between 0.65 and 0.77, providing an improvement between 0.01 and 0.04 against the current state of the art.
In our final empirical chapter, we look to answer our third research question; How can we detect important life events from other social media sites, such as Instagram?. We ask this question, as we believe Instagram to be a preferred environment to share personal life events. In this chapter, we extract a new dataset, targeting the following events: Getting Married, Having Children, Starting School, Graduation, and Buying a House. Our results find that our methodology provides F1 scores between 0.78, and 0.82, an improvement in F1 score between 0.01 and 0.04 against the current state of the art
- …