212,463 research outputs found

    Syn-QG: Syntactic and Shallow Semantic Rules for Question Generation

    Full text link
    Question Generation (QG) is fundamentally a simple syntactic transformation; however, many aspects of semantics influence what questions are good to form. We implement this observation by developing Syn-QG, a set of transparent syntactic rules leveraging universal dependencies, shallow semantic parsing, lexical resources, and custom rules which transform declarative sentences into question-answer pairs. We utilize PropBank argument descriptions and VerbNet state predicates to incorporate shallow semantic content, which helps generate questions of a descriptive nature and produce inferential and semantically richer questions than existing systems. In order to improve syntactic fluency and eliminate grammatically incorrect questions, we employ back-translation over the output of these syntactic rules. A set of crowd-sourced evaluations shows that our system can generate a larger number of highly grammatical and relevant questions than previous QG systems and that back-translation drastically improves grammaticality at a slight cost of generating irrelevant questions.Comment: Some of the results in the paper were incorrec

    Right Here Right Now (RHRN) pilot study: testing a method of near-real-time data collection on the social determinants of health

    Get PDF
    Background: Informing policy and practice with up-to-date evidence on the social determinants of health is an ongoing challenge. One limitation of traditional approaches is the time-lag between identification of a policy or practice need and availability of results. The Right Here Right Now (RHRN) study piloted a near-real-time data-collection process to investigate whether this gap could be bridged. Methods: A website was developed to facilitate the issue of questions, data capture and presentation of findings. Respondents were recruited using two distinct methods – a clustered random probability sample, and a quota sample from street stalls. Weekly four-part questions were issued by email, Short Messaging Service (SMS or text) or post. Quantitative data were descriptively summarised, qualitative data thematically analysed, and a summary report circulated two weeks after each question was issued. The pilot spanned 26 weeks. Results: It proved possible to recruit and retain a panel of respondents providing quantitative and qualitative data on a range of issues. The samples were subject to similar recruitment and response biases as more traditional data-collection approaches. Participants valued the potential to influence change, and stakeholders were enthusiastic about the findings generated, despite reservations about the lack of sample representativeness. Stakeholders acknowledged that decision-making processes are not flexible enough to respond to weekly evidence. Conclusion: RHRN produced a process for collecting near-real-time data for policy-relevant topics, although obtaining and maintaining representative samples was problematic. Adaptations were identified to inform a more sustainable model of near-real-time data collection and dissemination in the future

    Working with Legacy Media: A Lone Arranger\u27s First Steps

    Get PDF
    [Excerpt] In 2013, a naked hard drive from Fiji arriving in my small religious archives (an equivalent full-time staff of 2.5 – one archivist and two archives’ assistants) started me off on the path of digital preservation and, in particular, the digital forensics practices that are beneficial for archivists. With such a small staff, outsourced IT services, and no digital preservation policy in sight, it was time to start exploring how institutions of my size could manage legacy media and start planning for the born-digital archives that will continue to arrive. Since I hold a part-time position, I was able to undertake this exploration in my own time through the support provided by a scholarship from the Ian McLean Wards Memorial Trust in 2015

    BlogForever D2.6: Data Extraction Methodology

    Get PDF
    This report outlines an inquiry into the area of web data extraction, conducted within the context of blog preservation. The report reviews theoretical advances and practical developments for implementing data extraction. The inquiry is extended through an experiment that demonstrates the effectiveness and feasibility of implementing some of the suggested approaches. More specifically, the report discusses an approach based on unsupervised machine learning that employs the RSS feeds and HTML representations of blogs. It outlines the possibilities of extracting semantics available in blogs and demonstrates the benefits of exploiting available standards such as microformats and microdata. The report proceeds to propose a methodology for extracting and processing blog data to further inform the design and development of the BlogForever platform

    Applying semantic web technologies to knowledge sharing in aerospace engineering

    Get PDF
    This paper details an integrated methodology to optimise Knowledge reuse and sharing, illustrated with a use case in the aeronautics domain. It uses Ontologies as a central modelling strategy for the Capture of Knowledge from legacy docu-ments via automated means, or directly in systems interfacing with Knowledge workers, via user-defined, web-based forms. The domain ontologies used for Knowledge Capture also guide the retrieval of the Knowledge extracted from the data using a Semantic Search System that provides support for multiple modalities during search. This approach has been applied and evaluated successfully within the aerospace domain, and is currently being extended for use in other domains on an increasingly large scale

    Governing for Genuine Profit

    Get PDF
    Business corporations seek profit. That is, after subtracting cost, they maximize net revenue. Spillovers (both costs and benefits) involve trade-offs governing boards should make. Spillovers, especially when coupled with clumsy applications of discounted present value, distort a business' perception of profit. Today, businesses are buffeted by the old risks of recession and the new risks of terrorism. If modern society is to survive, then the seeds of terrorism and their fruit of tremendous loss must be contained. Accordingly, governing boards must propel businesses towards a paradigm of genuine profit. Governing boards must insist that their businesses prospect for positive feedback loops and implement a sustainable profit stream. In short, governing boards must insist that business be entrepreneurial.http://deepblue.lib.umich.edu/bitstream/2027.42/39918/3/wp533.pd

    Media play: new (and old) pedagogies with young chldren

    Get PDF
    Modern technologies mean that the principles of quality arts education are the same (as they ever were) and different. Discussion in this paper is based on a small research project that used art as pedagogy, art as research method and, for the young children participants, celebrated art for art's sake. The project was designed with two aims. Firstly, the authors were interested in how young children engage with media as a strand of the arts. This also informed some of their thinking around the debates over Information and Communication Technology (ICT) as a process for the production of a media text. Secondly, they were interested in the extent to which digital media could enable young children to make their learning visible

    What is an Analogue for the Semantic Web and Why is Having One Important?

    No full text
    This paper postulates that for the Semantic Web to grow and gain input from fields that will surely benefit it, it needs to develop an analogue that will help people not only understand what it is, but what the potential opportunities are that are enabled by these new protocols. The model proposed in the paper takes the way that Web interaction has been framed as a baseline to inform a similar analogue for the Semantic Web. While the Web has been represented as a Page + Links, the paper presents the argument that the Semantic Web can be conceptualized as a Notebook + Memex. The argument considers how this model also presents new challenges for fundamental human interaction with computing, and that hypertext models have much to contribute to this new understanding for distributed information systems
    corecore