212,463 research outputs found
Syn-QG: Syntactic and Shallow Semantic Rules for Question Generation
Question Generation (QG) is fundamentally a simple syntactic transformation;
however, many aspects of semantics influence what questions are good to form.
We implement this observation by developing Syn-QG, a set of transparent
syntactic rules leveraging universal dependencies, shallow semantic parsing,
lexical resources, and custom rules which transform declarative sentences into
question-answer pairs. We utilize PropBank argument descriptions and VerbNet
state predicates to incorporate shallow semantic content, which helps generate
questions of a descriptive nature and produce inferential and semantically
richer questions than existing systems. In order to improve syntactic fluency
and eliminate grammatically incorrect questions, we employ back-translation
over the output of these syntactic rules. A set of crowd-sourced evaluations
shows that our system can generate a larger number of highly grammatical and
relevant questions than previous QG systems and that back-translation
drastically improves grammaticality at a slight cost of generating irrelevant
questions.Comment: Some of the results in the paper were incorrec
Right Here Right Now (RHRN) pilot study: testing a method of near-real-time data collection on the social determinants of health
Background: Informing policy and practice with up-to-date evidence on the social determinants of health is an ongoing challenge. One limitation of traditional approaches is the time-lag between identification of a policy or practice need and availability of results. The Right Here Right Now (RHRN) study piloted a near-real-time data-collection process to investigate whether this gap could be bridged. Methods: A website was developed to facilitate the issue of questions, data capture and presentation of findings. Respondents were recruited using two distinct methods – a clustered random probability sample, and a quota sample from street stalls. Weekly four-part questions were issued by email, Short Messaging Service (SMS or text) or post. Quantitative data were descriptively summarised, qualitative data thematically analysed, and a summary report circulated two weeks after each question was issued. The pilot spanned 26 weeks. Results: It proved possible to recruit and retain a panel of respondents providing quantitative and qualitative data on a range of issues. The samples were subject to similar recruitment and response biases as more traditional data-collection approaches. Participants valued the potential to influence change, and stakeholders were enthusiastic about the findings generated, despite reservations about the lack of sample representativeness. Stakeholders acknowledged that decision-making processes are not flexible enough to respond to weekly evidence. Conclusion: RHRN produced a process for collecting near-real-time data for policy-relevant topics, although obtaining and maintaining representative samples was problematic. Adaptations were identified to inform a more sustainable model of near-real-time data collection and dissemination in the future
Working with Legacy Media: A Lone Arranger\u27s First Steps
[Excerpt] In 2013, a naked hard drive from Fiji arriving in my small religious archives (an equivalent full-time staff of 2.5 – one archivist and two archives’ assistants) started me off on the path of digital preservation and, in particular, the digital forensics practices that are beneficial for archivists. With such a small staff, outsourced IT services, and no digital preservation policy in sight, it was time to start exploring how institutions of my size could manage legacy media and start planning for the born-digital archives that will continue to arrive. Since I hold a part-time position, I was able to undertake this exploration in my own time through the support provided by a scholarship from the Ian McLean Wards Memorial Trust in 2015
BlogForever D2.6: Data Extraction Methodology
This report outlines an inquiry into the area of web data extraction, conducted within the context of blog preservation. The report reviews theoretical advances and practical developments for implementing data extraction. The inquiry is extended through an experiment that demonstrates the effectiveness and feasibility of implementing some of the suggested approaches. More specifically, the report discusses an approach based on unsupervised machine learning that employs the RSS feeds and HTML representations of blogs. It outlines the possibilities of extracting semantics available in blogs and demonstrates the benefits of exploiting available standards such as microformats and microdata. The report proceeds to propose a methodology for extracting and processing blog data to further inform the design and development of the BlogForever platform
Applying semantic web technologies to knowledge sharing in aerospace engineering
This paper details an integrated methodology to optimise Knowledge reuse and sharing, illustrated with a use case in the aeronautics domain. It uses Ontologies as a central modelling strategy for the Capture of Knowledge from legacy docu-ments via automated means, or directly in systems interfacing with Knowledge workers, via user-defined, web-based forms. The domain ontologies used for Knowledge Capture also guide the retrieval of the Knowledge extracted from the data using a Semantic Search System that provides support for multiple modalities during search. This approach has been applied and evaluated successfully within the aerospace domain, and is currently being extended for use in other domains on an increasingly large scale
Governing for Genuine Profit
Business corporations seek profit. That is, after subtracting cost, they maximize net revenue. Spillovers (both costs and benefits) involve trade-offs governing boards should make. Spillovers, especially when coupled with clumsy applications of discounted present value, distort a business' perception of profit. Today, businesses are buffeted by the old risks of recession and the new risks of terrorism. If modern society is to survive, then the seeds of terrorism and their fruit of tremendous loss must be contained. Accordingly, governing boards must propel businesses towards a paradigm of genuine profit. Governing boards must insist that their businesses prospect for positive feedback loops and implement a sustainable profit stream. In short, governing boards must insist that business be entrepreneurial.http://deepblue.lib.umich.edu/bitstream/2027.42/39918/3/wp533.pd
Media play: new (and old) pedagogies with young chldren
Modern technologies mean that the principles of quality arts education are the same (as they ever were) and different. Discussion in this paper is based on a small research project that used art as pedagogy, art as research method and, for the young children participants, celebrated art for art's sake. The project was designed with two aims. Firstly, the authors were interested in how young children engage with media as a strand of the arts. This also informed some of their thinking around the debates over Information and Communication Technology (ICT) as a process for the production of a media text. Secondly, they were interested in the extent to which digital media could enable young children to make their learning visible
What is an Analogue for the Semantic Web and Why is Having One Important?
This paper postulates that for the Semantic Web to grow and gain input from fields that will surely benefit it, it needs to develop an analogue that will help people not only understand what it is, but what the potential opportunities are that are enabled by these new protocols. The model proposed in the paper takes the way that Web interaction has been framed as a baseline to inform a similar analogue for the Semantic Web. While the Web has been represented as a Page + Links, the paper presents the argument that the Semantic Web can be conceptualized as a Notebook + Memex. The argument considers how this model also presents new challenges for fundamental human interaction with computing, and that hypertext models have much to contribute to this new understanding for distributed information systems
Recommended from our members
Using mobile RE tools to give end-users their own voice
Researchers highlight end-user involvement in system design as an important concept for developing useful and usable solutions. However, end-user involvement in software engineering is still an open-ended topic. Novel paradigms such as service-oriented computing strengthen the need for more active end-user involvement in order to provide systems that are tailored to individual end-user needs. Our work is based on the fact that the majority of end-users are familiar with mobile devices and use an increasing number of mobile applications. A mobile tool enabling end-user led requirements elicitation could be just one of many applications installed on end-users' mobile devices. In this paper, we present a framework of end-user involvement in requirements elicitation which motivates our research. The main contribution of our research is a tool-supported requirements elicitation approach allowing end-users to document needs in situ. Furthermore, we present first evaluation results to highlight the feasibility of on-site end-user led requirements elicitation
- …