13,709 research outputs found
Meaningfulness, the unsaid and translatability. Instead of an introduction
The present paper opens this topical issue on translation techniques by drawing a theoretical basis for the discussion of translational issues in a linguistic perspective. In order to forward an audience- oriented definition of translation, I will describe different forms of linguistic variability, highlighting how they present different difficulties to translators, with an emphasis on the semantic and communicative complexity that a source text can exhibit. The problem is then further discussed through a comparison between Quine's radically holistic position and the translatability principle supported by such semanticists as Katz. General translatability — at the expense of additional complexity — is eventually proposed as a possible synthesis of this debate. In describing the meaningfulness levels of source texts through Hjelmslevian semiotics, and his semiotic hierarchy in particular, the paper attempts to go beyond denotative semiotic, and reframe some translational issues in a connotative semiotic and metasemiotic perspective
Learning Sentence-internal Temporal Relations
In this paper we propose a data intensive approach for inferring
sentence-internal temporal relations. Temporal inference is relevant for
practical NLP applications which either extract or synthesize temporal
information (e.g., summarisation, question answering). Our method bypasses the
need for manual coding by exploiting the presence of markers like after", which
overtly signal a temporal relation. We first show that models trained on main
and subordinate clauses connected with a temporal marker achieve good
performance on a pseudo-disambiguation task simulating temporal inference
(during testing the temporal marker is treated as unseen and the models must
select the right marker from a set of possible candidates). Secondly, we assess
whether the proposed approach holds promise for the semi-automatic creation of
temporal annotations. Specifically, we use a model trained on noisy and
approximate data (i.e., main and subordinate clauses) to predict
intra-sentential relations present in TimeBank, a corpus annotated rich
temporal information. Our experiments compare and contrast several
probabilistic models differing in their feature space, linguistic assumptions
and data requirements. We evaluate performance against gold standard corpora
and also against human subjects
Summarising News Stories for Children
This paper proposes a system to automatically summarise news articles in a manner suitable for children by deriving and combining statistical ratings for how important, positively oriented and easy to read each sentence is. Our results demonstrate that this approach succeeds in generating summaries that are suitable for children, and that there is further scope for combining this extractive approach with abstractive methods used in text implification
Modus Ponens and the Logic of Decision
If modus ponens is valid, then you should take up smoking
Discourse Level Factors for Sentence Deletion in Text Simplification
This paper presents a data-driven study focusing on analyzing and predicting
sentence deletion -- a prevalent but understudied phenomenon in document
simplification -- on a large English text simplification corpus. We inspect
various document and discourse factors associated with sentence deletion, using
a new manually annotated sentence alignment corpus we collected. We reveal that
professional editors utilize different strategies to meet readability standards
of elementary and middle schools. To predict whether a sentence will be deleted
during simplification to a certain level, we harness automatically aligned data
to train a classification model. Evaluated on our manually annotated data, our
best models reached F1 scores of 65.2 and 59.7 for this task at the levels of
elementary and middle school, respectively. We find that discourse level
factors contribute to the challenging task of predicting sentence deletion for
simplification.Comment: Accepted in AAAI 2020. Adding more details on manual data annotatio
A Type-coherent, Expressive Representation as an Initial Step to Language Understanding
A growing interest in tasks involving language understanding by the NLP
community has led to the need for effective semantic parsing and inference.
Modern NLP systems use semantic representations that do not quite fulfill the
nuanced needs for language understanding: adequately modeling language
semantics, enabling general inferences, and being accurately recoverable. This
document describes underspecified logical forms (ULF) for Episodic Logic (EL),
which is an initial form for a semantic representation that balances these
needs. ULFs fully resolve the semantic type structure while leaving issues such
as quantifier scope, word sense, and anaphora unresolved; they provide a
starting point for further resolution into EL, and enable certain structural
inferences without further resolution. This document also presents preliminary
results of creating a hand-annotated corpus of ULFs for the purpose of training
a precise ULF parser, showing a three-person pairwise interannotator agreement
of 0.88 on confident annotations. We hypothesize that a divide-and-conquer
approach to semantic parsing starting with derivation of ULFs will lead to
semantic analyses that do justice to subtle aspects of linguistic meaning, and
will enable construction of more accurate semantic parsers.Comment: Accepted for publication at The 13th International Conference on
Computational Semantics (IWCS 2019
Staging Transformations for Multimodal Web Interaction Management
Multimodal interfaces are becoming increasingly ubiquitous with the advent of
mobile devices, accessibility considerations, and novel software technologies
that combine diverse interaction media. In addition to improving access and
delivery capabilities, such interfaces enable flexible and personalized dialogs
with websites, much like a conversation between humans. In this paper, we
present a software framework for multimodal web interaction management that
supports mixed-initiative dialogs between users and websites. A
mixed-initiative dialog is one where the user and the website take turns
changing the flow of interaction. The framework supports the functional
specification and realization of such dialogs using staging transformations --
a theory for representing and reasoning about dialogs based on partial input.
It supports multiple interaction interfaces, and offers sessioning, caching,
and co-ordination functions through the use of an interaction manager. Two case
studies are presented to illustrate the promise of this approach.Comment: Describes framework and software architecture for multimodal web
interaction managemen
- …