133 research outputs found
On the Linearity of Semantic Change: Investigating Meaning Variation via Dynamic Graph Models
We consider two graph models of semantic change. The first is a time-series
model that relates embedding vectors from one time period to embedding vectors
of previous time periods. In the second, we construct one graph for each word:
nodes in this graph correspond to time points and edge weights to the
similarity of the word's meaning across two time points. We apply our two
models to corpora across three different languages. We find that semantic
change is linear in two senses. Firstly, today's embedding vectors (= meaning)
of words can be derived as linear combinations of embedding vectors of their
neighbors in previous time periods. Secondly, self-similarity of words decays
linearly in time. We consider both findings as new laws/hypotheses of semantic
change.Comment: Published at ACL 2016, Berlin (short papers
Neural End-to-End Learning for Computational Argumentation Mining
We investigate neural techniques for end-to-end computational argumentation
mining (AM). We frame AM both as a token-based dependency parsing and as a
token-based sequence tagging problem, including a multi-task learning setup.
Contrary to models that operate on the argument component level, we find that
framing AM as dependency parsing leads to subpar performance results. In
contrast, less complex (local) tagging models based on BiLSTMs perform robustly
across classification scenarios, being able to catch long-range dependencies
inherent to the AM problem. Moreover, we find that jointly learning 'natural'
subtasks, in a multi-task learning setup, improves performance.Comment: To be published at ACL 201
- …