3 research outputs found
One Single Deep Bidirectional LSTM Network for Word Sense Disambiguation of Text Data
Due to recent technical and scientific advances, we have a wealth of
information hidden in unstructured text data such as offline/online narratives,
research articles, and clinical reports. To mine these data properly,
attributable to their innate ambiguity, a Word Sense Disambiguation (WSD)
algorithm can avoid numbers of difficulties in Natural Language Processing
(NLP) pipeline. However, considering a large number of ambiguous words in one
language or technical domain, we may encounter limiting constraints for proper
deployment of existing WSD models. This paper attempts to address the problem
of one-classifier-per-one-word WSD algorithms by proposing a single
Bidirectional Long Short-Term Memory (BLSTM) network which by considering
senses and context sequences works on all ambiguous words collectively.
Evaluated on SensEval-3 benchmark, we show the result of our model is
comparable with top-performing WSD algorithms. We also discuss how applying
additional modifications alleviates the model fault and the need for more
training data.Comment: 12 pages, 1 figure, to appear in the Proceedings of the 31st Canadian
Conference on Artificial Intelligence, 8-11 May, 2018, Toronto, Canad
Using Multi-Sense Vector Embeddings for Reverse Dictionaries
Popular word embedding methods such as word2vec and GloVe assign a single
vector representation to each word, even if a word has multiple distinct
meanings. Multi-sense embeddings instead provide different vectors for each
sense of a word. However, they typically cannot serve as a drop-in replacement
for conventional single-sense embeddings, because the correct sense vector
needs to be selected for each word. In this work, we study the effect of
multi-sense embeddings on the task of reverse dictionaries. We propose a
technique to easily integrate them into an existing neural network architecture
using an attention mechanism. Our experiments demonstrate that large
improvements can be obtained when employing multi-sense embeddings both in the
input sequence as well as for the target representation. An analysis of the
sense distributions and of the learned attention is provided as well.Comment: Accepted as long paper at the 13th International Conference on
Computational Semantics (IWCS 2019
A Novel Neural Sequence Model with Multiple Attentions for Word Sense Disambiguation
Word sense disambiguation (WSD) is a well researched problem in computational
linguistics. Different research works have approached this problem in different
ways. Some state of the art results that have been achieved for this problem
are by supervised models in terms of accuracy, but they often fall behind
flexible knowledge-based solutions which use engineered features as well as
human annotators to disambiguate every target word. This work focuses on
bridging this gap using neural sequence models incorporating the well-known
attention mechanism. The main gist of our work is to combine multiple
attentions on different linguistic features through weights and to provide a
unified framework for doing this. This weighted attention allows the model to
easily disambiguate the sense of an ambiguous word by attending over a suitable
portion of a sentence. Our extensive experiments show that multiple attention
enables a more versatile encoder-decoder model leading to state of the art
results.Comment: 9 pages, 3 Figures, Accepted as a conference paper in ICMLA 201