Prior Disambiguation of Word Tensors for Constructing Sentence Vectors

Abstract

Recent work has shown that compositional-distributional models using element-wise op-erations on contextual word vectors benefit from the introduction of a prior disambigua-tion step. The purpose of this paper is to generalise these ideas to tensor-based models, where relational words such as verbs and ad-jectives are represented by linear maps (higher order tensors) acting on a number of argu-ments (vectors). We propose disambiguation algorithms for a number of tensor-based mod-els, which we then test on a variety of tasks. The results show that disambiguation can pro-vide better compositional representation even for the case of tensor-based models. Further-more, we confirm previous findings regarding the positive effect of disambiguation on vec-tor mixture models, and we compare the ef-fectiveness of the two approaches.

Similar works

Full text

thumbnail-image

CiteSeerX

redirect
Last time updated on 29/10/2017

This paper was published in CiteSeerX.

Having an issue?

Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.