1,296 research outputs found
Cross-Lingual Lexico-Semantic Transfer in Language Learning
Lexico-semantic knowledge of our native language provides an initial foundation for second language learning. In this paper, we investigate whether and to what extent the lexico-semantic models of the native language (L1) are transferred to the second language (L2). Specifically, we focus on the problem of lexical choice and investigate it in the context of three typologically diverse languages: Russian, Spanish and English. We show that a statistical semantic model learned from L1 data improves automatic error detection in L2 for the speakers of the respective L1. Finally, we investigate whether the semantic model learned from a particular L1 is portable to other, typologically related languages.Ekaterina Kochmar’s research is supported by Cambridge English Language Assessment via the ALTA Institute. Ekaterina Shutova’s research is supported by the Leverhulme Trust Early Career Fellowship
Semantic access in number word translation - The role of crosslingual lexical similarity
The revised hierarchical model of bilingualism (e.g., Kroll & Stewart, 1994) assumes that second language (1,2) words primarily access semantics through their first language (L1) translation equivalents. Consequently, backward translation from L2 to L1 should not imply semantic access but occurs through lexical wordform associations. However, recent research with Dutch-French bilinguals showed that both backward and forward translation of number words yields a semantic number magnitude effect (Duyck & Brysbaert, 2004), providing evidence for strong form-to-meaning mappings of L2 number words. In two number-word translation experiments with Dutch-English-German trilinguals, the present study investigated whether semantic access in L1-L2 and L1-L3 number-word translation depends on lexical similarity of the languages involved. We found that backward translation from these more similar language pairs to L1 still yields a semantic magnitude effect, whereas forward translation does not, in contrast with the Dutch-French results of Duyck and Brysbaert (2004). We argue against a dual route model of word translation and suggest that the degree of semantic activation in translation depends on lexical form overlap between translation equivalents
BabelBERT: Massively Multilingual Transformers Meet a Massively Multilingual Lexical Resource
While pretrained language models (PLMs) primarily serve as general purpose
text encoders that can be fine-tuned for a wide variety of downstream tasks,
recent work has shown that they can also be rewired to produce high-quality
word representations (i.e., static word embeddings) and yield good performance
in type-level lexical tasks. While existing work primarily focused on lexical
specialization of PLMs in monolingual and bilingual settings, in this work we
expose massively multilingual transformers (MMTs, e.g., mBERT or XLM-R) to
multilingual lexical knowledge at scale, leveraging BabelNet as the readily
available rich source of multilingual and cross-lingual type-level lexical
knowledge. Concretely, we leverage BabelNet's multilingual synsets to create
synonym pairs across languages and then subject the MMTs (mBERT and XLM-R)
to a lexical specialization procedure guided by a contrastive objective. We
show that such massively multilingual lexical specialization brings massive
gains in two standard cross-lingual lexical tasks, bilingual lexicon induction
and cross-lingual word similarity, as well as in cross-lingual sentence
retrieval. Crucially, we observe gains for languages unseen in specialization,
indicating that the multilingual lexical specialization enables generalization
to languages with no lexical constraints. In a series of subsequent controlled
experiments, we demonstrate that the pretraining quality of word
representations in the MMT for languages involved in specialization has a much
larger effect on performance than the linguistic diversity of the set of
constraints. Encouragingly, this suggests that lexical tasks involving
low-resource languages benefit the most from lexical knowledge of resource-rich
languages, generally much more available
Discriminating between lexico-semantic relations with the specialization tensor model
We present a simple and effective feed-forward neural architecture for discriminating between lexico-semantic relations (synonymy, antonymy, hypernymy, and meronymy). Our Specialization Tensor Model (STM) simultaneously produces multiple different specializations of input distributional word vectors, tailored for predicting lexico-semantic relations for word pairs. STM outperforms more complex state-of-the-art architectures on two benchmark datasets and exhibits stable performance across languages. We also show that, if coupled with a lingual distributional space, the proposed model can transfer the prediction of lexico-semantic relations to a resource-lean target language without any training data
Recommended from our members
Cross-lingual semantic specialization via lexical relation induction
Semantic specialization integrates structured linguistic knowledge from external resources (such as lexical relations in WordNet) into pretrained distributional vectors in the form of constraints. However, this technique cannot be leveraged in many languages, because their structured external resources are typically incomplete or non-existent. To bridge this gap, we propose a novel method that transfers specialization from a resource-rich source language (English) to virtually any target language. Our specialization transfer comprises two crucial steps: 1) Inducing noisy constraints in the target language through automatic word translation; and 2) Filtering the noisy constraints via a state-of-the-art relation prediction model trained on the source language constraints. This allows us to specialize any set of distributional vectors in the target language with the refined constraints. We prove the effectiveness of our method through intrinsic word similarity evaluation in 8 languages, and with 3 downstream tasks in 5 languages: lexical simplification, dialog state tracking, and semantic textual similarity. The gains over the previous state-of-art specialization methods are substantial and consistent across languages. Our results also suggest that the transfer method is effective even for lexically distant source-target language pairs. Finally, as a by-product, our method produces lists of WordNet-style lexical relations in resource-poor languages
Cross-lingual semantic specialization via lexical relation induction
Semantic specialization integrates structured linguistic knowledge from external resources (such as lexical relations in WordNet) into pretrained distributional vectors in the form of constraints. However, this technique cannot be leveraged in many languages, because their structured external resources are typically incomplete or non-existent. To bridge this gap, we propose a novel method that transfers specialization from a resource-rich source language (English) to virtually any target language. Our specialization transfer comprises two crucial steps: 1) Inducing noisy constraints in the target language through automatic word translation; and 2) Filtering the noisy constraints via a state-of-the-art relation prediction model trained on the source language constraints. This allows us to specialize any set of distributional vectors in the target language with the refined constraints. We prove the effectiveness of our method through intrinsic word similarity evaluation in 8 languages, and with 3 downstream tasks in 5 languages: lexical simplification, dialog state tracking, and semantic textual similarity. The gains over the previous state-of-art specialization methods are substantial and consistent across languages. Our results also suggest that the transfer method is effective even for lexically distant source-target language pairs. Finally, as a by-product, our method produces lists of WordNet-style lexical relations in resource-poor languages
- …