285,947 research outputs found
Miracle’s 2005 Approach to Cross-lingual Information Retrieval
This paper presents the 2005 Miracle’s team approach to Bilingual and Multilingual Information Retrieval. In the multilingual track, we have concentrated our work on the merging process of the results of monolingual runs to get the multilingual overall result, relying on available translations. In the bilingual and multilingual tracks, we have used available translation resources, and in some cases we have using a combining approach
Multilingual Lexical Semantic Resources for Ontology Translation
We describe the integration of some multilingual language resources in ontological descriptions, with the purpose of providing ontologies, which are normally using concept labels in just one (natural) language, with multilingual facility in their design and use in the context of Semantic Web applications, supporting both the semantic annotation of textual documents with multilingual ontology labels and ontology extraction from multilingual text sources
Dublin City University at CLEF 2004: experiments in monolingual, bilingual and multilingual retrieval
The Dublin City University group participated in the monolingual, bilingual and multilingual retrieval tasks this year. The main focus of our investigation this year was extending our retrieval system to document languages other than English, and completing the multilingual task comprising four languages: English, French, Russian and Finnish. Results from our French monolingual experiments indicate that working in French is more effective for retrieval than adopting document and topic translation to English. However, comparison of our multilingual retrieval results using different topic and document translation reveals that this result does not extend to retrieved list merging for the multilingual task in a simple predictable way
Multilingual adaptive search for digital libraries
This paper describes a framework for Adaptive Multilingual Information Retrieval (AMIR) which allows multilingual resource discovery and delivery using on-the-fly machine translation of documents and queries. Result documents
are presented to the user in a contextualised manner. Challenges and affordances of both Adaptive and Multilingual IR, with a particular focus on Digital Libraries, are detailed. The framework components are motivated by a series of results from experiments on query logs and documents from The European Library. We conclude that factoring adaptivity and multilinguality aspects into the search process can enhance the user’s experience with online Digital Libraries
Combining Multiple Methods for the Automatic Construction of Multilingual WordNets
This paper explores the automatic construction of a multilingual Lexical
Knowledge Base from preexisting lexical resources. First, a set of automatic
and complementary techniques for linking Spanish words collected from
monolingual and bilingual MRDs to English WordNet synsets are described.
Second, we show how resulting data provided by each method is then combined to
produce a preliminary version of a Spanish WordNet with an accuracy over 85%.
The application of these combinations results on an increment of the extracted
connexions of a 40% without losing accuracy. Both coarse-grained (class level)
and fine-grained (synset assignment level) confidence ratios are used and
evaluated. Finally, the results for the whole process are presented.Comment: 7 pages, 4 postscript figure
GlobalTrait: Personality Alignment of Multilingual Word Embeddings
We propose a multilingual model to recognize Big Five Personality traits from
text data in four different languages: English, Spanish, Dutch and Italian. Our
analysis shows that words having a similar semantic meaning in different
languages do not necessarily correspond to the same personality traits.
Therefore, we propose a personality alignment method, GlobalTrait, which has a
mapping for each trait from the source language to the target language
(English), such that words that correlate positively to each trait are close
together in the multilingual vector space. Using these aligned embeddings for
training, we can transfer personality related training features from
high-resource languages such as English to other low-resource languages, and
get better multilingual results, when compared to using simple monolingual and
unaligned multilingual embeddings. We achieve an average F-score increase
(across all three languages except English) from 65 to 73.4 (+8.4), when
comparing our monolingual model to multilingual using CNN with personality
aligned embeddings. We also show relatively good performance in the regression
tasks, and better classification results when evaluating our model on a
separate Chinese dataset.Comment: Submitted and accepted to AAAI 2019 conferenc
- …
