30 research outputs found
Semantic Similarity in a Taxonomy: An Information-Based Measure and its Application to Problems of Ambiguity in Natural Language
This article presents a measure of semantic similarity in an IS-A taxonomy
based on the notion of shared information content. Experimental evaluation
against a benchmark set of human similarity judgments demonstrates that the
measure performs better than the traditional edge-counting approach. The
article presents algorithms that take advantage of taxonomic similarity in
resolving syntactic and semantic ambiguity, along with experimental results
demonstrating their effectiveness
WORD SENSE DISAMBIGUATION WITHIN A MULTILINGUAL FRAMEWORK
Word Sense Disambiguation (WSD) is the process of resolving the meaning of a
word unambiguously in a given natural language context. Within the scope of this
thesis, it is the process of marking text with explicit sense labels.
What constitutes a sense is a subject of great debate. An appealing perspective,
aims to define senses in terms of their multilingual correspondences, an idea explored
by several researchers, Dyvik (1998), Ide (1999), Resnik & Yarowsky (1999), and
Chugur, Gonzalo & Verdejo (2002) but to date it has not been given any practical
demonstration. This thesis is an empirical validation of these ideas of characterizing
word meaning using cross-linguistic correspondences. The idea is that word meaning
or word sense is quantifiable as much as it is uniquely translated in some language or
set of languages.
Consequently, we address the problem of WSD from a multilingual perspective;
we expand the notion of context to encompass multilingual evidence. We devise a
new approach to resolve word sense ambiguity in natural language, using a source of
information that was never exploited on a large scale for WSD before.
The core of the work presented builds on exploiting word correspondences across
languages for sense distinction. In essence, it is a practical and functional implementation
of a basic idea common to research interest in defining word meanings in
cross-linguistic terms.
We devise an algorithm, SALAAM for Sense Assignment Leveraging Alignment
And Multilinguality, that empirically investigates the feasibility and the validity of utilizing
translations for WSD. SALAAM is an unsupervised approach for word sense
tagging of large amounts of text given a parallel corpus — texts in translation — and
a sense inventory for one of the languages in the corpus. Using SALAAM, we obtain
large amounts of sense annotated data in both languages of the parallel corpus, simultaneously.
The quality of the tagging is rigorously evaluated for both languages of the
corpora.
The automatic unsupervised tagged data produced by SALAAM is further utilized
to bootstrap a supervised learning WSD system, in essence, combining supervised and
unsupervised approaches in an intelligent way to alleviate the resources acquisition
bottleneck for supervised methods. Essentially, SALAAM is extended as an unsupervised
approach for WSD within a learning framework; in many of the cases of the
words disambiguated, SALAAM coupled with the machine learning system rivals the
performance of a canonical supervised WSD system that relies on human tagged data
for training.
Realizing the fundamental role of similarity for SALAAM, we investigate different
dimensions of semantic similarity as it applies to verbs since they are relatively
more complex than nouns, which are the focus of the previous evaluations. We design
a human judgment experiment to obtain human ratings on verbs’ semantic similarity.
The obtained human ratings are cast as a reference point for comparing different
automated similarity measures that crucially rely on various sources of information.
Finally, a cognitively salient model integrating human judgments in SALAAM is proposed
as a means of improving its performance on sense disambiguation for verbs in
particular and other word types in general
Wiktionary: The Metalexicographic and the Natural Language Processing Perspective
Dictionaries are the main reference works for our understanding of language. They are used by humans and likewise by computational methods. So far, the compilation of dictionaries has almost exclusively been the profession of expert lexicographers. The ease of collaboration on the Web and the rising initiatives of collecting open-licensed knowledge, such as in Wikipedia, caused a new type of dictionary that is voluntarily created by large communities of Web users. This collaborative construction approach presents a new paradigm for lexicography that poses new research questions to dictionary research on the one hand and provides a very valuable knowledge source for natural language processing applications on the other hand. The subject of our research is Wiktionary, which is currently the largest collaboratively constructed dictionary project.
In the first part of this thesis, we study Wiktionary from the metalexicographic perspective. Metalexicography is the scientific study of lexicography including the analysis and criticism of dictionaries and lexicographic processes. To this end, we discuss three contributions related to this area of research: (i) We first provide a detailed analysis of Wiktionary and its various language editions and dictionary structures. (ii) We then analyze the collaborative construction process of Wiktionary. Our results show that the traditional phases of the lexicographic process do not apply well to Wiktionary, which is why we propose a novel process description that is based on the frequent and continual revision and discussion of the dictionary articles and the lexicographic instructions. (iii) We perform a large-scale quantitative comparison of Wiktionary and a number of other dictionaries regarding the covered languages, lexical entries, word senses, pragmatic labels, lexical relations, and translations. We conclude the metalexicographic perspective by finding that the collaborative Wiktionary is not an appropriate replacement for expert-built dictionaries due to its inconsistencies, quality flaws, one-fits-all-approach, and strong dependence on expert-built dictionaries. However, Wiktionary's rapid and continual growth, its high coverage of languages, newly coined words, domain-specific vocabulary and non-standard language varieties, as well as the kind of evidence based on the authors' intuition provide promising opportunities for both lexicography and natural language processing. In particular, we find that Wiktionary and expert-built wordnets and thesauri contain largely complementary entries.
In the second part of the thesis, we study Wiktionary from the natural language processing perspective with the aim of making available its linguistic knowledge for computational applications. Such applications require vast amounts of structured data with high quality. Expert-built resources have been found to suffer from insufficient coverage and high construction and maintenance cost, whereas fully automatic extraction from corpora or the Web often yields resources of limited quality. Collaboratively built encyclopedias present a viable solution, but do not cover well linguistically oriented knowledge as it is found in dictionaries. That is why we propose extracting linguistic knowledge from Wiktionary, which we achieve by the following three main contributions: (i) We propose the novel multilingual ontology OntoWiktionary that is created by extracting and harmonizing the weakly structured dictionary articles in Wiktionary. A particular challenge in this process is the ambiguity of semantic relations and translations, which we resolve by automatic word sense disambiguation methods. (ii) We automatically align Wiktionary with WordNet 3.0 at the word sense level. The largely complementary information from the two dictionaries yields an aligned resource with higher coverage and an enriched representation of word senses. (iii) We represent Wiktionary according to the ISO standard Lexical Markup Framework, which we adapt to the peculiarities of collaborative dictionaries. This standardized representation is of great importance for fostering the interoperability of resources and hence the dissemination of Wiktionary-based research. To this end, our work presents a foundational step towards the large-scale integrated resource UBY, which facilitates a unified access to a number of standardized dictionaries by means of a shared web interface for human users and an application programming interface for natural language processing applications. A user can, in particular, switch between and combine information from Wiktionary and other dictionaries without completely changing the software.
Our final resource and the accompanying datasets and software are publicly available and can be employed for multiple different natural language processing applications. It particularly fills the gap between the small expert-built wordnets and the large amount of encyclopedic knowledge from Wikipedia. We provide a survey of previous works utilizing Wiktionary, and we exemplify the usefulness of our work in two case studies on measuring verb similarity and detecting cross-lingual marketing blunders, which make use of our Wiktionary-based resource and the results of our metalexicographic study. We conclude the thesis by emphasizing the usefulness of collaborative dictionaries when being combined with expert-built resources, which bears much unused potential
Design of a Controlled Language for Critical Infrastructures Protection
We describe a project for the construction of controlled language for critical infrastructures protection (CIP). This project originates
from the need to coordinate and categorize the communications on CIP at the European level. These communications can be physically
represented by official documents, reports on incidents, informal communications and plain e-mail. We explore the application of
traditional library science tools for the construction of controlled languages in order to achieve our goal. Our starting point is an
analogous work done during the sixties in the field of nuclear science known as the Euratom Thesaurus.JRC.G.6-Security technology assessmen
Meaning refinement to improve cross-lingual information retrieval
Magdeburg, Univ., Fak. fĂĽr Informatik, Diss., 2012von Farag Ahme
The feminisation of agentives in French and Spanish speaking countries: a cross-linguistic and cross-continental comparison
Non-sexist writing guidelines have been produced since the middle of the 20th century but often cause controversy. Taking only one aspect of such language reform, the feminisation of agentives, the present study aims to compare two similarly-structured, grammatically-gendered languages, French and Spanish, with regard to the visibility of women in the print media.
After reviewing research that shows the use of masculine gendered agentives can induce, or reinforce, stereotypes which obscure female agency, prior studies of feminisation are classified by methodology and data source showing that little previous research has taken advantage of corpus techniques to analyse naturally occurring data, nor is there a significant body of contrastive research comparing feminisation strategies across languages or across countries with the same language. The collation of a cross-continental and cross-language corpus of media references to named people is therefore proposed and executed to allow both quantitative and qualitative analysis of naturally-occurring feminisations (or, indeed, their absence).
Using electronic techniques, a corpus of over 5,000 references to named individuals was collated from press websites in France, Spain, Canada and Argentina. The form of the agentives referring to women was compared to strategies suggested in the UN-produced guidelines on gender neutral language, for French and Spanish, and discrepancies were classified. Classification of the agentives’ morphology was also made, to assign a 'predicted' base gender to each agentive. Quantitative and qualitative analyses performed on the data then drive the discussion of similarities and differences in feminisation strategies, across the chosen languages and countries. The study shows that prestige agentives cause feminisation difficulties across both languages, independently of morphology, whilst also identifying issues that are specific to one language group or one area. Possible reasons for both the similarities and differences are suggested and in turn suggest areas for further research using similar, corpus-based techniques