899 research outputs found

    Objectivity, prescription, harmlessness, and drudgery: reflections of lexicographers in Slovenia

    Get PDF
    This contribution reports on a study that set out to paint as complete a picture as pos-sible of the context and content of modern Slovenian lexicography. We aimed to discern the phi-losophical underpinnings, the most noteworthy accomplishments, and the main projects of Slovenian dictionary work as presented by our seven subjects, who are all prominent members of the lexico-graphic community. We sought specialists who work on synchronic topics and concentrate more on the standard language and terminology rather than on dialectal variation and other lexico-graphic topics that are of more interest to scholars than to educated lay persons. The interview script consisted of thirteen narrative questions, designed to allow the interviewees to reflect in as much depth as possible on their daily practice as well as on their underlying vision of what lexi-cography or terminography is. This article discusses the development and influences of Slovenian lexicographic theory and presents part 1 of the results of this study: the views of the practicing lexicographers on whether they perceive their lexicographic work as drudgery and what they see as the essential nature of their role in society — how the dictionary maker can be a force for good and avoid any potential for harm.Keywords: harmless drudge, drudgery, harm, harmlessness, interview, lexicographer, lexicographic philosophy, lexicographic principles, monotony, repetitiveness, tediu

    Wiktionary: The Metalexicographic and the Natural Language Processing Perspective

    Get PDF
    Dictionaries are the main reference works for our understanding of language. They are used by humans and likewise by computational methods. So far, the compilation of dictionaries has almost exclusively been the profession of expert lexicographers. The ease of collaboration on the Web and the rising initiatives of collecting open-licensed knowledge, such as in Wikipedia, caused a new type of dictionary that is voluntarily created by large communities of Web users. This collaborative construction approach presents a new paradigm for lexicography that poses new research questions to dictionary research on the one hand and provides a very valuable knowledge source for natural language processing applications on the other hand. The subject of our research is Wiktionary, which is currently the largest collaboratively constructed dictionary project. In the first part of this thesis, we study Wiktionary from the metalexicographic perspective. Metalexicography is the scientific study of lexicography including the analysis and criticism of dictionaries and lexicographic processes. To this end, we discuss three contributions related to this area of research: (i) We first provide a detailed analysis of Wiktionary and its various language editions and dictionary structures. (ii) We then analyze the collaborative construction process of Wiktionary. Our results show that the traditional phases of the lexicographic process do not apply well to Wiktionary, which is why we propose a novel process description that is based on the frequent and continual revision and discussion of the dictionary articles and the lexicographic instructions. (iii) We perform a large-scale quantitative comparison of Wiktionary and a number of other dictionaries regarding the covered languages, lexical entries, word senses, pragmatic labels, lexical relations, and translations. We conclude the metalexicographic perspective by finding that the collaborative Wiktionary is not an appropriate replacement for expert-built dictionaries due to its inconsistencies, quality flaws, one-fits-all-approach, and strong dependence on expert-built dictionaries. However, Wiktionary's rapid and continual growth, its high coverage of languages, newly coined words, domain-specific vocabulary and non-standard language varieties, as well as the kind of evidence based on the authors' intuition provide promising opportunities for both lexicography and natural language processing. In particular, we find that Wiktionary and expert-built wordnets and thesauri contain largely complementary entries. In the second part of the thesis, we study Wiktionary from the natural language processing perspective with the aim of making available its linguistic knowledge for computational applications. Such applications require vast amounts of structured data with high quality. Expert-built resources have been found to suffer from insufficient coverage and high construction and maintenance cost, whereas fully automatic extraction from corpora or the Web often yields resources of limited quality. Collaboratively built encyclopedias present a viable solution, but do not cover well linguistically oriented knowledge as it is found in dictionaries. That is why we propose extracting linguistic knowledge from Wiktionary, which we achieve by the following three main contributions: (i) We propose the novel multilingual ontology OntoWiktionary that is created by extracting and harmonizing the weakly structured dictionary articles in Wiktionary. A particular challenge in this process is the ambiguity of semantic relations and translations, which we resolve by automatic word sense disambiguation methods. (ii) We automatically align Wiktionary with WordNet 3.0 at the word sense level. The largely complementary information from the two dictionaries yields an aligned resource with higher coverage and an enriched representation of word senses. (iii) We represent Wiktionary according to the ISO standard Lexical Markup Framework, which we adapt to the peculiarities of collaborative dictionaries. This standardized representation is of great importance for fostering the interoperability of resources and hence the dissemination of Wiktionary-based research. To this end, our work presents a foundational step towards the large-scale integrated resource UBY, which facilitates a unified access to a number of standardized dictionaries by means of a shared web interface for human users and an application programming interface for natural language processing applications. A user can, in particular, switch between and combine information from Wiktionary and other dictionaries without completely changing the software. Our final resource and the accompanying datasets and software are publicly available and can be employed for multiple different natural language processing applications. It particularly fills the gap between the small expert-built wordnets and the large amount of encyclopedic knowledge from Wikipedia. We provide a survey of previous works utilizing Wiktionary, and we exemplify the usefulness of our work in two case studies on measuring verb similarity and detecting cross-lingual marketing blunders, which make use of our Wiktionary-based resource and the results of our metalexicographic study. We conclude the thesis by emphasizing the usefulness of collaborative dictionaries when being combined with expert-built resources, which bears much unused potential

    Rules of engagement: Strategies used to enlist and retain underserved mothers in a mental health intervention

    Get PDF
    Patient engagement has been identified as both a goal and strategy to lower health care costs and improve health care outcomes. However, a lack of consensus and clarity exists as to how the process of patient engagement is implemented in clinical practice. Research addressing the underlying and crucial components of effective patient engagement is limited, leaving a significant gap as to how providers engage patients as active collaborators in their health and health care. This study provides specific, detailed insight and description into the processes through which advanced practice mental health nurses engaged low-income depressed mothers in a mental health intervention. The Interactive Care Model (ICM), a patient engagement framework, was used to examine and illuminate the key processes and partnership roles of patient engagement. Using a directed content analysis approach, we completed a secondary analysis of nursing narrative data using the 5 key processes and 7 partnership roles of the ICM to guide our analysis. The ICM demonstrated great utility in capturing the processes through which advanced practice nurses enlisted, engaged, and retained low-income depressed mothers in the mental health intervention. Additionally, the nursing narrative data provided specific detail and description as to how the ICM’s components were operationalized in practice. The ICM was validated by the nursing narrative data and provided sound organizational structure for the specific verbal and non-verbal engagement interventions nurses employed. Findings from this study can expand the knowledge base and understanding of the process of patient engagement and can help guide providers in executing behaviors that engage traditionally unengaged patients as active collaborators in their health and health care

    Collaboration in urban distribution of online grocery orders

    Get PDF
    Purpose: Population growth, urbanisation and the increased use of online shopping are some of the key challenges affecting the traditional logistics model. The purpose of this paper is to focus on the distribution of grocery products ordered online and the subsequent home delivery and click and collect services offered by online retailers to fulfil these orders. These services are unsustainable due to increased operational costs, carbon emissions, traffic and noise. The main objective of the research is to propose sustainable logistics models to reduce economic, environmental and social costs whilst maintaining service levels. Design/methodology/approach: The authors have a mixed methodology based on simulation and mathematical modelling to evaluate the proposed shared logistics model using: primary data from a major UK retailer, secondary data from online retailers and primary data from a consumer survey on preferences for receiving groceries purchased online. Integration of these three data sets serves as input to vehicle routing models that reveal the benefits from collaboration by solving individual distribution problems of two retailers first, followed by the joint distribution problem under single decision maker assumption. Findings: The benefits from collaboration could be more than 10 per cent in the distance travelled and 16 per cent in the time required to deliver the orders when two online grocery retailers collaborate in distribution activities. Originality/value: The collaborative model developed for the online grocery market incentivises retailers to switch from current unsustainable logistics models to the proposed collaborative models

    The EAGLES/ISLE initiative for setting standards: the Computational Lexicon Working Group for Multilingual Lexicons

    Get PDF
    ISLE (International Standards for Language Engineering), a transatlantic standards oriented initiative under the Human Language Technology (HLT) programme, is a continuation of the long standing EAGLES (Expert Advisory Group for Language Engineering Standards) initiative, carried out by European and American groups within the EU-US International Research Co-operation, supported by NSF and EC. The objective is to support HLT R&D international and national projects, and HLT industry, by developing and promoting widely agreed and urgently demanded HLT standards and guidelines for infrastructural language resources, tools, and HLT products. ISLE targets the areas of multilingual computational lexicons (MCL), natural interaction and multimodality (NIMM), and evaluation. For MCL, ISLE is working to: extend EAGLES work on lexical semantics, necessary to establish inter-language links; design standards for multilingual lexicons; develop a prototype tool to implement lexicon guidelines; create EAGLES-conformant sample lexicons and tag corpora for validation purposes; develop standardised evaluation procedures for lexicons. For NIMM, a rapidly innovating domain urgently requiring early standardisation, ISLE work is targeted to develop guidelines for: creation of NIMM data resources; interpretative annotation of NIMM data, including spoken dialogue; annotation of discourse phenomena. For evaluation, ISLE is working on: quality models for machine translation systems; maintenance of previous guidelines - in an ISO based framework. We concentrate in the paper on the Computational Lexicon Working Group, describing in detail the proposals of guidelines for the "Multilingual ISLE Lexical Entry" (MILE). We highlight some methodological principles applied in previous EAGLES, and followed in defining MILE. We also provide a description of the EU SIMPLE semantic lexicons built on the basis of previous EAGLES recommendations. Their importance is given by the fact that these lexicons are now enlarged to real-size lexicons within National Projects in 8 EU countries, thus building a really large infrastructural platform of harmonised lexicons in Europe. We will stress the relevance of standardised language resources also for the humanities applications. Numerous theories, approaches, systems are taken into account in ISLE, as any recommendation for harmonisation must build on the major contemporary approaches. Results will be widely disseminated, after validation in collaboration with EU and US HLT R&D projects, and industry. EAGLES work towards de facto standards has already allowed the field of Language Resources to establish broad consensus on key issues for some well-established areas - and will allow similar consensus to be achieved for other important areas through the ISLE project - providing thus a key opportunity for further consolidation and a basis for technological advance. EAGLES previous results in many areas have in fact already become de facto widely adopted standards, and EAGLES itself is a well-known trademark and a point of reference for HLT projects.Hosted by the Scholarly Text and Imaging Service (SETIS), the University of Sydney Library, and the Research Institute for Humanities and Social Sciences (RIHSS), the University of Sydney

    Language technologies for a multilingual Europe

    Get PDF
    This volume of the series “Translation and Multilingual Natural Language Processing” includes most of the papers presented at the Workshop “Language Technology for a Multilingual Europe”, held at the University of Hamburg on September 27, 2011 in the framework of the conference GSCL 2011 with the topic “Multilingual Resources and Multilingual Applications”, along with several additional contributions. In addition to an overview article on Machine Translation and two contributions on the European initiatives META-NET and Multilingual Web, the volume includes six full research articles. Our intention with this workshop was to bring together various groups concerned with the umbrella topics of multilingualism and language technology, especially multilingual technologies. This encompassed, on the one hand, representatives from research and development in the field of language technologies, and, on the other hand, users from diverse areas such as, among others, industry, administration and funding agencies. The Workshop “Language Technology for a Multilingual Europe” was co-organised by the two GSCL working groups “Text Technology” and “Machine Translation” (http://gscl.info) as well as by META-NET (http://www.meta-net.eu)

    Optimising resourcing skills to develop phraseological competence in legal translation: tasks and approaches

    Get PDF
    Based on a previous case study on common translation errors made by trainee translators when dealing with phraseological units in legal translation (Huertas Barros and Buendía 2017), this paper proposes some activities and approaches to minimise these errors and hence enhance students’ phraseological competence in this field. To this end, we first provide a description of the most representative legal resources available for translators, particularly for the English-Spanish language pair. Then we review some of the existing approaches that could develop students’ legal translation competence, particularly phraseological competence. For each type of error identified in our previous case study, we propose a set of research-based activities which could avoid such errors by maximising the use of legal resources. The emphasis is put on preliminary documentary research and effective use of corpora prior to the translation task. In order to mitigate translation errors, we propose an integrated approach combining task-based approaches with approaches based on critical discourse analysis, and problem-solving and decision-making. While these tasks have been designed for a semi-specialised legal text pertaining to a subdomain of Family Law (i.e. adoption), they can be easily applied to any other areas of subdomains of legal translation
    • …
    corecore