22 research outputs found

    Encoding of Variant Taxonomies in TEI

    Get PDF
    The inherent flexibility of the digital format has favored the rise of editions that enable access to every witness of a particular textual work. These types of editions might have different goals and seek to answer different research questions, but they usually coincide in drawing attention to the importance of textual variants. To maximize the computational analysis that may be practiced with the variants in different witnesses, a complex taxonomy that reflects the diversity of cases is required. Many scholars have followed the recommended TEI method for encoding types of variants—that is, through the attributes @cause or @type inside the element —while others find that method insufficient. These attributes are not able to enclose the hierarchy intrinsic to complicated taxonomies or the overlap of classes in an efficient way. However, the TEI Guidelines do offer a module that addresses this complex encoding issue: feature structures. The method proposed in this paper does not advocate for a controlled vocabulary to categorize types of variants. What it offers instead is a pliable encoding method that allows the editor to include multiple layers of information in each apparatus tagset

    A edición sinóptica e a súa aplicación ao estudo da variación lingüística na lírica galego-portuguesa

    Get PDF
    A edición sinóptica caracterízase por permitir o contraste de forma con forma, signo con signo, de todas as leccións dunha tradición textual. Por esta razón, é a tipoloxía ecdótica máis axeitada, desde o punto de vista metodolóxico, para a análise da variación lingüística, en canto permite o cotexo automático de todas as leccións transmitidas polos manuscritos. Promovemos a utilización deste tipo de edición, malia a súa escasa aplicación até o momento, porque o seu emprego supón a sistematización dunha ampla información que abre o camiño para futuras liñas de investigación, entre as que cómpre subliñar: identificación das diferentes mans que participan no proceso de copia e localización dos posíbeis scriptoria implicados na elaboración dos códices, caracterización lingüística dos diferentes manuscritos, identificación das leccións equipolentes e/ou das diverxencias de tradición na transmisión textual, e, en consecuencia, elaboración dunha collatio codicum rigoros

    Colación asistida por ordenador: estado de la cuestión y retos

    Get PDF
    Como advierte el profesor Blecua en su célebre manual de crítica textual, “la collatio codicum es la fase más ingrata y una de las más delicadas de todo el proceso editorial” (Blecua, 1983: 43). Tradicionalmente, esta ardua tarea se realizaba cotejando visualmente los diferentes testimonios y anotando las divergencias. A pesar de las ventajas que los métodos computacionales nos ofrecen (entre otras, mayor rapidez y, sobre todo, menos margen para el error) su implementación para esta tarea aún es escasa. En este artículo hacemos una revisión de los recursos ya existentes, analizando sus ventajas e inconvenientes, para terminar haciendo una propuesta alternativa a las herramientas que no requieren conocimientos sobre lenguajes de programación

    Using RDFa to link text and dictionary data for medieval French

    Get PDF
    Around the world, there is a wide range of traditional data manually collected for different scientific purposes. A small portion of this data has been digitised, but much of it remains less usable due to a lack of rich semantic models to enable humans and machines to understand, interpret and use these data. This paper presents ongoing work to build a semantic model to enrich and publish traditional data collection questionnaires in particular, and the historical data collection of the Bavarian Dialects in Austria in general. The use of cultural and linguistic concepts identified in the questionnaire questions allow for cultural exploration of the non-standard data (answers) of the collection. The approach focuses on capturing the semantics of the questionnaires dataset using domain analysis and schema analysis. This involves analysing the overall data collection process (domain analysis) and analysing the various schema used at different stages (schema analysis). By starting with modelling the data collection method, the focus is placed on the questionnaires as a gateway to understanding, interlinking and publishing the datasets. A model that describes the semantic structure of the main entities such as questionnaires, questions, answers and their relationships is presented

    Implemented to Be Shared: the WoPoss Annotation of Semantic Modality in a Latin Diachronic Corpus

    Get PDF
    The SNSF project A World of Possibilities [WoPoss] aims at tracking the evolution of modal meanings in the diachrony of the Latin language. Passages expressing modal notions of ‘necessity’, ‘possibility’ and ‘volition’ are annotated following a pipeline that combines both automatic and fine-grained manual annotation. Texts are first gathered from different online open access resources to create the initial dataset. Owing to the heterogeneity of formats and encodings, the files are regularized before the implementation of the automatic annotation of linguistic features. They are then uploaded to the annotation platform [INCEpTION] which has been previously customized for the fine-grained manual annotation [Bermúdez Sabel, in press]. This second type of annotation is carried out following the WoPoss guidelines [Dell’Oro, 2019] and it also involves checking the automatic annotation for curation. In a third stage the files are automatically enriched with metadata. In this paper, we focus on the first two phases of the workflow – i.e. the gathering and automatic annotation of the texts and the fine-grained manual annotation –, which have been the core tasks of the WoPoss team in the first year of the project (2019-2020)

    Navegación de corpus a través de anotaciones lingüísticas automáticas obtenidas por Procesamiento del Lenguaje Natural: de anecdótico a ecdótico

    Get PDF
    The article presents two case studies on the application of Natural Language Processing (NLP) technologies or Computational Linguistics to create corpus navigation interfaces. These interfaces help access relevant information for specific research questions in Social Sciences or Humanities. The paper also focuses on how these technologies for automatic text analysis can allow us to enrich scholarly digital editions. The theoretical framework that connects the aforementioned technologies with academic digital edition is described, and a reflection is made on the application of such technologies and interfaces to digital scholarly editing

    Retos, perspectivas y próximos pasos dentro de POSTDATA

    Get PDF
    Presentación en el HDH2017 de Málag

    Acuerdos y desacuerdos. Cómo hacer interoperables conceptos métricos complejos. El caso de la rima

    Get PDF
    Presentación sobre el tema en el Congreso HDH2017 de Málag

    Visualisation of semantic shifts: the case of modal markers

    Get PDF
    Poster submission for DH2020. This poster discusses the adequacy of semantic maps for the representation of semantic shifts and how previous models can be enriched by using visual cues to add more information. The poster presents a visualisation proposal applied to the diachronic study of modal markers in Latin

    Validation of a metadata application profile domain model

    Get PDF
    The development of Metadata Application Profiles is done in several phases. According to the Me4MAP method, one of these phases is the creation of the domain model. This paper reports the validation process of a domain model developed under the project POSTDATA - Poetry Standardization and Linked Open Data. The development of the domain model ran with two steps of construction and two of validation. The validation steps drew on the participation of specialists in European poetry and the use of real resources. On the first validation we used tables with information about resources related properties and for which the experts had to fill certain fields like, for examples, the values. The second validation used a XML framework to control the input of values in the model. The validation process allowed us to find and fix flaws in the domain model that would otherwise have been passed to the Description Set Profile and possibly would only be found after implementing the application profile in a real case.European Research Council(POSTDATA1)info:eu-repo/semantics/publishedVersio
    corecore