401 research outputs found
Encoding models for scholarly literature
We examine the issue of digital formats for document encoding, archiving and
publishing, through the specific example of "born-digital" scholarly journal
articles. We will begin by looking at the traditional workflow of journal
editing and publication, and how these practices have made the transition into
the online domain. We will examine the range of different file formats in which
electronic articles are currently stored and published. We will argue strongly
that, despite the prevalence of binary and proprietary formats such as PDF and
MS Word, XML is a far superior encoding choice for journal articles. Next, we
look at the range of XML document structures (DTDs, Schemas) which are in
common use for encoding journal articles, and consider some of their strengths
and weaknesses. We will suggest that, despite the existence of specialized
schemas intended specifically for journal articles (such as NLM), and more
broadly-used publication-oriented schemas such as DocBook, there are strong
arguments in favour of developing a subset or customization of the Text
Encoding Initiative (TEI) schema for the purpose of journal-article encoding;
TEI is already in use in a number of journal publication projects, and the
scale and precision of the TEI tagset makes it particularly appropriate for
encoding scholarly articles. We will outline the document structure of a
TEI-encoded journal article, and look in detail at suggested markup patterns
for specific features of journal articles
09051 Abstracts Collection -- Knowledge representation for intelligent music processing
From the twenty-fifth to the thirtieth of January, 2009, the
Dagstuhl Seminar 09051 on ``Knowledge representation for intelligent music
processing\u27\u27 was held in Schloss Dagstuhl~--~Leibniz Centre for Informatics.
During the seminar, several participants presented their current
research, and ongoing work and open problems were discussed. Abstracts
of the presentations and demos given during the seminar as well as
plenary presentations, reports of workshop discussions, results and
ideas are put together in this paper. The first section describes the
seminar topics and goals in general, followed by plenary `stimulus\u27
papers, followed by reports and abstracts arranged by workshop
followed finally by some concluding materials providing views of both
the seminar itself and also forward to the longer-term goals of the
discipline. Links to extended abstracts, full papers and supporting
materials are provided, if available.
The organisers thank David Lewis for editing these proceedings
When linguistics meets web technologies. Recent advances in modelling linguistic linked data
This article provides an up-to-date and comprehensive survey of models (including vocabularies, taxonomies and ontologies) used for representing linguistic linked data (LLD). It focuses on the latest developments in the area and both builds upon and complements previous works covering similar territory. The article begins with an overview of recent trends which have had an impact on linked data models and vocabularies, such as the growing influence of the FAIR guidelines, the funding of several major projects in which LLD is a key component, and the increasing importance of the relationship of the digital humanities with LLD. Next, we give an overview of some of the most well known vocabularies and models in LLD. After this we look at some of the latest developments in community standards and initiatives such as OntoLex-Lemon as well as recent work which has been in carried out in corpora and annotation and LLD including a discussion of the LLD metadata vocabularies META-SHARE and lime and language identifiers. In the following part of the paper we look at work which has been realised in a number of recent projects and which has a significant impact on LLD vocabularies and models
Graph Data-Models and Semantic Web Technologies in Scholarly Digital Editing
This volume is based on the selected papers presented at the Workshop on Scholarly Digital Editions, Graph Data-Models and Semantic Web Technologies, held at the Uni- versity of Lausanne in June 2019. The Workshop was organized by Elena Spadini (University of Lausanne) and Francesca Tomasi (University of Bologna), and spon- sored by the Swiss National Science Foundation through a Scientific Exchange grant, and by the Centre de recherche sur les lettres romandes of the University of Lausanne. The Workshop comprised two full days of vibrant discussions among the invited speakers, the authors of the selected papers, and other participants.1 The acceptance rate following the open call for papers was around 60%. All authors â both selected and invited speakers â were asked to provide a short paper two months before the Workshop. The authors were then paired up, and each pair exchanged papers. Paired authors prepared questions for one another, which were to be addressed during the talks at the Workshop; in this way, conversations started well before the Workshop itself. After the Workshop, the papers underwent a second round of peer-review before inclusion in this volume. This time, the relevance of the papers was not under discus- sion, but reviewers were asked to appraise specific aspects of each contribution, such as its originality or level of innovation, its methodological accuracy and knowledge of the literature, as well as more formal parameters such as completeness, clarity, and coherence. The bibliography of all of the papers is collected in the public Zotero group library GraphSDE20192, which has been used to generate the reference list for each contribution in this volume.
The invited speakers came from a wide range of backgrounds (academic, commer- cial, and research institutions) and represented the different actors involved in the remediation of our cultural heritage in the form of graphs and/or in a semantic web en- vironment. Georg Vogeler (University of Graz) and Ronald Haentjens Dekker (Royal Dutch Academy of Sciences, Humanities Cluster) brought the Digital Humanities research perspective; the work of Hans Cools and Roberta Laura Padlina (University of Basel, National Infrastructure for Editions), as well as of Tobias Schweizer and Sepi- deh Alassi (University of Basel, Digital Humanities Lab), focused on infrastructural challenges and the development of conceptual and software frameworks to support re- searchersâ needs; Michele Pasinâs contribution (Digital Science, Springer Nature) was informed by his experiences in both academic research, and in commercial technology companies that provide services for the scientific community.
The Workshop featured not only the papers of the selected authors and of the invited speakers, but also moments of discussion between interested participants. In addition to the common Q&A time, during the second day one entire session was allocated to working groups delving into topics that had emerged during the Workshop. Four working groups were created, with four to seven participants each, and each group presented a short report at the end of the session. Four themes were discussed: enhancing TEI from documents to data; ontologies for the Humanities; tools and infrastructures; and textual criticism. All of these themes are represented in this volume.
The Workshop would not have been of such high quality without the support of the members of its scientific committee: Gioele Barabucci, Fabio Ciotti, Claire Clivaz, Marion Rivoal, Greta Franzini, Simon Gabay, Daniel Maggetti, Frederike Neuber, Elena Pierazzo, Davide Picca, Michael Piotrowski, Matteo Romanello, MaĂŻeul Rouquette, Elena Spadini, Francesca Tomasi, Aris Xanthos â and, of course, the support of all the colleagues and administrative staff in Lausanne, who helped the Workshop to become a reality.
The final versions of these papers underwent a single-blind peer review process. We want to thank the reviewers: Helena Bermudez Sabel, Arianna Ciula, Marilena Daquino, Richard Hadden, Daniel Jeller, Tiziana Mancinelli, Davide Picca, Michael Piotrowski, Patrick Sahle, Raffaele Viglianti, Joris van Zundert, and others who preferred not to be named personally. Your input enhanced the quality of the volume significantly!
It is sad news that Hans Cools passed away during the production of the volume. We are proud to document a recent state of his work and will miss him and his ability to implement the vision of a digital scholarly edition based on graph data-models and semantic web technologies.
The production of the volume would not have been possible without the thorough copy-editing and proof reading by Lucy Emmerson and the support of the IDE team, in particular Bernhard Assmann, the TeX-master himself. This volume is sponsored by the University of Bologna and by the University of Lausanne.
Bologna, Lausanne, Graz, July 2021
Francesca Tomasi, Elena Spadini, Georg Vogele
The Practice and Benefit of Applying Digital Markup in Preserving Texts and Creating Digital Editions: A Poetical Analysis of a Blank-Verse Translation of Virgil\u27s Aeneid
Numerous examples of the digital scholarly edition exist online, and the genre is thriving in terms of interdisciplinary interest as well as support granted by funding agencies. Some editions are dedicated to the collection and representation of the life\u27s work of a single author, others to mass digitization and preservation of centuries\u27 worth of texts. Very few of these examples, however, approach the task of in-text interpretation through visualization. This project describes an approach to digital representation and investigates its potential benefit to scholars of various disciplines. It presents both a digital edition as well as a framework of justification surrounding said edition. In addition to composing this document as an XML file, I have digitized a 1794 English translation of Virgil\u27s Aeneid and used a customized digital markup schema based on the guidelines set forth by the Text Encoding Initiative to indicate a set of poetic figuresâsuch as simile and alliterationâwithin that text for analysis. While neither a translation project nor strictly a poetical analysis, this project and its unique approach to interpretive representation could prove of interest to scholars in several disciplines, including classics, digital scholarship, information management, and literary theory. The practice serves both as a case-in-point as well as an example method to replicate with future texts and projects
The role of humanities computing: experiences and challenges
Dieser Beitrag wurde anlĂ€sslich des dreiĂigjĂ€hrigen Bestehens der Abteilung Literarische und Dokumentarische Datenverarbeitung an der UniversitĂ€t TĂŒbingen verfasst. Er gibt einen Ăberblick ĂŒber die Entwicklung der Datenverarbeitung auf dem Gebiet der Geisteswissenschaften in diesem Zeitraum. ZunĂ€chst werden Erfahrungen mit der Datenverarbeitung in den Geisteswissenschaften referiert und aktuelle Entwicklungen auf diesem Gebiet vorgestellt. Dann werden Herausforderungen diskutiert, mit denen sich die Datenverarbeitung in den Geisteswissenschaften konfrontiert sieht. AbschlieĂend stellt der Verfasser die Bedeutung der Datenverarbeitung in den Geisteswissenschaften in den kommenden Jahren aus seiner Sicht dar. (ICEĂbers)'Due to the celebration of the thirtieth anniversary of the Department for Literary and Documentary Data Processing in Tuebingen this article is written. It gives an overview of humanities computing developments since the formation of this Research-Department. The paper is divided into three parts. First, the experiences in humanities computing are reviewed. For three purposes the author points out various aspects of the development and exploitation of scholarly materials using computers, considering some of the current work to create new tools for research. This chapter is followed by the discussion of some of the key challenges of this century, by that humanities computing and the scholarship, of which it is a part, are faced with. Finally, the author gives a summary of what in his opinion would be the key rotes of humanities computing in the future.' (author's abstract
Designing a Library of Components for Textual Scholarship
Il presente lavoro affronta e descrive temi legati all'applicazione di nuove tecnologie, di metodologie informatiche e di progettazione software volti allo sviluppo di strumenti innovativi per le Digital Humanities (DH), unâarea di studio caratterizzata da una forte interdisciplinaritĂ e da una continua evoluzione. In particolare, questo contributo definisce alcuni specifici requisiti relativi al dominio del Literary Computing e al settore del Digital Textual Scholarship. Conseguentemente, il contesto principale di elaborazione tratta documenti scritti in latino, greco e arabo, nonchĂ© testi in lingue moderne contenenti temi storici e filologici. L'attivitĂ di ricerca si concentra sulla progettazione di una libreria modulare (TSLib) in grado di operare su fonti ad elevato valore culturale, al fine di editarle, elaborarle, confrontarle, analizzarle, visualizzarle e ricercarle. La tesi si articola in cinque capitoli. Il capitolo 1 riassume il contesto del dominio applicativo e fornisce un quadro generale degli obiettivi e dei benefici della ricerca. Il capitolo 2 illustra alcuni importanti lavori e iniziative analoghe, insieme a una breve panoramica dei risultati piĂč significativi ottenuti nel settore delle DH. Il capitolo 3 ripercorre accuratamente e motiva il processo di progettazione messo a punto. Esso inizia con la descrizione dei principi tecnici adottati e mostra come essi vengono applicati al dominio d'interesse. Il capitolo continua definendo i requisiti, l'architettura e il modello del metodo proposto. Sono cosĂŹ evidenziati e discussi gli aspetti concernenti i design patterns e la progettazione delle Application Programming Interfaces (APIs). La parte finale del lavoro (capitolo 4) illustra i risultati ottenuti da concreti progetti di ricerca che, da un lato, hanno contribuito alla progettazione della libreria e, dall'altro, hanno avuto modo di sfruttarne gli sviluppi. Sono stati quindi discussi diversi temi: (a) l'acquisizione e la codifica del testo, (b) l'allineamento e la gestione delle varianti testuali, (c) le annotazioni multilivello. La tesi si conclude con alcune riflessioni e considerazioni indicando anche possibili percorsi d'indagine futuri (capitolo 5)
Development of a conceptual graphical user interface framework for the creation of XML metadata for digital archives
This dissertation is motivated by the DFG sponsored Jonas Cohn Archive digitization project at Steinheim-Institut whose aim was to preserve and provide digital access to structured handwritten historical archive material highlighting New Kantian philosophy scattered in the correspondence, diaries and private journals kept by and written to and by Jonas Cohn.
The dissertation describes a framework for processing and presenting multi-standard digital archive material. A set of standard markup schema and semantic bibliographic descriptions have been chosen to illustrate the multiple standard and hence semantic heterogeneous digital archiving process. The standards include Text Encoding Initiative (TEI), Metadata Encoding and Transmission Standard (METS) and Metadata Object Description Schema (MODS). The chosen standards best illustrate the structural contrast between the systematic archive, digitized archive and digitized text standards. Furthermore, combined digital preservation and presentation approaches offer not only the digitized texts but also metadata structured variably sized images of the archive documents enabling virtual visualization. State of the art applications focus solely on either one of the structural areas neglecting the compound idea of a virtual digital archive.
The content of this work describes the requirements analysis for managing multi-structured and therefore multi-standard digital archival artefacts in textual and image form. In addition to the architecture and design, an infrastructure suitable for processing, managing and presenting such scholarly archives is sought for recognition as a digital framework useful for the preservation and access to digitized cultural resources. The proposed solution therefore includes the instrumentation of a conglomerate of existing and novel XML technology for transformations based in a centralized application. The archive can then be managed via a client-server application thereby focusing archival activities on structured data collection and information preservation illustrated in the dissertation process by the:
âą Development of a prototype data model allowing the integration of the relevant markup schema
âą Implementation of a prototype client server application handling archive processing, management and presentation and based on the data model already mentioned
âą Development and implementation of a role archive access user interface
Furthermore as an infrastructural development serving expert archivists from the humanities, the dissertation explores methods of binding the existing XML metadata creation process to other programming languages. In doing so, one opens further for channels simplifying the metadata creation process by integrating the use of graphical user interfaces. To this end the java programming language, its swing and AWT graphical user interface libraries, associated relational persistency and enterprise client server architecture resemble a suitable environment for integrating XML metadata into main stream computing. Hence the implementation of Java XML Data Binding as part of the metadata creation framework is part and parcel of the proposed solution.Diese Arbeit geht hervor aus dem von der DFG geförderten Projekt zu Digitalisierung des Jonas Cohn Archivs im Steinheim-Institut, dessen Ziel es ist, eine strukturierte Auswahl von Handschriften des Philosophen Jonas Cohns in digitaler Form zu bewahren und den Zugang zu ihnen zu erleichtern.
Die Dissertation beschreibt ein Rahmenwerk fĂŒr die digitale Verarbeitung und PrĂ€sentation digitalisierter Archivinhalte und ihrer Metadaten, strukturiert anhand von mehr als einem Beschreibungsstandard. Eine Auswahl von Standard Markup Schemata und bibliographisch semantischen Beschreibungen wurde getroffen, um die Problematik darzustellen, die aus der BerĂŒcksichtigung mehrerer Standards und damit aus semantischer HeterogenitĂ€t des Digitalisierungsprozesses entsteht. Diese Auswahl umfasst unter anderem die Text Encoding Initiative (TEI), Metadata Encoding and Transmission Schema (METS) und Metadata Object Description Schema (MODS) als Beispiele fĂŒr Beschreibungsstandards. Diese Standards sind am besten geeignet, die strukturellen und semantischen Unterschiede zwischen den Standards eines systematisch und semantisch zu digitalisierenden Archivs darzustellen. ZusĂ€tzlich verbindet der Ansatz die digitale Bewahrung und PrĂ€sentation von digitalisierten Texten und von Metadaten strukturierter Bilder der Archivinhalte. Dies ermöglicht eine virtuelle PrĂ€sentation des digitalen Archivs. Eine groĂe Zahl bekannter Digitalisierungsanwendungen folgt nur einer der beiden Strukturierungsziele Bewahrung und PrĂ€sentation, wodurch der Ansatz eines vollstĂ€ndig virtuellen digitalen Archivs vernachlĂ€ssigt wird.
Der Schwerpunkt dieser Arbeit ist die Beschreibung einer Managementinfrastruktur fĂŒr die Erfassung und Auszeichnung von Multi-Standard Metadaten fĂŒr digitale Handschriftensammlungen. ZusĂ€tzlich zu der Architektur und dem Design wird nach einer geeigneten Infrastruktur gesucht fĂŒr die Erfassung, Verarbeitung und die PrĂ€sentation wissenschaftlicher Archive als digitales Rahmenwerk fĂŒr den Zugang zu und die Bewahrung von Kulturbesitz.
Die hier vorgeschlagene Lösung sieht deshalb die Nutzung bestehender und neuer XML Technologien vor, verknĂŒpft in einer zentralen Anwendung. So wird im Rahmen der Dissertation die Strukturierung des Archivs mittels einer Client-Server-Anwendung betrieben und die BewahrungsmaĂnahmen als Prozess herausgearbeitet. Die Arbeit verfolgt mehrere Zielsetzungen:
âą Die Entwicklung eines prototypischen Datenmodells mit der Einbindung relevanter Markup Schemata
âą Die Implementierung einer prototypischen Client Server Anwendung fĂŒr die Bearbeitung, Erfassung und PrĂ€sentation der Archive anhand des beschriebenen Datenmodells
âą Die Entwicklung, Implementierung und Bewertung einer Benutzerschnittstelle fĂŒr die Interaktion mit dem Rahmenwerk anhand einer Expertenevaluation
- âŠ