15,052 research outputs found
Metadata quality issues in learning repositories
Metadata lies at the heart of every digital repository project in the sense that it defines and drives the description of digital content stored in the repositories. Metadata allows content to be successfully stored, managed and retrieved but also preserved in the long-term. Despite the enormous importance of metadata in digital repositories, one that is widely recognized, studies indicate that what is defined as metadata quality, is relatively low in most cases of digital repositories. Metadata quality is loosely defined as "fitness for purpose" meaning that low quality of metadata means that metadata cannot fulfill its purpose which is to allow for the successful storage, management and retrieval of resources. In practice, low metadata quality leads to ineffective searches for content, ones that recall the wrong resources or even worse, no resources which makes them invisible to the intended user, that is the "client" of each digital repository. The present dissertation approaches this problem by proposing a comprehensive metadata quality assurance method, namely the Metadata Quality Assurance Certification Process (MQACP). The basic idea of this dissertation is to propose a set of methods that can be deployed throughout the lifecycle of a repository to ensure that metadata generated from content providers are of high quality. These methods have to be straightforward, simple to apply with measurable results. They also have to be adaptable with minimum effort so that they can be used in different contexts easily. This set of methods was described analytically, taking into account the actors needed to apply them, describing the tools needed and defining the anticipated outcomes. In order to test our proposal, we applied it on a Learning Federation of repositories, from day 1 of its existence until it reached its maturity and regular operation. We supported the metadata creation process throughout the different phases of the repositories involved by setting up specific experiments using the methods and tools of the MQACP. Throughout each phase, we measured the resulting metadata quality to certify that the anticipated improvement in metadata quality actually took place. Lastly, through these different phases, the cost of the MQACP application was measured to provide a comparison basis for future applications. Based on the success of this first application, we decided to validate the MQACP approach by applying it on another two cases of a Cultural and a Research Federation of repositories. This would allow us to prove the transferability of the approach to other cases the present some similarities with the initial one but mainly significant differences. The results showed that the MQACP was successfully adapted to the new contexts, with minimum adaptations needed, with similar results produced and also with comparable costs. In addition, looking closer at the common experiments carried out in each phase of each use case, we were able to identify interesting patterns in the behavior of content providers that can be further researched. The dissertation is completed with a set of future research directions that came out of the cases examined. These research directions can be explored in order to support the next version of the MQACP in terms of the methods deployed, the tools used to assess metadata quality as well as the cost analysis of the MQACP methods
Metadata quality issues in learning repositories
Metadata lies at the heart of every digital repository project in the sense that it defines and drives the description of digital content stored in the repositories. Metadata allows content to be successfully stored, managed and retrieved but also preserved in the long-term. Despite the enormous importance of metadata in digital repositories, one that is widely recognized, studies indicate that what is defined as metadata quality, is relatively low in most cases of digital repositories. Metadata quality is loosely defined as "fitness for purpose" meaning that low quality of metadata means that metadata cannot fulfill its purpose which is to allow for the successful storage, management and retrieval of resources. In practice, low metadata quality leads to ineffective searches for content, ones that recall the wrong resources or even worse, no resources which makes them invisible to the intended user, that is the "client" of each digital repository. The present dissertation approaches this problem by proposing a comprehensive metadata quality assurance method, namely the Metadata Quality Assurance Certification Process (MQACP). The basic idea of this dissertation is to propose a set of methods that can be deployed throughout the lifecycle of a repository to ensure that metadata generated from content providers are of high quality. These methods have to be straightforward, simple to apply with measurable results. They also have to be adaptable with minimum effort so that they can be used in different contexts easily. This set of methods was described analytically, taking into account the actors needed to apply them, describing the tools needed and defining the anticipated outcomes. In order to test our proposal, we applied it on a Learning Federation of repositories, from day 1 of its existence until it reached its maturity and regular operation. We supported the metadata creation process throughout the different phases of the repositories involved by setting up specific experiments using the methods and tools of the MQACP. Throughout each phase, we measured the resulting metadata quality to certify that the anticipated improvement in metadata quality actually took place. Lastly, through these different phases, the cost of the MQACP application was measured to provide a comparison basis for future applications. Based on the success of this first application, we decided to validate the MQACP approach by applying it on another two cases of a Cultural and a Research Federation of repositories. This would allow us to prove the transferability of the approach to other cases the present some similarities with the initial one but mainly significant differences. The results showed that the MQACP was successfully adapted to the new contexts, with minimum adaptations needed, with similar results produced and also with comparable costs. In addition, looking closer at the common experiments carried out in each phase of each use case, we were able to identify interesting patterns in the behavior of content providers that can be further researched. The dissertation is completed with a set of future research directions that came out of the cases examined. These research directions can be explored in order to support the next version of the MQACP in terms of the methods deployed, the tools used to assess metadata quality as well as the cost analysis of the MQACP methods
Quality assurance for digital learning object repositories: issues for the metadata creation process
Metadata enables users to find the resources they require, therefore it is an important component of any digital learning object repository. Much work has already been done within the learning technology community to assure metadata quality, focused on the development of metadata standards, specifications and vocabularies and their implementation within repositories. The metadata creation process has thus far been largely overlooked. There has been an assumption that metadata creation will be straightforward and that where machines cannot generate metadata effectively, authors of learning materials will be the most appropriate metadata creators. However, repositories are reporting difficulties in obtaining good quality metadata from their contributors, and it is becoming apparent that the issue of metadata creation warrants attention. This paper surveys the growing body of evidence, including three UK-based case studies, scopes the issues surrounding human-generated metadata creation and identifies questions for further investigation. Collaborative creation of metadata by resource authors and metadata specialists, and the design of tools and processes, are emerging as key areas for deeper research. Research is also needed into how end users will search learning object repositories
Recommended from our members
Social networking and open educational resources: updating quality assurance for e-learning excellence
Quality assurance approaches in higher education are well-established, but it is important to develop methods which are applicable to the domain of e-learning. The E-xcellence methodology (EADTU, 2009a) was therefore designed to assess the quality of e-learning in distance learning and blended learning contexts. The methodology is based around a set of benchmarks, supported by a practitioner handbook and a web-based ‘QuickScan’ self-evaluation tool. Experience shows that the E-xcellence methodology is particularly valuable for the process of improvement through collaborative internal review.
E-learning has evolved since the E-xcellence methodology was first developed. In particular, there is increasing awareness and use of open education resources (OERs) and social networking. However, these aspects were not explicit in the original E-xcellence resources. The E-xcellence Next project was therefore established to update the resources, incorporating these developments. To begin this process, a consultation was carried out among E-xcellence Next project members, followed by a participatory workshop on the themes of social networking and OERs. The E-xcellence resources were also used in a series of self-evaluation seminars held at European higher education institutions. Experience and feedback from these activities has been used to update the manual, the benchmarks and the QuickScan tool. The result is a set of quality assurance resources which encompass social networking, OERs and other recent developments in e-learning
A vision of quality in repositories of open educational resources
In the future, Open Educational Practices (OEP) will facilitate access to open materials by promoting collaboration among educators, who will share, reuse and evaluate digital pedagogical content using Repositories of Open Educational Resources (ROER)
KAPTUR: technical analysis report
Led by the Visual Arts Data Service (VADS) and funded by the JISC Managing Research Data programme (2011-13) KAPTUR will discover, create and pilot a sectoral model of best practice in the management of research data in the visual arts in collaboration with four institutional partners: Glasgow School of Art; Goldsmiths, University of London; University for the Creative Arts; and University of the Arts London.
This report is framed around the research question: which technical system is most suitable for managing visual arts research data?
The first stage involved a literature review including information gathered through attendance at meetings and events, and Internet research, as well as information on projects from the previous round of JISCMRD funding (2009-11).
During February and March 2012, the Technical Manager carried out interviews with the four KAPTUR Project Officers and also met with IT staff at each institution. This led to the creation of a user requirement document (Appendix A), which was then circulated to the project team for additional comments and feedback. The Technical Manager selected 17 systems to compare with the user requirement document (Appendix B). Five of the systems had similar scores so these were short-listed. The Technical Manager created an online form into which the Project Officers entered priority scores for each of the user requirements in order to calculate a more accurate score for each of the five short-listed systems (Appendix C) and this resulted in the choice of EPrints as the software for the KAPTUR project
The LIFE2 final project report
Executive summary: The first phase of LIFE (Lifecycle Information For E-Literature) made a major contribution to
understanding the long-term costs of digital preservation; an essential step in helping
institutions plan for the future. The LIFE work models the digital lifecycle and calculates the
costs of preserving digital information for future years. Organisations can apply this process
in order to understand costs and plan effectively for the preservation of their digital
collections
The second phase of the LIFE Project, LIFE2, has refined the LIFE Model adding three new
exemplar Case Studies to further build upon LIFE1. LIFE2 is an 18-month JISC-funded
project between UCL (University College London) and The British Library (BL), supported
by the LIBER Access and Preservation Divisions. LIFE2 began in March 2007, and
completed in August 2008.
The LIFE approach has been validated by a full independent economic review and has
successfully produced an updated lifecycle costing model (LIFE Model v2) and digital
preservation costing model (GPM v1.1). The LIFE Model has been tested with three further
Case Studies including institutional repositories (SHERPA-LEAP), digital preservation
services (SHERPA DP) and a comparison of analogue and digital collections (British Library
Newspapers). These Case Studies were useful for scenario building and have fed back into
both the LIFE Model and the LIFE Methodology.
The experiences of implementing the Case Studies indicated that enhancements made to the
LIFE Methodology, Model and associated tools have simplified the costing process. Mapping
a specific lifecycle to the LIFE Model isn’t always a straightforward process. The revised and
more detailed Model has reduced ambiguity. The costing templates, which were refined
throughout the process of developing the Case Studies, ensure clear articulation of both
working and cost figures, and facilitate comparative analysis between different lifecycles.
The LIFE work has been successfully disseminated throughout the digital preservation and
HE communities. Early adopters of the work include the Royal Danish Library, State
Archives and the State and University Library, Denmark as well as the LIFE2 Project partners.
Furthermore, interest in the LIFE work has not been limited to these sectors, with interest in
LIFE expressed by local government, records offices, and private industry. LIFE has also
provided input into the LC-JISC Blue Ribbon Task Force on the Economic Sustainability of
Digital Preservation.
Moving forward our ability to cost the digital preservation lifecycle will require further
investment in costing tools and models. Developments in estimative models will be needed to
support planning activities, both at a collection management level and at a later preservation
planning level once a collection has been acquired. In order to support these developments a
greater volume of raw cost data will be required to inform and test new cost models. This
volume of data cannot be supported via the Case Study approach, and the LIFE team would
suggest that a software tool would provide the volume of costing data necessary to provide a
truly accurate predictive model
Recommended from our members
Community Dimmensions of Learning Object Repositories. <i>Deliverable 1</i>: Report on Learning Communities and Repositories
Audit and Certification of Digital Repositories: Creating a Mandate for the Digital Curation Centre (DCC)
The article examines the issues surrounding the audit and certification of digital repositories in light of the work that the RLG/NARA Task Force did to draw up guidelines and the need for these guidelines to be validated.
- …