7,761 research outputs found

    Methodologies for the Automatic Location of Academic and Educational Texts on the Internet

    Get PDF
    Traditionally online databases of web resources have been compiled by a human editor, or though the submissions of authors or interested parties. Considerable resources are needed to maintain a constant level of input and relevance in the face of increasing material quantity and quality, and much of what is in databases is of an ephemeral nature. These pressures dictate that many databases stagnate after an initial period of enthusiastic data entry. The solution to this problem would seem to be the automatic harvesting of resources, however, this process necessitates the automatic classification of resources as ‘appropriate’ to a given database, a problem only solved by complex text content analysis. This paper outlines the component methodologies necessary to construct such an automated harvesting system, including a number of novel approaches. In particular this paper looks at the specific problems of automatically identifying academic research work and Higher Education pedagogic materials. Where appropriate, experimental data is presented from searches in the field of Geography as well as the Earth and Environmental Sciences. In addition, appropriate software is reviewed where it exists, and future directions are outlined

    Methodologies for the Automatic Location of Academic and Educational Texts on the Internet

    Get PDF
    Traditionally online databases of web resources have been compiled by a human editor, or though the submissions of authors or interested parties. Considerable resources are needed to maintain a constant level of input and relevance in the face of increasing material quantity and quality, and much of what is in databases is of an ephemeral nature. These pressures dictate that many databases stagnate after an initial period of enthusiastic data entry. The solution to this problem would seem to be the automatic harvesting of resources, however, this process necessitates the automatic classification of resources as ‘appropriate’ to a given database, a problem only solved by complex text content analysis. This paper outlines the component methodologies necessary to construct such an automated harvesting system, including a number of novel approaches. In particular this paper looks at the specific problems of automatically identifying academic research work and Higher Education pedagogic materials. Where appropriate, experimental data is presented from searches in the field of Geography as well as the Earth and Environmental Sciences. In addition, appropriate software is reviewed where it exists, and future directions are outlined

    Simple Strategies for Broadcasting Repository Resources

    Get PDF
    4th International Conference on Open RepositoriesThis presentation was part of the session : Conference PostersNSDL's data repository for STEM education is designed to provide organized access to digital educational materials through its online portal, NSDL.org. The resources held within the NSDL data repository along with their associated metadata can also be found through partner and external portals, often with high quality, pedagogical contextual information intact. Repositories are not, however, usually described as web broadcast devices for their holdings. Providing multiple contextual views of educational resources where users look for them underscores the idea that digital repositories can be systems for the management, preservation, discovery and reuse of rich resources within a domain that can also be pushed out from a repository into homes and classrooms through multiple channels. This presentation reviews two interrelated methods and usage data that support the concept of Ăą resource broadcastingĂą from the NSDL data repository as a method that takes advantage of the natural context of resources to encourage their additional use as stand-alone objects outside of specific discipline-oriented portals.National Science Foundatio

    BlogForever D5.2: Implementation of Case Studies

    Get PDF
    This document presents the internal and external testing results for the BlogForever case studies. The evaluation of the BlogForever implementation process is tabulated under the most relevant themes and aspects obtained within the testing processes. The case studies provide relevant feedback for the sustainability of the platform in terms of potential users’ needs and relevant information on the possible long term impact

    Enhancing Undergraduate AI Courses through Machine Learning Projects

    Full text link
    It is generally recognized that an undergraduate introductory Artificial Intelligence course is challenging to teach. This is, in part, due to the diverse and seemingly disconnected core topics that are typically covered. The paper presents work funded by the National Science Foundation to address this problem and to enhance the student learning experience in the course. Our work involves the development of an adaptable framework for the presentation of core AI topics through a unifying theme of machine learning. A suite of hands-on semester-long projects are developed, each involving the design and implementation of a learning system that enhances a commonly-deployed application. The projects use machine learning as a unifying theme to tie together the core AI topics. In this paper, we will first provide an overview of our model and the projects being developed and will then present in some detail our experiences with one of the projects – Web User Profiling which we have used in our AI class

    Scraping the Social? Issues in live social research

    Get PDF
    What makes scraping methodologically interesting for social and cultural research? This paper seeks to contribute to debates about digital social research by exploring how a ‘medium-specific’ technique for online data capture may be rendered analytically productive for social research. As a device that is currently being imported into social research, scraping has the capacity to re-structure social research, and this in at least two ways. Firstly, as a technique that is not native to social research, scraping risks to introduce ‘alien’ methodological assumptions into social research (such as an pre-occupation with freshness). Secondly, to scrape is to risk importing into our inquiry categories that are prevalent in the social practices enabled by the media: scraping makes available already formatted data for social research. Scraped data, and online social data more generally, tend to come with ‘external’ analytics already built-in. This circumstance is often approached as a ‘problem’ with online data capture, but we propose it may be turned into virtue, insofar as data formats that have currency in the areas under scrutiny may serve as a source of social data themselves. Scraping, we propose, makes it possible to render traffic between the object and process of social research analytically productive. It enables a form of ‘real-time’ social research, in which the formats and life cycles of online data may lend structure to the analytic objects and findings of social research. By way of a conclusion, we demonstrate this point in an exercise of online issue profiling, and more particularly, by relying on Twitter to profile the issue of ‘austerity’. Here we distinguish between two forms of real-time research, those dedicated to monitoring live content (which terms are current?) and those concerned with analysing the liveliness of issues (which topics are happening?)

    Web Science: expanding the notion of Computer Science

    No full text
    Academic disciplines which practice in the context of rapid external change face particular problems when seeking to maintain timely, current and relevant teaching programs. In different institutions faculty will tune and update individual component courses while more radical revisions are typically departmental-wide strategic responses to perceived needs. Internationally, the ACM has sought to define curriculum recommendations since the 1960s and recognizes the diversity of the computing disciplines with its 2005 overview volume. The consequent rolling program of revisions is demanding in terms of time and effort, but an inevitable response to the change inherent is our family of specialisms. Preparation for the Computer Curricula 2013 is underway, so it seems appropriate to ask what place Web Science will have in the curriculum landscape. Web Science has been variously described; the most concise definition being the ‘science of decentralized information systems’. Web science is fundamentally interdisciplinary encompassing the study of the technologies and engineering which constitute the Web, alongside emerging associated human, social and organizational practices. Furthermore, to date little teaching of Web Science is at undergraduate level. Some questions emerge - is Web Science a transient artifact? Can Web Science claim a place in the ACM family, Is Web Science an exotic relative with a home elsewhere? This paper discusses the role and place of Web Science in the context of the computing disciplines. It provides an account of work which has been established towards defining an initial curriculum for Web Science with plans for future developments utilizing novel methods to support and elaborate curriculum definition and review. The findings of a desk survey of existing related curriculum recommendations are presented. The paper concludes with recommendations for future activities which may help us determine whether we should expand the notion of computer science
    • 

    corecore