17,292 research outputs found
Overview of VideoCLEF 2009: New perspectives on speech-based multimedia content enrichment
VideoCLEF 2009 offered three tasks related to enriching video content for improved multimedia access in a multilingual environment. For each task, video data (Dutch-language television, predominantly documentaries) accompanied by speech recognition transcripts were provided.
The Subject Classification Task involved automatic tagging of videos with subject theme labels. The best performance was achieved by approaching subject tagging as an information retrieval task and using both speech recognition transcripts and archival metadata. Alternatively, classifiers were trained using either the training data provided or data collected from Wikipedia or via general Web search. The Affect Task involved detecting narrative peaks, defined as points where viewers perceive heightened dramatic tension. The task was carried out on the āBeeldenstormā collection containing 45 short-form documentaries on the visual arts. The best runs exploited affective vocabulary and audience directed speech. Other approaches included using topic changes, elevated speaking pitch, increased speaking intensity and radical visual changes. The Linking Task, also called āFinding Related Resources Across Languages,ā involved linking video to material on the same subject in a different language.
Participants were provided with a list of multimedia anchors (short video segments) in the Dutch-language āBeeldenstormā collection and were expected to return target pages drawn from English-language Wikipedia. The best performing methods used the transcript of the
speech spoken during the multimedia anchor to build a query to search an index of the Dutch language Wikipedia. The Dutch Wikipedia pages returned were used to identify related English pages. Participants also experimented with pseudo-relevance feedback, query translation and methods that targeted proper names
Metadata enrichment for digital heritage: users as co-creators
This paper espouses the concept of metadata enrichment through an expert and user-focused approach to metadata creation and management. To this end, it is argued the Web 2.0 paradigm enables users to be proactive metadata creators. As Shirky (2008, p.47) argues Web 2.0ās social tools enable āaction by loosely structured groups, operating without managerial direction and outside the profit motiveā. Lagoze (2010, p. 37) advises, āthe participatory nature of Web 2.0 should not be dismissed as just a popular phenomenon [or fad]ā. Carletti (2016) proposes a participatory digital cultural heritage approach where Web 2.0 approaches such as crowdsourcing can be sued to enrich digital cultural objects. It is argued that āheritage crowdsourcing, community-centred projects or other forms of public participationā. On the other hand, the new collaborative approaches of Web 2.0 neither negate nor replace contemporary standards-based metadata approaches. Hence, this paper proposes a mixed metadata approach where user created metadata augments expert-created metadata and vice versa. The metadata creation process no longer remains to be the sole prerogative of the metadata expert. The Web 2.0 collaborative environment would now allow users to participate in both adding and re-using metadata. The case of expert-created (standards-based, top-down) and user-generated metadata (socially-constructed, bottom-up) approach to metadata are complementary rather than mutually-exclusive. The two approaches are often mistakenly considered as dichotomies, albeit incorrectly (Gruber, 2007; Wright, 2007) .
This paper espouses the importance of enriching digital information objects with descriptions pertaining the about-ness of information objects. Such richness and diversity of description, it is argued, could chiefly be achieved by involving users in the metadata creation process. This paper presents the importance of the paradigm of metadata enriching and metadata filtering for the cultural heritage domain. Metadata enriching states that a priori metadata that is instantiated and granularly structured by metadata experts is continually enriched through socially-constructed (post-hoc) metadata, whereby users are pro-actively engaged in co-creating metadata. The principle also states that metadata that is enriched is also contextually and semantically linked and openly accessible. In addition, metadata filtering states that metadata resulting from implementing the principle of enriching should be displayed for users in line with their needs and convenience. In both enriching and filtering, users should be considered as prosumers, resulting in what is called collective metadata intelligence
How did the discussion go: Discourse act classification in social media conversations
We propose a novel attention based hierarchical LSTM model to classify
discourse act sequences in social media conversations, aimed at mining data
from online discussion using textual meanings beyond sentence level. The very
uniqueness of the task is the complete categorization of possible pragmatic
roles in informal textual discussions, contrary to extraction of
question-answers, stance detection or sarcasm identification which are very
much role specific tasks. Early attempt was made on a Reddit discussion
dataset. We train our model on the same data, and present test results on two
different datasets, one from Reddit and one from Facebook. Our proposed model
outperformed the previous one in terms of domain independence; without using
platform-dependent structural features, our hierarchical LSTM with word
relevance attention mechanism achieved F1-scores of 71\% and 66\% respectively
to predict discourse roles of comments in Reddit and Facebook discussions.
Efficiency of recurrent and convolutional architectures in order to learn
discursive representation on the same task has been presented and analyzed,
with different word and comment embedding schemes. Our attention mechanism
enables us to inquire into relevance ordering of text segments according to
their roles in discourse. We present a human annotator experiment to unveil
important observations about modeling and data annotation. Equipped with our
text-based discourse identification model, we inquire into how heterogeneous
non-textual features like location, time, leaning of information etc. play
their roles in charaterizing online discussions on Facebook
Knowledge Cartography for Controversies: The Iraq Debate
In analysing controversies and debatesāwhich would include reviewing a literature in order to plan research, or assessing intelligence to formulate policyāthere is no one worldview which can be mapped, for instance as a single, coherent concept map. The cartographic challenge is to show which facts are agreed and contested, and the different kinds of narrative links that use facts as evidence to define the nature of the problem, what to do about it, and why. We will use the debate around the invasion of Iraq to demonstrate the methodology of using a knowledge mapping tool to extract key ideas from source materials, in order to classify and connect them within and across a set of perspectives of interest to the analyst. We reflect on the value that this approach adds, and how it relates to other argument mapping approaches
Recommended from our members
A literature review of the use of Web 2.0 tools in Higher Education
This review focuses on the use of Web 2.0 tools in Higher Education. It provides a synthesis of the research literature in the field and a series of illustrative examples of how these tools are being used in learning and teaching. It draws out the perceived benefits that these new technologies appear to offer, and highlights some of the challenges and issues surrounding their use. The review forms the basis for a HE Academy funded project, āPeals in the Cloudā, which is exploring how Web 2.0 tools can be used to support evidence-based practices in learning and teaching. The project has also produced two in-depth case studies, which are reported elsewhere (Galley et al., 2010, Alevizou et al., 2010). The case studies focus on evaluation of a recently developed site for learning and teaching, Cloudworks, which harnesses Web 2.0 functionality to facilitate the sharing and discussion of educational practice. The case studies aim to explore to what extent the Web 2.0 affordances of the site are successfully promoting the sharing of ideas, as well as scholarly reflections, on learning and teaching
Cloudworks: social networking for learning design
Why do some social networking services work and others fail? Can we apply the best of Web 2.0 principles to an educational context? More specifically can we use this as a means of shifting teaching practice to a culture of sharing learning ideas and designs? Can we harness the potential of technologies to create more engaging learning experiences for students? These are the key questions this paper addresses. We describe how we are using the concept of 'object-orientated social networking' to underpin the creation of a social networking tool, Cloudworks, for sharing learning ideas and designs
Semantic modelling of user interests based on cross-folksonomy analysis
The continued increase in Web usage, in particular participation in folksonomies, reveals a trend towards a more dynamic and interactive Web where individuals can organise and share resources. Tagging has emerged as the de-facto standard for the organisation of such resources, providing a versatile and reactive knowledge management mechanism that users find easy to use and understand. It is common nowadays for users to have multiple profiles in various folksonomies, thus distributing their tagging activities. In this paper, we present a method for the automatic consolidation of user profiles across two popular social networking sites, and subsequent semantic modelling of their interests utilising Wikipedia as a multi-domain model. We evaluate how much can be learned from such sites, and in which domains the knowledge acquired is focussed. Results show that far richer interest profiles can be generated for users when multiple tag-clouds are combine
A lightweight web video model with content and context descriptions for integration with linked data
The rapid increase of video data on the Web has warranted an urgent need for effective representation, management and retrieval of web videos. Recently, many studies have been carried out for ontological representation of videos, either using domain dependent or generic schemas such as MPEG-7, MPEG-4, and COMM. In spite of their extensive coverage and sound theoretical grounding, they are yet to be widely used by users. Two main possible reasons are the complexities involved and a lack of tool support. We propose a lightweight video content model for content-context description and integration. The uniqueness of the model is that it tries to model the emerging social context to describe and interpret the video. Our approach is grounded on exploiting easily extractable evolving contextual metadata and on the availability of existing data on the Web. This enables representational homogeneity and a firm basis for information integration among semantically-enabled data sources. The model uses many existing schemas to describe various ontology classes and shows the scope of interlinking with the Linked Data cloud
- ā¦