2,051 research outputs found
A Novel Multimodal Approach for Studying the Dynamics of Curiosity in Small Group Learning
Curiosity is a vital metacognitive skill in educational contexts, leading to
creativity, and a love of learning. And while many school systems increasingly
undercut curiosity by teaching to the test, teachers are increasingly
interested in how to evoke curiosity in their students to prepare them for a
world in which lifelong learning and reskilling will be more and more
important. One aspect of curiosity that has received little attention, however,
is the role of peers in eliciting curiosity. We present what we believe to be
the first theoretical framework that articulates an integrated socio-cognitive
account of curiosity that ties observable behaviors in peers to underlying
curiosity states. We make a bipartite distinction between individual and
interpersonal functions that contribute to curiosity, and multimodal behaviors
that fulfill these functions. We validate the proposed framework by leveraging
a longitudinal latent variable modeling approach. Findings confirm a positive
predictive relationship between the latent variables of individual and
interpersonal functions and curiosity, with the interpersonal functions
exercising a comparatively stronger influence. Prominent behavioral
realizations of these functions are also discovered in a data-driven manner. We
instantiate the proposed theoretical framework in a set of strategies and
tactics that can be incorporated into learning technologies to indicate, evoke,
and scaffold curiosity. This work is a step towards designing learning
technologies that can recognize and evoke moment-by-moment curiosity during
learning in social contexts and towards a more complete multimodal learning
analytics. The underlying rationale is applicable more generally for developing
computer support for other metacognitive and socio-emotional skills.Comment: arXiv admin note: text overlap with arXiv:1704.0748
Semantic multimedia modelling & interpretation for annotation
The emergence of multimedia enabled devices, particularly the incorporation of cameras in mobile phones, and the accelerated revolutions in the low cost storage devices, boosts the multimedia data production rate drastically. Witnessing such an iniquitousness of digital images and videos, the research community has been projecting the issue of its significant utilization and management. Stored in monumental multimedia corpora, digital data need to be retrieved and organized in an intelligent way, leaning on the rich semantics involved. The utilization of these image and video collections demands proficient image and video annotation and retrieval techniques. Recently, the multimedia research community is progressively veering its emphasis to the personalization of these media. The main impediment in the image and video analysis is the semantic gap, which is the discrepancy among a user’s high-level interpretation of an image and the video and the low level computational interpretation of it. Content-based image and video annotation systems are remarkably susceptible to the semantic gap due to their reliance on low-level visual features for delineating semantically rich image and video contents. However, the fact is that the visual similarity is not semantic similarity, so there is a demand to break through this dilemma through an alternative way. The semantic gap can be narrowed by counting high-level and user-generated information in the annotation. High-level descriptions of images and or videos are more proficient of capturing the semantic meaning of multimedia content, but it is not always applicable to collect this information. It is commonly agreed that the problem of high level semantic annotation of multimedia is still far from being answered. This dissertation puts forward approaches for intelligent multimedia semantic extraction for high level annotation. This dissertation intends to bridge the gap between the visual features and semantics. It proposes a framework for annotation enhancement and refinement for the object/concept annotated images and videos datasets. The entire theme is to first purify the datasets from noisy keyword and then expand the concepts lexically and commonsensical to fill the vocabulary and lexical gap to achieve high level semantics for the corpus. This dissertation also explored a novel approach for high level semantic (HLS) propagation through the images corpora. The HLS propagation takes the advantages of the semantic intensity (SI), which is the concept dominancy factor in the image and annotation based semantic similarity of the images. As we are aware of the fact that the image is the combination of various concepts and among the list of concepts some of them are more dominant then the other, while semantic similarity of the images are based on the SI and concept semantic similarity among the pair of images. Moreover, the HLS exploits the clustering techniques to group similar images, where a single effort of the human experts to assign high level semantic to a randomly selected image and propagate to other images through clustering. The investigation has been made on the LabelMe image and LabelMe video dataset. Experiments exhibit that the proposed approaches perform a noticeable improvement towards bridging the semantic gap and reveal that our proposed system outperforms the traditional systems
Online learning of personalised human activity recognition models from user-provided annotations
PhD ThesisIn Human Activity Recognition (HAR), supervised and semi-supervised training are
important tools for devising parametric activity models. For the best modelling performance,
large amounts of annotated personalised sample data are typically required.
Annotating often represents the bottleneck in the overall modelling process as it usually
involves retrospective analysis of experimental ground truth, like video footage. These
approaches typically neglect that prospective users of HAR systems are themselves
key sources of ground truth for their own activities.
This research therefore involves the users of HAR monitors in the annotation process.
The process relies solely on users' short term memory and engages with them to
parsimoniously provide annotations for their own activities as they unfold. E ects
of user input are optimised by using Online Active Learning (OAL) to identify the
most critical annotations which are expected to lead to highly optimal HAR model
performance gains.
Personalised HAR models are trained from user-provided annotations as part of the
evaluation, focusing mainly on objective model accuracy. The OAL approach is contrasted
with Random Selection (RS) { a naive method which makes uninformed annotation
requests. A range of simulation-based annotation scenarios demonstrate that
using OAL brings bene ts in terms of HAR model performance over RS. Additionally,
a mobile application is implemented and deployed in a naturalistic context to collect
annotations from a panel of human participants. The deployment is proof that the
method can truly run in online mode and it also shows that considerable HAR model
performance gains can be registered even under realistic conditions.
The ndings from this research point to the conclusion that online learning from userprovided
annotations is a valid solution to the problem of constructing personalised
HAR models
- …