65 research outputs found
MOOCs Meet Measurement Theory: A Topic-Modelling Approach
This paper adapts topic models to the psychometric testing of MOOC students
based on their online forum postings. Measurement theory from education and
psychology provides statistical models for quantifying a person's attainment of
intangible attributes such as attitudes, abilities or intelligence. Such models
infer latent skill levels by relating them to individuals' observed responses
on a series of items such as quiz questions. The set of items can be used to
measure a latent skill if individuals' responses on them conform to a Guttman
scale. Such well-scaled items differentiate between individuals and inferred
levels span the entire range from most basic to the advanced. In practice,
education researchers manually devise items (quiz questions) while optimising
well-scaled conformance. Due to the costly nature and expert requirements of
this process, psychometric testing has found limited use in everyday teaching.
We aim to develop usable measurement models for highly-instrumented MOOC
delivery platforms, by using participation in automatically-extracted online
forum topics as items. The challenge is to formalise the Guttman scale
educational constraint and incorporate it into topic models. To favour topics
that automatically conform to a Guttman scale, we introduce a novel
regularisation into non-negative matrix factorisation-based topic modelling. We
demonstrate the suitability of our approach with both quantitative experiments
on three Coursera MOOCs, and with a qualitative survey of topic
interpretability on two MOOCs by domain expert interviews.Comment: 12 pages, 9 figures; accepted into AAAI'201
Methods for Ordinal Peer Grading
MOOCs have the potential to revolutionize higher education with their wide
outreach and accessibility, but they require instructors to come up with
scalable alternates to traditional student evaluation. Peer grading -- having
students assess each other -- is a promising approach to tackling the problem
of evaluation at scale, since the number of "graders" naturally scales with the
number of students. However, students are not trained in grading, which means
that one cannot expect the same level of grading skills as in traditional
settings. Drawing on broad evidence that ordinal feedback is easier to provide
and more reliable than cardinal feedback, it is therefore desirable to allow
peer graders to make ordinal statements (e.g. "project X is better than project
Y") and not require them to make cardinal statements (e.g. "project X is a
B-"). Thus, in this paper we study the problem of automatically inferring
student grades from ordinal peer feedback, as opposed to existing methods that
require cardinal peer feedback. We formulate the ordinal peer grading problem
as a type of rank aggregation problem, and explore several probabilistic models
under which to estimate student grades and grader reliability. We study the
applicability of these methods using peer grading data collected from a real
class -- with instructor and TA grades as a baseline -- and demonstrate the
efficacy of ordinal feedback techniques in comparison to existing cardinal peer
grading methods. Finally, we compare these peer-grading techniques to
traditional evaluation techniques.Comment: Submitted to KDD 201
Recommended from our members
Investigating Different Types of Assessment in Massive Open Online Courses
The current technological era has largely influenced the development of learning environments. As a result, there are new opportunities for teaching, learning and assessment. The emergence of Massive Open Online Courses (MOOCs) in particular, has attracted the attention of higher education institutions and course designers. MOOCs may provide the opportunity to thousands of students to learn from anywhere and at their convenience. Assessment is a component of the learning environment that drives student learning. However, only a small proportion of existing literature on assessment investigates its use for the enhancement of educational growth as most of the literature is concerned with how to use assessment for purposes of grading and ranking (Rowntree, 1987). Assessment has a double role in learning by both motivating students to study in order to undertake it, but also providing the necessary feedback on their performance so that students can track their learning progress (Rowntree, 1987).
Research in MOOCs is currently growing, focusing on different aspects such as the “questionable course quality, high dropout rate, unavailable course credits, complex copyright, limited hardware and ineffective assessments” (Chen, 2014). Assessment in MOOCs has been mostly investigated from a perspective that is looking at: how the grading load can be diminished by adopting automated techniques, the aims of each technique, and finally new potential approaches that will be able to assess high-level cognition. Summing up, researchers are currently testing tools that will be automatically scoring essays and giving feedback to learners in an effective way (see Balfour, 2013). However, the learners’ voice and standpoint about the different assessment types in the MOOCs context is inconclusive in the current literature and there is need for more research.
This study explores learners’ views on assessment types in Massive Open Online Courses, whether any of these has an impact on their enrolment and completion of a course and in what aspects each type of assessment is effective in supporting their learning experience. Auto-assessment, peer-assessment and self-assessment are the types under investigation as they are frequently used in MOOCs and therefore are the most commonly discussed in literature (see Balfour, 2013, Suen, 2013, Wilkowski et al, 2014). The study draws upon literature on assessment in general and on assessment in MOOCs in particular. The concept of online communities, i.e. the learners that appear in MOOCs will also be discussed in detail.
Online ethnographic approaches are employed to explore the issue in question by using online interviewing and observation methods. Thematic analysis is carried out using a sample of 12 MOOCs participants from online interviews and 13 posts of online observations. The outcome of this qualitative research study reveals that even though participants identify benefits in peer assessment, there is a preference for automated assessment since it is an already known, clear type of assessment for them. Moreover, self-assessment is not popular by participants. Learners’ comments also reveal that a clear guidance for assessment helps them to carry out peer assessment more effectively. Some learners also consider that the combination of assessment types may also have a positive effect on students’ learning as each of them serves a different purpose
Economic behavior of information acquisition: Impact of peer grading in MOOCs
A critical issue in operating massive open online courses (MOOCs) is the scalability of providing feedback. Because it is not feasible for instructors to grade a large number of students’ assignments, MOOCs use peer grading systems. This study investigates the efficacy of that practice when student graders are rational economic agents. We characterize grading as a process of (a) acquiring information to assess an assignment’s quality and (b) reporting a score. This process entails a tradeoff between the cost of acquiring information and the benefits of accurate grading. Because the true quality is not observable, any measure of inaccuracy must reference the actions of other graders, which motivates student graders to behave strategically. We present the unique equilibrium information level and reporting strategy of a homogeneous group of student graders and then examine the outcome of peer grading. We show how both the peer grading structure and the nature of MOOC courses affect peer grading accuracy, and we identify conditions under which the process fails. There is a systematic grading bias toward the mean, which discourages students from learning. To improve current practice, we introduce a scale-shift grading scheme, theoretically examine how it can improve grading accuracy and adjust grading bias and discuss how it can be practically implemented
Music Learning with Massive Open Online Courses
Steels, Luc et al.-- Editors: Luc SteelsMassive Open Online Courses, known as MOOCs, have arisen as the logical consequence of marrying long-distance education with the web and social media. MOOCs were confidently predicted by advanced thinkers decades ago. They are undoubtedly here to stay, and provide a valuable resource for learners and teachers alike.
This book focuses on music as a domain of knowledge, and has three objectives: to introduce the phenomenon of MOOCs; to present ongoing research into making MOOCs more effective and better adapted to the needs of teachers and learners; and finally to present the first steps towards 'social MOOCs’, which support the creation of learning communities in which interactions between learners go beyond correcting each other's assignments. Social MOOCs try to mimic settings for humanistic learning, such as workshops, small choirs, or groups participating in a Hackathon, in which students aided by somebody acting as a tutor learn by solving problems and helping each other.
The papers in this book all discuss steps towards social MOOCs; their foundational pedagogy, platforms to create learning communities, methods for assessment and social feedback and concrete experiments. These papers are organized into five sections: background; the role of feedback; platforms for learning communities; experiences with social MOOCs; and looking backwards and looking forward.
Technology is not a panacea for the enormous challenges facing today's educators and learners, but this book will be of interest to all those striving to find more effective and humane learning opportunities for a larger group of students.Funded by the European Commission's OpenAIRE2020 project.Peer reviewe
A Comprehensive Exploration of Personalized Learning in Smart Education: From Student Modeling to Personalized Recommendations
With the development of artificial intelligence, personalized learning has
attracted much attention as an integral part of intelligent education. China,
the United States, the European Union, and others have put forward the
importance of personalized learning in recent years, emphasizing the
realization of the organic combination of large-scale education and
personalized training. The development of a personalized learning system
oriented to learners' preferences and suited to learners' needs should be
accelerated. This review provides a comprehensive analysis of the current
situation of personalized learning and its key role in education. It discusses
the research on personalized learning from multiple perspectives, combining
definitions, goals, and related educational theories to provide an in-depth
understanding of personalized learning from an educational perspective,
analyzing the implications of different theories on personalized learning, and
highlighting the potential of personalized learning to meet the needs of
individuals and to enhance their abilities. Data applications and assessment
indicators in personalized learning are described in detail, providing a solid
data foundation and evaluation system for subsequent research. Meanwhile, we
start from both student modeling and recommendation algorithms and deeply
analyze the cognitive and non-cognitive perspectives and the contribution of
personalized recommendations to personalized learning. Finally, we explore the
challenges and future trajectories of personalized learning. This review
provides a multidimensional analysis of personalized learning through a more
comprehensive study, providing academics and practitioners with cutting-edge
explorations to promote continuous progress in the field of personalized
learning.Comment: 82 pages,5 figure
- …