363 research outputs found

    Supportive technologies for group discussion in MOOCs

    Get PDF
    A key hurdle that prevents MOOCs from reaching their transformative potential in terms of making valuable learning experiences available to the masses is providing support for students to make use of the resources they can provide for each other. This paper lays the foundation for meeting this challenge by beginning with a case study and computational modeling of social interaction data. The analysis yields new knowledge that informs design and development of novel, real-time support for building healthy learning communities that foster a high level of engagement and learning. We conclude by suggesting specific areas for potential impact of new technology

    Personalisation in MOOCs: a critical literature review

    No full text
    The advent and rise of Massive Open Online Courses (MOOCs) have brought many issues to the area of educational technology. Researchers in the field have been addressing these issues such as pedagogical quality of MOOCs, high attrition rates, and sustainability of MOOCs. However, MOOCs personalisation has not been subject of the wide discussions around MOOCs. This paper presents a critical literature survey and analysis of the available literature on personalisation in MOOCs to identify the needs, the current states and efforts to personalise learning in MOOCs. The findings illustrate that there is a growing attention to personalisation to improve learners’ individual learning experiences in MOOCs. In order to implement personalised services, personalised learning path, personalised assessment and feedback, personalised forum thread and recommendation service for related learning materials or learning tasks are commonly applied

    A Computational Method for Enabling Teaching-Learning Process in Huge Online Courses and Communities

    Get PDF
    Massive Open Online Courses and e-learning represent the future of the teaching-learning processes through the development of Information and Communication Technologies. They are the response to the new education needs of society. However, this future also presents many challenges such as the processing of online forums when a huge number of messages are generated. These forums provide an excellent platform for learning and connecting students of the subject, but the difficulties in following and searching the vast volume of information that they generate may produce the opposite effect. In this paper, we propose a computational method for enabling the educational process in huge online learning communities. This method analyses the forum information through Natural Language Processing techniques and extract the main topics discussed. The results generated improves the management of the forums, increases the effectiveness of the teachers’ explanations and reduces the time spent by students to follow the course. The proposal has been complemented with a real case study that shows promising results.This paper has been partially supported by the MESOLAP (TIN2010-14860), GEODAS-BI (TIN2012-37493-C03-03), DIIM2.0 (PROMETEOII/2014/001) and RESCATA (TIN2015-65100-R) projects from the Spanish Ministry of Education and Competitiveness

    Crowdsourcing Cognitive Presence: A Quantitative Content Analysis of a K12 Educator MOOC Discussion Forum

    Get PDF
    Massively Open Online Courses (MOOCs) offer participants opportunities to engage with content and discussion forums similar to other online courses. Pedagogical components of MOOCs and the nature of learning are worth of examining due to issues involving scale, interaction and the role of the instructor (Ross, Sinclair, Know, Bayne & McLeod, 2014). The Community of Inquiry (CoI) framework provides a basis for measuring cognitive presence in online discussion forums. As voluntary point of entry to a community of learners, it is important to consider the nature of participant contributions in terms of cognitive presence. This study focused on an educator MOOC because MOOCs have been proposed as an efficient vehicle for providing professional development due to the significant self-identification of participants as educators (Ho et al. 2014). Participant attributes have been categorized, however the discussion forum is difficult to study on a massive scale (Kizilcec, Piech, & Schulz, 2013). Automated measures of cognitive presence may not provide the full view of learning behaviors implicit in messages posted to the forums (Wong, Pursel, Divinsky & Jansen, 2015). To address this gap, the forum messages were hand-coded and analyzed using quantitative content analysis (Neuendorf, 2002). The study found that the measure of exploration increased over the duration of the course. Viewing cognitive presence over time provided a new metaphor for explaining the proportions of cognitive presence in the discussion forum of an educator MOOC. This finding suggests that increased instructor presence during the later stages of the course may increase cognitive presence over time (Akyol & Garrison, 2007; Garrison & Cleveland-Innes, 2005)

    Setting The Pace: Examining Cognitive Processing in MOOC Discussion Forums With Automatic Text Analysis

    Get PDF
    Learning analytics focuses on extracting meaning from large amounts of data. One of the largest datasets in education comes from Massive Open Online Courses (MOOCs) that typically feature enrollments in the tens of thousands. Analyzing MOOC discussion forums presents logistical issues, resulting chiefly from the size of the dataset, which can create challenges for understanding and adequately describing student behaviors. Utilizing automatic text analysis, this study built a hierarchical linear model that examines the influence of the pacing condition of a massive open online course (MOOC), whether it is self-paced or instructor-paced, on the demonstration of cognitive processing in a HarvardX MOOC. The analysis of 2,423 discussion posts generated by 671 students revealed the number of dictionary words used were positively associated with cognitive processing while analytical thinking and clout was negatively associated. We found that none of the student background information (gender, education), status of the course engagement (explored or completed), or the course pace (self-paced versus instructor paced) significantly influenced the cognitive processing of the postings

    Solving the imbalanced data issue: automatic urgency detection for instructor assistance in MOOC discussion forums

    Get PDF
    In MOOCs, identifying urgent comments on discussion forums is an ongoing challenge. Whilst urgent comments require immediate reactions from instructors, to improve interaction with their learners, and potentially reducing drop-out rates—the task is difficult, as truly urgent comments are rare. From a data analytics perspective, this represents a highly unbalanced (sparse) dataset. Here, we aim to automate the urgent comments identification process, based on fine-grained learner modelling—to be used for automatic recommendations to instructors. To showcase and compare these models, we apply them to the first gold standard dataset for Urgent iNstructor InTErvention (UNITE), which we created by labelling FutureLearn MOOC data. We implement both benchmark shallow classifiers and deep learning. Importantly, we not only compare, for the first time for the unbalanced problem, several data balancing techniques, comprising text augmentation, text augmentation with undersampling, and undersampling, but also propose several new pipelines for combining different augmenters for text augmentation. Results show that models with undersampling can predict most urgent cases; and 3X augmentation + undersampling usually attains the best performance. We additionally validate the best models via a generic benchmark dataset (Stanford). As a case study, we showcase how the naïve Bayes with count vector can adaptively support instructors in answering learner questions/comments, potentially saving time or increasing efficiency in supporting learners. Finally, we show that the errors from the classifier mirrors the disagreements between annotators. Thus, our proposed algorithms perform at least as well as a ‘super-diligent’ human instructor (with the time to consider all comments)
    corecore