37,667 research outputs found
Lecturersâ and Studentsâ Experiences with an Automated Programming Assessment System
Assessment of source code in university education has become an integral part of grading students and providing them valuable feedback on their developed software solutions. Thereby, lecturers have to deal with a rapidly growing number of students from heterogeneous fields of study, a shortage of lecturers, a highly dynamic set of learning objectives and technologies, and the need for more targeted student support. To meet these challenges, the use of an automated programming assessment system (APAS) to support traditional teaching is a promising solution. This paper examines this trend by analyzing the experiences of lecturers and students at various universities with an APAS and its impact over the course of a semester. In doing so, we conducted a total number of 30 expert interviews with end users, including 15 lecturers and 15 students, from four different universities within the same country. The results discuss the experiences of lecturers and students and highlight challenges that should be addressed in future research
Introductory programming: a systematic literature review
As computing becomes a mainstream discipline embedded in the school curriculum and acts as an enabler for an increasing range of academic disciplines in higher education, the literature on introductory programming is growing. Although there have been several reviews that focus on specific aspects of introductory programming, there has been no broad overview of the literature exploring recent trends across the breadth of introductory programming.
This paper is the report of an ITiCSE working group that conducted a systematic review in order to gain an overview of the introductory programming literature. Partitioning the literature into papers addressing the student, teaching, the curriculum, and assessment, we explore trends, highlight advances in knowledge over the past 15 years, and indicate possible directions for future research
Early experiences of computerâaided assessment and administration when teaching computer programming
This paper describes early experiences with the Ceilidh system currently being piloted at over 30 institutions of higher education. Ceilidh is a courseâmanagement system for teaching computer programming whose core is an autoâassessment facility. This facility automatically marks students programs from a range of perspectives, and may be used in an iterative manner, enabling students to work towards a target level of attainment. Ceilidh also includes extensive courseâadministration and progressâmonitoring facilities, as well as support for other forms of assessment including shortâanswer marking and the collation of essays for later handâmarking. The paper discusses the motivation for developing Ceilidh, outlines its major facilities, then summarizes experiences of developing and actually using it at the coalâface over three years of teaching
An integrated approach to courseware
Software engineering is becoming increasingly important as an engineering discipline, and its teaching in universities and other higher education institutions should be of high quality. In this paper we describe a tool (BOSS â the Boss Online Submission System) which aids the education of software engineers. BOSS allows students to submit programming assignments online, and to run black-box tests on their programs prior to submission. Instructors can use BOSS to assist in marking such assignments by allowing submitted programs to be tested against multiple data sets. We describe how BOSS helps in the teaching of specific conceptual aspects of software engineering, and how it addresses some of the practical issues involved in teaching large student numbers in a pedagogically neutral manner
Framework to Enhance Teaching and Learning in System Analysis and Unified Modelling Language
Cowling, MA ORCiD: 0000-0003-1444-1563; Munoz Carpio, JC ORCiD: 0000-0003-0251-5510Systems Analysis modelling is considered foundational for Information and Communication Technology (ICT) students, with introductory and advanced units included in nearly all ICT and computer science degrees. Yet despite this, novice systems analysts (learners) find modelling and systems thinking quite difficult to learn and master. This makes the process of teaching the fundamentals frustrating and time intensive. This paper will discuss the foundational problems that learners face when learning Systems Analysis modelling. Through a systematic literature review, a framework will be proposed based on the key problems that novice learners experience. In this proposed framework, a sequence of activities has been developed to facilitate understanding of the requirements, solutions and incremental modelling. An example is provided illustrating how the framework could be used to incorporate visualization and gaming elements into a Systems Analysis classroom; therefore, improving motivation and learning. Through this work, a greater understanding of the approach to teaching modelling within the computer science classroom will be provided, as well as a framework to guide future teaching activities
First Year Computer Science Projects at Coventry University:Activity-led integrative team projects with continuous assessment.
We describe the group projects undertaken by first year undergraduate
Computer Science students at Coventry University. These are integrative course
projects: designed to bring together the topics from the various modules
students take, to apply them as a coherent whole. They follow an activity-led
approach, with students given a loose brief and a lot of freedom in how to
develop their project.
We outline the new regulations at Coventry University which eases the use of
such integrative projects. We then describe our continuous assessment approach:
where students earn a weekly mark by demonstrating progress to a teacher as an
open presentation to the class. It involves a degree of self and peer
assessment and allows for an assessment of group work that is both fair, and
seen to be fair. It builds attendance, self-study / continuous engagement
habits, public speaking / presentation skills, and rewards group members for
making meaningful individual contributions.Comment: 4 pages. Accepted for presentation at CEP2
Alternative model for the administration and analysis of research-based assessments
Research-based assessments represent a valuable tool for both instructors and
researchers interested in improving undergraduate physics education. However,
the historical model for disseminating and propagating conceptual and
attitudinal assessments developed by the physics education research (PER)
community has not resulted in widespread adoption of these assessments within
the broader community of physics instructors. Within this historical model,
assessment developers create high quality, validated assessments, make them
available for a wide range of instructors to use, and provide minimal (if any)
support to assist with administration or analysis of the results. Here, we
present and discuss an alternative model for assessment dissemination, which is
characterized by centralized data collection and analysis. This model provides
a greater degree of support for both researchers and instructors in order to
more explicitly support adoption of research-based assessments. Specifically,
we describe our experiences developing a centralized, automated system for an
attitudinal assessment we previously created to examine students'
epistemologies and expectations about experimental physics. This system
provides a proof-of-concept that we use to discuss the advantages associated
with centralized administration and data collection for research-based
assessments in PER. We also discuss the challenges that we encountered while
developing, maintaining, and automating this system. Ultimately, we argue that
centralized administration and data collection for standardized assessments is
a viable and potentially advantageous alternative to the default model
characterized by decentralized administration and analysis. Moreover, with the
help of online administration and automation, this model can support the
long-term sustainability of centralized assessment systems.Comment: 7 pages, 1 figure, accepted in Phys. Rev. PE
Teaching Data Science
We describe an introductory data science course, entitled Introduction to
Data Science, offered at the University of Illinois at Urbana-Champaign. The
course introduced general programming concepts by using the Python programming
language with an emphasis on data preparation, processing, and presentation.
The course had no prerequisites, and students were not expected to have any
programming experience. This introductory course was designed to cover a wide
range of topics, from the nature of data, to storage, to visualization, to
probability and statistical analysis, to cloud and high performance computing,
without becoming overly focused on any one subject. We conclude this article
with a discussion of lessons learned and our plans to develop new data science
courses.Comment: 10 pages, 4 figures, International Conference on Computational
Science (ICCS 2016
- âŠ