125 research outputs found
Using item response theory to explore the psychometric properties of extended matching questions examination in undergraduate medical education
BACKGROUND:
As assessment has been shown to direct learning, it is critical that the examinations developed to test clinical competence in medical undergraduates are valid and reliable. The use of extended matching questions (EMQ) has been advocated to overcome some of the criticisms of using multiple-choice questions to test factual and applied knowledge.
METHODS:
We analysed the results from the Extended Matching Questions Examination taken by 4th year undergraduate medical students in the academic year 2001 to 2002. Rasch analysis was used to examine whether the set of questions used in the examination mapped on to a unidimensional scale, the degree of difficulty of questions within and between the various medical and surgical specialties and the pattern of responses within individual questions to assess the impact of the distractor options.
RESULTS:
Analysis of a subset of items and of the full examination demonstrated internal construct validity and the absence of bias on the majority of questions. Three main patterns of response selection were identified.
CONCLUSION:
Modern psychometric methods based upon the work of Rasch provide a useful approach to the calibration and analysis of EMQ undergraduate medical assessments. The approach allows for a formal test of the unidimensionality of the questions and thus the validity of the summed score. Given the metric calibration which follows fit to the model, it also allows for the establishment of items banks to facilitate continuity and equity in exam standards
A latent trait look at pretest-posttest validation of criterion-referenced test items
Since Cox and Vargas (1966) introduced their pretest-posttest validity index for criterion-referenced test items, a great number of additions and modifications have followed. All are based on the idea of gain scoring; that is, they are computed from the differences between proportions of pretest and posttest item responses. Although the method is simple and generally considered as the prototype of criterion-referenced item analysis, it has many and serious disadvantages. Some of these go back to the fact that it leads to indices based on a dual test administration- and population-dependent item p values. Others have to do with the global information about the discriminating power that these indices provide, the implicit weighting they suppose, and the meaningless maximization of posttest scores they lead to. Analyzing the pretest-posttest method from a latent trait point of view, it is proposed to replace indices like Cox and Vargas’ Dpp by an evaluation of the item information function for the mastery score. An empirical study was conducted to compare the differences in item selection between both methods
Learning physical examination skills outside timetabled training sessions: what happens and why?
Lack of published studies on students’ practice behaviour of physical examination skills outside timetabled training sessions inspired this study into what activities medical students undertake to improve their skills and factors influencing this. Six focus groups of a total of 52 students from Years 1–3 using a pre-established interview guide. Interviews were recorded, transcribed and analyzed using qualitative methods. The interview guide was based on questionnaire results; overall response rate for Years 1–3 was 90% (n = 875). Students report a variety of activities to improve their physical examination skills. On average, students devote 20% of self-study time to skill training with Year 1 students practising significantly more than Year 3 students. Practice patterns shift from just-in-time learning to a longitudinal selfdirected approach. Factors influencing this change are assessment methods and simulated/real patients. Learning resources used include textbooks, examination guidelines, scientific articles, the Internet, videos/DVDs and scoring forms from previous OSCEs. Practising skills on fellow students happens at university rooms or at home. Also family and friends were mentioned to help. Simulated/real patients stimulated students to practise of physical examination skills, initially causing confusion and anxiety about skill performance but leading to increased feelings of competence. Difficult or enjoyable skills stimulate students to practise. The strategies students adopt to master physical examination skills outside timetabled training sessions are self-directed. OSCE assessment does have influence, but learning takes place also when there is no upcoming assessment. Simulated and real patients provide strong incentives to work on skills. Early patient contacts make students feel more prepared for clinical practice
Nothing Lasts Forever: Environmental Discourses on the Collapse of Past Societies
The study of the collapse of past societies raises many questions for the theory and practice of archaeology. Interest in collapse extends as well into the natural sciences and environmental and sustainability policy. Despite a range of approaches to collapse, the predominant paradigm is environmental collapse, which I argue obscures recognition of the dynamic role of social processes that lie at the heart of human communities. These environmental discourses, together with confusion over terminology and the concepts of collapse, have created widespread aporia about collapse and resulted in the creation of mixed messages about complex historical and social processes
The problem of assessing problem solving: can comparative judgement help?
This definitive version of this paper is available at Springerlink: http:dx.doi.org/10.1007/s10649-015-9607-1School mathematics examination papers are typically dominated by short, structured items
that fail to assess sustained reasoning or problem solving. A contributory factor to this
situation is the need for student work to be marked reliably by a large number of markers of
varied experience and competence. We report a study that tested an alternative approach to
assessment, called comparative judgement, which may represent a superior method for
assessing open-ended questions that encourage a range of unpredictable responses. An
innovative problem solving examination paper was specially designed by examiners,
evaluated by mathematics teachers, and administered to 750 secondary school students of
varied mathematical achievement. The students’ work was then assessed by mathematics
education experts using comparative judgement as well as a specially designed, resourceintensive
marking procedure. We report two main findings from the research. First, the
examination paper writers, when freed from the traditional constraint of producing a mark
scheme, designed questions that were less structured and more problem-based than is typical
in current school mathematics examination papers. Second, the comparative judgement
approach to assessing the student work proved successful by our measures of inter-rater
reliability and validity. These findings open new avenues for how school mathematics, and
indeed other areas of the curriculum, might be assessed in the future
- …