17 research outputs found

    Rating scale development: a multistage exploratory sequential design

    Get PDF
    The project chosen to showcase the application of the exploratory sequential design in second/ foreign (L2) language assessment comes from the context of rating scale development and focuses on the development of a set of scales for a suite of high-stakes L2 speaking tests. The assessment of speaking requires assigning scores to a speech sample in a systematic fashion by focusing on explicitly defined criteria which describe different levels of performance (Ginther 2013). Rating scales are the instruments used in this evaluation process, and they can be either holistic (i.e. providing a global overall assessment) or analytic (i.e. providing an independent evaluations for a number of assessment criteria, e.g. Grammar, Vocabulary, Organisation, etc.). The discussion in this chapter is framed within the context of rating scales in speaking assessment. However, it is worth noting that the principles espoused, stages employed and decisions taken during the development process have wider applicability to performance assessment in general

    A comparison of holistic, analytic, and part marking models in speaking assessment

    Get PDF
    This mixed methods study examined holistic, analytic, and part marking models (MMs) in terms of their measurement properties and impact on candidate CEFR classifications in a semi-direct online speaking test. Speaking performances of 240 candidates were first marked holistically and by part (phase 1). On the basis of phase 1 findings – which suggested stronger measurement properties for the part MM – phase 2 focused on a comparison of part and analytic MMs. Speaking performances of 400 candidates were rated analytically and by part during that phase. Raters provided open comments on their marking experiences. Results suggested a significant impact of MM; approximately 30% and 50% of candidates in phases 1 and 2 respectively were awarded different (adjacent) CEFR levels depending on the choice of MM used to assign scores. There was a trend of higher CEFR levels with the holistic MM and lower CEFR levels with the part MM. While strong correlations were found between all pairings of MMs, further analyses revealed important differences. The part MM was shown to display superior measurement qualities particularly in allowing raters to make finer distinctions between different speaking ability levels. These findings have implications for the scoring validity of speaking tests

    Exploring the use of video-conferencing technology in the assessment of spoken language: a mixed-methods study

    Get PDF
    This research explores how internet-based video-conferencing technology can be used to deliver and conduct a speaking test, and what similarities and differences can be discerned between the standard and computer-mediated face-to-face modes. The context of the study is a high-stakes speaking test, and the motivation for the research is the need for test providers to keep under constant review the extent to which their tests are accessible and fair to a wide constituency of test takers. The study examines test-takers’ scores and linguistic output, and examiners’ test administration and rating behaviors across the two modes. A convergent parallel mixed-methods research design was used, analyzing test-takers’ scores and language functions elicited, examiners’ written comments, feedback questionnaires and verbal reports, as well as observation notes taken by researchers. While the two delivery modes generated similar test score outcomes, some differences were observed in test-takers’ functional output and the behavior of examiners who served as both raters and interlocutors

    Developing tools for learning oriented assessment of interactional competence: bridging theory and practice

    Get PDF
    In this paper we report on a project in which we developed tools to support the classroom assessment of learners’ interactional competence (IC) and provided learning oriented feedback in the context of preparation for a high-stakes face-to-face speaking test.  Six trained examiners provided stimulated verbal reports (n=72) on 12 paired interactions, focusing on interactional features of candidates’ performance. We thematically analyzed the verbal reports to inform a draft checklist and materials, which were then trialled by four experienced teachers. Informed by both data sources, the final product comprised (a) a detailed IC checklist with nine main categories and over 50 sub-categories, accompanying detailed description of each area and feedback to learners, which teachers can adapt to suit their teaching and testing contexts, and (b) a concise IC checklist with four categories and bite-sized feedback for real-time classroom assessment. IC, a key aspect of face-to-face communication, is under-researched and under-explored in second/foreign language teaching, learning, and assessment contexts. This in-depth treatment of it, therefore, stands to contribute to learning contexts through raising teachers’ and learners’ awareness of micro-level features of the construct, and to assessment contexts through developing a more comprehensive understanding of the construct

    Introduction

    No full text

    Scoring validity

    No full text

    Exploring language frameworks: proceedings of the ALTE KrakĂłw Conference, July 2011

    No full text
    This volume explores the impact of language frameworks on learning, teaching and assessment, viewed from the perspective of policies, procedures and challenges. It brings together a selection of edited papers, based on presentations given at the 4th International Conference of the Association of Language Testers in Europe (ALTE) held in KrakĂłw, Poland, in July 2011. The selected papers focus on the conference's core themes as follows: the effect of frameworks on teaching, learning and assessment; the value of frameworks for teachers, learners and language policymakers; the contribution of frameworks towards describing particular languages

    Measured constructs: a history of Cambridge English Examinations, 1913-2012

    No full text
    This volume establishes the salient conceptualizations of language constructs with reference to the development of language pedagogy and applied linguistics in the UK over the period 1913 to 2012 and compares these with the constructs underlying the tasks employed in English language testing by a leading UK examining board over the same period. It documents for the first time the historical development of our understanding of the English language construct and shows how English language testing in this period has been shaped by socio-economic considerations, language teaching as well as by socio-cognitive theoretical advances in applied linguistics
    corecore