AUTHOR=Wahlen Andreas , Kuhn Christiane , Zlatkin-Troitschanskaia Olga , Gold Christian , Zesch Torsten , Horbach Andrea TITLE=Automated Scoring of Teachers’ Pedagogical Content Knowledge – A Comparison Between Human and Machine Scoring JOURNAL=Frontiers in Education VOLUME=5 YEAR=2020 URL=https://www.frontiersin.org/journals/education/articles/10.3389/feduc.2020.00149 DOI=10.3389/feduc.2020.00149 ISSN=2504-284X ABSTRACT=

To validly assess teachers’ pedagogical content knowledge (PCK), performance-based tasks with open-response formats are required. Automated scoring is considered an appropriate approach to reduce the resource-intensity of human scoring and to achieve more consistent scoring results than human raters. The focus is on the comparability of human and automated scoring of PCK for economics teachers. The answers of (prospective) teachers (N = 852) to six open-response tasks from a standardized and validated test were scored by two trained human raters and the engine “Educational SCoRIng Toolkit” (ESCRITO). The average agreement between human and computer ratings, κw = 0.66, suggests a convergent validity of the scoring results. The results of the single-sector variance analysis show a significant influence of the answers for each homogeneous subgroup (students = 460, trainees = 230, in-service teachers = 162) on the automated scoring. Findings are discussed in terms of implications for the use of automated scoring in educational assessment and its potentials and limitations.