Despite the importance of writing texts in school, teachers´ competence in assessing the quality of students’ texts seems to be limited with respect to interrater reliability, i.e. objectivity. However, it is unclear whether the reason for this lies in the challenging task itself (assessing text quality) or is a matter of teachers’ lack of expertise (which could be improved by better teacher training). In this study, groups of presumed experts, teachers, and novices rated the overall quality of 20 students´ texts. In addition, they rated the importance of different component properties of texts for text quality assessments. Their ratings of text quality/importance of criteria were compared within the framework of the expert-novice paradigm. A many-facet Rasch model analysis indicated that neither teachers nor any of the other expert groups met predefined expertise criteria. All groups’ diagnostic competences were comparable to novices’ competences. We argue that more effort must be undertaken to identify manifest criteria that define good texts and are suitable for use in school.
Stichwort
assessing text qualityteachers’ competencesmany-facet Rasch model analysiscomposition competenceexpert-novice paradigm