9791221501063-28.pdf

Most measures of interrater agreement are defined for ratings regarding a group of targets, each rated by the same group of raters (e.g., the agreement of raters who assess on a rating scale the language proficiency of a corpus of argumentative written texts). However, there are situations in which...

Πλήρης περιγραφή

Λεπτομέρειες βιβλιογραφικής εγγραφής
Γλώσσα:English
Έκδοση: Firenze University Press, Genova University Press 2023
Διαθέσιμο Online:https://books.fupress.com/doi/capitoli/979-12-215-0106-3_28
id oapen-20.500.12657-74900
record_format dspace
spelling oapen-20.500.12657-749002023-08-03T17:59:39Z Chapter Measures of interrater agreement when each target is evaluated by a different group of raters BOVE, Giuseppe Interrater agreement Ordinal data Teacher evaluation bic Book Industry Communication::J Society & social sciences Most measures of interrater agreement are defined for ratings regarding a group of targets, each rated by the same group of raters (e.g., the agreement of raters who assess on a rating scale the language proficiency of a corpus of argumentative written texts). However, there are situations in which agreement between ratings regards a group of targets where each target is evaluated by a different group of raters, like for instance when teachers in a school are evaluated by a questionnaire administered to all the pupils (students) in the classroom. In these situations, a first approach is to evaluate the level of agreement for the whole group of targets by the ANOVA one-way random model. A second approach is to apply subject-specific indices of interrater agreement like rWG, which represents the observed variance in ratings compared to the variance of a theoretical distribution representing no agreement (i.e., the null distribution). Both these approaches are not appropriate for ordinal or nominal scales. In this paper, an index is proposed to evaluate the agreement between raters for each single target (subject or object) on an ordinal scale, and to obtain also a global measure of the interrater agreement for the whole group of cases evaluated. The index is not affected by the possible concentration of ratings on a very small number of levels of the scale, like it happens for the measures based on the ANOVA approach, and it does not depend on the definition of a null distributions like rWG. The main features of the proposal will be illustrated in a study for the assessment of learning teacher behavior in classroom collected in a research conducted in 2018 at Roma Tre University. 2023-08-03T15:06:08Z 2023-08-03T15:06:08Z 2023 chapter ONIX_20230803_9791221501063_96 2704-5846 9791221501063 https://library.oapen.org/handle/20.500.12657/74900 eng Proceedings e report application/pdf Attribution 4.0 International 9791221501063-28.pdf https://books.fupress.com/doi/capitoli/979-12-215-0106-3_28 Firenze University Press, Genova University Press ASA 2022 Data-Driven Decision Making 10.36253/979-12-215-0106-3.28 10.36253/979-12-215-0106-3.28 9223d3ac-6fd2-44c9-bb99-5b98ca9d2fad 863aa499-dbee-4191-9a14-3b5d5ef9e635 9791221501063 134 6 Florence open access
institution OAPEN
collection DSpace
language English
description Most measures of interrater agreement are defined for ratings regarding a group of targets, each rated by the same group of raters (e.g., the agreement of raters who assess on a rating scale the language proficiency of a corpus of argumentative written texts). However, there are situations in which agreement between ratings regards a group of targets where each target is evaluated by a different group of raters, like for instance when teachers in a school are evaluated by a questionnaire administered to all the pupils (students) in the classroom. In these situations, a first approach is to evaluate the level of agreement for the whole group of targets by the ANOVA one-way random model. A second approach is to apply subject-specific indices of interrater agreement like rWG, which represents the observed variance in ratings compared to the variance of a theoretical distribution representing no agreement (i.e., the null distribution). Both these approaches are not appropriate for ordinal or nominal scales. In this paper, an index is proposed to evaluate the agreement between raters for each single target (subject or object) on an ordinal scale, and to obtain also a global measure of the interrater agreement for the whole group of cases evaluated. The index is not affected by the possible concentration of ratings on a very small number of levels of the scale, like it happens for the measures based on the ANOVA approach, and it does not depend on the definition of a null distributions like rWG. The main features of the proposal will be illustrated in a study for the assessment of learning teacher behavior in classroom collected in a research conducted in 2018 at Roma Tre University.
title 9791221501063-28.pdf
spellingShingle 9791221501063-28.pdf
title_short 9791221501063-28.pdf
title_full 9791221501063-28.pdf
title_fullStr 9791221501063-28.pdf
title_full_unstemmed 9791221501063-28.pdf
title_sort 9791221501063-28.pdf
publisher Firenze University Press, Genova University Press
publishDate 2023
url https://books.fupress.com/doi/capitoli/979-12-215-0106-3_28
_version_ 1799945252869505024