Low Inter-Rater Reliability of a High Stakes Performance Assessment of Teacher Candidates

The Performance Assessment for California Teachers (PACT) is a high stakes summative assessment that was designed to measure pre-service teacher readiness. We examined the inter-rater reliability (IRR) of trained PACT evaluators who rated 19 candidates. As measured by Cohen’s weighted kappa, the ove...

Full description

Bibliographic Details
Main Authors: Scott A. Lyness, Kent Peterson, Kenneth Yates
Format: Article
Language:English
Published: MDPI AG 2021-10-01
Series:Education Sciences
Subjects:
Online Access:https://www.mdpi.com/2227-7102/11/10/648
_version_ 1797514773419524096
author Scott A. Lyness
Kent Peterson
Kenneth Yates
author_facet Scott A. Lyness
Kent Peterson
Kenneth Yates
author_sort Scott A. Lyness
collection DOAJ
description The Performance Assessment for California Teachers (PACT) is a high stakes summative assessment that was designed to measure pre-service teacher readiness. We examined the inter-rater reliability (IRR) of trained PACT evaluators who rated 19 candidates. As measured by Cohen’s weighted kappa, the overall IRR estimate was 0.17 (poor strength of agreement). IRR estimates ranged from −0.29 (worse than expected by chance) to 0.54 (moderate strength of agreement); all were below the standard of 0.70 for consensus agreement. Follow-up interviews of 10 evaluators revealed possible reasons we observed low IRR, such as departures from established PACT scoring protocol, and lack of, or inconsistent, use of a scoring aid document. Evaluators reported difficulties scoring the materials that candidates submitted, particularly the use of Academic Language. Cognitive Task Analysis (CTA) is suggested as a method to improve IRR in the PACT and other teacher performance assessments such as the edTPA.
first_indexed 2024-03-10T06:36:19Z
format Article
id doaj.art-24b8fc1cc78d42fc94ed094eb2241795
institution Directory Open Access Journal
issn 2227-7102
language English
last_indexed 2024-03-10T06:36:19Z
publishDate 2021-10-01
publisher MDPI AG
record_format Article
series Education Sciences
spelling doaj.art-24b8fc1cc78d42fc94ed094eb22417952023-11-22T18:01:15ZengMDPI AGEducation Sciences2227-71022021-10-01111064810.3390/educsci11100648Low Inter-Rater Reliability of a High Stakes Performance Assessment of Teacher CandidatesScott A. Lyness0Kent Peterson1Kenneth Yates2Rossier School of Education, University of Southern California, Los Angeles, CA 90089, USARossier School of Education, University of Southern California, Los Angeles, CA 90089, USARossier School of Education, University of Southern California, Los Angeles, CA 90089, USAThe Performance Assessment for California Teachers (PACT) is a high stakes summative assessment that was designed to measure pre-service teacher readiness. We examined the inter-rater reliability (IRR) of trained PACT evaluators who rated 19 candidates. As measured by Cohen’s weighted kappa, the overall IRR estimate was 0.17 (poor strength of agreement). IRR estimates ranged from −0.29 (worse than expected by chance) to 0.54 (moderate strength of agreement); all were below the standard of 0.70 for consensus agreement. Follow-up interviews of 10 evaluators revealed possible reasons we observed low IRR, such as departures from established PACT scoring protocol, and lack of, or inconsistent, use of a scoring aid document. Evaluators reported difficulties scoring the materials that candidates submitted, particularly the use of Academic Language. Cognitive Task Analysis (CTA) is suggested as a method to improve IRR in the PACT and other teacher performance assessments such as the edTPA.https://www.mdpi.com/2227-7102/11/10/648inter-rater reliabilitypreservice teacher performance assessmentPACTedTPAweighted kappacognitive task analysis
spellingShingle Scott A. Lyness
Kent Peterson
Kenneth Yates
Low Inter-Rater Reliability of a High Stakes Performance Assessment of Teacher Candidates
Education Sciences
inter-rater reliability
preservice teacher performance assessment
PACT
edTPA
weighted kappa
cognitive task analysis
title Low Inter-Rater Reliability of a High Stakes Performance Assessment of Teacher Candidates
title_full Low Inter-Rater Reliability of a High Stakes Performance Assessment of Teacher Candidates
title_fullStr Low Inter-Rater Reliability of a High Stakes Performance Assessment of Teacher Candidates
title_full_unstemmed Low Inter-Rater Reliability of a High Stakes Performance Assessment of Teacher Candidates
title_short Low Inter-Rater Reliability of a High Stakes Performance Assessment of Teacher Candidates
title_sort low inter rater reliability of a high stakes performance assessment of teacher candidates
topic inter-rater reliability
preservice teacher performance assessment
PACT
edTPA
weighted kappa
cognitive task analysis
url https://www.mdpi.com/2227-7102/11/10/648
work_keys_str_mv AT scottalyness lowinterraterreliabilityofahighstakesperformanceassessmentofteachercandidates
AT kentpeterson lowinterraterreliabilityofahighstakesperformanceassessmentofteachercandidates
AT kennethyates lowinterraterreliabilityofahighstakesperformanceassessmentofteachercandidates