ERIC Number: ED218317
Record Type: RIE
Publication Date: 1980
Reference Count: 0
Problems in Stabilizing the Judgment Process.
Measurement problems which jeopardize the reliability and validity of competency-based writing assessments are analyzed. Methods to stabilize rating criteria and readers' application of them are necessary. Most writing assessment programs use guidelines from norm-referenced test methodology. Use of this method of criteria application based on ranking within-occasion endangers or precludes between-occasion uniformity. Judgment stability within a session and across sessions are other problems of measurement. The instability of ratings has been a major weakness of writing skills measures. Two indicators of rating variability are discussed. Rater drift is the rater's progressive deviation within a scoring session from previously shared criteria. Scale instability is the differential application of criteria by raters in different scoring sessions. During scale development and validation, assessments should collect separate ratings on component text features that comprise a total score. Rating methods should intersperse periodic checks. Frequency of checks and nature of feedback on scoring accuracy are important. Sound writing assessment requires scale stability. (Author/DWH)
Publication Type: Reports - Research; Guides - Non-Classroom
Education Level: N/A
Sponsor: National Inst. of Education (ED), Washington, DC.
Authoring Institution: California Univ., Los Angeles. Center for the Study of Evaluation.