Publication Date
| In 2015 | 0 |
| Since 2014 | 0 |
| Since 2011 (last 5 years) | 11 |
| Since 2006 (last 10 years) | 17 |
| Since 1996 (last 20 years) | 24 |
Descriptor
| Scoring | 27 |
| Writing Evaluation | 21 |
| Computer Assisted Testing | 9 |
| Student Evaluation | 9 |
| Essay Tests | 8 |
| Essays | 8 |
| Higher Education | 7 |
| Validity | 7 |
| Writing Instruction | 6 |
| Writing Tests | 6 |
| More ▼ | |
Source
| Assessing Writing | 27 |
Author
| James, Cindy L. | 2 |
| Ramineni, Chaitanya | 2 |
| Weigle, Sara Cushing | 2 |
| Brown, Gavin T. L. | 1 |
| Burgin, John | 1 |
| Carrell, Patricia L. | 1 |
| Childs, Ruth | 1 |
| Condon, William | 1 |
| Deane, Paul | 1 |
| Decker, Emily | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 27 |
| Reports - Research | 17 |
| Reports - Evaluative | 8 |
| Guides - Non-Classroom | 1 |
| Reports - Descriptive | 1 |
Education Level
| Elementary Secondary Education | 8 |
| Higher Education | 6 |
| Postsecondary Education | 5 |
| Elementary Education | 3 |
| Grade 6 | 1 |
| High Schools | 1 |
Audience
| Teachers | 3 |
| Policymakers | 1 |
Showing 1 to 15 of 27 results
Klobucar, Andrew; Elliot, Norbert; Deess, Perry; Rudniy, Oleksandr; Joshi, Kamal – Assessing Writing, 2013
This study investigated the use of automated essay scoring (AES) to identify at-risk students enrolled in a first-year university writing course. An application of AES, the "Criterion"[R] Online Writing Evaluation Service was evaluated through a methodology focusing on construct modelling, response processes, disaggregation, extrapolation,…
Descriptors: Writing Evaluation, Scoring, Writing Instruction, Essays
Ramineni, Chaitanya; Williamson, David M. – Assessing Writing, 2013
In this paper, we provide an overview of psychometric procedures and guidelines Educational Testing Service (ETS) uses to evaluate automated essay scoring for operational use. We briefly describe the e-rater system, the procedures and criteria used to evaluate e-rater, implications for a range of potential uses of e-rater, and directions for…
Descriptors: Educational Testing, Guidelines, Scoring, Psychometrics
Condon, William – Assessing Writing, 2013
Automated Essay Scoring (AES) has garnered a great deal of attention from the rhetoric and composition/writing studies community since the Educational Testing Service began using e-rater[R] and the "Criterion"[R] Online Writing Evaluation Service as products in scoring writing tests, and most of the responses have been negative. While the…
Descriptors: Measurement, Psychometrics, Evaluation Methods, Educational Testing
Weigle, Sara Cushing – Assessing Writing, 2013
This article presents considerations for using automated scoring systems to evaluate second language writing. A distinction is made between English language learners in English-medium educational systems and those studying English in their own countries for a variety of purposes, and between learning-to-write and writing-to-learn in a second…
Descriptors: Scoring, Second Language Learning, Second Languages, English Language Learners
Deane, Paul – Assessing Writing, 2013
This paper examines the construct measured by automated essay scoring (AES) systems. AES systems measure features of the text structure, linguistic structure, and conventional print form of essays; as such, the systems primarily measure text production skills. In the current state-of-the-art, AES provide little direct evidence about such matters…
Descriptors: Scoring, Essays, Text Structure, Writing (Composition)
Ramineni, Chaitanya – Assessing Writing, 2013
In this paper, I describe the design and evaluation of automated essay scoring (AES) models for an institution's writing placement program. Information was gathered on admitted student writing performance at a science and technology research university in the northeastern United States. Under timed conditions, first-year students (N = 879) were…
Descriptors: Validity, Comparative Analysis, Internet, Student Placement
Wiseman, Cynthia S. – Assessing Writing, 2012
The decision-making behaviors of 8 raters when scoring 39 persuasive and 39 narrative essays written by second language learners were examined, first using Rasch analysis and then, through think aloud protocols. Results based on Rasch analysis and think aloud protocols recorded by raters as they were scoring holistically and analytically suggested…
Descriptors: Self Concept, Protocol Analysis, Scoring, Item Response Theory
Serviss, Tricia – Assessing Writing, 2012
Drawing upon archival materials, I describe the history, design, and assessment of literacy tests from early 20th century New York state. Practitioners working with these early standardized writing tests grappled with tensions created by public Nativist sentiment, the legislation of "literacy," and calls to score the tests in standardized yet…
Descriptors: Literacy, Writing Tests, Standardized Tests, Scoring
Harsch, Claudia; Martin, Guido – Assessing Writing, 2012
We explore how a local rating scale can be based on the Common European Framework CEF-proficiency scales. As part of the scale validation (Alderson, 1991; Lumley, 2002), we examine which adaptations are needed to turn CEF-proficiency descriptors into a rating scale for a local context, and to establish a practicable method to revise the initial…
Descriptors: Rating Scales, Validity, Media Adaptation, Feedback (Response)
DiPardo, Anne; Storms, Barbara A.; Selland, Makenzie – Assessing Writing, 2011
This paper describes the process by which a rubric development team affiliated with the National Writing Project negotiated difficulties and dilemmas concerning an analytic scoring category initially termed Voice and later renamed Stance. Although these labels reference an aspect of student writing that many teachers value, the challenge of…
Descriptors: Student Evaluation, Scoring Rubrics, Scoring, Educational Assessment
Penny, James A.; Johnson, Robert L. – Assessing Writing, 2011
When multiple raters score a writing sample, on occasion they will award discrepant scores. To report a single score to the examinee, some method of resolving those differences must be applied to the ratings before an operational score can be reported. Several forms of resolving score discrepancies have been described in the literature. Initial…
Descriptors: Monte Carlo Methods, Scores, Academic Achievement, Models
Gebril, Atta – Assessing Writing, 2010
Integrated tasks are currently employed in a number of L2 exams since they are perceived as an addition to the writing-only task type. Given this trend, the current study investigates composite score generalizability of both reading-to-write and writing-only tasks. For this purpose, a multivariate generalizability analysis is used to investigate…
Descriptors: Scoring, Scores, Second Language Instruction, Writing Evaluation
McCurry, Doug – Assessing Writing, 2010
This article considers the claim that machine scoring of writing test responses agrees with human readers as much as humans agree with other humans. These claims about the reliability of machine scoring of writing are usually based on specific and constrained writing tasks, and there is reason for asking whether machine scoring of writing requires…
Descriptors: Writing Tests, Scoring, Interrater Reliability, Computer Assisted Testing
Jeffery, Jill V. – Assessing Writing, 2009
Persistent gaps between optimistic state and pessimistic national academic performance assessment results are increasingly leading to calls for unified national standards in the US. Critics argue that these gaps reveal vast differences in how proficiency is conceptualized; however, little is known about how conceptualizations compare among…
Descriptors: Student Evaluation, Performance Based Assessment, National Standards, Performance Tests
Burgin, John; Hughes, Gail D. – Assessing Writing, 2009
The authors explored the credibility of using informal reading inventories and writing samples for 138 students (K-4) to evaluate the effectiveness of a summer literacy program. Running Records (a measure of a child's reading level) and teacher experience during daily reading instruction were used to estimate the reliability of the more formal…
Descriptors: Informal Reading Inventories, Multiple Choice Tests, Program Effectiveness, Scoring
Previous Page | Next Page ยป
Pages: 1 | 2
Peer reviewed
Direct link
