Publication Date
| In 2015 | 0 |
| Since 2014 | 3 |
| Since 2011 (last 5 years) | 9 |
| Since 2006 (last 10 years) | 20 |
| Since 1996 (last 20 years) | 50 |
Descriptor
| Test Scoring Machines | 156 |
| Scoring | 72 |
| Computer Assisted Testing | 38 |
| Higher Education | 32 |
| Evaluation Methods | 26 |
| Testing | 26 |
| Automation | 24 |
| Computer Programs | 23 |
| Computers | 22 |
| Educational Testing | 22 |
| More ▼ | |
Source
Author
| Anderson, Paul S. | 6 |
| Attali, Yigal | 6 |
| Shermis, Mark D. | 6 |
| Bejar, Isaac I. | 5 |
| Burstein, Jill | 4 |
| Clauser, Brian E. | 4 |
| Williamson, David M. | 4 |
| Bridgeman, Brent | 3 |
| Clyman, Stephen G. | 3 |
| Herdegen, Mary M., Ed. | 3 |
| More ▼ | |
Publication Type
Education Level
| Higher Education | 8 |
| Elementary Secondary Education | 5 |
| Postsecondary Education | 5 |
| Elementary Education | 4 |
| Grade 6 | 2 |
| Grade 8 | 2 |
| Secondary Education | 2 |
| Early Childhood Education | 1 |
| Grade 3 | 1 |
| Grade 4 | 1 |
| More ▼ | |
Audience
| Practitioners | 8 |
| Researchers | 5 |
| Teachers | 2 |
| Policymakers | 1 |
Showing 1 to 15 of 156 results
Brooks, Thomas; O'Malley, Kimberly; Ragland, Shelley; Young, Michael; Kirkpatrick, Rob – Journal of Educational Research, 2014
The authors compared the performance of third-grade students testing on answer sheets with those testing on machine-scored test booklets. The 1,832 students in the nationally representative sample were assigned at the campus level to complete the Stanford Achievement Test Series, Tenth Edition in 1 of 4 conditions: (a) Form A answer sheet, (b)…
Descriptors: Grade 3, Answer Sheets, Performance Based Assessment, Test Scoring Machines
Attali, Yigal – Educational Testing Service, 2011
The e-rater[R] automated essay scoring system is used operationally in the scoring of TOEFL iBT[R] independent essays. Previous research has found support for a 3-factor structure of the e-rater features. This 3-factor structure has an attractive hierarchical linguistic interpretation with a word choice factor, a grammatical convention within a…
Descriptors: Essay Tests, Language Tests, Test Scoring Machines, Automation
Attali, Yigal – Educational Testing Service, 2011
This paper proposes an alternative content measure for essay scoring, based on the "difference" in the relative frequency of a word in high-scored versus low-scored essays. The "differential word use" (DWU) measure is the average of these differences across all words in the essay. A positive value indicates the essay is using vocabulary more…
Descriptors: Scoring, Essay Tests, Word Frequency, Content Analysis
Sherry, Michael B. – Contemporary Issues in Technology and Teacher Education (CITE Journal), 2014
Teacher candidates have few opportunities during their teacher preparation coursework to investigate practices associated with eliciting and responding to student writing. This article describes an attempt to address this problem with a searchable online digital archive of student writing, with and without teachers' written feedback, as well…
Descriptors: Student Writing Models, Database Design, Search Strategies, Elementary Secondary Education
Haro, Elizabeth K.; Haro, Luis S. – Journal of Chemical Education, 2014
The multiple-choice question (MCQ) is the foundation of knowledge assessment in K-12, higher education, and standardized entrance exams (including the GRE, MCAT, and DAT). However, standard MCQ exams are limited with respect to the types of questions that can be asked when there are only five choices. MCQs offering additional choices more…
Descriptors: Multiple Choice Tests, Coding, Scoring Rubrics, Test Scoring Machines
Bridgeman, Brent; Trapani, Catherine; Attali, Yigal – Applied Measurement in Education, 2012
Essay scores generated by machine and by human raters are generally comparable; that is, they can produce scores with similar means and standard deviations, and machine scores generally correlate as highly with human scores as scores from one human correlate with scores from another human. Although human and machine essay scores are highly related…
Descriptors: Scoring, Essay Tests, College Entrance Examinations, High Stakes Tests
Balogh, Jennifer; Bernstein, Jared; Cheng, Jian; Van Moere, Alistair; Townshend, Brent; Suzuki, Masanori – Educational and Psychological Measurement, 2012
A two-part experiment is presented that validates a new measurement tool for scoring oral reading ability. Data collected by the U.S. government in a large-scale literacy assessment of adults were analyzed by a system called VersaReader that uses automatic speech recognition and speech processing technologies to score oral reading fluency. In the…
Descriptors: Reading Fluency, Measures (Individuals), Scoring, Reading Ability
Liao, Chen-Huei; Kuo, Bor-Chen; Pai, Kai-Chih – Turkish Online Journal of Educational Technology - TOJET, 2012
Automated scoring by means of Latent Semantic Analysis (LSA) has been introduced lately to improve the traditional human scoring system. The purposes of the present study were to develop a LSA-based assessment system to evaluate children's Chinese sentence construction skills and to examine the effectiveness of LSA-based automated scoring function…
Descriptors: Foreign Countries, Program Effectiveness, Scoring, Personality
Young, Jeffrey R. – Education Digest: Essential Readings Condensed for Quick Review, 2011
The best way to eliminate grade inflation is to take professors out of the grading process: Replace them with professional evaluators who never meet students and don't worry that students will punish harsh grades with poor reviews. That's the argument made by leaders of Western Governors University, which has hired 300 adjunct professors who do…
Descriptors: Grading, Evaluators, Outsourcing, Computers
Attali, Yigal; Bridgeman, Brent; Trapani, Catherine – Journal of Technology, Learning, and Assessment, 2010
A generic approach in automated essay scoring produces scores that have the same meaning across all prompts, existing or new, of a writing assessment. This is accomplished by using a single set of linguistic indicators (or features), a consistent way of combining and weighting these features into essay scores, and a focus on features that are not…
Descriptors: Writing Evaluation, Writing Tests, Scoring, Test Scoring Machines
McCurry, Doug – Assessing Writing, 2010
This article considers the claim that machine scoring of writing test responses agrees with human readers as much as humans agree with other humans. These claims about the reliability of machine scoring of writing are usually based on specific and constrained writing tasks, and there is reason for asking whether machine scoring of writing requires…
Descriptors: Writing Tests, Scoring, Interrater Reliability, Computer Assisted Testing
Johnson, Martin; Nadas, Rita – Learning, Media and Technology, 2009
Within large scale educational assessment agencies in the UK, there has been a shift towards assessors marking digitally scanned copies rather than the original paper scripts that were traditionally used. This project uses extended essay examination scripts to consider whether the mode in which an essay is read potentially influences the…
Descriptors: Reading Comprehension, Educational Assessment, Internet, Essay Tests
Coniam, David – ReCALL, 2009
This paper describes a study of the computer essay-scoring program BETSY. While the use of computers in rating written scripts has been criticised in some quarters for lacking transparency or lack of fit with how human raters rate written scripts, a number of essay rating programs are available commercially, many of which claim to offer comparable…
Descriptors: Writing Tests, Scoring, Foreign Countries, Interrater Reliability
Shermis, Mark D.; Shneyderman, Aleksandr; Attali, Yigal – Assessment in Education: Principles, Policy & Practice, 2008
This study was designed to examine the extent to which "content" accounts for variance in scores assigned in automated essay scoring protocols. Specifically it was hypothesised that certain writing genre would emphasise content more than others. Data were drawn from 1668 essays calibrated at two grade levels (6 and 8) using "e-rater[TM]", an…
Descriptors: Predictor Variables, Test Scoring Machines, Essays, Grade 8
Scharber, Cassandra; Dexter, Sara; Riedel, Eric – Journal of Technology, Learning, and Assessment, 2008
The purpose of this research is to analyze preservice teachers' use of and reactions to an automated essay scorer used within an online, case-based learning environment called ETIPS. Data analyzed include post-assignment surveys, a user log of students' actions within the cases, instructor-assigned scores on final essays, and interviews with four…
Descriptors: Test Scoring Machines, Essays, Student Experience, Automation

Peer reviewed
Direct link
