NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing all 14 results
Peer reviewed Peer reviewed
Direct linkDirect link
Papanastasiou, Elena C. – Practical Assessment, Research & Evaluation, 2015
If good measurement depends in part on the estimation of accurate item characteristics, it is essential that test developers become aware of discrepancies that may exist on the item parameters before and after item review. The purpose of this study was to examine the answer changing patterns of students while taking paper-and-pencil multiple…
Descriptors: Psychometrics, Difficulty Level, Test Items, Multiple Choice Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E.; Seo, Dong Gi – Practical Assessment, Research & Evaluation, 2014
This article provides a brief overview and comparison of three conditional growth percentile methods; student growth percentiles, percentile rank residuals, and a nonparametric matching method. These approaches seek to describe student growth in terms of the relative percentile ranking of a student in relationship to students that had the same…
Descriptors: Academic Achievement, Achievement Gains, Evaluation Methods, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Hathcoat, John D. – Practical Assessment, Research & Evaluation, 2013
The semantics, or meaning, of validity is a fluid concept in educational and psychological testing. Contemporary controversies surrounding this concept appear to stem from the proper location of validity. Under one view, validity is a property of score-based inferences and entailed uses of test scores. This view is challenged by the…
Descriptors: Test Validity, Educational Testing, Psychological Testing, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Dadey, Nathan; Briggs, Derek C. – Practical Assessment, Research & Evaluation, 2012
A vertical scale, in principle, provides a common metric across tests with differing difficulties (e.g., spanning multiple grades) so that statements of "absolute" growth can be made. This paper compares 16 states' 2007-2008 effect size growth trends on vertically scaled reading and math assessments across grades 3 to 8. Two patterns common in…
Descriptors: Meta Analysis, Scaling, Effect Size, Reading Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Schafer, William D.; Coverdale, Bradley J.; Luxenberg, Harlan; Jin, Ying – Practical Assessment, Research & Evaluation, 2011
There are relatively few examples of quantitative approaches to quality control in educational assessment and accountability contexts. Among the several techniques that are used in other fields, Shewart charts have been found in a few instances to be applicable in educational settings. This paper describes Shewart charts and gives examples of how…
Descriptors: Charts, Quality Control, Educational Assessment, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Lyren, Per-Erik – Practical Assessment, Research & Evaluation, 2009
The added value of reporting subscores on a college admission test (SweSAT) was examined in this study. Using a CTT-derived objective method for determining the value of reporting subscores, it was concluded that there is added value in reporting section scores (Verbal/Quantitative) as well as subtest scores. These results differ from a study of…
Descriptors: College Entrance Examinations, Scores, Test Theory, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Wiberg, Marie; Sundstrom, Anna – Practical Assessment, Research & Evaluation, 2009
A common problem in predictive validity studies in the educational and psychological fields, e.g. in educational and employment selection, is restriction in range of the predictor variables. There are several methods for correcting correlations for restriction of range. The aim of this paper was to examine the usefulness of two approaches to…
Descriptors: Predictive Validity, Predictor Variables, Correlation, Mathematics
Peer reviewed Peer reviewed
Direct linkDirect link
Judd, Wallace – Practical Assessment, Research & Evaluation, 2009
Over the past twenty years in performance testing a specific item type with distinguishing characteristics has arisen time and time again. It's been invented independently by dozens of test development teams. And yet this item type is not recognized in the research literature. This article is an invitation to investigate the item type, evaluate…
Descriptors: Test Items, Test Format, Evaluation, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Cawthon, Stephanie W.; Ho, Eching; Patel, Puja G.; Potvin, Deborah C.; Trundt, Katherine M. – Practical Assessment, Research & Evaluation, 2009
Students with disabilities frequently use accommodations to participate in large-scale, standardized assessments. Accommodations can include changes to the administration of the test, such as extended time, changes to the test items, such as read aloud, or changes to the student's response, such as the use of a scribe. Some accommodations or…
Descriptors: Test Items, Student Evaluation, Test Validity, Student Characteristics
Peer reviewed Peer reviewed
Direct linkDirect link
DiStefano, Christine; Zhu, Min; Mindrila, Diana – Practical Assessment, Research & Evaluation, 2009
Following an exploratory factor analysis, factor scores may be computed and used in subsequent analyses. Factor scores are composite variables which provide information about an individual's placement on the factor(s). This article discusses popular methods to create factor scores under two different classes: refined and non-refined. Strengths and…
Descriptors: Factor Structure, Factor Analysis, Researchers, Scores
Peer reviewed Peer reviewed
Stansfield, Charles W. – Practical Assessment, Research & Evaluation, 2002
Studied the effects of linguistic simplification, a test accommodation designed for students of limited English proficiency (LEP), using data from the Delaware state science assessment program for grades 4 and 6. Findings for 11,306 non-LEP and 109 LEP students show that tests and items can be simplified linguistically without compromising score…
Descriptors: Elementary School Students, Intermediate Grades, Limited English Speaking, Science Tests
Peer reviewed Peer reviewed
La Marca, Paul M. – Practical Assessment, Research & Evaluation, 2001
Provides an overview of the concept of alignment and the role it plays in assessment and accountability systems. Discusses some methodological issues affecting the study of alignment and explores the relationship between alignment and test score interpretation. Alignment is not only a methodological requirement but also an ethical requirement.…
Descriptors: Accountability, Educational Assessment, Elementary Secondary Education, Ethics
Peer reviewed Peer reviewed
Russell, Michael – Practical Assessment, Research & Evaluation, 2000
Discusses the advantages and disadvantages of three commonly used methods for summarizing change in test scores: (1) change in percentile rank; (2) scale or raw score change; and (3) percent change. In general, these methods are insensitive to the measurement scale on which scores are expressed and perform mathematical operations that are…
Descriptors: Change, Measurement Techniques, Scores
Peer reviewed Peer reviewed
Russell, Michael – Practical Assessment, Research & Evaluation, 2000
Describes two less commonly used methods to summarize changes in test scores: (1) standardized growth estimates; and (2) effect sizes. Combines aspects of these approaches and applies them to the Iowa Test of Basic Skills to demonstrate the usefulness of a third method, termed "Expected Growth Size." Provides a template to calculate expected…
Descriptors: Change, Effect Size, Estimation (Mathematics), Scores