Publication Date
| In 2015 | 0 |
| Since 2014 | 0 |
| Since 2011 (last 5 years) | 5 |
| Since 2006 (last 10 years) | 13 |
| Since 1996 (last 20 years) | 13 |
Descriptor
| Grade 4 | 7 |
| Mathematics Tests | 5 |
| Item Response Theory | 4 |
| Test Bias | 4 |
| Test Items | 4 |
| Academic Accommodations… | 3 |
| Disabilities | 3 |
| Factor Structure | 3 |
| Grade 5 | 3 |
| Learning Disabilities | 3 |
| More ▼ | |
Source
| Applied Measurement in… | 13 |
Author
| Lee, Yoonsun | 2 |
| Taylor, Catherine S. | 2 |
| Abedi, Jamal | 1 |
| Alvarez, Karina | 1 |
| Benson, Jeri | 1 |
| Cho, Hyun-Jeong | 1 |
| Cline, Frederick | 1 |
| Cook, Linda | 1 |
| Domaleski, Christopher S. | 1 |
| Dossey, John | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 13 |
| Reports - Research | 8 |
| Reports - Evaluative | 3 |
| Reports - Descriptive | 2 |
Education Level
| Grade 4 | 13 |
| Grade 3 | 5 |
| Grade 5 | 5 |
| Grade 7 | 5 |
| Grade 8 | 5 |
| Elementary Education | 4 |
| Elementary Secondary Education | 4 |
| Grade 6 | 4 |
| Grade 10 | 2 |
| Grade 12 | 1 |
| More ▼ | |
Audience
Showing all 13 results
Kachchaf, Rachel; Solano-Flores, Guillermo – Applied Measurement in Education, 2012
We examined how rater language background affects the scoring of short-answer, open-ended test items in the assessment of English language learners (ELLs). Four native English and four native Spanish-speaking certified bilingual teachers scored 107 responses of fourth- and fifth-grade Spanish-speaking ELLs to mathematics items administered in…
Descriptors: Error of Measurement, English Language Learners, Scoring, Bilingual Teachers
Taylor, Catherine S.; Lee, Yoonsun – Applied Measurement in Education, 2012
This was a study of differential item functioning (DIF) for grades 4, 7, and 10 reading and mathematics items from state criterion-referenced tests. The tests were composed of multiple-choice and constructed-response items. Gender DIF was investigated using POLYSIBTEST and a Rasch procedure. The Rasch procedure flagged more items for DIF than did…
Descriptors: Test Bias, Gender Differences, Reading Tests, Mathematics Tests
Cho, Hyun-Jeong; Lee, Jaehoon; Kingston, Neal – Applied Measurement in Education, 2012
This study examined the validity of test accommodation in third-eighth graders using differential item functioning (DIF) and mixture IRT models. Two data sets were used for these analyses. With the first data set (N = 51,591) we examined whether item type (i.e., story, explanation, straightforward) or item features were associated with item…
Descriptors: Testing Accommodations, Test Bias, Item Response Theory, Validity
Engelhard, George, Jr.; Fincher, Melissa; Domaleski, Christopher S. – Applied Measurement in Education, 2011
This study examines the effects of two test administration accommodations on the mathematics performance of students within the context of a large-scale statewide assessment. The two test administration accommodations were resource guides and calculators. A stratified random sample of schools was selected to represent the demographic…
Descriptors: Testing Accommodations, Disabilities, High Stakes Tests, Program Effectiveness
Van Nijlen, Daniel; Janssen, Rianne – Applied Measurement in Education, 2011
The distinction between quantitative and qualitative differences in mastery is essential when monitoring student progress and is crucial for instructional interventions to deal with learning difficulties. Mixture item response theory (IRT) models can provide a convenient way to make the distinction between quantitative and qualitative differences…
Descriptors: Spelling, Indo European Languages, Vowels, Verbal Tests
Wise, Lauress L. – Applied Measurement in Education, 2010
The articles in this special issue make two important contributions to our understanding of the impact of accommodations on test score validity. First, they illustrate a variety of methods for collection and rigorous analyses of empirical data that can supplant expert judgment of the impact of accommodations. These methods range from internal…
Descriptors: Reading Achievement, Educational Assessment, Test Reliability, Learning Disabilities
Taylor, Catherine S.; Lee, Yoonsun – Applied Measurement in Education, 2010
Item response theory (IRT) methods are generally used to create score scales for large-scale tests. Research has shown that IRT scales are stable across groups and over time. Most studies have focused on items that are dichotomously scored. Now Rasch and other IRT models are used to create scales for tests that include polytomously scored items.…
Descriptors: Measures (Individuals), Item Response Theory, Robustness (Statistics), Item Analysis
Laitusis, Cara Cahalan – Applied Measurement in Education, 2010
This study examined the impact of a read-aloud accommodation on standardized test scores of reading comprehension at grades 4 and 8. Under a repeated measures design, students with and without reading-based learning disabilities took both a standard administration and a read-aloud administration of a reading comprehension test. Results show that…
Descriptors: Learning Disabilities, Standardized Tests, Scores, Academic Accommodations (Disabilities)
Cook, Linda; Eignor, Daniel; Sawaki, Yasuyo; Steinberg, Jonathan; Cline, Frederick – Applied Measurement in Education, 2010
This study compared the underlying factors measured by a state standards-based grade 4 English-Language Arts (ELA) assessment given to several groups of students. The focus of the research was to gather evidence regarding whether or not the tests measured the same construct or constructs for students without disabilities who took the test under…
Descriptors: Language Arts, Educational Assessment, Grade 4, State Standards
Lane, Suzanne; Zumbo, Bruno D.; Abedi, Jamal; Benson, Jeri; Dossey, John; Elliott, Stephen N.; Kane, Michael; Linn, Robert; Paredes-Ziker, Cindy; Rodriguez, Michael; Schraw, Gregg; Slattery, Jean; Thomas, Veronica; Willhoft, Joe – Applied Measurement in Education, 2009
Given the changing landscape of educational accountability at the local, state, and national levels, and the changes in the uses of the National Assessment of Educational Progress (NAEP), including the evolving uses of NAEP as a policy tool to interpret state assessment and accountability systems, an explicit statement of the current and potential…
Descriptors: National Competency Tests, Academic Achievement, Accountability, Test Validity
Penfield, Randall D.; Alvarez, Karina; Lee, Okhee – Applied Measurement in Education, 2009
The assessment of differential item functioning (DIF) in polytomous items addresses between-group differences in measurement properties at the item level, but typically does not inform which score levels may be involved in the DIF effect. The framework of differential step functioning (DSF) addresses this issue by examining between-group…
Descriptors: Test Bias, Classification, Test Items, Criteria
Kim, Do-Hong; Schneider, Christina; Siskind, Theresa – Applied Measurement in Education, 2009
This study examined the extent to which the underlying factor structure of the 2005 South Carolina Palmetto Achievement Challenge Tests (PACT) in science for grades 3, 4, and 5 was equivalent for students who were administered the test in a regular (standard) or accommodated form. Three accommodation groups were of interest: students who received…
Descriptors: Testing Accommodations, Science Tests, Elementary School Science, Measurement
Tong, Ye; Kolen, Michael J. – Applied Measurement in Education, 2007
A number of vertical scaling methodologies were examined in this article. Scaling variations included data collection design, scaling method, item response theory (IRT) scoring procedure, and proficiency estimation method. Vertical scales were developed for Grade 3 through Grade 8 for 4 content areas and 9 simulated datasets. A total of 11 scaling…
Descriptors: Achievement Tests, Scaling, Methods, Item Response Theory

Peer reviewed
Direct link
