Publication Date
| In 2015 | 0 |
| Since 2014 | 35 |
| Since 2011 (last 5 years) | 121 |
| Since 2006 (last 10 years) | 214 |
| Since 1996 (last 20 years) | 319 |
Descriptor
| Foreign Countries | 68 |
| Student Evaluation | 65 |
| Academic Achievement | 53 |
| Scores | 45 |
| Educational Assessment | 44 |
| Evaluation Methods | 41 |
| Test Items | 37 |
| Comparative Analysis | 34 |
| Validity | 34 |
| Elementary Secondary Education | 30 |
| More ▼ | |
Author
| Baker, Eva L. | 10 |
| Gearhart, Maryl | 5 |
| Niemi, David | 5 |
| Shavelson, Richard J. | 5 |
| Wilson, Mark | 5 |
| Borko, Hilda | 4 |
| Furtak, Erin Marie | 4 |
| Martinez, Jose Felipe | 4 |
| Roeser, Robert W. | 4 |
| Sireci, Stephen G. | 4 |
| More ▼ | |
Publication Type
Education Level
| Elementary Secondary Education | 57 |
| Higher Education | 49 |
| Elementary Education | 43 |
| Postsecondary Education | 29 |
| Grade 4 | 24 |
| Middle Schools | 23 |
| Secondary Education | 23 |
| Grade 5 | 21 |
| Intermediate Grades | 14 |
| Grade 8 | 13 |
| More ▼ | |
Audience
| Administrators | 1 |
| Researchers | 1 |
| Teachers | 1 |
Showing 46 to 60 of 319 results
Meyer, J. Patrick; Cash, Anne H.; Mashburn, Andrew – Educational Assessment, 2011
Student-teacher interactions are dynamic relationships that change and evolve over the course of a school year. Measuring classroom quality through observations that focus on these interactions presents challenges when observations are conducted throughout the school year. Variability in observed scores could reflect true changes in the quality of…
Descriptors: Observation, Reliability, Teacher Student Relationship, Error of Measurement
Liu, Ou Lydia; Lee, Hee-Sun; Linn, Marcia C. – Educational Assessment, 2011
Both multiple-choice and constructed-response items have known advantages and disadvantages in measuring scientific inquiry. In this article we explore the function of explanation multiple-choice (EMC) items and examine how EMC items differ from traditional multiple-choice and constructed-response items in measuring scientific reasoning. A group…
Descriptors: Science Tests, Multiple Choice Tests, Responses, Test Items
Shaw, Emily J.; Mattern, Krista D.; Patterson, Brian F. – Educational Assessment, 2011
Despite the similarities that researchers note between the cognitive processes and knowledge involved in reading and writing, there are students who are much stronger readers than writers and those who are much stronger writers than readers. The addition of the writing section to the SAT provides an opportunity to examine whether certain groups of…
Descriptors: College Entrance Examinations, Critical Reading, Reading Tests, Writing Tests
Reed, Deborah K. – Educational Assessment, 2011
This narrative synthesis reviews the psychometric properties of commercially and publicly available retell instruments used to assess the reading comprehension of students in grades K-12. Eleven instruments met selection criteria and were systematically coded for data related to the administration procedures, scoring procedures, and technical…
Descriptors: Reading Comprehension, Elementary Secondary Education, Construct Validity, Validity
Anderson, Daniel; Lai, Cheng-Fei; Alonzo, Julie; Tindal, Gerald – Educational Assessment, 2011
Students with disabilities participate in two major measurement systems. The Individuals with Disabilities Education Act emphasizes working within a Response to Intervention (RTI) framework to identify and monitor the progress of low-performing students. Persistent low-performing students also may be eligible for some form of an alternate…
Descriptors: Curriculum Based Assessment, Alternative Assessment, Learning Disabilities, Legislation
Furtak, Erin Marie; Hardy, Ilonca; Beinbrech, Christina; Shavelson, Richard J.; Shemwell, Jonathan T. – Educational Assessment, 2010
This article adapts the Evidence-Based Reasoning (EBR) Framework (Brown, Furtak, Timms, Nagashima, & Wilson, this issue) to create a coding system for assessing argumentation in science classroom discourse. The instrument, "Evidence-Based Reasoning in Science Classroom Discourse", is intended to provide a means for measuring the quality of EBR in…
Descriptors: Science Education, Logical Thinking, Thinking Skills, Evidence
Brown, Nathaniel J. S.; Nagashima, Sam O.; Fu, Alice; Timms, Michael; Wilson, Mark – Educational Assessment, 2010
The Evidence-Based Reasoning Assessment System (EBRAS) brings together advances in modeling scientific reasoning and assessment design to guide the development of written assessment items that target, disentangle, and elicit evidence of the multiple proficiencies underlying scientific argumentation. In this study, the EBRAS was used to assess the…
Descriptors: Measures (Individuals), Evidence, Persuasive Discourse, Logical Thinking
Shemwell, Jonathan T.; Furtak, Erin Marie – Educational Assessment, 2010
One way to frame science classroom discussion is to engage students in scientific argumentation, an important discourse format within science aimed at coordinating empirical evidence and scientific theory. Framing discussion as scientific argumentation gives clear priority to contributions that are sustained by evidence. We question whether this…
Descriptors: Discussion (Teaching Technique), Persuasive Discourse, Evidence, Classroom Communication
Brown, Nathaniel J. S.; Furtak, Erin Marie; Timms, Michael; Nagashima, Sam O.; Wilson, Mark – Educational Assessment, 2010
Recent science education reforms have emphasized the importance of students engaging with and reasoning from evidence to develop scientific explanations. A number of studies have created frameworks based on Toulmin's (1958/2003) argument pattern, whereas others have developed systems for assessing the quality of students' reasoning to support…
Descriptors: Science Education, Logical Thinking, Thinking Skills, Evidence
Hardy, Ilonca; Kloetzer, Birgit; Moeller, Kornelia; Sodian, Beate – Educational Assessment, 2010
Students' ability to participate in scientific discourse and to appropriately use empirical evidence to support beliefs or conclusions has been consistently stated as a core goal of science education. In the present study, we analyzed the quality of scientific reasoning in elementary school science classrooms, using the Evidence-Based Reasoning…
Descriptors: Elementary School Science, Logical Thinking, Thinking Skills, Concept Formation
Karelitz, Tzur M.; Parrish, Deborah Montgomery; Yamada, Hiroyuki; Wilson, Mark – Educational Assessment, 2010
Assessment systems that track children's progress across time need to be sensitive to the variegated nature of development. Although instruments are commonly designed to assess behaviors within a specific age range, some children advance slower or faster than others and, as a result, often show behaviors from a younger or older age group. This…
Descriptors: Age Groups, Inferences, Test Validity, Test Reliability
Wise, Steven L.; DeMars, Christine E. – Educational Assessment, 2010
Educational program assessment studies often use data from low-stakes tests to provide evidence of program quality. The validity of scores from such tests, however, is potentially threatened by examinee noneffort. This study investigated the extent to which one type of noneffort--rapid-guessing behavior--distorted the results from three types of…
Descriptors: Validity, Program Evaluation, Guessing (Tests), Motivation
Petridou, Alexandra; Williams, Julian – Educational Assessment, 2010
The person-fit literature assumes that aberrant response patterns could be a sign of person mismeasurement, but this assumption has rarely, if ever, been empirically investigated before. We explore the validity of test responses and measures of 10-year-old examinees whose response patterns on a commercial standardized paper-and-pencil mathematics…
Descriptors: Validity, Measurement, Response Style (Tests), Scores
Kim, Do-Hong; Huynh, Huynh – Educational Assessment, 2010
This study investigated whether scores obtained from the online and paper-and-pencil administrations of the statewide end-of-course English test were equivalent for students with and without disabilities. Score comparability was evaluated by examining equivalence of factor structure (measurement invariance) and differential item and bundle…
Descriptors: Computer Assisted Testing, Language Tests, English, Scores
Young, John W.; Steinberg, Jonathan; Cline, Fred; Stone, Elizabeth; Martiniello, Maria; Ling, Guangming; Cho, Yeonsuk – Educational Assessment, 2010
To date, assessment validity research on non-native English speaking students in the United States has focused exclusively on those who are presently English language learners (ELLs). However, little, if any, research has been conducted on two other sizable groups of language minority students: (a) bilingual or multilingual students who were…
Descriptors: Test Validity, English (Second Language), Multilingualism, Bilingualism

Peer reviewed
Direct link
