Publication Date
| In 2024 | 54 |
| Since 2023 | 103 |
| Since 2020 (last 5 years) | 282 |
| Since 2015 (last 10 years) | 625 |
| Since 2005 (last 20 years) | 1408 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 109 |
| Practitioners | 107 |
| Teachers | 45 |
| Administrators | 25 |
| Policymakers | 24 |
| Counselors | 12 |
| Parents | 7 |
| Students | 7 |
| Support Staff | 4 |
| Community | 2 |
Location
| California | 60 |
| Canada | 58 |
| United States | 52 |
| Turkey | 47 |
| Australia | 42 |
| Florida | 34 |
| Germany | 26 |
| Netherlands | 25 |
| China | 24 |
| Texas | 24 |
| United Kingdom (England) | 21 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
Buckley, Jack, Ed.; Letukas, Lynn, Ed.; Wildavsky, Ben, Ed. – Johns Hopkins University Press, 2018
For more than seventy-five years, standardized tests have been considered a vital tool for gauging students' readiness for college. However, few people--including students, parents, teachers, and policy makers--understand how tests like the SAT or ACT are used in admissions decisions. Once touted as the best way to compare students from diverse…
Descriptors: Student Evaluation, Standardized Tests, College Entrance Examinations, Admission Criteria
Turkan, Azmi; Cetin, Bayram – Journal of Education and Practice, 2017
Validity and reliability are among the most crucial characteristics of a test. One of the steps to make sure that a test is valid and reliable is to examine the bias in test items. The purpose of this study was to examine the bias in 2012 Placement Test items in terms of gender variable using Rasch Model in Turkey. The sample of this study was…
Descriptors: Item Response Theory, Gender Differences, Test Bias, Test Items
Martinková, Patricia; Drabinová, Adéla; Liaw, Yuan-Ling; Sanders, Elizabeth A.; McFarland, Jenny L.; Price, Rebecca M. – CBE - Life Sciences Education, 2017
We provide a tutorial on differential item functioning (DIF) analysis, an analytic method useful for identifying potentially biased items in assessments. After explaining a number of methodological approaches, we test for gender bias in two scenarios that demonstrate why DIF analysis is crucial for developing assessments, particularly because…
Descriptors: Test Bias, Test Items, Gender Bias, Science Tests
Kaspersen, Eivind; Pepin, Birgit; Sikko, Svein Arne – International Journal of Research & Method in Education, 2017
Several attempts have been made to measure and categorize beliefs and practices of mathematics teachers [Swan, M. 2006. "Designing and Using Research Instruments to Describe the Beliefs and Practices of Mathematics Teachers." "Research in Education" 75 (1): 58-70]. One of the reasons for measuring both beliefs and practices is…
Descriptors: Student Teacher Attitudes, Student Teachers, Beliefs, Mathematics Instruction
Maeda, Yukiko; Yoon, So Yoon – Journal of Psychoeducational Assessment, 2016
We investigated the extent to which the observed gender differences in mental rotation ability among the 2,468 freshmen studying engineering at a Midwest public university attributed to the gender bias of a test. The Revised Purdue Spatial Visualization Tests: Visualization of Rotations (Revised PSVT:R) is a spatial test frequently used to measure…
Descriptors: Gender Differences, Spatial Ability, College Freshmen, Engineering Education
Orosco, Michael J. – International Journal of Science and Mathematics Education, 2016
The psychometric properties of a 10-item math motivation scale were empirically validated with an independent sample consisting of 182 elementary-school students. Analysis of the model dimensionality supported a one-factor structure fit. Item parameter estimates from a Classical Test Theory framework revealed that most items were highly…
Descriptors: Psychometrics, Student Motivation, Mathematics Instruction, Elementary School Students
Ahmadi, Alireza; Bazvand, Ali Darabi – Iranian Journal of Language Teaching Research, 2016
Differential Item Functioning (DIF) exists when examinees of equal ability from different groups have different probabilities of successful performance in a certain item. This study examined gender differential item functioning across the PhD Entrance Exam of TEFL (PEET) in Iran, using both logistic regression (LR) and one-parameter item response…
Descriptors: Test Bias, Gender Bias, College Entrance Examinations, English (Second Language)
Smarter Balanced Assessment Consortium, 2019
The Smarter Balanced Assessment Consortium (Smarter Balanced) strives to provide every student with a positive and productive assessment experience, generating results that are a fair and accurate estimate of each student's achievement. Further, Smarter Balanced is building on a framework of accessibility for all students, including English…
Descriptors: Student Evaluation, Evaluation Methods, English Language Learners, Students with Disabilities
Naumann, Alexander; Hochweber, Jan; Hartig, Johannes – Journal of Educational Measurement, 2014
Students' performance in assessments is commonly attributed to more or less effective teaching. This implies that students' responses are significantly affected by instruction. However, the assumption that outcome measures indeed are instructionally sensitive is scarcely investigated empirically. In the present study, we propose a…
Descriptors: Test Bias, Longitudinal Studies, Hierarchical Linear Modeling, Test Items
Beinicke, Andrea; Pässler, Katja; Hell, Benedikt – International Journal for Educational and Vocational Guidance, 2014
The study investigates consequences of eliminating items showing gender-specific differential item functioning (DIF) on the psychometric structure of a standard RIASEC interest inventory. Holland's hexagonal model was tested for structural invariance using a confirmatory methodological approach (confirmatory factor analysis and randomization…
Descriptors: Test Bias, Gender Differences, Vocational Interests, Interest Inventories
Wells, Craig S.; Hambleton, Ronald K.; Kirkpatrick, Robert; Meng, Yu – Applied Measurement in Education, 2014
The purpose of the present study was to develop and evaluate two procedures flagging consequential item parameter drift (IPD) in an operational testing program. The first procedure was based on flagging items that exhibit a meaningful magnitude of IPD using a critical value that was defined to represent barely tolerable IPD. The second procedure…
Descriptors: Test Items, Test Bias, Equated Scores, Item Response Theory
Davis, Julius; Martin, Danny Bernard – Journal of Urban Mathematics Education, 2018
Couched within a larger critique of assessment practices and how they are used to stigmatize African American children, the authors examine teachers' instructional practices in response to demands of increasing test scores. Many mathematics teachers might be unaware of how these test-driven instructional practices can simultaneously reflect…
Descriptors: Racial Bias, African American Children, Whites, Children
Achieve, Inc., 2018
This document describes the features of a statewide summative science assessment that has been designed to embody standards based on the "Framework for K-12 Science Education," such as the Next Generation Science Standards (NGSS)--to reflect its intent, grounded in the specific expectations of three-dimensional standards. Importantly,…
Descriptors: Summative Evaluation, Science Tests, Standardized Tests, National Standards
Frick, Hannah; Strobl, Carolin; Zeileis, Achim – Educational and Psychological Measurement, 2015
Rasch mixture models can be a useful tool when checking the assumption of measurement invariance for a single Rasch model. They provide advantages compared to manifest differential item functioning (DIF) tests when the DIF groups are only weakly correlated with the manifest covariates available. Unlike in single Rasch models, estimation of Rasch…
Descriptors: Item Response Theory, Test Bias, Comparative Analysis, Scores
Tay, Louis; Vermunt, Jeroen K.; Wang, Chun – International Journal of Testing, 2013
We evaluate the item response theory with covariates (IRT-C) procedure for assessing differential item functioning (DIF) without preknowledge of anchor items (Tay, Newman, & Vermunt, 2011). This procedure begins with a fully constrained baseline model, and candidate items are tested for uniform and/or nonuniform DIF using the Wald statistic.…
Descriptors: Item Response Theory, Test Bias, Models, Statistical Analysis

Direct link
Peer reviewed
