NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 310 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Deschênes, Marie-France; Dionne, Éric; Dorion, Michelle; Grondin, Julie – Practical Assessment, Research & Evaluation, 2023
The use of the aggregate scoring method for scoring concordance tests requires the weighting of test items to be derived from the performance of a group of experts who take the test under the same conditions as the examinees. However, the average score of experts constituting the reference panel remains a critical issue in the use of these tests.…
Descriptors: Scoring, Tests, Evaluation Methods, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Effatpanah, Farshad; Baghaei, Purya – Practical Assessment, Research & Evaluation, 2023
Item response theory (IRT) refers to a family of mathematical models which describe the relationship between latent continuous variables (attributes or characteristics) and their manifestations (dichotomous/polytomous observed outcomes or responses) with regard to a set of item characteristics. Researchers typically use parametric IRT (PIRT)…
Descriptors: Item Response Theory, Feedback (Response), Mathematical Models, Item Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Clark, Amy; Kobrin, Jennifer L.; Karvonen, Meagan; Hirt, Ashley – Practical Assessment, Research & Evaluation, 2023
Large-scale summative assessment results are typically used for program-evaluation and resource-allocation purposes; however, stakeholders increasingly desire results from large-scale K-12 assessments that inform instruction. Because large-scale summative results are usually delivered after the end of the school year, teacher use of results is…
Descriptors: Data Use, Diagnostic Tests, Decision Making, Summative Evaluation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Evans, Carla M. – Practical Assessment, Research & Evaluation, 2023
Large-scale performance assessment programs are a longstanding reform tool. However, standard setting can be a challenge for assessment programs that use primarily non-standardized assessments. The purpose of this paper is to extend this field of research by explaining the standard setting methodology applied to one more recent instantiation of a…
Descriptors: Standard Setting, Accountability, Performance Based Assessment, State Programs
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Yoo, Hanwook; Wolf, Mikyung Kim; Ballard, Laura D. – Practical Assessment, Research & Evaluation, 2023
As the theme of the 2022 annual meeting of the American Education Research Association, cultivating equitable education systems has gained renewed attention amid an increasingly diverse society. However, systemic inequalities persist for traditionally underserved student populations. As a way to better address diverse students' needs, it is of…
Descriptors: Comparative Analysis, Native Language, English Language Learners, Multilingualism
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Wolkowitz, Amanda A.; Foley, Brett; Zurn, Jared – Practical Assessment, Research & Evaluation, 2023
The purpose of this study is to introduce a method for converting scored 4-option multiple-choice (MC) items into scored 3-option MC items without re-pretesting the 3-option MC items. This study describes a six-step process for achieving this goal. Data from a professional credentialing exam was used in this study and the method was applied to 24…
Descriptors: Multiple Choice Tests, Test Items, Accuracy, Test Format
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Tan, Teck Kiang – Practical Assessment, Research & Evaluation, 2023
Researchers often have hypotheses concerning the state of affairs in the population from which they sampled their data to compare group means. The classical frequentist approach provides one way of carrying out hypothesis testing using ANOVA to state the null hypothesis that there is no difference in the means and proceed with multiple comparisons…
Descriptors: Comparative Analysis, Hypothesis Testing, Statistical Analysis, Guidelines
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Wan, Siyu; Keller, Lisa A. – Practical Assessment, Research & Evaluation, 2023
Statistical process control (SPC) charts have been widely used in the field of educational measurement. The cumulative sum (CUSUM) is an established SPC method to detect aberrant responses for educational assessments. There are many studies that investigated the performance of CUSUM in different test settings. This paper describes the CUSUM…
Descriptors: Visual Aids, Educational Assessment, Evaluation Methods, Item Response Theory
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Metsämuuronen, Jari – Practical Assessment, Research & Evaluation, 2023
Traditional estimators of reliability such as coefficients alpha, theta, omega, and rho (maximal reliability) are prone to give radical underestimates of reliability for the tests common when testing educational achievement. These tests are often structured by widely deviating item difficulties. This is a typical pattern where the traditional…
Descriptors: Test Reliability, Achievement Tests, Computation, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Pentecost, Thomas C.; Raker, Jeffery R.; Murphy, Kristen L. – Practical Assessment, Research & Evaluation, 2023
Using multiple versions of an assessment has the potential to introduce item environment effects. These types of effects result in version dependent item characteristics (i.e., difficulty and discrimination). Methods to detect such effects and resulting implications are important for all levels of assessment where multiple forms of an assessment…
Descriptors: Item Response Theory, Test Items, Test Format, Science Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Hsu, Chia-Ling; Chen, Yi-Hsin; Wu, Yi-Jhen – Practical Assessment, Research & Evaluation, 2023
Correct specifications of hierarchical attribute structures in analyses using diagnostic classification models (DCMs) are pivotal because misspecifications can lead to biased parameter estimations and inaccurate classification profiles. This research is aimed to demonstrate DCM analyses with various hierarchical attribute structures via Bayesian…
Descriptors: Bayesian Statistics, Computation, International Assessment, Achievement Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Elizabeth R. Thomas; Robyn K. Pinilla; Leanne R. Ketterlin-Geller; Cassandra Hatfield – Practical Assessment, Research & Evaluation, 2023
Cognitive interviews play an important role in articulating the intended construct of educational assessments. This paper describes the iterative development of protocols for cognitive interviews with kindergarten through second-grade children to understand how their spatial reasoning skill development aligns with intended constructs. We describe…
Descriptors: Interviews, Kindergarten, Grade 1, Grade 2
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Finch, Holmes – Practical Assessment, Research & Evaluation, 2022
Researchers in many disciplines work with ranking data. This data type is unique in that it is often deterministic in nature (the ranks of items "k"-1 determine the rank of item "k"), and the difference in a pair of rank scores separated by "k" units is equivalent regardless of the actual values of the two ranks in…
Descriptors: Data Analysis, Statistical Inference, Models, College Faculty
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ludlow, Larry H.; O'Keefe, Theresa; Braun, Henry; Anghel, Ella; Szendey, Olivia; Matz, Christina; Howell, Burton – Practical Assessment, Research & Evaluation, 2022
Development of purpose is an important goal of post-secondary education. This study advances the measurement of purpose by (a) enriching the construct through incorporating the facet of horizon; (b) providing a framework for Rasch/Guttman Scenario score interpretation; and (c) providing evidence of convergent, divergent, and known groups validity.
Descriptors: Higher Education, Role of Education, Measurement, Item Response Theory
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Metsämuuronen, Jari – Practical Assessment, Research & Evaluation, 2022
This article discusses visual techniques for detecting test items that would be optimal to be selected to the final compilation on the one hand and, on the other hand, to out-select those items that would lower the quality of the compilation. Some classic visual tools are discussed, first, in a practical manner in diagnosing the logical,…
Descriptors: Test Items, Item Analysis, Item Response Theory, Cutting Scores
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  21