NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 380 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Cui, Zhongmin – Measurement: Interdisciplinary Research and Perspectives, 2022
Although many educational and psychological tests are labeled as computerized adaptive test (CAT), not all tests show the same level of adaptivity -- some tests might not have much adaptation because of various constraints imposed by test developers. Researchers have proposed some indices to measure the amount of adaption for an adaptive test.…
Descriptors: Adaptive Testing, Computer Assisted Testing, Measurement Techniques
Peer reviewed Peer reviewed
Direct linkDirect link
Cole, Ki; Paek, Insu – Measurement: Interdisciplinary Research and Perspectives, 2022
Statistical Analysis Software (SAS) is a widely used tool for data management analysis across a variety of fields. The procedure for item response theory (PROC IRT) is one to perform unidimensional and multidimensional item response theory (IRT) analysis for dichotomous and polytomous data. This review provides a summary of the features of PROC…
Descriptors: Item Response Theory, Computer Software, Item Analysis, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E.; McBride, James R. – Measurement: Interdisciplinary Research and Perspectives, 2022
A common practical challenge is how to assign ability estimates to all incorrect and all correct response patterns when using item response theory (IRT) models and maximum likelihood estimation (MLE) since ability estimates for these types of responses equal -8 or +8. This article uses a simulation study and data from an operational K-12…
Descriptors: Scores, Adaptive Testing, Computer Assisted Testing, Test Length
Peer reviewed Peer reviewed
Direct linkDirect link
Wheeler, Jordan M.; Engelhard, George; Wang, Jue – Measurement: Interdisciplinary Research and Perspectives, 2022
Objectively scoring constructed-response items on educational assessments has long been a challenge due to the use of human raters. Even well-trained raters using a rubric can inaccurately assess essays. Unfolding models measure rater's scoring accuracy by capturing the discrepancy between criterion and operational ratings by placing essays on an…
Descriptors: Accuracy, Scoring, Statistical Analysis, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Schumacker, Randall E.; Wind, Stefanie A.; Holmes, Lauren F. – Measurement: Interdisciplinary Research and Perspectives, 2021
A variety of resources are available from which researchers can identify measurement instruments, including peer-reviewed journal articles, collections of technical information about published instruments, and electronic databases that are sponsored by universities, testing organizations, and other groups. Although these resources are widespread,…
Descriptors: Measurement Techniques, Journal Articles, Databases, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Sun, Ting; Kim, Stella Yun – Measurement: Interdisciplinary Research and Perspectives, 2021
In many large testing programs, equipercentile equating has been widely used under a random groups design to adjust test difficulty between forms. However, one thorny issue occurs with equipercentile equating when a particular score has no observed frequency. The purpose of this study is to suggest and evaluate six potential methods in…
Descriptors: Equated Scores, Test Length, Sample Size, Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Dimitrov, Dimiter M.; Atanasov, Dimitar V. – Measurement: Interdisciplinary Research and Perspectives, 2021
This study offers an approach to test equating under the latent D-scoring method (DSM-L) using the nonequivalent groups with anchor tests (NEAT) design. The accuracy of the test equating was examined via a simulation study under a 3 × 3 design by two conditions: group ability at three levels and test difficulty at three levels. The results for…
Descriptors: Equated Scores, Scoring, Test Items, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Huber, Chuck; Marcoulides, George A.; Pusic, Martin; Menold, Natalja – Measurement: Interdisciplinary Research and Perspectives, 2021
A readily and widely applicable procedure is discussed that can be used to point and interval estimate the probabilities of particular responses on polytomous items at pre-specified points along underlying latent continua. The items are assumed thereby to be part of unidimensional multi-component measuring instruments that may contain also binary…
Descriptors: Probability, Computation, Test Items, Responses
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A.; Pusic, Martin – Measurement: Interdisciplinary Research and Perspectives, 2021
An interval estimation procedure is discussed that can be used to evaluate the probability of a particular response for a binary or binary scored item at a pre-specified point along an underlying latent continuum. The item is assumed to: (a) be part of a unidimensional multi-component measuring instrument that may contain also polytomous items,…
Descriptors: Item Response Theory, Computation, Probability, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Liao, Manqian; Patton, Jeffrey; Yan, Ray; Jiao, Hong – Measurement: Interdisciplinary Research and Perspectives, 2021
Item harvesters who memorize, record and share test items can jeopardize the validity and fairness of credentialing tests. Item harvesting behaviors are difficult to detect by the existing statistical modeling approaches due to the absence of operational definitions and the idiosyncratic nature of human behaviors. Motivated to detect the…
Descriptors: Data Analysis, Cheating, Identification, Behavior Patterns
Peer reviewed Peer reviewed
Direct linkDirect link
Tomek, Sara; Robinson, Cecil – Measurement: Interdisciplinary Research and Perspectives, 2021
Typical longitudinal growth models assume constant functional growth over time. However, there are often conditions where trajectories may not be constant over time. For example, trajectories of psychological behaviors may vary based on a participant's age, or conversely, participants may experience an intervention that causes trajectories to…
Descriptors: Growth Models, Statistical Analysis, Hierarchical Linear Modeling, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Cai, Liuhan; Albano, Anthony D.; Roussos, Louis A. – Measurement: Interdisciplinary Research and Perspectives, 2021
Multistage testing (MST), an adaptive test delivery mode that involves algorithmic selection of predefined item modules rather than individual items, offers a practical alternative to linear and fully computerized adaptive testing. However, interactions across stages between item modules and examinee groups can lead to challenges in item…
Descriptors: Adaptive Testing, Test Items, Item Response Theory, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Olvera Astivia, Oscar L. – Measurement: Interdisciplinary Research and Perspectives, 2021
Partially specified correlation matrices (not to be confused with matrices with missing data or EM-correlation matrices) can appear in research settings such as integrative data analyses, quantitative systematic reviews or whenever the study design only allows for the collection of certain variables. Although approaches to fill in these missing…
Descriptors: Correlation, Matrices, Statistical Analysis, Research Problems
Peer reviewed Peer reviewed
Direct linkDirect link
Zheng, Xiaying; Yang, Ji Seung – Measurement: Interdisciplinary Research and Perspectives, 2021
The purpose of this paper is to briefly introduce two most common applications of multiple group item response theory (IRT) models, namely detecting differential item functioning (DIF) analysis and nonequivalent group score linking with a simultaneous calibration. We illustrate how to conduct those analyses using the "Stata" item…
Descriptors: Item Response Theory, Test Bias, Computer Software, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Camilli, Gregory – Measurement: Interdisciplinary Research and Perspectives, 2021
After 25 years with small to moderate gains in performance in mathematics, scores on the National Assessment of Educational Progress (NAEP) main assessment declined between 2013 and 2015 in Grades 4 and 8. Previous research has suggested the decline may be linked to the implementation of the "Common Core" state standards. In this…
Descriptors: National Competency Tests, Mathematics Tests, Common Core State Standards, Mathematics Achievement
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  26