Publication Date
| In 2015 | 2 |
| Since 2014 | 4 |
| Since 2011 (last 5 years) | 9 |
| Since 2006 (last 10 years) | 15 |
| Since 1996 (last 20 years) | 16 |
Descriptor
| Item Response Theory | 16 |
| Test Items | 11 |
| Difficulty Level | 5 |
| Models | 5 |
| Computation | 4 |
| Multiple Choice Tests | 4 |
| Reading Tests | 4 |
| Accuracy | 3 |
| Adaptive Testing | 3 |
| Classification | 3 |
| More ▼ | |
Source
| Practical Assessment,… | 16 |
Author
| Baghaei, Purya | 2 |
| Han, Kyung T. | 2 |
| Rudner, Lawrence M. | 2 |
| Alves, Cecilia | 1 |
| Bao, Han | 1 |
| Briggs, Derek C. | 1 |
| Bucak, S. Deniz | 1 |
| Carstensen, Claus H. | 1 |
| Cor, Ken | 1 |
| Dadey, Nathan | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 16 |
| Reports - Research | 8 |
| Reports - Descriptive | 6 |
| Reports - Evaluative | 2 |
Education Level
| Junior High Schools | 2 |
| Middle Schools | 2 |
| Elementary Education | 1 |
| Elementary Secondary Education | 1 |
| Grade 3 | 1 |
| Grade 4 | 1 |
| Grade 5 | 1 |
| Grade 6 | 1 |
| Grade 7 | 1 |
| Grade 8 | 1 |
| More ▼ | |
Audience
Showing 1 to 15 of 16 results
Baghaei, Purya; Kubinger, Klaus D. – Practical Assessment, Research & Evaluation, 2015
The present paper gives a general introduction to the linear logistic test model (Fischer, 1973), an extension of the Rasch model with linear constraints on item parameters, along with eRm (an R package to estimate different types of Rasch models; Mair, Hatzinger, & Mair, 2014) functions to estimate the model and interpret its parameters. The…
Descriptors: Item Response Theory, Models, Test Validity, Hypothesis Testing
Papanastasiou, Elena C. – Practical Assessment, Research & Evaluation, 2015
If good measurement depends in part on the estimation of accurate item characteristics, it is essential that test developers become aware of discrepancies that may exist on the item parameters before and after item review. The purpose of this study was to examine the answer changing patterns of students while taking paper-and-pencil multiple…
Descriptors: Psychometrics, Difficulty Level, Test Items, Multiple Choice Tests
Han, Kyung T.; Guo, Fanmin – Practical Assessment, Research & Evaluation, 2014
The full-information maximum likelihood (FIML) method makes it possible to estimate and analyze structural equation models (SEM) even when data are partially missing, enabling incomplete data to contribute to model estimation. The cornerstone of FIML is the missing-at-random (MAR) assumption. In (unidimensional) computerized adaptive testing…
Descriptors: Maximum Likelihood Statistics, Structural Equation Models, Data, Computer Assisted Testing
Stoffel, Heather; Raymond, Mark R.; Bucak, S. Deniz; Haist, Steven A. – Practical Assessment, Research & Evaluation, 2014
Previous research on the impact of text and formatting changes on test-item performance has produced mixed results. This matter is important because it is generally acknowledged that "any" change to an item requires that it be recalibrated. The present study investigated the effects of seven classes of stylistic changes on item…
Descriptors: Test Construction, Test Items, Standardized Tests, Physicians
Baghaei, Purya; Carstensen, Claus H. – Practical Assessment, Research & Evaluation, 2013
Standard unidimensional Rasch models assume that persons with the same ability parameters are comparable. That is, the same interpretation applies to persons with identical ability estimates as regards the underlying mental processes triggered by the test. However, research in cognitive psychology shows that persons at the same trait level may…
Descriptors: Item Response Theory, Models, Reading Comprehension, Reading Tests
Han, Kyung T. – Practical Assessment, Research & Evaluation, 2012
For several decades, the "three-parameter logistic model" (3PLM) has been the dominant choice for practitioners in the field of educational measurement for modeling examinees' response data from multiple-choice (MC) items. Past studies, however, have pointed out that the c-parameter of 3PLM should not be interpreted as a guessing parameter. This…
Descriptors: Statistical Analysis, Models, Multiple Choice Tests, Guessing (Tests)
Gadermann, Anne M.; Guhn, Martin; Zumbo, Bruno D. – Practical Assessment, Research & Evaluation, 2012
This paper provides a conceptual, empirical, and practical guide for estimating ordinal reliability coefficients for ordinal item response data (also referred to as Likert, Likert-type, ordered categorical, or rating scale item responses). Conventionally, reliability coefficients, such as Cronbach's alpha, are calculated using a Pearson…
Descriptors: Likert Scales, Rating Scales, Reliability, Computation
Dadey, Nathan; Briggs, Derek C. – Practical Assessment, Research & Evaluation, 2012
A vertical scale, in principle, provides a common metric across tests with differing difficulties (e.g., spanning multiple grades) so that statements of "absolute" growth can be made. This paper compares 16 states' 2007-2008 effect size growth trends on vertically scaled reading and math assessments across grades 3 to 8. Two patterns common in…
Descriptors: Meta Analysis, Scaling, Effect Size, Reading Tests
Thompson, Nathan A. – Practical Assessment, Research & Evaluation, 2011
Computerized classification testing (CCT) is an approach to designing tests with intelligent algorithms, similar to adaptive testing, but specifically designed for the purpose of classifying examinees into categories such as "pass" and "fail." Like adaptive testing for point estimation of ability, the key component is the termination criterion,…
Descriptors: Adaptive Testing, Computer Assisted Testing, Classification, Probability
Huynh, Huynh; Meyer, Patrick – Practical Assessment, Research & Evaluation, 2010
The first part of this paper describes the use of the robust z[subscript R] statistic to link test forms using the Rasch (or one-parameter logistic) model. The procedure is then extended to the two-parameter and three-parameter logistic and two-parameter partial credit (2PPC) models. A real set of data was used to illustrate the extension. The…
Descriptors: Item Response Theory, Statistics, Educational Assessment, Test Items
Shin, Seon-Hi – Practical Assessment, Research & Evaluation, 2009
This study investigated the impact of the coding scheme on IRT-based true score equating under a common-item nonequivalent groups design. Two different coding schemes under investigation were carried out by assigning either a zero or a blank to a missing item response in the equating data. The investigation involved a comparison study using actual…
Descriptors: True Scores, Equated Scores, Item Response Theory, Coding
Rudner, Lawrence M. – Practical Assessment, Research & Evaluation, 2009
This paper describes and evaluates the use of measurement decision theory (MDT) to classify examinees based on their item response patterns. The model has a simple framework that starts with the conditional probabilities of examinees in each category or mastery state responding correctly to each item. The presented evaluation investigates: (1) the…
Descriptors: Classification, Scoring, Item Response Theory, Measurement
Judd, Wallace – Practical Assessment, Research & Evaluation, 2009
Over the past twenty years in performance testing a specific item type with distinguishing characteristics has arisen time and time again. It's been invented independently by dozens of test development teams. And yet this item type is not recognized in the research literature. This article is an invitation to investigate the item type, evaluate…
Descriptors: Test Items, Test Format, Evaluation, Item Analysis
Cor, Ken; Alves, Cecilia; Gierl, Mark – Practical Assessment, Research & Evaluation, 2009
While linear programming is a common tool in business and industry, there have not been many applications in educational assessment and only a handful of individuals have been actively involved in conducting psychometric research in this area. Perhaps this is due, at least in part, to the complexity of existing software packages. This article…
Descriptors: Educational Assessment, Psychometrics, Mathematical Applications, Test Construction
Bao, Han; Dayton, C. Mitchell; Hendrickson, Amy B. – Practical Assessment, Research & Evaluation, 2009
When testlet effects and item idiosyncratic features are both considered to be the reasons of DIF in educational tests using testlets (Wainer & Kiely, 1987) or item bundles (Rosenbaum, 1988), it is interesting to investigate the phenomena of DIF amplification and cancellation due to the interactive effects of these two factors. This research…
Descriptors: Test Bias, Reading Tests, Item Response Theory, Test Items
Previous Page | Next Page ยป
Pages: 1 | 2
Peer reviewed
Direct link
