Publication Date
| In 2015 | 2 |
| Since 2014 | 4 |
| Since 2011 (last 5 years) | 13 |
| Since 2006 (last 10 years) | 20 |
| Since 1996 (last 20 years) | 21 |
Descriptor
| Test Items | 21 |
| Item Response Theory | 11 |
| Difficulty Level | 7 |
| Models | 6 |
| Statistical Analysis | 5 |
| Adaptive Testing | 4 |
| Computer Assisted Testing | 4 |
| Foreign Countries | 4 |
| Multiple Choice Tests | 4 |
| Reading Tests | 4 |
| More ▼ | |
Source
| Practical Assessment,… | 21 |
Author
| Baghaei, Purya | 2 |
| Buckendahl, Chad W. | 2 |
| Han, Kyung T. | 2 |
| Asmundson, Gordon J. G. | 1 |
| Bao, Han | 1 |
| Becker, Kirk A. | 1 |
| Bergstrom, Betty A. | 1 |
| Bucak, S. Deniz | 1 |
| Carleton, R. Nicholas | 1 |
| Carstensen, Claus H. | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 21 |
| Reports - Research | 10 |
| Reports - Descriptive | 7 |
| Reports - Evaluative | 4 |
Education Level
| Junior High Schools | 4 |
| Middle Schools | 4 |
| Elementary Secondary Education | 3 |
| Secondary Education | 3 |
| Elementary Education | 2 |
| Grade 8 | 2 |
| Higher Education | 2 |
| Postsecondary Education | 2 |
| Grade 3 | 1 |
| Grade 5 | 1 |
| More ▼ | |
Audience
Showing 1 to 15 of 21 results
Baghaei, Purya; Kubinger, Klaus D. – Practical Assessment, Research & Evaluation, 2015
The present paper gives a general introduction to the linear logistic test model (Fischer, 1973), an extension of the Rasch model with linear constraints on item parameters, along with eRm (an R package to estimate different types of Rasch models; Mair, Hatzinger, & Mair, 2014) functions to estimate the model and interpret its parameters. The…
Descriptors: Item Response Theory, Models, Test Validity, Hypothesis Testing
Papanastasiou, Elena C. – Practical Assessment, Research & Evaluation, 2015
If good measurement depends in part on the estimation of accurate item characteristics, it is essential that test developers become aware of discrepancies that may exist on the item parameters before and after item review. The purpose of this study was to examine the answer changing patterns of students while taking paper-and-pencil multiple…
Descriptors: Psychometrics, Difficulty Level, Test Items, Multiple Choice Tests
Han, Kyung T.; Guo, Fanmin – Practical Assessment, Research & Evaluation, 2014
The full-information maximum likelihood (FIML) method makes it possible to estimate and analyze structural equation models (SEM) even when data are partially missing, enabling incomplete data to contribute to model estimation. The cornerstone of FIML is the missing-at-random (MAR) assumption. In (unidimensional) computerized adaptive testing…
Descriptors: Maximum Likelihood Statistics, Structural Equation Models, Data, Computer Assisted Testing
Stoffel, Heather; Raymond, Mark R.; Bucak, S. Deniz; Haist, Steven A. – Practical Assessment, Research & Evaluation, 2014
Previous research on the impact of text and formatting changes on test-item performance has produced mixed results. This matter is important because it is generally acknowledged that "any" change to an item requires that it be recalibrated. The present study investigated the effects of seven classes of stylistic changes on item…
Descriptors: Test Construction, Test Items, Standardized Tests, Physicians
Becker, Kirk A.; Bergstrom, Betty A. – Practical Assessment, Research & Evaluation, 2013
The need for increased exam security, improved test formats, more flexible scheduling, better measurement, and more efficient administrative processes has caused testing agencies to consider converting the administration of their exams from paper-and-pencil to computer-based testing (CBT). Many decisions must be made in order to provide an optimal…
Descriptors: Testing, Models, Testing Programs, Program Administration
Baghaei, Purya; Carstensen, Claus H. – Practical Assessment, Research & Evaluation, 2013
Standard unidimensional Rasch models assume that persons with the same ability parameters are comparable. That is, the same interpretation applies to persons with identical ability estimates as regards the underlying mental processes triggered by the test. However, research in cognitive psychology shows that persons at the same trait level may…
Descriptors: Item Response Theory, Models, Reading Comprehension, Reading Tests
Han, Kyung T. – Practical Assessment, Research & Evaluation, 2012
For several decades, the "three-parameter logistic model" (3PLM) has been the dominant choice for practitioners in the field of educational measurement for modeling examinees' response data from multiple-choice (MC) items. Past studies, however, have pointed out that the c-parameter of 3PLM should not be interpreted as a guessing parameter. This…
Descriptors: Statistical Analysis, Models, Multiple Choice Tests, Guessing (Tests)
Carleton, R. Nicholas; Thibodeau, Michel A.; Osborne, Jason W.; Asmundson, Gordon J. G. – Practical Assessment, Research & Evaluation, 2012
The present study was designed to test for item order effects by measuring four distinct constructs that contribute substantively to anxiety-related psychopathology (i.e., anxiety sensitivity, fear of negative evaluation, injury/illness sensitivity, and intolerance of uncertainty). Participants (n = 999; 71% women) were randomly assigned to…
Descriptors: Anxiety, Test Items, Serial Ordering, Measures (Individuals)
Huebner, Alan – Practical Assessment, Research & Evaluation, 2012
Computerized classification tests (CCTs) often use sequential item selection which administers items according to maximizing psychometric information at a cut point demarcating passing and failing scores. This paper illustrates why this method of item selection leads to the overexposure of a significant number of items, and the performances of…
Descriptors: Computer Assisted Testing, Classification, Test Items, Sequential Approach
Buckendahl, Chad W.; Davis-Becker, Susan L. – Practical Assessment, Research & Evaluation, 2012
The consequences associated with the uses and interpretations of scores for many credentialing testing programs have important implications for a range of stakeholders. Within licensure settings specifically, results from examination programs are often one of the final steps in the process of assessing whether individuals will be allowed to enter…
Descriptors: Licensing Examinations (Professions), Test Items, Dentistry, Minimum Competency Testing
Thompson, Nathan A. – Practical Assessment, Research & Evaluation, 2011
Computerized classification testing (CCT) is an approach to designing tests with intelligent algorithms, similar to adaptive testing, but specifically designed for the purpose of classifying examinees into categories such as "pass" and "fail." Like adaptive testing for point estimation of ability, the key component is the termination criterion,…
Descriptors: Adaptive Testing, Computer Assisted Testing, Classification, Probability
Pibal, Florian; Cesnik, Hermann S. – Practical Assessment, Research & Evaluation, 2011
When administering tests across grades, vertical scaling is often employed to place scores from different tests on a common overall scale so that test-takers' progress can be tracked. In order to be able to link the results across grades, however, common items are needed that are included in both test forms. In the literature there seems to be no…
Descriptors: Scaling, Test Items, Equated Scores, Reading Tests
Schafer, William D.; Coverdale, Bradley J.; Luxenberg, Harlan; Jin, Ying – Practical Assessment, Research & Evaluation, 2011
There are relatively few examples of quantitative approaches to quality control in educational assessment and accountability contexts. Among the several techniques that are used in other fields, Shewart charts have been found in a few instances to be applicable in educational settings. This paper describes Shewart charts and gives examples of how…
Descriptors: Charts, Quality Control, Educational Assessment, Statistical Analysis
Huynh, Huynh; Meyer, Patrick – Practical Assessment, Research & Evaluation, 2010
The first part of this paper describes the use of the robust z[subscript R] statistic to link test forms using the Rasch (or one-parameter logistic) model. The procedure is then extended to the two-parameter and three-parameter logistic and two-parameter partial credit (2PPC) models. A real set of data was used to illustrate the extension. The…
Descriptors: Item Response Theory, Statistics, Educational Assessment, Test Items
Buckendahl, Chad W.; Ferdous, Abdullah A.; Gerrow, Jack – Practical Assessment, Research & Evaluation, 2010
Many testing programs face the practical challenge of having limited resources to conduct comprehensive standard setting studies. Some researchers have suggested that replicating a group's recommended cut score on a full-length test may be possible by using a subset of the items. However, these studies were based on simulated data. This study…
Descriptors: Cutting Scores, Test Items, Standard Setting (Scoring), Methods
Previous Page | Next Page ยป
Pages: 1 | 2
Peer reviewed
Direct link
