Publication Date
| In 2015 | 0 |
| Since 2014 | 1 |
| Since 2011 (last 5 years) | 33 |
| Since 2006 (last 10 years) | 181 |
| Since 1996 (last 20 years) | 361 |
Descriptor
| Scores | 99 |
| Correlation | 98 |
| Factor Analysis | 91 |
| Factor Structure | 84 |
| Measures (Individuals) | 77 |
| Comparative Analysis | 76 |
| Psychometrics | 76 |
| Test Items | 73 |
| Item Response Theory | 69 |
| Mathematical Models | 65 |
| More ▼ | |
Source
| Educational and Psychological… | 561 |
Author
| Marcoulides, George A. | 12 |
| Algina, James | 10 |
| Kromrey, Jeffrey D. | 8 |
| Thompson, Bruce | 8 |
| Wang, Wen-Chung | 8 |
| Keselman, H. J. | 6 |
| Penfield, Randall D. | 6 |
| Dodd, Barbara G. | 5 |
| Finney, Sara J. | 5 |
| Hancock, Gregory R. | 5 |
| More ▼ | |
Publication Type
| Journal Articles | 561 |
| Reports - Evaluative | 561 |
| Reports - Research | 32 |
| Speeches/Meeting Papers | 25 |
| Information Analyses | 5 |
| Opinion Papers | 3 |
| Reports - Descriptive | 2 |
| Book/Product Reviews | 1 |
| Reference Materials - General | 1 |
Education Level
| Higher Education | 27 |
| High Schools | 9 |
| Elementary Secondary Education | 8 |
| Elementary Education | 6 |
| Grade 4 | 4 |
| Grade 8 | 4 |
| Postsecondary Education | 4 |
| Secondary Education | 4 |
| Grade 7 | 3 |
| Grade 9 | 3 |
| More ▼ | |
Audience
Showing 1 to 15 of 561 results
Hayduk, Leslie – Educational and Psychological Measurement, 2014
Researchers using factor analysis tend to dismiss the significant ill fit of factor models by presuming that if their factor model is close-to-fitting, it is probably close to being properly causally specified. Close fit may indeed result from a model being close to properly causally specified, but close-fitting factor models can also be seriously…
Descriptors: Factor Analysis, Goodness of Fit, Factor Structure, Structural Equation Models
Raykov, Tenko; Dimitrov, Dimiter M.; von Eye, Alexander; Marcoulides, George A. – Educational and Psychological Measurement, 2013
A latent variable modeling method for evaluation of interrater agreement is outlined. The procedure is useful for point and interval estimation of the degree of agreement among a given set of judges evaluating a group of targets. In addition, the approach allows one to test for identity in underlying thresholds across raters as well as to identify…
Descriptors: Interrater Reliability, Models, Statistical Analysis, Computation
Liu, Yang; Maydeu-Olivares, Alberto – Educational and Psychological Measurement, 2013
Local dependence (LD) for binary IRT models can be diagnosed using Chen and Thissen's bivariate X[superscript 2] statistic and the score test statistics proposed by Glas and Suarez-Falcon, and Liu and Thissen. Alternatively, LD can be assessed using general purpose statistics such as bivariate residuals or Maydeu-Olivares and Joe's M[subscript r]…
Descriptors: Item Response Theory, Statistical Analysis, Models, Goodness of Fit
Raykov, Tenko; Marcoulides, George A.; Lee, Chun-Lung; Chang, Chi – Educational and Psychological Measurement, 2013
This note is concerned with a latent variable modeling approach for the study of differential item functioning in a multigroup setting. A multiple-testing procedure that can be used to evaluate group differences in response probabilities on individual items is discussed. The method is readily employed when the aim is also to locate possible…
Descriptors: Test Bias, Statistical Analysis, Models, Hypothesis Testing
Raykov, Tenko; Lee, Chun-Lung; Marcoulides, George A.; Chang, Chi – Educational and Psychological Measurement, 2013
The relationship between saturated path-analysis models and their fit to data is revisited. It is demonstrated that a saturated model need not fit perfectly or even well a given data set when fit to the raw data is examined, a criterion currently frequently overlooked by researchers utilizing path analysis modeling techniques. The potential of…
Descriptors: Structural Equation Models, Goodness of Fit, Path Analysis, Correlation
Tian, Wei; Cai, Li; Thissen, David; Xin, Tao – Educational and Psychological Measurement, 2013
In item response theory (IRT) modeling, the item parameter error covariance matrix plays a critical role in statistical inference procedures. When item parameters are estimated using the EM algorithm, the parameter error covariance matrix is not an automatic by-product of item calibration. Cai proposed the use of Supplemented EM algorithm for…
Descriptors: Item Response Theory, Computation, Matrices, Statistical Inference
Yang, Ji Seung; Hansen, Mark; Cai, Li – Educational and Psychological Measurement, 2012
Traditional estimators of item response theory scale scores ignore uncertainty carried over from the item calibration process, which can lead to incorrect estimates of the standard errors of measurement (SEMs). Here, the authors review a variety of approaches that have been applied to this problem and compare them on the basis of their statistical…
Descriptors: Item Response Theory, Scores, Statistical Analysis, Comparative Analysis
Moses, Tim; Kim, Sooyeon – Educational and Psychological Measurement, 2012
In this study, a ranking strategy was evaluated for comparing subgroups' change using identical, equated, and nonidentical measures. Four empirical data sets were evaluated, each of which contained examinees' scores on two occasions, where the two occasions' scores were obtained on a single identical measure, on two equated tests, and on two…
Descriptors: Testing, Change, Scores, Measures (Individuals)
Patarapichayatham, Chalie; Kamata, Akihito; Kanjanawasee, Sirichai – Educational and Psychological Measurement, 2012
Model specification issues on the cross-level two-way differential item functioning model were previously investigated by Patarapichayatham et al. (2009). Their study clarified that an incorrect model specification can easily lead to biased estimates of key parameters. The objective of this article is to provide further insights on the issue by…
Descriptors: Test Bias, Models, Bayesian Statistics, Statistical Analysis
Thomas, D. Roland; Zumbo, Bruno D. – Educational and Psychological Measurement, 2012
There is such doubt in research practice about the reliability of difference scores that granting agencies, journal editors, reviewers, and committees of graduate students' theses have been known to deplore their use. This most maligned index can be used in studies of change, growth, or perhaps discrepancy between two measures taken on the same…
Descriptors: Statistical Analysis, Reliability, Scores, Change
Debelak, Rudolf; Arendasy, Martin – Educational and Psychological Measurement, 2012
A new approach to identify item clusters fitting the Rasch model is described and evaluated using simulated and real data. The proposed method is based on hierarchical cluster analysis and constructs clusters of items that show a good fit to the Rasch model. It thus gives an estimate of the number of independent scales satisfying the postulates of…
Descriptors: Test Items, Factor Analysis, Evaluation Methods, Simulation
Jiao, Hong; Liu, Junhui; Haynie, Kathleen; Woo, Ada; Gorham, Jerry – Educational and Psychological Measurement, 2012
This study explored the impact of partial credit scoring of one type of innovative items (multiple-response items) in a computerized adaptive version of a large-scale licensure pretest and operational test settings. The impacts of partial credit scoring on the estimation of the ability parameters and classification decisions in operational test…
Descriptors: Test Items, Computer Assisted Testing, Measures (Individuals), Scoring
Chang, Shu-Ren; Plake, Barbara S.; Kramer, Gene A.; Lien, Shu-Mei – Educational and Psychological Measurement, 2011
This study examined the amount of time that different ability-level examinees spend on questions they answer correctly or incorrectly across different pretest item blocks presented on a fixed-length, time-restricted computerized adaptive testing (CAT). Results indicate that different ability-level examinees require different amounts of time to…
Descriptors: Evidence, Test Items, Reaction Time, Adaptive Testing
Frey, Andreas; Seitz, Nicki-Nils – Educational and Psychological Measurement, 2011
The usefulness of multidimensional adaptive testing (MAT) for the assessment of student literacy in the Programme for International Student Assessment (PISA) was examined within a real data simulation study. The responses of N = 14,624 students who participated in the PISA assessments of the years 2000, 2003, and 2006 in Germany were used to…
Descriptors: Adaptive Testing, Literacy, Academic Achievement, Achievement Tests
Brown, Anna; Maydeu-Olivares, Alberto – Educational and Psychological Measurement, 2011
Multidimensional forced-choice formats can significantly reduce the impact of numerous response biases typically associated with rating scales. However, if scored with classical methodology, these questionnaires produce ipsative data, which lead to distorted scale relationships and make comparisons between individuals problematic. This research…
Descriptors: Item Response Theory, Models, Questionnaires, Measurement Techniques

Peer reviewed
Direct link
