NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 59 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Hayes, Timothy; Usami, Satoshi – Educational and Psychological Measurement, 2020
Recently, quantitative researchers have shown increased interest in two-step factor score regression (FSR) approaches to structural model estimation. A particularly promising approach proposed by Croon involves first extracting factor scores for each latent factor in a larger model, then correcting the variance-covariance matrix of the factor…
Descriptors: Regression (Statistics), Structural Equation Models, Statistical Bias, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Ippel, Lianne; Magis, David – Educational and Psychological Measurement, 2020
In dichotomous item response theory (IRT) framework, the asymptotic standard error (ASE) is the most common statistic to evaluate the precision of various ability estimators. Easy-to-use ASE formulas are readily available; however, the accuracy of some of these formulas was recently questioned and new ASE formulas were derived from a general…
Descriptors: Item Response Theory, Error of Measurement, Accuracy, Standards
Peer reviewed Peer reviewed
Direct linkDirect link
Donadello, Ivan; Spoto, Andrea; Sambo, Francesco; Badaloni, Silvana; Granziol, Umberto; Vidotto, Giulio – Educational and Psychological Measurement, 2017
The clinical assessment of mental disorders can be a time-consuming and error-prone procedure, consisting of a sequence of diagnostic hypothesis formulation and testing aimed at restricting the set of plausible diagnoses for the patient. In this article, we propose a novel computerized system for the adaptive testing of psychological disorders.…
Descriptors: Adaptive Testing, Mental Disorders, Computer Assisted Testing, Psychological Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Conger, Anthony J. – Educational and Psychological Measurement, 2017
Drawing parallels to classical test theory, this article clarifies the difference between rater accuracy and reliability and demonstrates how category marginal frequencies affect rater agreement and Cohen's kappa. Category assignment paradigms are developed: comparing raters to a standard (index) versus comparing two raters to one another…
Descriptors: Interrater Reliability, Evaluators, Accuracy, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A. – Educational and Psychological Measurement, 2016
The frequently neglected and often misunderstood relationship between classical test theory and item response theory is discussed for the unidimensional case with binary measures and no guessing. It is pointed out that popular item response models can be directly obtained from classical test theory-based models by accounting for the discrete…
Descriptors: Test Theory, Item Response Theory, Models, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Guo, Jiin-Huarng; Luh, Wei-Ming – Educational and Psychological Measurement, 2008
This study proposes an approach for determining appropriate sample size for Welch's F test when unequal variances are expected. Given a certain maximum deviation in population means and using the quantile of F and t distributions, there is no need to specify a noncentrality parameter and it is easy to estimate the approximate sample size needed…
Descriptors: Sample Size, Monte Carlo Methods, Statistical Analysis, Mathematical Formulas
Peer reviewed Peer reviewed
Direct linkDirect link
Rupp, Andre A.; Zumbo, Bruno D. – Educational and Psychological Measurement, 2006
One theoretical feature that makes item response theory (IRT) models those of choice for many psychometric data analysts is parameter invariance, the equality of item and examinee parameters from different examinee populations or measurement conditions. In this article, using the well-known fact that item and examinee parameters are identical only…
Descriptors: Psychometrics, Probability, Simulation, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Bernaards, Coen A.; Jennrich, Robert I. – Educational and Psychological Measurement, 2005
Almost all modern rotation of factor loadings is based on optimizing a criterion, for example, the quartimax criterion for quartimax rotation. Recent advancements in numerical methods have led to general orthogonal and oblique algorithms for optimizing essentially any rotation criterion. All that is required for a specific application is a…
Descriptors: Computer Software, Factor Analysis, Evaluation Methods, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Raju, Nambury S.; Oshima, T.C. – Educational and Psychological Measurement, 2005
Two new prophecy formulas for estimating item response theory (IRT)-based reliability of a shortened or lengthened test are proposed. Some of the relationships between the two formulas, one of which is identical to the well-known Spearman-Brown prophecy formula, are examined and illustrated. The major assumptions underlying these formulas are…
Descriptors: Item Response Theory, Test Reliability, Evaluation Methods, Computation
Peer reviewed Peer reviewed
Zimmerman, Donald W. – Educational and Psychological Measurement, 1983
A definition of test validity as the ratio of a covariance term to a variance term, analogous to the classical definition of test reliability, is proposed. When error scores on distinct tests are uncorrelated, the proposed definition coincides with the usual one, but it remains meaningful when error scores are correlated. (Author/BW)
Descriptors: Definitions, Mathematical Formulas, Mathematical Models, Test Theory
Peer reviewed Peer reviewed
Conger, Anthony J. – Educational and Psychological Measurement, 1980
Reliability maximizing weights are related to theoretically specified true score scaling weights to show a constant relationship that is invariant under separate linear tranformations on each variable in the system. Test theoretic relations should be derived for the most general model available and not for unnecessarily constrained models.…
Descriptors: Mathematical Formulas, Scaling, Test Reliability, Test Theory
Peer reviewed Peer reviewed
Fowler, Robert L. – Educational and Psychological Measurement, 1987
This paper develops a general method for comparing treatment magnitudes for research employing multiple treatment fixed effects analysis of variance designs, which may be used for main effects with any number of levels without regard to directionality. (Author/BS)
Descriptors: Analysis of Variance, Comparative Analysis, Effect Size, Hypothesis Testing
Peer reviewed Peer reviewed
Gordon, Leonard V. – Educational and Psychological Measurement, 1973
A simple shortcut procedure for analysis of variance is presented using the means, standard deviations, and number of cases in each sample directly. (Author/NE)
Descriptors: Analysis of Variance, Hypothesis Testing, Mathematical Formulas, Statistical Analysis
Peer reviewed Peer reviewed
Magee, Kevin N.; Overall, John E. – Educational and Psychological Measurement, 1992
Formulae for estimating individual rater reliabilities from analysis of treatment effects are presented and evaluated. Monte Carlo methods illustrate the formulae. Results indicate that large sample sizes, large true treatment effects, and large differences in the actual reliabilities of raters are required for the approach to be useful. (SLD)
Descriptors: Effect Size, Estimation (Mathematics), Experimental Groups, Mathematical Formulas
Peer reviewed Peer reviewed
Kuder, Frederic – Educational and Psychological Measurement, 1991
Recommendations are made for the appropriate use and identification of traditional Kuder-Richardson formulas for the estimation of reliability. "Alpha" should be used for reliabilities estimated for tests or scales composed of items yielding scores distributed on more than two points. (SLD)
Descriptors: Estimation (Mathematics), Evaluation Methods, Mathematical Formulas, Scores
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4