Publication Date

In 2022 | 0 |

Since 2021 | 0 |

Since 2018 (last 5 years) | 2 |

Since 2013 (last 10 years) | 5 |

Since 2003 (last 20 years) | 9 |

Descriptor

Source

Educational and Psychological… | 59 |

Author

Vegelius, Jan | 4 |

Conger, Anthony J. | 3 |

Raju, Nambury S. | 3 |

Algina, James | 2 |

Charter, Richard A. | 2 |

Fowler, Robert L. | 2 |

Terrell, Colin D. | 2 |

Wilcox, Rand R. | 2 |

Aiken, Lewis R. | 1 |

Alexander, Ralph A. | 1 |

Badaloni, Silvana | 1 |

More ▼ |

Publication Type

Journal Articles | 58 |

Reports - Research | 46 |

Reports - Evaluative | 8 |

Guides - Non-Classroom | 7 |

Numerical/Quantitative Data | 3 |

Reference Materials - General | 2 |

Reports - Descriptive | 2 |

Opinion Papers | 1 |

Education Level

Audience

Location

Laws, Policies, & Programs

Assessments and Surveys

Fundamental Interpersonal… | 1 |

Rokeach Value Survey | 1 |

What Works Clearinghouse Rating

Hayes, Timothy; Usami, Satoshi – Educational and Psychological Measurement, 2020

Recently, quantitative researchers have shown increased interest in two-step factor score regression (FSR) approaches to structural model estimation. A particularly promising approach proposed by Croon involves first extracting factor scores for each latent factor in a larger model, then correcting the variance-covariance matrix of the factor…

Descriptors: Regression (Statistics), Structural Equation Models, Statistical Bias, Correlation

Ippel, Lianne; Magis, David – Educational and Psychological Measurement, 2020

In dichotomous item response theory (IRT) framework, the asymptotic standard error (ASE) is the most common statistic to evaluate the precision of various ability estimators. Easy-to-use ASE formulas are readily available; however, the accuracy of some of these formulas was recently questioned and new ASE formulas were derived from a general…

Descriptors: Item Response Theory, Error of Measurement, Accuracy, Standards

Donadello, Ivan; Spoto, Andrea; Sambo, Francesco; Badaloni, Silvana; Granziol, Umberto; Vidotto, Giulio – Educational and Psychological Measurement, 2017

The clinical assessment of mental disorders can be a time-consuming and error-prone procedure, consisting of a sequence of diagnostic hypothesis formulation and testing aimed at restricting the set of plausible diagnoses for the patient. In this article, we propose a novel computerized system for the adaptive testing of psychological disorders.…

Descriptors: Adaptive Testing, Mental Disorders, Computer Assisted Testing, Psychological Evaluation

Conger, Anthony J. – Educational and Psychological Measurement, 2017

Drawing parallels to classical test theory, this article clarifies the difference between rater accuracy and reliability and demonstrates how category marginal frequencies affect rater agreement and Cohen's kappa. Category assignment paradigms are developed: comparing raters to a standard (index) versus comparing two raters to one another…

Descriptors: Interrater Reliability, Evaluators, Accuracy, Statistical Analysis

Raykov, Tenko; Marcoulides, George A. – Educational and Psychological Measurement, 2016

The frequently neglected and often misunderstood relationship between classical test theory and item response theory is discussed for the unidimensional case with binary measures and no guessing. It is pointed out that popular item response models can be directly obtained from classical test theory-based models by accounting for the discrete…

Descriptors: Test Theory, Item Response Theory, Models, Correlation

Guo, Jiin-Huarng; Luh, Wei-Ming – Educational and Psychological Measurement, 2008

This study proposes an approach for determining appropriate sample size for Welch's F test when unequal variances are expected. Given a certain maximum deviation in population means and using the quantile of F and t distributions, there is no need to specify a noncentrality parameter and it is easy to estimate the approximate sample size needed…

Descriptors: Sample Size, Monte Carlo Methods, Statistical Analysis, Mathematical Formulas

Rupp, Andre A.; Zumbo, Bruno D. – Educational and Psychological Measurement, 2006

One theoretical feature that makes item response theory (IRT) models those of choice for many psychometric data analysts is parameter invariance, the equality of item and examinee parameters from different examinee populations or measurement conditions. In this article, using the well-known fact that item and examinee parameters are identical only…

Descriptors: Psychometrics, Probability, Simulation, Item Response Theory

Bernaards, Coen A.; Jennrich, Robert I. – Educational and Psychological Measurement, 2005

Almost all modern rotation of factor loadings is based on optimizing a criterion, for example, the quartimax criterion for quartimax rotation. Recent advancements in numerical methods have led to general orthogonal and oblique algorithms for optimizing essentially any rotation criterion. All that is required for a specific application is a…

Descriptors: Computer Software, Factor Analysis, Evaluation Methods, Statistical Analysis

Raju, Nambury S.; Oshima, T.C. – Educational and Psychological Measurement, 2005

Two new prophecy formulas for estimating item response theory (IRT)-based reliability of a shortened or lengthened test are proposed. Some of the relationships between the two formulas, one of which is identical to the well-known Spearman-Brown prophecy formula, are examined and illustrated. The major assumptions underlying these formulas are…

Descriptors: Item Response Theory, Test Reliability, Evaluation Methods, Computation

Peer reviewed

Zimmerman, Donald W. – Educational and Psychological Measurement, 1983

A definition of test validity as the ratio of a covariance term to a variance term, analogous to the classical definition of test reliability, is proposed. When error scores on distinct tests are uncorrelated, the proposed definition coincides with the usual one, but it remains meaningful when error scores are correlated. (Author/BW)

Descriptors: Definitions, Mathematical Formulas, Mathematical Models, Test Theory

Peer reviewed

Conger, Anthony J. – Educational and Psychological Measurement, 1980

Reliability maximizing weights are related to theoretically specified true score scaling weights to show a constant relationship that is invariant under separate linear tranformations on each variable in the system. Test theoretic relations should be derived for the most general model available and not for unnecessarily constrained models.…

Descriptors: Mathematical Formulas, Scaling, Test Reliability, Test Theory

Peer reviewed

Fowler, Robert L. – Educational and Psychological Measurement, 1987

This paper develops a general method for comparing treatment magnitudes for research employing multiple treatment fixed effects analysis of variance designs, which may be used for main effects with any number of levels without regard to directionality. (Author/BS)

Descriptors: Analysis of Variance, Comparative Analysis, Effect Size, Hypothesis Testing

Peer reviewed

Gordon, Leonard V. – Educational and Psychological Measurement, 1973

A simple shortcut procedure for analysis of variance is presented using the means, standard deviations, and number of cases in each sample directly. (Author/NE)

Descriptors: Analysis of Variance, Hypothesis Testing, Mathematical Formulas, Statistical Analysis

Peer reviewed

Magee, Kevin N.; Overall, John E. – Educational and Psychological Measurement, 1992

Formulae for estimating individual rater reliabilities from analysis of treatment effects are presented and evaluated. Monte Carlo methods illustrate the formulae. Results indicate that large sample sizes, large true treatment effects, and large differences in the actual reliabilities of raters are required for the approach to be useful. (SLD)

Descriptors: Effect Size, Estimation (Mathematics), Experimental Groups, Mathematical Formulas

Peer reviewed

Kuder, Frederic – Educational and Psychological Measurement, 1991

Recommendations are made for the appropriate use and identification of traditional Kuder-Richardson formulas for the estimation of reliability. "Alpha" should be used for reliabilities estimated for tests or scales composed of items yielding scores distributed on more than two points. (SLD)

Descriptors: Estimation (Mathematics), Evaluation Methods, Mathematical Formulas, Scores