NotesFAQContact Us
Search Tips
Showing all 9 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Jordan, Pascal; Spiess, Martin – Educational and Psychological Measurement, 2019
Factor loadings and item discrimination parameters play a key role in scale construction. A multitude of heuristics regarding their interpretation are hardwired into practice--for example, neglecting low loadings and assigning items to exactly one scale. We challenge the common sense interpretation of these parameters by providing counterexamples…
Descriptors: Test Construction, Test Items, Item Response Theory, Factor Structure
Peer reviewed Peer reviewed
Direct linkDirect link
France, Stephen L.; Batchelder, William H. – Educational and Psychological Measurement, 2015
Cultural consensus theory (CCT) is a data aggregation technique with many applications in the social and behavioral sciences. We describe the intuition and theory behind a set of CCT models for continuous type data using maximum likelihood inference methodology. We describe how bias parameters can be incorporated into these models. We introduce…
Descriptors: Maximum Likelihood Statistics, Test Items, Difficulty Level, Test Theory
Peer reviewed Peer reviewed
Aiken, Lewis R. – Educational and Psychological Measurement, 1996
This article describes a set of 11 menu-driven procedures written in BASICA for MS-DOS based microcomputers for constructing several types of rating scales, attitude scales, and checklists, and for scoring responses to the constructed instruments. The uses of the program are described in detail. (SLD)
Descriptors: Attitude Measures, Check Lists, Computer Assisted Testing, Computer Software
Peer reviewed Peer reviewed
Sax, Gilbert – Educational and Psychological Measurement, 1996
Using various Latin square and incomplete Latin square formats, the Fields test formats provide a novel way of presenting tests to students using machine scoreable answer sheets that can be item analyzed. Items can be constructed to help students acquire knowledge or to measure the attainment of course objectives. (SLD)
Descriptors: Answer Sheets, Item Analysis, Measures (Individuals), Scoring
Peer reviewed Peer reviewed
Fan, Xitao; Chen, Michael – Educational and Psychological Measurement, 2000
Provides a sample of seven published studies in different disciplines that inappropriately generalized reliability coefficients involving several raters to scores generated by a single rater. Score reliability when only one rater is used for scoring is lower than the score reliability for which two raters are used. (SLD)
Descriptors: Interrater Reliability, Research Reports, Scores, Scoring
Peer reviewed Peer reviewed
Ludlow, Larry H.; O'Leary, Michael – Educational and Psychological Measurement, 1999
Focuses on the practical effects of using different statistical treatments with omitted and not-reached items in an item-response theory application. The strategy selected for scoring such items has considerable impact on the interpretation of results for individual or group-level assessments. (Author/SLD)
Descriptors: Data Analysis, Item Response Theory, Scoring, Test Interpretation
Peer reviewed Peer reviewed
Luecht, Richard M. – Educational and Psychological Measurement, 1987
Test Pac, a test scoring and analysis computer program for moderate-sized sample designs using dichotomous response items, performs comprehensive item analyses and multiple reliability estimates. It also performs single-facet generalizability analysis of variance, single-parameter item response theory analyses, test score reporting, and computer…
Descriptors: Computer Assisted Testing, Computer Software, Computer Software Reviews, Item Analysis
Peer reviewed Peer reviewed
Wilcox, Rand R. – Educational and Psychological Measurement, 1981
This paper describes and compares procedures for estimating the reliability of proficiency tests that are scored with latent structure models. Results suggest that the predictive estimate is the most accurate of the procedures. (Author/BW)
Descriptors: Criterion Referenced Tests, Scoring, Test Reliability
Peer reviewed Peer reviewed
Aiken, Lewis R. – Educational and Psychological Measurement, 1987
Formulas for transforming scores and statistics on a rating scale having any number of categories to a scale having a different number of categories are described. To illustrate the use of one of the formulas, the means and variances of items on six forms of a course evaluation questionnaire were compared. (Author/LMO)
Descriptors: Mathematical Models, Rating Scales, Scoring, Statistical Studies