NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing all 12 results
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Sooyeon; Livingston, Samuel A. – Journal of Educational Measurement, 2010
Score equating based on small samples of examinees is often inaccurate for the examinee populations. We conducted a series of resampling studies to investigate the accuracy of five methods of equating in a common-item design. The methods were chained equipercentile equating of smoothed distributions, chained linear equating, chained mean equating,…
Descriptors: Equated Scores, Test Items, Item Sampling, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Livingston, Samuel A.; Kim, Sooyeon – Journal of Educational Measurement, 2009
This article suggests a method for estimating a test-score equating relationship from small samples of test takers. The method does not require the estimated equating transformation to be linear. Instead, it constrains the estimated equating curve to pass through two pre-specified end points and a middle point determined from the data. In a…
Descriptors: Measurement, Measurement Techniques, Psychometrics, Sample Size
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1973
Article commented on a study by Harris, who presented formulas for the variance of errors of estimation (of a true score from an observed score) and the variance of errors of prediction (of an observed score from an observed score on a parallel test). (Author/RK)
Descriptors: Criterion Referenced Tests, Measurement, Norm Referenced Tests, Test Reliability
Peer reviewed Peer reviewed
Dorans, Neil J.; Livingston, Samuel A. – Journal of Educational Measurement, 1987
This study investigated the hypothesis that females who score high on the Mathematical portion of Scholastic Aptitude Test do so because they have high verbal skills, whereas some males score high on the mathematics despite their relatively low verbal skills. Evidence for and against the hypothesis was observed. (Author/JAZ)
Descriptors: College Entrance Examinations, Females, High Schools, Hypothesis Testing
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1982
To set a standard on the "beardedness" test (see TM 507 062) the probability that a student with a specific score will be judged as bearded must be estimated for each test score. To get an unbiased estimate of that probability, a representative sample of students at each test score level must be chosen. (BW)
Descriptors: Cutting Scores, Evaluation Methods, Graduation Requirements, Minimum Competency Testing
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1982
For tests used to make pass/fail decisions, the relevant standard error of measurement (SEM) is the SEM at the passing score. If the test is highly stratified, this SEM should be estimated by a split-halves approach. A formula and its derivation are provided. (Author)
Descriptors: Cutting Scores, Error of Measurement, Estimation (Mathematics), Mathematical Formulas
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1972
A reliability coefficient for criterion-referenced tests is developed from the assumptions of classical test theory. The coefficient is based on deviations of scores from the criterion score, rather than from the mean. (Author/CK)
Descriptors: Criterion Referenced Tests, Error of Measurement, Mathematical Applications, Norm Referenced Tests
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1972
This article is a reply to a previous paper (see TM 500 488) interpreting Livingston's original article (see TM 500 487). (CK)
Descriptors: Criterion Referenced Tests, Error of Measurement, Norm Referenced Tests, Test Construction
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1972
Author replies to article TM 500 559. (MB)
Descriptors: Criterion Referenced Tests, Measurement Techniques, Norm Referenced Tests, Scoring
Peer reviewed Peer reviewed
Livingston, Samuel A.; Lewis, Charles – Journal of Educational Measurement, 1995
A method is presented for estimating the accuracy and consistency of classifications based on test scores. The reliability of the score is used to estimate effective test length in terms of discrete items. The true-score distribution is estimated by fitting a four-parameter beta model. (SLD)
Descriptors: Classification, Estimation (Mathematics), Scores, Statistical Distributions
Peer reviewed Peer reviewed
Livingston, Samuel A.; Wingersky, Marilyn A. – Journal of Educational Measurement, 1979
Procedures are described for studying the reliability of decisions based on specific passing scores with tests made up of discrete items and designed to measure continuous rather than categorical traits. These procedures are based on the estimation of the joint distribution of true scores and observed scores. (CTM)
Descriptors: Cutting Scores, Decision Making, Efficiency, Error of Measurement
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1993
The extent to which log-linear smoothing could improve the accuracy of common-item equating by the chained equipercentile method in small samples of examinees was investigated with responses from a 100-item test and 93,283 examinees. Smoothing reduced the sample size required for a given degree of accuracy. (SLD)
Descriptors: Advanced Placement Programs, Equated Scores, Estimation (Mathematics), High School Students