NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 13 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Breland, Hunter; Lee, Yong-Won; Muraki, Eiji – Educational and Psychological Measurement, 2005
Eighty-three Test of English as a Foreign Language (TOEFL) writing prompts administered via computer-based testing between July 1998 and August 2000 were examined for differences attributable to the response mode (handwriting or word processing) chosen by examinees. Differences were examined statistically using polytomous logistic regression. A…
Descriptors: Evaluation Methods, Word Processing, Handwriting, Effect Size
Peer reviewed Peer reviewed
Aguinis, Herman; And Others – Educational and Psychological Measurement, 1996
A program designed to administer questionnaires on IBM and IBM-compatible personal computers is described. The program prompts subjects to indicate responses by clicking on a graphic line segment or entering a numeric value. The program enhances accuracy in estimating moderating effects by overcoming transcriptional errors and scale coarseness.…
Descriptors: Computer Assisted Testing, Computer Software, Computer Software Evaluation, Estimation (Mathematics)
Peer reviewed Peer reviewed
Aiken, Lewis R. – Educational and Psychological Measurement, 1996
This article describes a set of 11 menu-driven procedures written in BASICA for MS-DOS based microcomputers for constructing several types of rating scales, attitude scales, and checklists, and for scoring responses to the constructed instruments. The uses of the program are described in detail. (SLD)
Descriptors: Attitude Measures, Check Lists, Computer Assisted Testing, Computer Software
Peer reviewed Peer reviewed
Berry, Kenneth J.; Mielke, Paul W., Jr. – Educational and Psychological Measurement, 1997
Describes a FORTRAN software program that calculates the probability of an observed difference between agreement measures obtained from two independent sets of raters. An example illustrates the use of the DIFFER program in evaluating undergraduate essays. (Author/SLD)
Descriptors: Comparative Analysis, Computer Software, Evaluation Methods, Higher Education
Peer reviewed Peer reviewed
Sawilowsky, Shlomo S. – Educational and Psychological Measurement, 2000
B. Thompson and T. Vacha-Haase have examined the statement "the reliability of the test" with emphasis on the following three words: (1) the first "the"; (2) "test"; and (3) the second "the." This discussion focuses instead on the word "reliability." (Author)
Descriptors: Generalization, Meta Analysis, Psychometrics, Reliability
Peer reviewed Peer reviewed
Williams, Richard H. – Educational and Psychological Measurement, 1992
A QuickBASIC program is described that can be used to conduct Dunnett-like comparisons on a set of independent proportions. The statistical method used, based on Levy's statistical method (1975), is an extension of the well-known Dunnett test for multiple comparisons on means. (Author/SLD)
Descriptors: Comparative Analysis, Computer Software, Computer Software Reviews, Control Groups
Peer reviewed Peer reviewed
Thompson, Bruce – Educational and Psychological Measurement, 2001
Introduces the three subsequent articles from this special section, which extend the discussion of future prospects for progress in the reporting and interpreting of effect sizes by researchers. The authors of these pieces represent diverse views. (SLD)
Descriptors: Effect Size, Hypothesis Testing, Psychology, Research Reports
Peer reviewed Peer reviewed
Mick, David Glen – Educational and Psychological Measurement, 2000
Suggest that the call for more pointed graduate education and more affirmative journal policies on replication-extension made by R. Hubbard and P. Ryan is useful, although inadequate and probably pointless. Statistical significance testing appears to be here to stay despite the charge that it is of "marginal scientific value." (SLD)
Descriptors: Graduate Study, Higher Education, Meta Analysis, Psychological Testing
Peer reviewed Peer reviewed
Stewart, David W. – Educational and Psychological Measurement, 2000
Suggests that replication research and meta-analysis are not substitutes for statistical significance testing, but rather, like measures of effect size, they are complements to statistical significance testing. Significance testing does provide a means for determining what might be usefully replicated. (SLD)
Descriptors: Effect Size, Meta Analysis, Psychological Testing, Scholarly Journals
Peer reviewed Peer reviewed
Kover, Arthur J. – Educational and Psychological Measurement, 2000
The Hubbard and Ryan article is a little ingenuous in its implications for action. Both meta-analyses and replication have problems of their own; each requires careful attention. Good measurement emphasizes proper sampling techniques and using whatever means possible to analyze data. (SLD)
Descriptors: Meta Analysis, Psychological Testing, Scholarly Journals, Statistical Significance
Peer reviewed Peer reviewed
Winer, Russell S. – Educational and Psychological Measurement, 2000
Agrees with R. Hubbard and P. Ryan that statistical significance testing has had a negative impact in that some users have closed their minds to alternative approaches to conducting research. In marketing, the alternatives are not completely satisfactory, however, and researchers are likely to continue to rely on statistical significance testing.…
Descriptors: Meta Analysis, Psychological Testing, Scholarly Journals, Statistical Significance
Peer reviewed Peer reviewed
Hansen, Joe B.; Patton, Michael Quinn – Educational and Psychological Measurement, 1994
Two reviews consider the program-evaluation standards of the Joint Committee on Standards for Educational Evaluations. This edition makes the complexity, risks, and power of evaluation clear through descriptive writing and case illustrations. The organization and content of the standards document are described. (SLD)
Descriptors: Case Studies, Educational Assessment, Evaluation Methods, Evaluation Utilization
Peer reviewed Peer reviewed
Luecht, Richard M. – Educational and Psychological Measurement, 1987
Test Pac, a test scoring and analysis computer program for moderate-sized sample designs using dichotomous response items, performs comprehensive item analyses and multiple reliability estimates. It also performs single-facet generalizability analysis of variance, single-parameter item response theory analyses, test score reporting, and computer…
Descriptors: Computer Assisted Testing, Computer Software, Computer Software Reviews, Item Analysis