NotesFAQContact Us
Collection
Advanced
Search Tips
Source
Educational and Psychological…56
Audience
Location
Canada1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 56 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Xiaowen; Jane Rogers, H. – Educational and Psychological Measurement, 2022
Test fairness is critical to the validity of group comparisons involving gender, ethnicities, culture, or treatment conditions. Detection of differential item functioning (DIF) is one component of efforts to ensure test fairness. The current study compared four treatments for items that have been identified as showing DIF: deleting, ignoring,…
Descriptors: Item Analysis, Comparative Analysis, Culture Fair Tests, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Zijlmans, Eva A. O.; Tijmstra, Jesper; van der Ark, L. Andries; Sijtsma, Klaas – Educational and Psychological Measurement, 2018
Reliability is usually estimated for a total score, but it can also be estimated for item scores. Item-score reliability can be useful to assess the repeatability of an individual item score in a group. Three methods to estimate item-score reliability are discussed, known as method MS, method [lambda][subscript 6], and method CA. The item-score…
Descriptors: Test Items, Test Reliability, Correlation, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Stanley, Leanne M.; Edwards, Michael C. – Educational and Psychological Measurement, 2016
The purpose of this article is to highlight the distinction between the reliability of test scores and the fit of psychometric measurement models, reminding readers why it is important to consider both when evaluating whether test scores are valid for a proposed interpretation and/or use. It is often the case that an investigator judges both the…
Descriptors: Test Reliability, Goodness of Fit, Scores, Patients
Peer reviewed Peer reviewed
Direct linkDirect link
Zhang, Xijuan; Savalei, Victoria – Educational and Psychological Measurement, 2016
Many psychological scales written in the Likert format include reverse worded (RW) items in order to control acquiescence bias. However, studies have shown that RW items often contaminate the factor structure of the scale by creating one or more method factors. The present study examines an alternative scale format, called the Expanded format,…
Descriptors: Factor Structure, Psychological Testing, Alternative Assessment, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Keeley, Jared W.; English, Taylor; Irons, Jessica; Henslee, Amber M. – Educational and Psychological Measurement, 2013
Many measurement biases affect student evaluations of instruction (SEIs). However, two have been relatively understudied: halo effects and ceiling/floor effects. This study examined these effects in two ways. To examine the halo effect, using a videotaped lecture, we manipulated specific teacher behaviors to be "good" or "bad"…
Descriptors: Robustness (Statistics), Test Bias, Course Evaluation, Student Evaluation of Teacher Performance
Peer reviewed Peer reviewed
Direct linkDirect link
Erford, Bradley T.; Klein, Lauren – Educational and Psychological Measurement, 2007
The Slosson-Diagnostic Math Screener (S-DMS) was designed to help identify students in Grades 1 to 8 at risk for mathematics failure. Internal consistency, test-retest reliability, item analysis, decision efficiency, convergent validity, and factorial validity of all five levels of the S-DMS were studied using 20 independent samples of students…
Descriptors: Grade 1, Test Validity, Item Analysis, Test Reliability
Peer reviewed Peer reviewed
Kuncel, Ruth Boutin; Fiske, Donald W. – Educational and Psychological Measurement, 1974
Four hypotheses regarding stability of response process and response in personality testing are tested and supported. (RC)
Descriptors: College Students, Item Analysis, Personality Measures, Response Style (Tests)
Peer reviewed Peer reviewed
Head, Mary K.; And Others – Educational and Psychological Measurement, 1974
Presents details of construction and initial validation of a likert scale for assessing attitude toward six categories of school life with particular emphasis on school lunch. (Author/RC)
Descriptors: Elementary Secondary Education, Item Analysis, Lunch Programs, Rating Scales
Peer reviewed Peer reviewed
Pearce, W. Parnett; Wiebe, Bernie – Educational and Psychological Measurement, 1975
Determines the reliability and discrimination values of each item; whether items judged highly intimate by Jourard discriminate between high and low disclosers more than items judged lowly intimate; and sex differences in disclosure. (Author)
Descriptors: College Students, Item Analysis, Questionnaires, Self Concept
Peer reviewed Peer reviewed
Neill, John A.; Jackson, Douglas N. – Educational and Psychological Measurement, 1976
Illustrates a multivariate approach to item analysis. Previous formulation is extended by investigating techniques simultaneously taking into account scale variance with the goal of reducing the average correlation between scales. Study examines problems in determining optimum values for combinations of item parameters selected for personality…
Descriptors: Correlation, Factor Structure, Item Analysis, Personality Measures
Peer reviewed Peer reviewed
Carroll, C. Dennis – Educational and Psychological Measurement, 1976
A computer program for item evaluation, reliability estimation, and test scoring is described. The program contains a variable format procedure allowing flexible input of responses. Achievement tests and affective scales may be analyzed. (Author)
Descriptors: Achievement Tests, Affective Measures, Computer Programs, Item Analysis
Peer reviewed Peer reviewed
Abbott, Robert D.; Perkins, David – Educational and Psychological Measurement, 1978
The development and implementation in a psychology department of a set of student rating-of-instruction items was discussed. The results of item descriptive statistics, correlational, and principal component analysis supported the construct validity of the items. (Author)
Descriptors: College Faculty, Factor Analysis, Higher Education, Item Analysis
Peer reviewed Peer reviewed
Lovett, Hubert T. – Educational and Psychological Measurement, 1978
The validity of five methods of estimating the reliability of criterion-referenced tests was evaluated across nine conditions of variability among item means. The results were analyzed by analysis of variance, the Newman-Keuls test, and a nonparametric procedure. There was a tendency for all of the methods to be conservative. (Author/JKS)
Descriptors: Analysis of Variance, Criterion Referenced Tests, Item Analysis, Nonparametric Statistics
Peer reviewed Peer reviewed
Serlin, Ronald C.; Kaiser, Henry F. – Educational and Psychological Measurement, 1978
When multiple-choice tests are scored in the usual manner, giving each correct answer one point, information concerning response patterns is lost. A method for utilizing this information is suggested. An example is presented and compared with two conventional methods of scoring. (Author/JKS)
Descriptors: Correlation, Factor Analysis, Item Analysis, Multiple Choice Tests
Peer reviewed Peer reviewed
Andrulis, Richard S.; And Others – Educational and Psychological Measurement, 1978
The effects of repeaters (testees included in both administrations of two forms of a test) on the test equating process are examined. It is shown that repeaters do effect test equating and tend to lower the cutoff point for passing the test. (JKS)
Descriptors: Cutting Scores, Equated Scores, Item Analysis, Scoring
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4