NotesFAQContact Us
Collection
Advanced
Search Tips
Source
Educational and Psychological…48
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 48 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
von Davier, Matthias; Tyack, Lillian; Khorramdel, Lale – Educational and Psychological Measurement, 2023
Automated scoring of free drawings or images as responses has yet to be used in large-scale assessments of student achievement. In this study, we propose artificial neural networks to classify these types of graphical responses from a TIMSS 2019 item. We are comparing classification accuracy of convolutional and feed-forward approaches. Our…
Descriptors: Scoring, Networks, Artificial Intelligence, Elementary Secondary Education
Peer reviewed Peer reviewed
Direct linkDirect link
Menold, Natalja; Raykov, Tenko – Educational and Psychological Measurement, 2022
The possible dependency of criterion validity on item formulation in a multicomponent measuring instrument is examined. The discussion is concerned with evaluation of the differences in criterion validity between two or more groups (populations/subpopulations) that have been administered instruments with items having differently formulated item…
Descriptors: Test Items, Measures (Individuals), Test Validity, Difficulty Level
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Xiaowen; Jane Rogers, H. – Educational and Psychological Measurement, 2022
Test fairness is critical to the validity of group comparisons involving gender, ethnicities, culture, or treatment conditions. Detection of differential item functioning (DIF) is one component of efforts to ensure test fairness. The current study compared four treatments for items that have been identified as showing DIF: deleting, ignoring,…
Descriptors: Item Analysis, Comparative Analysis, Culture Fair Tests, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A. – Educational and Psychological Measurement, 2019
This note discusses the merits of coefficient alpha and their conditions in light of recent critical publications that miss out on significant research findings over the past several decades. That earlier research has demonstrated the empirical relevance and utility of coefficient alpha under certain empirical circumstances. The article highlights…
Descriptors: Test Validity, Test Reliability, Test Items, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Stanley, Leanne M.; Edwards, Michael C. – Educational and Psychological Measurement, 2016
The purpose of this article is to highlight the distinction between the reliability of test scores and the fit of psychometric measurement models, reminding readers why it is important to consider both when evaluating whether test scores are valid for a proposed interpretation and/or use. It is often the case that an investigator judges both the…
Descriptors: Test Reliability, Goodness of Fit, Scores, Patients
Peer reviewed Peer reviewed
Direct linkDirect link
Zhang, Xijuan; Savalei, Victoria – Educational and Psychological Measurement, 2016
Many psychological scales written in the Likert format include reverse worded (RW) items in order to control acquiescence bias. However, studies have shown that RW items often contaminate the factor structure of the scale by creating one or more method factors. The present study examines an alternative scale format, called the Expanded format,…
Descriptors: Factor Structure, Psychological Testing, Alternative Assessment, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Sideridis, Georgios D. – Educational and Psychological Measurement, 2016
The purpose of the present studies was to test the hypothesis that the psychometric characteristics of ability scales may be significantly distorted if one accounts for emotional factors during test taking. Specifically, the present studies evaluate the effects of anxiety and motivation on the item difficulties of the Rasch model. In Study 1, the…
Descriptors: Learning Disabilities, Test Validity, Measures (Individuals), Hierarchical Linear Modeling
Peer reviewed Peer reviewed
Direct linkDirect link
Yoo, Jin Eun – Educational and Psychological Measurement, 2009
This Monte Carlo study investigates the beneficiary effect of including auxiliary variables during estimation of confirmatory factor analysis models with multiple imputation. Specifically, it examines the influence of sample size, missing rates, missingness mechanism combinations, missingness types (linear or convex), and the absence or presence…
Descriptors: Monte Carlo Methods, Research Methodology, Test Validity, Factor Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Bing, Mark N.; Stewart, Susan M.; Davison, H. Kristl – Educational and Psychological Measurement, 2009
Handheld calculators have been used on the job for more than 30 years, yet the degree to which these devices can affect performance on employment tests of mathematical ability has not been thoroughly examined. This study used a within-subjects research design (N = 167) to investigate the effects of calculator use on test score reliability, test…
Descriptors: Calculators, Mathematics Tests, Occupational Tests, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Young-Sun; Grossman, Jennifer; Krishnan, Anita – Educational and Psychological Measurement, 2008
This study examined the cultural relevance of adult attachment within a Korean sample (N = 390) using Rasch rating scale modeling. The psychometric properties of scores from the Korean version of the Revised Experiences in Close Relationships, comprised of two subscales of Anxiety (self) and Avoidance (other), were assessed. Results obtained from…
Descriptors: Cultural Relevance, Attachment Behavior, Rating Scales, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Worley, Jody A.; Vassar, Matt; Wheeler, Denna L.; Barnes, Laura L. B. – Educational and Psychological Measurement, 2008
This study provides a summary of 45 exploratory and confirmatory factor-analytic studies that examined the internal structure of scores obtained from the Maslach Burnout Inventory (MBI). It highlights characteristics of the studies that account for differences in reporting of the MBI factor structure. This approach includes an examination of the…
Descriptors: Burnout, Factor Structure, Meta Analysis, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Yi-Hsuan; Ip, Edward H.; Fuh, Cheng-Der – Educational and Psychological Measurement, 2008
Although computerized adaptive tests have enjoyed tremendous growth, solutions for important problems remain unavailable. One problem is the control of item exposure rate. Because adaptive algorithms are designed to select optimal items, they choose items with high discriminating power. Thus, these items are selected more often than others,…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Items, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Fortunato, Vincent J.; LeBourgeois, Monique K.; Harsh, John – Educational and Psychological Measurement, 2008
This article describes the development of a measure of adult sleep quality: the Adult Sleep-Wake Scale (ADSWS). The ADSWS is a self-report pencil-and-paper measure of sleep quality consisting of five behavioral dimensions (Going to Bed, Falling Asleep, Maintaining Sleep, Reinitiating Sleep, and Returning to Wakefulness). Data were collected from…
Descriptors: Construct Validity, Test Validity, Sleep, Personality Traits
Peer reviewed Peer reviewed
Direct linkDirect link
Kong, Xiaojing J.; Wise, Steven L.; Bhola, Dennison S. – Educational and Psychological Measurement, 2007
This study compared four methods for setting item response time thresholds to differentiate rapid-guessing behavior from solution behavior. Thresholds were either (a) common for all test items, (b) based on item surface features such as the amount of reading required, (c) based on visually inspecting response time frequency distributions, or (d)…
Descriptors: Test Items, Reaction Time, Timed Tests, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Feldt, Leonard S.; Kim, Seonghoon – Educational and Psychological Measurement, 2006
Researchers sometimes need a statistical test of the hypothesis that two values of Cronbach's alpha reliability coefficient are equal. The situation may involve scores from two different measures administered to independent random samples or from the same measure administered to random samples from two different populations. Feldt derived a test…
Descriptors: Individual Testing, Test Items, Sample Size, Scores
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4