NotesFAQContact Us
Collection
Advanced
Search Tips
Source
Educational and Psychological…37
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 37 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Hsiao, Yu-Yu; Kwok, Oi-Man; Lai, Mark H. C. – Educational and Psychological Measurement, 2018
Path models with observed composites based on multiple items (e.g., mean or sum score of the items) are commonly used to test interaction effects. Under this practice, researchers generally assume that the observed composites are measured without errors. In this study, we reviewed and evaluated two alternative methods within the structural…
Descriptors: Error of Measurement, Testing, Scores, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Andersson, Björn; Xin, Tao – Educational and Psychological Measurement, 2018
In applications of item response theory (IRT), an estimate of the reliability of the ability estimates or sum scores is often reported. However, analytical expressions for the standard errors of the estimators of the reliability coefficients are not available in the literature and therefore the variability associated with the estimated reliability…
Descriptors: Item Response Theory, Test Reliability, Test Items, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Luo, Yong; Jiao, Hong – Educational and Psychological Measurement, 2018
Stan is a new Bayesian statistical software program that implements the powerful and efficient Hamiltonian Monte Carlo (HMC) algorithm. To date there is not a source that systematically provides Stan code for various item response theory (IRT) models. This article provides Stan code for three representative IRT models, including the…
Descriptors: Bayesian Statistics, Item Response Theory, Probability, Computer Software
Casabianca, Jodi M.; Lockwood, J. R.; McCaffrey, Daniel F. – Educational and Psychological Measurement, 2015
Observations and ratings of classroom teaching and interactions collected over time are susceptible to trends in both the quality of instruction and rater behavior. These trends have potential implications for inferences about teaching and for study design. We use scores on the Classroom Assessment Scoring System-Secondary (CLASS-S) protocol from…
Descriptors: Scores, Middle School Teachers, Teacher Effectiveness, Teacher Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Jones, W. Paul – Educational and Psychological Measurement, 2014
A study in a university clinic/laboratory investigated adaptive Bayesian scaling as a supplement to interpretation of scores on the Mini-IPIP. A "probability of belonging" in categories of low, medium, or high on each of the Big Five traits was calculated after each item response and continued until all items had been used or until a…
Descriptors: Personality Traits, Personality Measures, Bayesian Statistics, Clinics
Peer reviewed Peer reviewed
Direct linkDirect link
Mashburn, Andrew J.; Meyer, J. Patrick; Allen, Joseph P.; Pianta, Robert C. – Educational and Psychological Measurement, 2014
Observational methods are increasingly being used in classrooms to evaluate the quality of teaching. Operational procedures for observing teachers are somewhat arbitrary in existing measures and vary across different instruments. To study the effect of different observation procedures on score reliability and validity, we conducted an experimental…
Descriptors: Observation, Teacher Evaluation, Reliability, Validity
Peer reviewed Peer reviewed
Direct linkDirect link
von Eye, Alexander; Wiedermann, Wolfgang – Educational and Psychological Measurement, 2014
Approaches to determining direction of dependence in nonexperimental data are based on the relation between higher-than second-order moments on one side and correlation and regression models on the other. These approaches have experienced rapid development and are being applied in contexts such as research on partner violence, attention deficit…
Descriptors: Statistical Analysis, Factor Analysis, Structural Equation Models, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Dowdy, Erin; Nylund-Gibson, Karen; Felix, Erika D.; Morovati, Diane; Carnazzo, Katherine W.; Dever, Bridget V. – Educational and Psychological Measurement, 2014
The practice of screening students to identify behavioral and emotional risk is gaining momentum, with limited guidance regarding the frequency with which screenings should occur. Screening frequency decisions are influenced by the stability of the constructs assessed and changes in risk status over time. This study investigated the 4-year…
Descriptors: Screening Tests, Risk, Behavior Disorders, Emotional Disturbances
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Pohl, Steffi – Educational and Psychological Measurement, 2013
A method for examining common factor variance in multiple-component measuring instruments is outlined. The procedure is based on an application of the latent variable modeling methodology and is concerned with evaluating observed variance explained by a global factor and by one or more additional component-specific factors. The approach furnishes…
Descriptors: Statistical Analysis, Factor Structure, Scores, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Moses, Tim; Kim, Sooyeon – Educational and Psychological Measurement, 2012
In this study, a ranking strategy was evaluated for comparing subgroups' change using identical, equated, and nonidentical measures. Four empirical data sets were evaluated, each of which contained examinees' scores on two occasions, where the two occasions' scores were obtained on a single identical measure, on two equated tests, and on two…
Descriptors: Testing, Change, Scores, Measures (Individuals)
Peer reviewed Peer reviewed
Direct linkDirect link
Thomas, D. Roland; Zumbo, Bruno D. – Educational and Psychological Measurement, 2012
There is such doubt in research practice about the reliability of difference scores that granting agencies, journal editors, reviewers, and committees of graduate students' theses have been known to deplore their use. This most maligned index can be used in studies of change, growth, or perhaps discrepancy between two measures taken on the same…
Descriptors: Statistical Analysis, Reliability, Scores, Change
Peer reviewed Peer reviewed
Direct linkDirect link
Yang, Ji Seung; Hansen, Mark; Cai, Li – Educational and Psychological Measurement, 2012
Traditional estimators of item response theory scale scores ignore uncertainty carried over from the item calibration process, which can lead to incorrect estimates of the standard errors of measurement (SEMs). Here, the authors review a variety of approaches that have been applied to this problem and compare them on the basis of their statistical…
Descriptors: Item Response Theory, Scores, Statistical Analysis, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Feldt, Leonard S. – Educational and Psychological Measurement, 2011
This article presents a simple, computer-assisted method of determining the extent to which increases in reliability increase the power of the "F" test of equality of means. The method uses a derived formula that relates the changes in the reliability coefficient to changes in the noncentrality of the relevant "F" distribution. A readily available…
Descriptors: Scores, Reliability, Statistical Analysis, Change
Peer reviewed Peer reviewed
Direct linkDirect link
Teo, Timothy – Educational and Psychological Measurement, 2010
This article describes the development, validation, and measurement invariance of scores from a survey designed to measure preservice teachers' reported acceptance of technology. Drawing from conceptual models in the areas of information systems and technology acceptance, a five-factor Technology Acceptance Measure for Preservice Teachers (TAMPST)…
Descriptors: Preservice Teachers, Computer Attitudes, Attitude Measures, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Hulpia, Hester; Devos, Geert; Rosseel, Yves – Educational and Psychological Measurement, 2009
Systematic quantitative research on measuring distributed leadership is scarce. In this study, the Distributed Leadership Inventory (DLI) was developed and evaluated to investigate leadership team characteristics and distribution of leadership functions between formally designed leadership positions in large secondary schools. The DLI was…
Descriptors: Factor Structure, Factor Analysis, Statistical Analysis, Leadership
Previous Page | Next Page »
Pages: 1  |  2  |  3