NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing 106 to 120 of 3,486 results
Peer reviewed Peer reviewed
Direct linkDirect link
Kobrin, Jennifer L.; Kim, YoungKoung; Sackett, Paul R. – Educational and Psychological Measurement, 2012
There is much debate on the merits and pitfalls of standardized tests for college admission, with questions regarding the format (multiple-choice vs. constructed response), cognitive complexity, and content of these assessments (achievement vs. aptitude) at the forefront of the discussion. This study addressed these questions by investigating the…
Descriptors: Grade Point Average, Standardized Tests, Predictive Validity, Predictor Variables
Peer reviewed Peer reviewed
Direct linkDirect link
Moses, Tim; Kim, Sooyeon – Educational and Psychological Measurement, 2012
In this study, a ranking strategy was evaluated for comparing subgroups' change using identical, equated, and nonidentical measures. Four empirical data sets were evaluated, each of which contained examinees' scores on two occasions, where the two occasions' scores were obtained on a single identical measure, on two equated tests, and on two…
Descriptors: Testing, Change, Scores, Measures (Individuals)
Peer reviewed Peer reviewed
Direct linkDirect link
Finch, W. Holmes; French, Brian F. – Educational and Psychological Measurement, 2012
Effect size use has been increasing in the past decade in many research areas. Confidence intervals associated with effect sizes are encouraged to be reported. Prior work has investigated the performance of confidence interval estimation with Cohen's d. This study extends this line of work to the analysis of variance case with more than two…
Descriptors: Computation, Statistical Analysis, Effect Size, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Santelices, Maria Veronica; Wilson, Mark – Educational and Psychological Measurement, 2012
The relationship between differential item functioning (DIF) and item difficulty on the SAT is such that more difficult items tended to exhibit DIF in favor of the focal group (usually minority groups). These results were reported by Kulick and Hu, and Freedle and have been enthusiastically discussed by more recent literature. Examining the…
Descriptors: Test Bias, Test Items, Difficulty Level, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Patarapichayatham, Chalie; Kamata, Akihito; Kanjanawasee, Sirichai – Educational and Psychological Measurement, 2012
Model specification issues on the cross-level two-way differential item functioning model were previously investigated by Patarapichayatham et al. (2009). Their study clarified that an incorrect model specification can easily lead to biased estimates of key parameters. The objective of this article is to provide further insights on the issue by…
Descriptors: Test Bias, Models, Bayesian Statistics, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Thomas, D. Roland; Zumbo, Bruno D. – Educational and Psychological Measurement, 2012
There is such doubt in research practice about the reliability of difference scores that granting agencies, journal editors, reviewers, and committees of graduate students' theses have been known to deplore their use. This most maligned index can be used in studies of change, growth, or perhaps discrepancy between two measures taken on the same…
Descriptors: Statistical Analysis, Reliability, Scores, Change
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, HwaYoung; Dodd, Barbara G. – Educational and Psychological Measurement, 2012
This study investigated item exposure control procedures under various combinations of item pool characteristics and ability distributions in computerized adaptive testing based on the partial credit model. Three variables were manipulated: item pool characteristics (120 items for each of easy, medium, and hard item pools), two ability…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Ability
Peer reviewed Peer reviewed
Direct linkDirect link
Lakin, Joni M.; Lai, Emily R. – Educational and Psychological Measurement, 2012
For educators seeking to differentiate instruction, cognitive ability tests sampling multiple content domains, including verbal, quantitative, and nonverbal reasoning, provide superior information about student strengths and weaknesses compared with unidimensional reasoning measures. However, these ability tests have not been fully evaluated with…
Descriptors: Aptitude Tests, Nonverbal Ability, Cognitive Ability, Verbal Ability
Peer reviewed Peer reviewed
Direct linkDirect link
Dimitrov, Dimiter M.; Atanasov, Dimitar V. – Educational and Psychological Measurement, 2012
Many models of cognitive diagnosis, including the "least squares distance model" (LSDM), work under the "conjunctive" assumption that a correct item response occurs when all latent attributes required by the item are correctly performed. This article proposes a "disjunctive" version of the LSDM under which the correct item response occurs when "at…
Descriptors: Least Squares Statistics, Models, Item Response Theory, Cognitive Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Cheng, Ying; Yuan, Ke-Hai; Liu, Cheng – Educational and Psychological Measurement, 2012
Reliability of test scores is one of the most pervasive psychometric concepts in measurement. Reliability coefficients based on a unifactor model for continuous indicators include maximal reliability rho and an unweighted sum score-based omega, among many others. With increasing popularity of item response theory, a parallel reliability measure pi…
Descriptors: Reliability, Factor Analysis, Psychometrics, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Green, Samuel B.; Levy, Roy; Thompson, Marilyn S.; Lu, Min; Lo, Wen-Juo – Educational and Psychological Measurement, 2012
A number of psychometricians have argued for the use of parallel analysis to determine the number of factors. However, parallel analysis must be viewed at best as a heuristic approach rather than a mathematically rigorous one. The authors suggest a revision to parallel analysis that could improve its accuracy. A Monte Carlo study is conducted to…
Descriptors: Monte Carlo Methods, Factor Structure, Data Analysis, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Yan; Zumbo, Bruno D. – Educational and Psychological Measurement, 2012
There is a lack of research on the effects of outliers on the decisions about the number of factors to retain in an exploratory factor analysis, especially for outliers arising from unintended and unknowingly included subpopulations. The purpose of the present research was to investigate how outliers from an unintended and unknowingly included…
Descriptors: Factor Analysis, Factor Structure, Evaluation Research, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Debelak, Rudolf; Arendasy, Martin – Educational and Psychological Measurement, 2012
A new approach to identify item clusters fitting the Rasch model is described and evaluated using simulated and real data. The proposed method is based on hierarchical cluster analysis and constructs clusters of items that show a good fit to the Rasch model. It thus gives an estimate of the number of independent scales satisfying the postulates of…
Descriptors: Test Items, Factor Analysis, Evaluation Methods, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Walker, Cindy M.; Zhang, Bo; Banks, Kathleen; Cappaert, Kevin – Educational and Psychological Measurement, 2012
The purpose of this simulation study was to establish general effect size guidelines for interpreting the results of differential bundle functioning (DBF) analyses using simultaneous item bias test (SIBTEST). Three factors were manipulated: number of items in a bundle, test length, and magnitude of uniform differential item functioning (DIF)…
Descriptors: Test Bias, Test Length, Simulation, Guidelines
Peer reviewed Peer reviewed
Direct linkDirect link
Balogh, Jennifer; Bernstein, Jared; Cheng, Jian; Van Moere, Alistair; Townshend, Brent; Suzuki, Masanori – Educational and Psychological Measurement, 2012
A two-part experiment is presented that validates a new measurement tool for scoring oral reading ability. Data collected by the U.S. government in a large-scale literacy assessment of adults were analyzed by a system called VersaReader that uses automatic speech recognition and speech processing technologies to score oral reading fluency. In the…
Descriptors: Reading Fluency, Measures (Individuals), Scoring, Reading Ability
Pages: 1  |  ...  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  12  |  ...  |  233