NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing 16 to 30 of 278 results
Peer reviewed Peer reviewed
Direct linkDirect link
Ferrett, Helen L.; Carey, Paul D.; Baufeldt, Angela L.; Cuzen, Natalie L.; Conradie, Simone; Dowling, Tessa; Stein, Dan J.; Thomas, Kevin G. F. – International Journal of Testing, 2014
Because of their global clinical utility, phonemic fluency tests are frequently incorporated into neuropsychological assessment batteries. However, in heterogeneous societies their use is complicated by the lack of careful attention to using letters of equivalent difficulty across languages, and the paucity of norms stratified by relevant…
Descriptors: Foreign Countries, Phonemes, Language Fluency, Alphabets
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, HyeSun; Geisinger, Kurt F. – International Journal of Testing, 2014
Differential item functioning (DIF) analysis is important in terms of test fairness. While DIF analyses have mainly been conducted with manifest grouping variables, such as gender or race/ethnicity, it has been recently claimed that not only the grouping variables but also contextual variables pertaining to examinees should be considered in DIF…
Descriptors: Test Bias, Gender Differences, Regression (Statistics), Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Zilberberg, Anna; Finney, Sara J.; Marsh, Kimberly R.; Anderson, Robin D. – International Journal of Testing, 2014
Given worldwide prevalence of low-stakes testing for monitoring educational quality and students' progress through school (e.g., Trends in International Mathematics and Science Study, Program for International Student Assessment), interpretability of resulting test scores is of global concern. The nonconsequential nature of low-stakes tests…
Descriptors: Student Attitudes, Student Motivation, Test Validity, Accountability
Peer reviewed Peer reviewed
Direct linkDirect link
Rios, Joseph A.; Sireci, Stephen G. – International Journal of Testing, 2014
The International Test Commission's "Guidelines for Translating and Adapting Tests" (2010) provide important guidance on developing and evaluating tests for use across languages. These guidelines are widely applauded, but the degree to which they are followed in practice is unknown. The objective of this study was to perform a…
Descriptors: Guidelines, Translation, Adaptive Testing, Second Languages
Peer reviewed Peer reviewed
Direct linkDirect link
Engelhard, George, Jr.; Kobrin, Jennifer L.; Wind, Stefanie A. – International Journal of Testing, 2014
The purpose of this study is to explore patterns in model-data fit related to subgroups of test takers from a large-scale writing assessment. Using data from the SAT, a calibration group was randomly selected to represent test takers who reported that English was their best language from the total population of test takers (N = 322,011). A…
Descriptors: College Entrance Examinations, Writing Tests, Goodness of Fit, English
Peer reviewed Peer reviewed
Direct linkDirect link
Hess, Brian J.; Johnston, Mary M.; Lipner, Rebecca S. – International Journal of Testing, 2013
Current research on examination response time has focused on tests comprised of traditional multiple-choice items. Consequently, the impact of other innovative or complex item formats on examinee response time is not understood. The present study used multilevel growth modeling to investigate examinee characteristics associated with response time…
Descriptors: Test Items, Test Format, Reaction Time, Individual Characteristics
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Sooyeon; Moses, Tim – International Journal of Testing, 2013
The major purpose of this study is to assess the conditions under which single scoring for constructed-response (CR) items is as effective as double scoring in the licensure testing context. We used both empirical datasets of five mixed-format licensure tests collected in actual operational settings and simulated datasets that allowed for the…
Descriptors: Scoring, Test Format, Licensing Examinations (Professions), Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
In'nami, Yo; Koizumi, Rie – International Journal of Testing, 2013
The importance of sample size, although widely discussed in the literature on structural equation modeling (SEM), has not been widely recognized among applied SEM researchers. To narrow this gap, we focus on second language testing and learning studies and examine the following: (a) Is the sample size sufficient in terms of precision and power of…
Descriptors: Structural Equation Models, Sample Size, Second Language Instruction, Monte Carlo Methods
Peer reviewed Peer reviewed
Direct linkDirect link
DeMars, Christine E. – International Journal of Testing, 2013
This tutorial addresses possible sources of confusion in interpreting trait scores from the bifactor model. The bifactor model may be used when subscores are desired, either for formative feedback on an achievement test or for theoretically different constructs on a psychological test. The bifactor model is often chosen because it requires fewer…
Descriptors: Test Interpretation, Scores, Models, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Fine, Saul – International Journal of Testing, 2013
While psychological tests are used extensively in Israel, the current controls over testing practices in Israel deserve some attention. Specifically, unlike in some European countries and the United States, (a) no specific certifications are offered to Israeli psychologists in the area of testing; (b) Israeli psychologists are not obligated to…
Descriptors: Foreign Countries, Psychological Testing, Psychologists, Attitudes
Peer reviewed Peer reviewed
Direct linkDirect link
Pacico, Juliana Cerentini; Zanon, Cristian; Bastianello, Micheline Roat; Reppold, Caroline Tozzi; Hutz, Claudio Simon – International Journal of Testing, 2013
The objective of this study was to adapt and gather validity evidence for a Brazilian sample version of the Hope Index and to verify if cultural differences would produce different results than those found in the United States. In this study, we present a set of analyses that together comprise a comprehensive validity argument for the use of a…
Descriptors: Foreign Countries, Cognitive Tests, Content Validity, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Tay, Louis; Vermunt, Jeroen K.; Wang, Chun – International Journal of Testing, 2013
We evaluate the item response theory with covariates (IRT-C) procedure for assessing differential item functioning (DIF) without preknowledge of anchor items (Tay, Newman, & Vermunt, 2011). This procedure begins with a fully constrained baseline model, and candidate items are tested for uniform and/or nonuniform DIF using the Wald statistic.…
Descriptors: Item Response Theory, Test Bias, Models, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Oliveri, Maria Elena; Ercikan, Kadriye; Zumbo, Bruno – International Journal of Testing, 2013
In this study, we investigated differential item functioning (DIF) and its sources using a latent class (LC) modeling approach. Potential sources of LC DIF related to instruction and teacher-related variables were investigated using substantive and three statistical approaches: descriptive discriminant function, multinomial logistic regression,…
Descriptors: Test Bias, Test Items, Multivariate Analysis, Discriminant Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Kruyen, Peter M.; Emons, Wilco H. M.; Sijtsma, Klaas – International Journal of Testing, 2013
To efficiently assess multiple psychological constructs and to minimize the burden on respondents, psychologists increasingly use shortened versions of existing tests. However, compared to the longer test, a shorter test version may have a substantial impact on the reliability and the validity of the test scores in psychological research and…
Descriptors: Test Length, Psychological Testing, Test Use, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Roivainen, Eka – International Journal of Testing, 2013
To study the concept of national IQ profile, we compared U.S. and Finnish WAIS, WAIS-R, and WAIS III nonverbal and working memory subtest norms. The U.S. standardization samples had consistently higher scores on the Coding and Digit span subtests, while the Finnish samples had higher scores on the Block design subtest. No stable cross-national…
Descriptors: Intelligence Tests, Profiles, Cultural Influences, Nonverbal Tests
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  19