NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing 61 to 75 of 278 results
Peer reviewed Peer reviewed
Direct linkDirect link
Zhang, Mo; Williamson, David M.; Breyer, F. Jay; Trapani, Catherine – International Journal of Testing, 2012
This article describes two separate, related studies that provide insight into the effectiveness of "e-rater" score calibration methods based on different distributional targets. In the first study, we developed and evaluated a new type of "e-rater" scoring model that was cost-effective and applicable under conditions of absent human rating and…
Descriptors: Automation, Scoring, Models, Essay Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Young-Sun; Park, Yoon Soo; Taylan, Didem – International Journal of Testing, 2011
Studies of international mathematics achievement such as the Trends in Mathematics and Science Study (TIMSS) have employed classical test theory and item response theory to rank individuals within a latent ability continuum. Although these approaches have provided insights into comparisons between countries, they have yet to examine how specific…
Descriptors: Mathematics Achievement, Achievement Tests, Models, Cognitive Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Hopfenbeck, Therese N.; Maul, Andrew – International Journal of Testing, 2011
The aim of this study was to investigate response-process based evidence for the validity of the Programme for International Student Assessment's (PISA) self-report questionnaire scales as measures of specific psychological constructs, with a focus on scales meant to measure inclination toward specific learning strategies. Cognitive interviews (N…
Descriptors: Student Reaction, Learning Strategies, Validity, Questionnaires
Peer reviewed Peer reviewed
Direct linkDirect link
Xu, Lihua; Barnes, Laura L. B. – International Journal of Testing, 2011
Measurement invariance of the 8-factor Inventory of School Motivation (McInerney & Sinclair, 1991) between American and Chinese college students was tested using single-group and multi-group confirmatory factor analysis. A Mandarin Chinese version of the ISM was developed for this study. Comparisons of latent means were conducted when warranted by…
Descriptors: College Students, Factor Analysis, Positive Reinforcement, Mandarin Chinese
Peer reviewed Peer reviewed
Direct linkDirect link
Klinger, Don A.; Rogers, W. Todd – International Journal of Testing, 2011
The intent of this study was to examine the views of teachers regarding the appropriateness of the purposes and uses of the provincial assessments in Alberta and Ontario and the seriousness of the concerns raised about these assessments. These provinces represent educational jurisdictions that use large-scale assessments within a low-stakes…
Descriptors: Testing Programs, Educational Improvement, Measures (Individuals), Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Puhan, Gautam – International Journal of Testing, 2011
This study examined the effect of including or excluding repeaters on the equating process and results. New forms of two tests were equated to their respective old forms using either all examinees or only the first timer examinees in the new form sample. Results showed that for both tests used in this study, including or excluding repeaters in the…
Descriptors: Equated Scores, Educational Testing, Student Evaluation, Sample Size
Peer reviewed Peer reviewed
Direct linkDirect link
Stark, Stephen; Chernyshenko, Oleksandr S. – International Journal of Testing, 2011
This article delves into a relatively unexplored area of measurement by focusing on adaptive testing with unidimensional pairwise preference items. The use of such tests is becoming more common in applied non-cognitive assessment because research suggests that this format may help to reduce certain types of rater error and response sets commonly…
Descriptors: Test Length, Simulation, Adaptive Testing, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Brown, Allison R.; Finney, Sara J. – International Journal of Testing, 2011
The current study examined whether psychological reactance differs across compliant and non-compliant examinees. Given the lack of consensus regarding the factor structure and scoring of the Hong Psychological Reactance Scale (HPRS), its factor structure was evaluated and subsequently tested for measurement invariance (configural, metric, and…
Descriptors: Testing, Factor Structure, Measures (Individuals), Compliance (Psychology)
Peer reviewed Peer reviewed
Direct linkDirect link
Ong, Yoke Mooi; Williams, Julian Scott; Lamprianou, Iasonas – International Journal of Testing, 2011
The aims of this study are (a) to examine the sources of differential functioning by gender via differential bundle functioning (DBF) in mathematics assessment and (b) to use DBF to explore whether the differential functioning displayed is construct-relevant or construct-irrelevant. Three qualitatively different areas, namely curriculum domains,…
Descriptors: Test Bias, Gender Differences, Gender Bias, Mathematics Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Lyren, Per-Erik; Hambleton, Ronald K. – International Journal of Testing, 2011
The equal ability distribution assumption associated with the equivalent groups equating design was investigated in the context of a selection test for admission to higher education. The purpose was to assess the consequences for the test-takers in terms of receiving improperly high or low scores compared to their peers, and to find strong…
Descriptors: Evidence, Test Items, Ability Grouping, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Viglione, Donald J.; Perry, William; Giromini, Luciano; Meyer, Gregory J. – International Journal of Testing, 2011
We used multiple regression to calculate a new Ego Impairment Index (EII-3). The aim was to incorporate changes in the component variables and distribution of the number of responses as found in the new Rorschach Performance Assessment System, while sustaining the validity and reliability of previous EIIs. The EII-3 formula was derived from a…
Descriptors: Test Items, Self Concept, Validity, Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
D'Agostino, Jerome; Karpinski, Aryn; Welsh, Megan – International Journal of Testing, 2011
After a test is developed, most content validation analyses shift from ascertaining domain definition to studying domain representation and relevance because the domain is assumed to be set once a test exists. We present an approach that allows for the examination of alternative domain structures based on extant test items. In our example based on…
Descriptors: Expertise, Test Items, Mathematics Tests, Factor Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Magis, David; Raiche, Gilles; Beland, Sebastien; Gerard, Paul – International Journal of Testing, 2011
We present an extension of the logistic regression procedure to identify dichotomous differential item functioning (DIF) in the presence of more than two groups of respondents. Starting from the usual framework of a single focal group, we propose a general approach to estimate the item response functions in each group and to test for the presence…
Descriptors: Language Skills, Identification, Foreign Countries, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Xie, Qin – International Journal of Testing, 2011
This study examined test takers' perception of assessment demand and its impact on the measurement of intended constructs. More than 800 test takers took a pre- and a posttest of College English Test Band 4 and filled in a perception questionnaire to report the skills they perceive as necessary for answering the test. The study found test takers…
Descriptors: College English, Reading Tests, Essay Tests, Academic Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Davis-Becker, Susan L.; Buckendahl, Chad W.; Gerrow, Jack – International Journal of Testing, 2011
Throughout the world, cut scores are an important aspect of a high-stakes testing program because they are a key operational component of the interpretation of test scores. One method for setting standards that is prevalent in educational testing programs--the Bookmark method--is intended to be a less cognitively complex alternative to methods…
Descriptors: Standard Setting (Scoring), Cutting Scores, Educational Testing, Licensing Examinations (Professions)
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  19