NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing 1 to 15 of 86 results
Peer reviewed Peer reviewed
Direct linkDirect link
Häggström, Jenny; Wiberg, Marie – Journal of Educational Measurement, 2014
The selection of bandwidth in kernel equating is important because it has a direct impact on the equated test scores. The aim of this article is to examine the use of double smoothing when selecting bandwidths in kernel equating and to compare double smoothing with the commonly used penalty method. This comparison was made using both an equivalent…
Descriptors: Equated Scores, Data Analysis, Comparative Analysis, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Albano, Anthony D. – Journal of Educational Measurement, 2013
In many testing programs it is assumed that the context or position in which an item is administered does not have a differential effect on examinee responses to the item. Violations of this assumption may bias item response theory estimates of item and person parameters. This study examines the potentially biasing effects of item position. A…
Descriptors: Test Items, Item Response Theory, Test Format, Questioning Techniques
Peer reviewed Peer reviewed
Direct linkDirect link
Belov, Dmitry I. – Journal of Educational Measurement, 2013
The development of statistical methods for detecting test collusion is a new research direction in the area of test security. Test collusion may be described as large-scale sharing of test materials, including answers to test items. Current methods of detecting test collusion are based on statistics also used in answer-copying detection.…
Descriptors: Cheating, Computer Assisted Testing, Adaptive Testing, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Wen-Chung; Jin, Kuan-Yu; Qiu, Xue-Lan; Wang, Lei – Journal of Educational Measurement, 2012
In some tests, examinees are required to choose a fixed number of items from a set of given items to answer. This practice creates a challenge to standard item response models, because more capable examinees may have an advantage by making wiser choices. In this study, we developed a new class of item response models to account for the choice…
Descriptors: Item Response Theory, Test Items, Selection, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Zwick, Rebecca; Himelfarb, Igor – Journal of Educational Measurement, 2011
Research has often found that, when high school grades and SAT scores are used to predict first-year college grade-point average (FGPA) via regression analysis, African-American and Latino students, are, on average, predicted to earn higher FGPAs than they actually do. Under various plausible models, this phenomenon can be explained in terms of…
Descriptors: Socioeconomic Status, Grades (Scholastic), Error of Measurement, White Students
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Changjiang; Gierl, Mark J. – Journal of Educational Measurement, 2011
The purpose of this study is to apply the attribute hierarchy method (AHM) to a subset of SAT critical reading items and illustrate how the method can be used to promote cognitive diagnostic inferences. The AHM is a psychometric procedure for classifying examinees' test item responses into a set of attribute mastery patterns associated with…
Descriptors: Reading Comprehension, Test Items, Critical Reading, Protocol Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Jinghua; Sinharay, Sandip; Holland, Paul W.; Curley, Edward; Feigenbaum, Miriam – Journal of Educational Measurement, 2011
This study explores an anchor that is different from the traditional miniature anchor in test score equating. In contrast to a traditional "mini" anchor that has the same spread of item difficulties as the tests to be equated, the studied anchor, referred to as a "midi" anchor (Sinharay & Holland), has a smaller spread of item difficulties than…
Descriptors: Equated Scores, Case Studies, College Entrance Examinations, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Culpepper, Steven A.; Davenport, Ernest C. – Journal of Educational Measurement, 2009
Previous research notes the importance of understanding racial/ethnic differential prediction of college grades across multiple institutions. Institutional variation in selection indices is especially important given some states' laws governing public institutions' admissions decisions. This paper employed multilevel moderated multiple regression…
Descriptors: Prediction, College Students, Grades (Scholastic), Race
Peer reviewed Peer reviewed
Direct linkDirect link
Zwick, Rebecca; Greif Green, Jennifer – Journal of Educational Measurement, 2007
In studies of the SAT, correlations of SAT scores, high school grades, and socioeconomic factors (SES) are usually obtained using a university as the unit of analysis. This approach obscures an important structural aspect of the data: The high school grades received by a given institution come from a large number of high schools, all of which have…
Descriptors: Organizations (Groups), High School Students, Grades (Scholastic), Grading
Peer reviewed Peer reviewed
Direct linkDirect link
Ariel, Adelaide; van der Linden, Wim J.; Veldkamp, Bernard P. – Journal of Educational Measurement, 2006
Item-pool management requires a balancing act between the input of new items into the pool and the output of tests assembled from it. A strategy for optimizing item-pool management is presented that is based on the idea of a periodic update of an optimal blueprint for the item pool to tune item production to test assembly. A simulation study with…
Descriptors: Item Banks, Simulation, Interaction, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Jinghua; Cahn, Miriam F.; Dorans, Neil J. – Journal of Educational Measurement, 2006
The College Board's SAT[R] data are used to illustrate how the score equity assessment (SEA) can help inform the program about equatability. SEA is used to examine whether the content change(s) to the revised new SAT result in differential linking functions across gender groups. Results of population sensitivity analyses are reported on the…
Descriptors: Aptitude Tests, Comparative Analysis, Gender Differences, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Finch, Holmes; Habing, Brian – Journal of Educational Measurement, 2005
This study examines the performance of a new method for assessing and characterizing dimensionality in test data using the NOHARM model, and comparing it with DETECT. Dimensionality assessment is carried out using two goodness-of-fit statistics that are compared to reference X[2] distributions. A Monte Carlo study is used with item parameters…
Descriptors: Program Effectiveness, Monte Carlo Methods, Item Response Theory, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Bridgeman, Brent; Trapani, Catherine; Curley, Edward – Journal of Educational Measurement, 2004
The impact of allowing more time for each question on the SAT I: Reasoning Test scores was estimated by embedding sections with a reduced number of questions into the standard 30-minute equating section of two national test administrations. Thus, for example, questions were deleted from a verbal section that contained 35 questions to produce forms…
Descriptors: College Entrance Examinations, Test Length, Scores
Peer reviewed Peer reviewed
Gallagher, Ann; Bridgeman, Brent; Cahalan, Cara – Journal of Educational Measurement, 2002
Examined data from several national testing programs to determine whether the change from paper-based administration to computer-based tests influences group differences in performance. Results from four college and graduate entrance examinations and a professional licensing test show that African Americans and, to a lesser degree, Hispanics,…
Descriptors: Blacks, College Entrance Examinations, Computer Assisted Testing, Ethnic Groups
Peer reviewed Peer reviewed
Dorans, Neil J. – Journal of Educational Measurement, 2002
Describes the process used to produce the conversions that take scores from the original Scholastic Assessment Test (SAT) scales to recentered scales in which the reference group scores are centered near the midpoint of the score-reporting range. Also describes the performance of the 1990 reference group and discusses issues related to…
Descriptors: College Entrance Examinations, High School Students, High Schools, Scores
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6