NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing 1 to 15 of 135 results
Peer reviewed Peer reviewed
Direct linkDirect link
Meyer, J. Patrick; Setzer, J. Carl – Journal of Educational Measurement, 2009
Recent changes to federal guidelines for the collection of data on race and ethnicity allow respondents to select multiple race categories. Redefining race subgroups in this manner poses problems for research spanning both sets of definitions. NAEP long-term trends have used the single-race subgroup definitions for over thirty years. Little is…
Descriptors: Elementary Secondary Education, Federal Legislation, Simulation, Maximum Likelihood Statistics
Peer reviewed Peer reviewed
Smits, Niels; Mellenbergh, Gideon J.; Vorst, Harrie C. M. – Journal of Educational Measurement, 2002
Studied the use of grade point average as a missing data technique for unavailable grades in school grade records. Results of 2 studies, 1 involving 2,080 Dutch college students, and the other involving 446 Dutch college freshmen, show that alternative missing grade methods produced better and more consistent predictions than other missing data…
Descriptors: College Students, Foreign Countries, Grade Point Average, Grades (Scholastic)
Peer reviewed Peer reviewed
Stocking, Martha L.; Lawrence, Ida; Feigenbaum, Miriam; Jirele, Thomas; Lewis, Charles; Van Essen, Thomas – Journal of Educational Measurement, 2002
Constructed four different kinds of test sections using three methods of test assembly that incorporate the goals of simultaneous moderation of the impact of gender, African American status, and Hispanic-American status, resulting in 10 test forms completed by at least 7,000 test takers per form. Discusses the effects of moderating impact in this…
Descriptors: Black Students, Higher Education, Hispanic American Students, Sex Differences
Peer reviewed Peer reviewed
Vispoel, Walter P.; Hendrickson, Amy B.; Bleiler, Timothy – Journal of Educational Measurement, 2000
Evaluated the effectiveness of vocabulary computerized adaptive tests (CATs) with restricted review in a live testing setting involving 242 college students in which special efforts were made to increase test efficiency and reduce the possibility of obtaining positively biased proficiency estimates. Results suggest the efficacy of allowing limited…
Descriptors: Adaptive Testing, Attitudes, College Students, Computer Assisted Testing
Peer reviewed Peer reviewed
Allalouf, Avi; Hambleton, Ronald K.; Sireci, Stephen G. – Journal of Educational Measurement, 1999
Focused on whether differential item functioning (DIF) is related to item type in translated test items and the causes of DIF using data from an Israeli college entrance test in Hebrew and a Russian translation. Results from 24,304 college applicants indicate that 34% of items functioned differently across items. (SLD)
Descriptors: College Applicants, College Entrance Examinations, Foreign Countries, Hebrew
Peer reviewed Peer reviewed
Bennett, Randy Elliot; Morley, Mary; Quardt, Dennis; Rock, Donald A.; Singley, Mark K.; Katz, Irvin R.; Nhouyvanisvong, Adisack – Journal of Educational Measurement, 1999
Evaluated a computer-delivered response type for measuring quantitative skill, the "Generating Examples" (GE) response type, which presents under-determined problems that can have many right answers. Results from 257 graduate students and applicants indicate that GE scores are reasonably reliable, but only moderately related to Graduate Record…
Descriptors: College Applicants, Computer Assisted Testing, Graduate Students, Graduate Study
Peer reviewed Peer reviewed
Clauser, Brian E.; Clyman, Stephen G.; Swanson, David B. – Journal of Educational Measurement, 1999
Two studies focused on aspects of the rating process in performance assessment. The first, which involved 15 raters and about 400 medical students, made the "committee" facet of raters working in groups explicit, and the second, which involved about 200 medical students and four raters, made the "rating-occasion" facet explicit. (SLD)
Descriptors: Error Patterns, Evaluation Methods, Evaluators, Higher Education
Peer reviewed Peer reviewed
Vispoel, Walter P.; Rocklin, Thomas R.; Wang, Tianyou; Bleiler, Timothy – Journal of Educational Measurement, 1999
Investigated the effectiveness of H. Wainer's (1993) strategy for obtaining positively biased ability estimates when examinees can review and change answers on computerized adaptive tests. Results, based on simulation and testing data from 87 college students, show that the Wainer strategy sometimes yields inflated ability estimates and sometimes…
Descriptors: Ability, College Students, Computer Assisted Testing, Higher Education
Peer reviewed Peer reviewed
Vispoel, Walter P. – Journal of Educational Measurement, 1998
Studied effects of administration mode [computer adaptive test (CAT) versus self-adaptive test (SAT)], item-by-item answer feedback, and test anxiety on results from computerized vocabulary tests taken by 293 college students. CATs were more reliable than SATs, and administration time was less when feedback was provided. (SLD)
Descriptors: Adaptive Testing, College Students, Computer Assisted Testing, Feedback
Peer reviewed Peer reviewed
Enright, Mary K.; Rock, Donald A.; Bennett, Randy Elliot – Journal of Educational Measurement, 1998
Examined alternative-item types and section configurations for improving the discriminant and convergent validity of the Graduate Record Examination (GRE) general test using a computer-based test given to 388 examinees who had taken the GRE previously. Adding new variations of logical meaning appeared to decrease discriminant validity. (SLD)
Descriptors: Admission (School), College Entrance Examinations, College Students, Computer Assisted Testing
Peer reviewed Peer reviewed
Vispoel, Walter P. – Journal of Educational Measurement, 1998
Compared results from computer-adaptive and self-adaptive tests under conditions in which item review was and was not permitted for 379 college students. Results suggest that, when given the opportunity, most examinees will change answers, but usually only to a small portion of items, resulting in some benefit to the test taker. (SLD)
Descriptors: Adaptive Testing, College Students, Computer Assisted Testing, Higher Education
Peer reviewed Peer reviewed
Allalouf, Avi; Ben-Shakhar, Gershon – Journal of Educational Measurement, 1998
Examined how coaching affects the predictive validity and fairness of scholastic aptitude tests. A coached (n=271) and uncoached (n=95) group were compared. Comparison revealed that although coaching enhanced scores on the Israeli Psychometric Entrance Test by about 25% of a standard deviation, it did not create a prediction bias or affect…
Descriptors: College Entrance Examinations, High School Students, High Schools, Higher Education
Peer reviewed Peer reviewed
Vispoel, Walter P.; And Others – Journal of Educational Measurement, 1997
Efficiency, precision, and concurrent validity of results from adaptive and fixed-item music listening tests were studied using: (1) 2,200 simulated examinees; (2) 204 live examinees; and (3) 172 live examinees. Results support the usefulness of adaptive tests for measuring skills that require aurally produced items. (SLD)
Descriptors: Adaptive Testing, Adults, College Students, Comparative Analysis
Peer reviewed Peer reviewed
Clauser, Brian E.; Nungester, Ronald J.; Mazor, Kathleen; Ripkey, Douglas – Journal of Educational Measurement, 1996
Compared the results of differential item functioning (DIF) analysis with matching based on the total test score, matching based on subtest scores, or multivariate matching using multiple subtest scores. Results using 2,000 responses from medical students suggest that matching on multiple subtest scores may be superior to the other methods. (SLD)
Descriptors: Higher Education, Item Bias, Medical Education, Medical Students
Peer reviewed Peer reviewed
Janssen, Rianne; De Boeck, Paul – Journal of Educational Measurement, 1996
Multiple regression analysis shows that both a response-production component and an evaluation component are involved in answers to a free-response synonym task by 299 Belgian college students. Format differences between the multiple choice evaluation task and the synonym task are explained in terms of verbal abilities measured. (SLD)
Descriptors: College Students, Evaluation Methods, Higher Education, Multiple Choice Tests
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9