NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing 151 to 165 of 520 results
Peer reviewed Peer reviewed
Direct linkDirect link
Shumate, Steven R.; Surles, James; Johnson, Robert L.; Penny, Jim – Applied Measurement in Education, 2007
Increasingly, assessment practitioners use generalizability coefficients to estimate the reliability of scores from performance tasks. Little research, however, examines the relation between the estimation of generalizability coefficients and the number of rubric scale points and score distributions. The purpose of the present research is to…
Descriptors: Generalizability Theory, Monte Carlo Methods, Measures (Individuals), Program Effectiveness
Peer reviewed Peer reviewed
Direct linkDirect link
Breland, Hunter; Lee, Yong-Won – Applied Measurement in Education, 2007
The objective of the present investigation was to examine the comparability of writing prompts for different gender groups in the context of the computer-based Test of English as a Foreign Language[TM] (TOEFL[R]-CBT). A total of 87 prompts administered from July 1998 through March 2000 were analyzed. An extended version of logistic regression for…
Descriptors: Learning Theories, Writing Evaluation, Writing Tests, Second Language Learning
Peer reviewed Peer reviewed
Direct linkDirect link
Hogan, Thomas P.; Murphy, Gavin – Applied Measurement in Education, 2007
We determined the recommendations for preparing and scoring constructed-response (CR) test items in 25 sources (textbooks and chapters) on educational and psychological measurement. The project was similar to Haladyna's (2004) analysis for multiple-choice items. We identified 12 recommendations for preparing CR items given by multiple sources,…
Descriptors: Test Items, Scoring, Test Construction, Educational Indicators
Peer reviewed Peer reviewed
Direct linkDirect link
Luecht, Richard; Brumfield, Terry; Breithaupt, Krista – Applied Measurement in Education, 2006
This article describes multistage tests and some practical test development considerations related to the design and implementation of a multistage test, using the Uniform CPA (certified public accountant) Examination as a case study. The article further discusses the use of automated test assembly procedures in an operational context to produce…
Descriptors: Test Construction, Case Studies, Accounting
Peer reviewed Peer reviewed
Direct linkDirect link
Jodoin, Michael G.; Zenisky, April; Hambleton, Ronald K. – Applied Measurement in Education, 2006
Many credentialing agencies today are either administering their examinations by computer or are likely to be doing so in the coming years. Unfortunately, although several promising computer-based test designs are available, little is known about how well they function in examination settings. The goal of this study was to compare fixed-length…
Descriptors: Computers, Test Results, Psychometrics, Computer Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Chuah, Siang Chee; Drasgow, Fritz; Luecht, Richard – Applied Measurement in Education, 2006
Adaptive tests offer the advantages of reduced test length and increased accuracy in ability estimation. However, adaptive tests require large pools of precalibrated items. This study looks at the development of an item pool for 1 type of adaptive administration: the computer-adaptive sequential test. An important issue is the sample size required…
Descriptors: Test Length, Sample Size, Adaptive Testing, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Hambleton, Ronald K.; Xing, Dehui – Applied Measurement in Education, 2006
Now that many credentialing exams are being routinely administered by computer, new computer-based test designs, along with item response theory models, are being aggressively researched to identify specific designs that can increase the decision consistency and accuracy of pass-fail decisions. The purpose of this study was to investigate the…
Descriptors: Test Construction, Objective Tests, Item Response Theory, Feedback
Peer reviewed Peer reviewed
Direct linkDirect link
Banks, Kathleen – Applied Measurement in Education, 2006
The purpose of this article is to present a working definition of the term "culture," as well as to describe and demonstrate a comprehensive framework for evaluating hypotheses about cultural bias in educational testing. The framework is demonstrated using 5th-grade reading and language arts data from the Terra Nova test (CTB/McGraw-Hill, 1999).…
Descriptors: Test Bias, Educational Testing, Test Items, Hispanic Americans
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L. – Applied Measurement in Education, 2006
In low-stakes testing, the motivation levels of examinees are often a matter of concern to test givers because a lack of examinee effort represents a direct threat to the validity of the test data. This study investigated the use of response time to assess the amount of examinee effort received by individual test items. In 2 studies, it was found…
Descriptors: Computer Assisted Testing, Motivation, Test Validity, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Brookhart, Susan M.; Walsh, Janet M.; Zientarski, Wayne A. – Applied Measurement in Education, 2006
Motivation and effort patterns associated with achievement on classroom assessments in middle-school science and social studies were studied with a sample of 223 8th graders in different classroom assessment environments. Classroom assessment environments were characterized by student perceptions of the importance and value of assessment tasks,…
Descriptors: Student Motivation, Educational Assessment, Middle Schools, Science Education
Peer reviewed Peer reviewed
Direct linkDirect link
Custer, Michael; Omar, Md Hafidz; Pomplun, Mark – Applied Measurement in Education, 2006
This study compared vertical scaling results for the Rasch model from BILOG-MG and WINSTEPS. The item and ability parameters for the simulated vocabulary tests were scaled across 11 grades; kindergarten through 10th. Data were based on real data and were simulated under normal and skewed distribution assumptions. WINSTEPS and BILOG-MG were each…
Descriptors: Models, Scaling, Computer Software, Vocabulary
Peer reviewed Peer reviewed
Direct linkDirect link
Penfield, Randall D. – Applied Measurement in Education, 2006
This study applied the maximum expected information (MEI) and the maximum posterior-weighted information (MPI) approaches of computer adaptive testing item selection to the case of a test using polytomous items following the partial credit model. The MEI and MPI approaches are described. A simulation study compared the efficiency of ability…
Descriptors: Bayesian Statistics, Adaptive Testing, Computer Assisted Testing, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Mroch, Andrew A.; Bolt, Daniel M. – Applied Measurement in Education, 2006
Recently, nonparametric methods have been proposed that provide a dimensionally based description of test structure for tests with dichotomous items. Because such methods are based on different notions of dimensionality than are assumed when using a psychometric model, it remains unclear whether these procedures might lead to a different…
Descriptors: Simulation, Comparative Analysis, Psychometrics, Methods Research
Peer reviewed Peer reviewed
Direct linkDirect link
von Schrader, Sarah; Ansley, Timothy – Applied Measurement in Education, 2006
Much has been written concerning the potential group differences in responding to multiple-choice achievement test items. This discussion has included references to possible disparities in tendency to omit such test items. When test scores are used for high-stakes decision making, even small differences in scores and rankings that arise from male…
Descriptors: Gender Differences, Multiple Choice Tests, Achievement Tests, Grade 3
Peer reviewed Peer reviewed
Direct linkDirect link
Huynh, Huynh; Barton, Karen E. – Applied Measurement in Education, 2006
This study examined the effect of oral administration accommodations on test structure and student performance on the Reading test of the South Carolina High School Exit Examination (HSEE). The examination was given at Grade 10 and was untimed; hence, students were permitted as much time as they needed to answer all the questions. Three groups of…
Descriptors: Reading Tests, Exit Examinations, Learning Disabilities, Academic Achievement
Pages: 1  |  ...  |  7  |  8  |  9  |  10  |  11  |  12  |  13  |  14  |  15  |  ...  |  35