Publication Date
| In 2015 | 1 |
| Since 2014 | 2 |
| Since 2011 (last 5 years) | 6 |
| Since 2006 (last 10 years) | 11 |
| Since 1996 (last 20 years) | 11 |
Descriptor
| College Students | 4 |
| Test Items | 4 |
| College Entrance Examinations | 3 |
| Computer Assisted Testing | 3 |
| Item Response Theory | 3 |
| Test Bias | 3 |
| Achievement Tests | 2 |
| Correlation | 2 |
| Equated Scores | 2 |
| Grade Point Average | 2 |
| More ▼ | |
Source
| Applied Measurement in… | 11 |
Author
| Wise, Steven L. | 2 |
| Antal, Judit | 1 |
| Beatty, Adam S. | 1 |
| Bridgeman, Brent | 1 |
| Burton, Nancy | 1 |
| Chauvin, Sheila W. | 1 |
| Cline, Frederick | 1 |
| Imus, Anna | 1 |
| Kiger, Thomas B. | 1 |
| Kim, Brian | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 11 |
| Reports - Research | 7 |
| Reports - Evaluative | 3 |
| Reports - Descriptive | 1 |
Education Level
| Higher Education | 11 |
| Postsecondary Education | 11 |
| High Schools | 1 |
| Secondary Education | 1 |
Audience
Showing all 11 results
Suh, Youngsuk; Talley, Anna E. – Applied Measurement in Education, 2015
This study compared and illustrated four differential distractor functioning (DDF) detection methods for analyzing multiple-choice items. The log-linear approach, two item response theory-model-based approaches with likelihood ratio tests, and the odds ratio approach were compared to examine the congruence among the four DDF detection methods.…
Descriptors: Test Bias, Multiple Choice Tests, Test Items, Methods
Steedle, Jeffrey T. – Applied Measurement in Education, 2014
Possible lack of motivation is a perpetual concern when tests have no stakes attached to performance. Specifically, the validity of test score interpretations may be compromised when examinees are unmotivated to exert their best efforts. Motivation filtering, a procedure that filters out apparently unmotivated examinees, was applied to the…
Descriptors: College Outcomes Assessment, Student Motivation, Sampling, Validity
Setzer, J. Carl; Wise, Steven L.; van den Heuvel, Jill R.; Ling, Guangming – Applied Measurement in Education, 2013
Assessment results collected under low-stakes testing situations are subject to effects of low examinee effort. The use of computer-based testing allows researchers to develop new ways of measuring examinee effort, particularly using response times. At the item level, responses can be classified as exhibiting either rapid-guessing behavior or…
Descriptors: Testing, Guessing (Tests), Reaction Time, Test Items
Shen, Winny; Sackett, Paul R.; Kuncel, Nathan R.; Beatty, Adam S.; Rigdon, Jana L.; Kiger, Thomas B. – Applied Measurement in Education, 2012
Previous research has demonstrated that cognitive test validities are generalizable and predictive of academic performance across situations. However, even after accounting for statistical artifacts (e.g., sampling error, range restriction, criterion reliability), substantial variability often remains around estimates of cognitive test-performance…
Descriptors: College Entrance Examinations, Standardized Tests, Test Validity, Institutional Characteristics
Imus, Anna; Schmitt, Neal; Kim, Brian; Oswald, Frederick L.; Merritt, Stephanie; Wrestring, Alyssa Friede – Applied Measurement in Education, 2011
Investigations of differential item functioning (DIF) have been conducted mostly on ability tests and have found little evidence of easily interpretable differences across various demographic subgroups. In this study, we examined the degree to which DIF in biographical data items referencing academically relevant background, experiences, and…
Descriptors: Test Bias, Gender Differences, Racial Differences, Biographical Inventories
Liu, Ou Lydia – Applied Measurement in Education, 2011
The TOEFL[R] iBT has increased the length of each reading passage to better approximate academic reading at North American universities, resulting in a reduction in the number of passages on the reading section of the test. One of the concerns brought about by this change is whether the decrease in topic variety increases the likelihood that an…
Descriptors: Language Tests, Reading Tests, English (Second Language), Test Bias
Kim, HeeKyoung; Kolen, Michael J. – Applied Measurement in Education, 2010
Test equating might be affected by including in the equating analyses examinees who have taken the test previously. This study evaluated the effect of including such repeaters on Medical College Admission Test (MCAT) equating using a population invariance approach. Three-parameter logistic (3-PL) item response theory (IRT) true score and…
Descriptors: Repetition, Equated Scores, College Entrance Examinations, Medical Schools
Livingston, Samuel A.; Antal, Judit – Applied Measurement in Education, 2010
A simultaneous equating of four new test forms to each other and to one previous form was accomplished through a complex design incorporating seven separate equating links. Each new form was linked to the reference form by four different paths, and each path produced a different score conversion. The procedure used to resolve these inconsistencies…
Descriptors: Measurement Techniques, Measurement, Educational Assessment, Educational Testing
Thompson, James J.; Yang, Tong; Chauvin, Sheila W. – Applied Measurement in Education, 2009
In some professions, speed and accuracy are as important as acquired requisite knowledge and skills. The availability of computer-based testing now facilitates examination of these two important aspects of student performance. We found that student response times in a conventional non-speeded multiple-choice test, at both the global and individual…
Descriptors: Reaction Time, Test Items, Student Reaction, Multiple Choice Tests
Bridgeman, Brent; Burton, Nancy; Cline, Frederick – Applied Measurement in Education, 2009
Descriptions of validity results based solely on correlation coefficients or percent of the variance accounted for are not merely difficult to interpret, they are likely to be misinterpreted. Predictors that apparently account for a small percent of the variance may actually be highly important from a practical perspective. This study combined two…
Descriptors: Predictive Validity, College Entrance Examinations, Graduate Study, Grade Point Average
Wise, Steven L.; Pastor, Dena A.; Kong, Xiaojing J. – Applied Measurement in Education, 2009
Previous research has shown that rapid-guessing behavior can degrade the validity of test scores from low-stakes proficiency tests. This study examined, using hierarchical generalized linear modeling, examinee and item characteristics for predicting rapid-guessing behavior. Several item characteristics were found significant; items with more text…
Descriptors: Guessing (Tests), Achievement Tests, Correlation, Test Items

Peer reviewed
Direct link
