Publication Date
| In 2015 | 0 |
| Since 2014 | 6 |
| Since 2011 (last 5 years) | 12 |
| Since 2006 (last 10 years) | 19 |
| Since 1996 (last 20 years) | 23 |
Descriptor
| Statistical Analysis | 23 |
| Foreign Countries | 9 |
| Item Response Theory | 7 |
| Models | 7 |
| Psychometrics | 7 |
| Test Bias | 6 |
| Test Items | 6 |
| Scores | 5 |
| English | 4 |
| College Students | 3 |
| More ▼ | |
Source
| International Journal of… | 23 |
Author
| Anagnostopoulos, D. C. | 1 |
| Barnes, Laura L. B. | 1 |
| Baufeldt, Angela L. | 1 |
| Behrens, John T. | 1 |
| Bodkin-Andrews, Gawaian H. | 1 |
| Bradshaw, Laine P. | 1 |
| Carey, Paul D. | 1 |
| Carpenter, Allison | 1 |
| Cohen, Allan S. | 1 |
| Conradie, Simone | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 23 |
| Reports - Research | 16 |
| Reports - Descriptive | 4 |
| Reports - Evaluative | 3 |
| Information Analyses | 2 |
| Guides - Non-Classroom | 1 |
Education Level
| Higher Education | 5 |
| Postsecondary Education | 5 |
| Elementary Secondary Education | 3 |
| High Schools | 2 |
| Secondary Education | 2 |
| Elementary Education | 1 |
| Grade 2 | 1 |
| Grade 4 | 1 |
| Preschool Education | 1 |
Audience
Showing 1 to 15 of 23 results
Jurich, Daniel P.; Bradshaw, Laine P. – International Journal of Testing, 2014
The assessment of higher-education student learning outcomes is an important component in understanding the strengths and weaknesses of academic and general education programs. This study illustrates the application of diagnostic classification models, a burgeoning set of statistical models, in assessing student learning outcomes. To facilitate…
Descriptors: College Outcomes Assessment, Classification, Statistical Analysis, Models
Quaiser-Pohl, Claudia; Neuburger, Sarah; Heil, Martin; Jansen, Petra; Schmelter, Andrea – International Journal of Testing, 2014
This article presents a reanalysis of the data of 862 second and fourth graders collected in two previous studies, focusing on the influence of method (psychometric vs. chronometric) and stimulus type on the gender difference in mental-rotation accuracy. The children had to solve mental-rotation tasks with animal pictures, letters, or cube…
Descriptors: Foreign Countries, Gender Differences, Accuracy, Age Differences
Ferrett, Helen L.; Carey, Paul D.; Baufeldt, Angela L.; Cuzen, Natalie L.; Conradie, Simone; Dowling, Tessa; Stein, Dan J.; Thomas, Kevin G. F. – International Journal of Testing, 2014
Because of their global clinical utility, phonemic fluency tests are frequently incorporated into neuropsychological assessment batteries. However, in heterogeneous societies their use is complicated by the lack of careful attention to using letters of equivalent difficulty across languages, and the paucity of norms stratified by relevant…
Descriptors: Foreign Countries, Phonemes, Language Fluency, Alphabets
Lee, HyeSun; Geisinger, Kurt F. – International Journal of Testing, 2014
Differential item functioning (DIF) analysis is important in terms of test fairness. While DIF analyses have mainly been conducted with manifest grouping variables, such as gender or race/ethnicity, it has been recently claimed that not only the grouping variables but also contextual variables pertaining to examinees should be considered in DIF…
Descriptors: Test Bias, Gender Differences, Regression (Statistics), Statistical Analysis
Rios, Joseph A.; Sireci, Stephen G. – International Journal of Testing, 2014
The International Test Commission's "Guidelines for Translating and Adapting Tests" (2010) provide important guidance on developing and evaluating tests for use across languages. These guidelines are widely applauded, but the degree to which they are followed in practice is unknown. The objective of this study was to perform a…
Descriptors: Guidelines, Translation, Adaptive Testing, Second Languages
Engelhard, George, Jr.; Kobrin, Jennifer L.; Wind, Stefanie A. – International Journal of Testing, 2014
The purpose of this study is to explore patterns in model-data fit related to subgroups of test takers from a large-scale writing assessment. Using data from the SAT, a calibration group was randomly selected to represent test takers who reported that English was their best language from the total population of test takers (N = 322,011). A…
Descriptors: College Entrance Examinations, Writing Tests, Goodness of Fit, English
Tay, Louis; Vermunt, Jeroen K.; Wang, Chun – International Journal of Testing, 2013
We evaluate the item response theory with covariates (IRT-C) procedure for assessing differential item functioning (DIF) without preknowledge of anchor items (Tay, Newman, & Vermunt, 2011). This procedure begins with a fully constrained baseline model, and candidate items are tested for uniform and/or nonuniform DIF using the Wald statistic.…
Descriptors: Item Response Theory, Test Bias, Models, Statistical Analysis
Duong, Minh Q.; von Davier, Alina A. – International Journal of Testing, 2012
Test equating is a statistical procedure for adjusting for test form differences in difficulty in a standardized assessment. Equating results are supposed to hold for a specified target population (Kolen & Brennan, 2004; von Davier, Holland, & Thayer, 2004) and to be (relatively) independent of the subpopulations from the target population (see…
Descriptors: Ability Grouping, Difficulty Level, Psychometrics, Statistical Analysis
Gattamorta, Karina A.; Penfield, Randall D.; Myers, Nicholas D. – International Journal of Testing, 2012
Measurement invariance is a common consideration in the evaluation of the validity and fairness of test scores when the tested population contains distinct groups of examinees, such as examinees receiving different forms of a translated test. Measurement invariance in polytomous items has traditionally been evaluated at the item-level,…
Descriptors: Foreign Countries, Psychometrics, Test Bias, Test Items
Gierl, Mark J.; Lai, Hollis – International Journal of Testing, 2012
Automatic item generation represents a relatively new but rapidly evolving research area where cognitive and psychometric theories are used to produce tests that include items generated using computer technology. Automatic item generation requires two steps. First, test development specialists create item models, which are comparable to templates…
Descriptors: Foreign Countries, Psychometrics, Test Construction, Test Items
Hopfenbeck, Therese N.; Maul, Andrew – International Journal of Testing, 2011
The aim of this study was to investigate response-process based evidence for the validity of the Programme for International Student Assessment's (PISA) self-report questionnaire scales as measures of specific psychological constructs, with a focus on scales meant to measure inclination toward specific learning strategies. Cognitive interviews (N…
Descriptors: Student Reaction, Learning Strategies, Validity, Questionnaires
Xu, Lihua; Barnes, Laura L. B. – International Journal of Testing, 2011
Measurement invariance of the 8-factor Inventory of School Motivation (McInerney & Sinclair, 1991) between American and Chinese college students was tested using single-group and multi-group confirmatory factor analysis. A Mandarin Chinese version of the ISM was developed for this study. Comparisons of latent means were conducted when warranted by…
Descriptors: College Students, Factor Analysis, Positive Reinforcement, Mandarin Chinese
Bodkin-Andrews, Gawaian H.; Ha, My Trinh; Craven, Rhonda G.; Yeung, Alexander Seesing – International Journal of Testing, 2010
This investigation reports on the cross-cultural equivalence testing of the Self-Description Questionnaire II (short version; SDQII-S) for Indigenous and non-Indigenous Australian secondary student samples. A variety of statistical analysis techniques were employed to assess the psychometric properties of the SDQII-S for both the Indigenous and…
Descriptors: Indigenous Populations, Disadvantaged, Testing, Measures (Individuals)
Item Equivalence in English and Chinese Translation of a Cognitive Development Test for Preschoolers
He, Wei; Wolfe, Edward W. – International Journal of Testing, 2010
This article reports the results of a study of potential sources of item nonequivalence between English and Chinese language versions of a cognitive development test for preschool-aged children. Items were flagged for potential nonequivalence through statistical and judgment-based procedures, and the relationship between flag status and item…
Descriptors: Preschool Children, Mandarin Chinese, Cognitive Development, Item Analysis
Wells, Craig S.; Cohen, Allan S.; Patton, Jeffrey – International Journal of Testing, 2009
A primary concern with testing differential item functioning (DIF) using a traditional point-null hypothesis is that a statistically significant result does not imply that the magnitude of DIF is of practical interest. Similarly, for a given sample size, a non-significant result does not allow the researcher to conclude the item is free of DIF. To…
Descriptors: Test Bias, Test Items, Statistical Analysis, Hypothesis Testing
Previous Page | Next Page ยป
Pages: 1 | 2
Peer reviewed
Direct link
