Publication Date
| In 2015 | 0 |
| Since 2014 | 0 |
| Since 2011 (last 5 years) | 1 |
| Since 2006 (last 10 years) | 5 |
| Since 1996 (last 20 years) | 9 |
Descriptor
| Test Items | 5 |
| Mathematics Tests | 4 |
| National Competency Tests | 4 |
| Scores | 3 |
| Comparative Analysis | 2 |
| Evaluation Methods | 2 |
| Item Response Theory | 2 |
| Mathematics Achievement | 2 |
| Scoring | 2 |
| Test Construction | 2 |
| More ▼ | |
Source
| Applied Measurement in… | 9 |
Author
| Sireci, Stephen G. | 9 |
| Hambleton, Ronald K. | 3 |
| Wells, Craig S. | 3 |
| Zenisky, April L. | 3 |
| Baldwin, Su | 1 |
| Berberoglu, Giray | 1 |
| Han, Kyung T. | 1 |
| Hauger, Jeffrey B. | 1 |
| Jirka, Stephen | 1 |
| Karatonis, Ana | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 9 |
| Reports - Research | 7 |
| Information Analyses | 1 |
| Reports - Descriptive | 1 |
Education Level
| Elementary Secondary Education | 3 |
| Grade 8 | 2 |
| Grade 12 | 1 |
| High Schools | 1 |
Audience
Showing all 9 results
Han, Kyung T.; Wells, Craig S.; Sireci, Stephen G. – Applied Measurement in Education, 2012
Item parameter drift (IPD) occurs when item parameter values change from their original value over time. IPD may pose a serious threat to the fairness and validity of test score interpretations, especially when the goal of the assessment is to measure growth or improvement. In this study, we examined the effect of multidirectional IPD (i.e., some…
Descriptors: Item Response Theory, Test Items, Scaling, Methods
Wells, Craig S.; Baldwin, Su; Hambleton, Ronald K.; Sireci, Stephen G.; Karatonis, Ana; Jirka, Stephen – Applied Measurement in Education, 2009
Score equity assessment is an important analysis to ensure inferences drawn from test scores are comparable across subgroups of examinees. The purpose of the present evaluation was to assess the extent to which the Grade 8 NAEP Math and Reading assessments for 2005 were equivalent across selected states. More specifically, the present study…
Descriptors: National Competency Tests, Test Bias, Equated Scores, Grade 8
Sireci, Stephen G.; Hauger, Jeffrey B.; Wells, Craig S.; Shea, Christine; Zenisky, April L. – Applied Measurement in Education, 2009
The National Assessment Governing Board used a new method to set achievement level standards on the 2005 Grade 12 NAEP Math test. In this article, we summarize our independent evaluation of the process used to set these standards. The evaluation data included observations of the standard-setting meeting, observations of advisory committee meetings…
Descriptors: Advisory Committees, Mathematics Tests, Standard Setting, National Competency Tests
Zenisky, April L.; Hambleton, Ronald K.; Sireci, Stephen G. – Applied Measurement in Education, 2009
How a testing agency approaches score reporting can have a significant impact on the perception of that assessment and the usefulness of the information among intended users and stakeholders. Too often, important decisions about reporting test data are left to the end of the test development cycle, but by considering the audience(s) and the kinds…
Descriptors: National Competency Tests, Scores, Test Results, Information Dissemination
Hambleton, Ronald K.; Sireci, Stephen G.; Smith, Zachary R. – Applied Measurement in Education, 2009
In this study, we mapped achievement levels from the National Assessment of Educational Progress (NAEP) onto the score scales for selected assessments from the Trends in International Mathematics and Science Study (TIMSS) and the Program for International Student Achievement (PISA). The mapping was conducted on NAEP, TIMSS, and PISA Mathematics…
Descriptors: National Competency Tests, Mathematics Achievement, Mathematics Tests, Comparative Analysis
Peer reviewedKeller, Lisa A.; Swaminathan, Hariharan; Sireci, Stephen G. – Applied Measurement in Education, 2003
Evaluated two strategies for scoring context-dependent test items: ignoring the depending and scoring dichotomously or modeling the dependence through polytomous scoring. Results for data from 38,965 examinees taking a professional examination show that dichotomous scoring may overestimate test information, but polytomous scoring may underestimate…
Descriptors: Adults, Licensing Examinations (Professions), Scoring, Test Items
Peer reviewedZenisky, April L.; Sireci, Stephen G. – Applied Measurement in Education, 2002
Reviews and illustrates some of the current technological developments in computer-based testing, focusing on novel item formats and automated scoring methodologies. The review shows a number of innovations being researched and implemented. (SLD)
Descriptors: Educational Innovation, Educational Technology, Elementary Secondary Education, Large Scale Assessment
Peer reviewedSireci, Stephen G.; Berberoglu, Giray – Applied Measurement in Education, 2000
Studied a method for investigating the equivalence of translated-adapted items using bilingual test takers through item response theory. Results from an English-Turkish course evaluation form completed by 688 Turkish students indicate that the methodology is effective in flagging items that function differentially across languages and informing…
Descriptors: Bilingualism, College Students, Evaluation Methods, Higher Education
Peer reviewedSireci, Stephen G.; Robin, Frederic; Patelis, Thanos – Applied Measurement in Education, 1999
Presents a procedure for standard setting that involves the cluster analysis of test takers to discover examinee groups that are useful for envisioning marginally competent performance or defining borderline or contrasting groups. Illustrates use of the procedure with a statewide mathematics test, and concludes that cluster analysis is useful in…
Descriptors: Cluster Analysis, Mathematics Tests, Standard Setting (Scoring), Standards

Direct link
