Publication Date
| In 2015 | 1 |
| Since 2014 | 10 |
| Since 2011 (last 5 years) | 77 |
| Since 2006 (last 10 years) | 133 |
| Since 1996 (last 20 years) | 202 |
Descriptor
| Test Length | 439 |
| Test Items | 132 |
| Item Response Theory | 118 |
| Test Construction | 118 |
| Test Reliability | 93 |
| Sample Size | 79 |
| Test Validity | 78 |
| Computer Assisted Testing | 76 |
| Adaptive Testing | 75 |
| Simulation | 71 |
| More ▼ | |
Source
Author
| Hambleton, Ronald K. | 14 |
| Wang, Wen-Chung | 8 |
| Livingston, Samuel A. | 6 |
| Sijtsma, Klaas | 6 |
| Wainer, Howard | 6 |
| Wilcox, Rand R. | 6 |
| Gessaroli, Marc E. | 5 |
| Lewis, Charles | 5 |
| Reckase, Mark D. | 5 |
| Weiss, David J. | 5 |
| More ▼ | |
Publication Type
Showing 1 to 15 of 439 results
He, Wei; Reckase, Mark D. – Educational and Psychological Measurement, 2014
For computerized adaptive tests (CATs) to work well, they must have an item pool with sufficient numbers of good quality items. Many researchers have pointed out that, in developing item pools for CATs, not only is the item pool size important but also the distribution of item parameters and practical considerations such as content distribution…
Descriptors: Item Banks, Test Length, Computer Assisted Testing, Adaptive Testing
Gelfand, Jessica T.; Christie, Robert E.; Gelfand, Stanley A. – Journal of Speech, Language, and Hearing Research, 2014
Purpose: Speech recognition may be analyzed in terms of recognition probabilities for perceptual wholes (e.g., words) and parts (e.g., phonemes), where j or the j-factor reveals the number of independent perceptual units required for recognition of the whole (Boothroyd, 1968b; Boothroyd & Nittrouer, 1988; Nittrouer & Boothroyd, 1990). For…
Descriptors: Phonemes, Word Recognition, Vowels, Syllables
Liang, Tie; Wells, Craig S.; Hambleton, Ronald K. – Journal of Educational Measurement, 2014
As item response theory has been more widely applied, investigating the fit of a parametric model becomes an important part of the measurement process. There is a lack of promising solutions to the detection of model misfit in IRT. Douglas and Cohen introduced a general nonparametric approach, RISE (Root Integrated Squared Error), for detecting…
Descriptors: Item Response Theory, Measurement Techniques, Nonparametric Statistics, Models
Lathrop, Quinn N.; Cheng, Ying – Journal of Educational Measurement, 2014
When cut scores for classifications occur on the total score scale, popular methods for estimating classification accuracy (CA) and classification consistency (CC) require assumptions about a parametric form of the test scores or about a parametric response model, such as item response theory (IRT). This article develops an approach to estimate CA…
Descriptors: Cutting Scores, Classification, Computation, Nonparametric Statistics
Straat, J. Hendrik; van der Ark, L. Andries; Sijtsma, Klaas – Educational and Psychological Measurement, 2014
An automated item selection procedure in Mokken scale analysis partitions a set of items into one or more Mokken scales, if the data allow. Two algorithms are available that pursue the same goal of selecting Mokken scales of maximum length: Mokken's original automated item selection procedure (AISP) and a genetic algorithm (GA). Minimum…
Descriptors: Sampling, Test Items, Effect Size, Scaling
Runco, Mark A.; Walczyk, Jeffrey John; Acar, Selcuk; Cowger, Ernest L.; Simundson, Melissa; Tripp, Sunny – Journal of Creative Behavior, 2014
This article describes an empirical refinement of the "Runco Ideational Behavior Scale" (RIBS). The RIBS seems to be associated with divergent thinking, and the potential for creative thinking, but it was possible that its validity could be improved. With this in mind, three new scales were developed and the unique benefit (or…
Descriptors: Behavior Rating Scales, Creative Thinking, Test Validity, Psychometrics
Lee, Jihyun; Paek, Insu – Journal of Psychoeducational Assessment, 2014
Likert-type rating scales are still the most widely used method when measuring psychoeducational constructs. The present study investigates a long-standing issue of identifying the optimal number of response categories. A special emphasis is given to categorical data, which were generated by the Item Response Theory (IRT) Graded-Response Modeling…
Descriptors: Likert Scales, Responses, Item Response Theory, Classification
Yang, Sophie Xin; Jowett, Sophia – Measurement in Physical Education and Exercise Science, 2013
The Coach-Athlete Relationship Questionnaire was developed to effectively measure affective, cognitive, and behavioral aspects, represented by the interpersonal constructs of closeness, commitment, and complementarity, of the quality of the relationship within the context of sport coaching. The current study sought to determine the internal…
Descriptors: Foreign Countries, Athletes, Athletic Coaches, Interpersonal Relationship
Earl, Kerry – Distance Education, 2013
Tertiary educators (and their institutions) are turning to technology to help meet increased demands in a changing environment. Assessment is one area where such moves are being made. This paper reminds us that assessment is more than a summative check on student knowledge and skills, it is an experience and part of the communication, and…
Descriptors: Online Courses, Student Attitudes, Assignments, Test Format
Watanabe, Yoshinori – Language Testing, 2013
This article describes the National Center Test for University Admissions, a unified national test in Japan, which is taken by 500,000 students every year. It states that implementation of the Center Test began in 1990, with the English component consisting only of the written section until 2005, when the listening section was first implemented…
Descriptors: College Admission, Foreign Countries, College Entrance Examinations, English (Second Language)
Hacker, Jason; Carr, Andrea; Abrams, Matthew; Brown, Steven D. – Journal of Career Assessment, 2013
Prior research using a 167-item measure of career indecision (Career Indecision Profile-167 [CIP-167]) has suggested that career choice difficulties may be associated with four major sources of career indecision: neuroticism/negative affectivity, choice/commitment anxiety, lack of readiness, and interpersonal conflicts. The purpose of this study…
Descriptors: Career Choice, Decision Making, Measures (Individuals), Factor Structure
Svetina, Dubravka – Educational and Psychological Measurement, 2013
The purpose of this study was to investigate the effect of complex structure on dimensionality assessment in noncompensatory multidimensional item response models using dimensionality assessment procedures based on DETECT (dimensionality evaluation to enumerate contributing traits) and NOHARM (normal ogive harmonic analysis robust method). Five…
Descriptors: Item Response Theory, Statistical Analysis, Computation, Test Length
Arens, A. Katrin; Yeung, Alexander Seeshing; Craven, Rhonda G.; Hasselhorn, Marcus – International Journal of Research & Method in Education, 2013
This study aims to develop a short German version of the Self Description Questionnaire (SDQ I-GS) in order to present a robust economical instrument for measuring German preadolescents' multidimensional self-concept. A full German version of the SDQ I (SDQ I-G) that maintained the original structure and thus length of the English original…
Descriptors: Foreign Countries, Questionnaires, Test Construction, Test Length
Socha, Alan; DeMars, Christine E. – Educational and Psychological Measurement, 2013
Modeling multidimensional test data with a unidimensional model can result in serious statistical errors, such as bias in item parameter estimates. Many methods exist for assessing the dimensionality of a test. The current study focused on DIMTEST. Using simulated data, the effects of sample size splitting for use with the ATFIND procedure for…
Descriptors: Sample Size, Test Length, Correlation, Test Format
Harrell-Williams, Leigh M.; Wolfe, Edward W. – Educational and Psychological Measurement, 2013
Most research on confirmatory factor analysis using information-based fit indices (Akaike information criterion [AIC], Bayesian information criteria [BIC], bias-corrected AIC [AICc], and consistent AIC [CAIC]) has used a structural equation modeling framework. Minimal research has been done concerning application of these indices to item response…
Descriptors: Correlation, Goodness of Fit, Test Length, Item Response Theory

Peer reviewed
Direct link
