Publication Date
| In 2015 | 0 |
| Since 2014 | 4 |
| Since 2011 (last 5 years) | 9 |
| Since 2006 (last 10 years) | 22 |
| Since 1996 (last 20 years) | 48 |
Descriptor
| Rating Scales | 244 |
| Test Validity | 84 |
| Higher Education | 58 |
| Test Reliability | 54 |
| Factor Analysis | 44 |
| Test Construction | 39 |
| Factor Structure | 37 |
| Measurement Techniques | 36 |
| Attitude Measures | 32 |
| College Students | 30 |
| More ▼ | |
Source
| Educational and Psychological… | 244 |
Author
| Aiken, Lewis R. | 9 |
| Abbott, Robert D. | 3 |
| Alliger, George M. | 3 |
| Knapp, Robert R. | 3 |
| Ludlow, Larry H. | 3 |
| Schriesheim, Chester A. | 3 |
| Wang, Wen-Chung | 3 |
| Chafouleas, Sandra M. | 2 |
| Christ, Theodore J. | 2 |
| Cicchetti, Domenic V. | 2 |
| More ▼ | |
Publication Type
| Journal Articles | 156 |
| Reports - Research | 116 |
| Reports - Evaluative | 32 |
| Reports - Descriptive | 10 |
| Speeches/Meeting Papers | 2 |
| Tests/Questionnaires | 2 |
| Book/Product Reviews | 1 |
| Numerical/Quantitative Data | 1 |
Education Level
| Higher Education | 7 |
| Elementary Education | 3 |
| Grade 3 | 2 |
| Grade 6 | 2 |
| Grade 2 | 1 |
| Grade 4 | 1 |
| Grade 5 | 1 |
| High Schools | 1 |
| Middle Schools | 1 |
| Postsecondary Education | 1 |
| More ▼ | |
Audience
| Practitioners | 1 |
Showing 1 to 15 of 244 results
Attali, Yigal – Educational and Psychological Measurement, 2014
This article presents a comparative judgment approach for holistically scored constructed response tasks. In this approach, the grader rank orders (rather than rate) the quality of a small set of responses. A prior automated evaluation of responses guides both set formation and scaling of rankings. Sets are formed to have similar prior scores and…
Descriptors: Responses, Item Response Theory, Scores, Rating Scales
Plieninger, Hansjörg; Meiser, Thorsten – Educational and Psychological Measurement, 2014
Response styles, the tendency to respond to Likert-type items irrespective of content, are a widely known threat to the reliability and validity of self-report measures. However, it is still debated how to measure and control for response styles such as extreme responding. Recently, multiprocess item response theory models have been proposed that…
Descriptors: Validity, Item Response Theory, Rating Scales, Models
Jin, Kuan-Yu; Wang, Wen-Chung – Educational and Psychological Measurement, 2014
Extreme response style (ERS) is a systematic tendency for a person to endorse extreme options (e.g., strongly disagree, strongly agree) on Likert-type or rating-scale items. In this study, we develop a new class of item response theory (IRT) models to account for ERS so that the target latent trait is free from the response style and the tendency…
Descriptors: Item Response Theory, Research Methodology, Bayesian Statistics, Response Style (Tests)
Thomas, Troy D.; Abts, Koen; Vander Weyden, Patrick – Educational and Psychological Measurement, 2014
This article investigates the effect of the rural-urban divide on mean response styles (RSs) and their relationships with the sociodemographic characteristics of the respondents. It uses the Representative Indicator Response Style Means and Covariance Structure (RIRSMACS) method and data from Guyana--a developing country in the Caribbean. The…
Descriptors: Rural Urban Differences, Response Style (Tests), Demography, Social Characteristics
Keeley, Jared W.; English, Taylor; Irons, Jessica; Henslee, Amber M. – Educational and Psychological Measurement, 2013
Many measurement biases affect student evaluations of instruction (SEIs). However, two have been relatively understudied: halo effects and ceiling/floor effects. This study examined these effects in two ways. To examine the halo effect, using a videotaped lecture, we manipulated specific teacher behaviors to be "good" or "bad"…
Descriptors: Robustness (Statistics), Test Bias, Course Evaluation, Student Evaluation of Teacher Performance
Raykov, Tenko; Pohl, Steffi – Educational and Psychological Measurement, 2013
A procedure for examining essential unidimensionality in multicomponent measuring instruments is discussed. The method is based on an application of latent variable modeling and is concerned with the extent to which a common factor for all components of a given scale accounts for their correlations. The approach provides point and interval…
Descriptors: Measures (Individuals), Statistical Analysis, Factor Structure, Correlation
Zhu, Xiaowen; Stone, Clement A. – Educational and Psychological Measurement, 2012
This study examined the relative effectiveness of Bayesian model comparison methods in selecting an appropriate graded response (GR) model for performance assessment applications. Three popular methods were considered: deviance information criterion (DIC), conditional predictive ordinate (CPO), and posterior predictive model checking (PPMC). Using…
Descriptors: Bayesian Statistics, Item Response Theory, Comparative Analysis, Models
Brown, Anna; Maydeu-Olivares, Alberto – Educational and Psychological Measurement, 2011
Multidimensional forced-choice formats can significantly reduce the impact of numerous response biases typically associated with rating scales. However, if scored with classical methodology, these questionnaires produce ipsative data, which lead to distorted scale relationships and make comparisons between individuals problematic. This research…
Descriptors: Item Response Theory, Models, Questionnaires, Measurement Techniques
Preston, Kathleen; Reise, Steven; Cai, Li; Hays, Ron D. – Educational and Psychological Measurement, 2011
The authors used a nominal response item response theory model to estimate category boundary discrimination (CBD) parameters for items drawn from the Emotional Distress item pools (Depression, Anxiety, and Anger) developed in the Patient-Reported Outcomes Measurement Information Systems (PROMIS) project. For polytomous items with ordered response…
Descriptors: Item Response Theory, Models, Item Banks, Rating Scales
Granberg-Rademacker, J. Scott – Educational and Psychological Measurement, 2010
The extensive use of survey instruments in the social sciences has long created debate and concern about validity of outcomes, especially among instruments that gather ordinal-level data. Ordinal-level survey measurement of concepts that could be measured at the interval or ratio level produce errors because respondents are forced to truncate or…
Descriptors: Intervals, Rating Scales, Surveys, Markov Processes
Liu, Yan; Wu, Amery D.; Zumbo, Bruno D. – Educational and Psychological Measurement, 2010
In a recent Monte Carlo simulation study, Liu and Zumbo showed that outliers can severely inflate the estimates of Cronbach's coefficient alpha for continuous item response data--visual analogue response format. Little, however, is known about the effect of outliers for ordinal item response data--also commonly referred to as Likert, Likert-type,…
Descriptors: Reliability, Computation, Monte Carlo Methods, Rating Scales
Christ, Theodore J.; Riley-Tillman, T. Chris; Chafouleas, Sandra M.; Boice, Christina H. – Educational and Psychological Measurement, 2010
Generalizability theory was used to examine the generalizability and dependability of outcomes from two single-item Direct Behavior Rating (DBR) scales: DBR of actively manipulating and DBR of visually distracted. DBR is a behavioral assessment tool with specific instrumentation and procedures that can be used by a variety of service delivery…
Descriptors: Generalizability Theory, Student Behavior, Data Collection, Student Evaluation
Marshall, Seth J.; Wodrich, David L.; Gorin, Joanna S. – Educational and Psychological Measurement, 2009
This study examined psychometric properties of the Tempe Sorting Task (TST), a new measure of executive function (EF) for children. To increase the meaningfulness of test score interpretations, an age-appropriate construct was employed to incorporate Denckla's description of EF. Multiple measures of EF, including the TST, were collected for…
Descriptors: Cognitive Tests, Cognitive Processes, Children, Attention Deficit Hyperactivity Disorder
Liu, Ou Lydia; Minsky, Jennifer; Ling, Guangming; Kyllonen, Patrick – Educational and Psychological Measurement, 2009
In an effort to standardize academic application procedures, the authors developed the Standardized Letters of Recommendation (SLR) to capture important cognitive and noncognitive qualities of graduate school candidates. The SLR, which consists of seven scales, is applied to an intern-selection scenario. Both professor ratings (n = 414) during the…
Descriptors: Rating Scales, Reliability, Validity, Item Response Theory
Chafouleas, Sandra M.; Christ, Theodore J.; Riley-Tillman, T. Chris – Educational and Psychological Measurement, 2009
Generalizability theory is used to examine the impact of scaling gradients on a single-item Direct Behavior Rating (DBR). A DBR refers to a type of rating scale used to efficiently record target behavior(s) following an observation occasion. Variance components associated with scale gradients are estimated using a random effects design for persons…
Descriptors: Generalizability Theory, Undergraduate Students, Scaling, Rating Scales

Peer reviewed
Direct link
