NotesFAQContact Us
Collection
Advanced
Search Tips
50 Years of ERIC
50 Years of ERIC
The Education Resources Information Center (ERIC) is celebrating its 50th Birthday! First opened on May 15th, 1964 ERIC continues the long tradition of ongoing innovation and enhancement.

Learn more about the history of ERIC here. PDF icon

Showing 166 to 180 of 278 results
Peer reviewed Peer reviewed
Direct linkDirect link
Ryan, Thomas G. – International Journal of Testing, 2006
Arguably, performance assessment is an integral element of all human educative endeavors. It is the centerpiece of the standards-based reforms movement of the 1990s and continues to be a dominant feature of curriculum planning today. Using performance assessment as a lens, this article addresses the concerns and cautions offered by seminal…
Descriptors: Curriculum Development, Performance Based Assessment, Criticism, Academic Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Leeson, Heidi V. – International Journal of Testing, 2006
In addition to the potential that computer-based testing (CBT) offers, empirical evidence has found that identical computerized and paper-and-pencil tests have not produced equivalent test-taker performance. Referred to as the "mode effect," previous literature has identified many factors that may be responsible for such differential performance.…
Descriptors: Literature Reviews, Computer Assisted Testing, Educational Technology, Performance Based Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Wiberg, Marie – International Journal of Testing, 2006
A simulation study of a sequential computerized mastery test is carried out with items modeled with the 3 parameter logistic item response theory model. The examinees' responses are either identically distributed, not identically distributed, or not identically distributed together with estimation errors in the item characteristics. The…
Descriptors: Test Length, Computer Simulation, Mastery Tests, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A. – International Journal of Testing, 2006
A structural equation modeling approach to scale reliability evaluation can be employed to estimate generalizability theory indexes in settings where sampling of subjects and conditions is carried out. In one- and two-facet crossed designs, it is demonstrated how this method can be used to obtain estimates of relative generalizability…
Descriptors: Computation, Generalizability Theory, Structural Equation Models, Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Austin, David W.; Carlbring, Per; Richards, Jeffrey C.; Andersson, Gerhard – International Journal of Testing, 2006
This study assessed the degree of equivalence between paper and Internet administration of three measures of panic and agoraphobia-related cognition and behavior: Body Sensations Questionnaire (BSQ), Agoraphobic Cognitions Questionnaire (ACQ), and Mobility Inventory (MI). Participants were 110 people with panic disorder who had registered for an…
Descriptors: Comparative Analysis, Behavior Disorders, Anxiety, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Cheong, Yuk Fai – International Journal of Testing, 2006
This article considers and illustrates a strategy to study effects of school context on differential item functioning (DIF) in large-scale assessment. The approach employs a hierarchical generalized linear modeling framework to (a) detect DIF, and (b) identify school-level correlates of the between-group differences in item performance. To…
Descriptors: Context Effect, Test Bias, Causal Models, Educational Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Marsh, Herbert W.; Hau, Kit-Tai; Artelt, Cordula; Baumert, Jurgen; Peschar, Jules L. – International Journal of Testing, 2006
Through a rigorous process of selecting educational psychology's most useful affective constructs, the Organisation for Economic Co-operation and Development (OECD) constructed the Students' Approaches to Learning (SAL) instrument, which requires only 10 min to measure 14 factors that assess self-regulated learning strategies, self-beliefs,…
Descriptors: Measurement Techniques, Educational Psychology, Psychometrics, Cross Cultural Studies
Peer reviewed Peer reviewed
Direct linkDirect link
Bryant, Damon U.; Wooten, William – International Journal of Testing, 2006
The purpose of this study was to demonstrate how cognitive and measurement principles can be integrated to create an essentially unidimensional test. Two studies were conducted. In Study 1, test questions were created by using the feature integration theory of attention to develop a cognitive model of performance and then manipulating complexity…
Descriptors: Test Construction, Cognitive Measurement, Theories, Attention
Peer reviewed Peer reviewed
Direct linkDirect link
Arce-Ferrer, Alvaro J. – International Journal of Testing, 2006
The goal of this study is to investigate how features of a rating scale developed for English-speaking populations interact with Spanish-speaking respondents' response styles and functional categories of judgment. A sample of 400 Spanish-speaking students took a translated scale and a scaling task developed to measure response sets and functional…
Descriptors: Item Response Theory, Multidimensional Scaling, Rating Scales, Spanish Speaking
Peer reviewed Peer reviewed
Direct linkDirect link
Stricker, Lawrence J.; Wilder, Gita Z.; Bridgeman, Brent – International Journal of Testing, 2006
The aim of this study was to assess test takers' attitudes and beliefs about an admissions test used extensively in graduate schools of business in the United States, the Graduate Management Admission Test (GMAT), and the relationships of these attitudes and beliefs to test performance. A set of attitude and belief items was administered by…
Descriptors: Computer Assisted Testing, Test Wiseness, Gender Differences, Ethnic Groups
Peer reviewed Peer reviewed
Direct linkDirect link
Ross, Steven J.; Okabe, Junko – International Journal of Testing, 2006
Test validity is predicated on there being a lack of bias in tasks, items, or test content. It is well-known that factors such as test candidates' mother tongue, life experiences, and socialization practices of the wider community may serve to inject subtle interactions between individuals' background and the test content. When the gender of the…
Descriptors: Gender Bias, Language Tests, Test Validity, Reading Comprehension
Peer reviewed Peer reviewed
Direct linkDirect link
Wu, Amery D.; Ercikan, Kadriye – International Journal of Testing, 2006
Identifying the sources of differential item functioning (DIF) in international assessments is very challenging, because such sources are often nebulous and intertwined. Even though researchers frequently focus on test translation and content area, few actually go beyond these factors to investigate other cultural sources of DIF. This article…
Descriptors: Test Bias, Cultural Influences, Case Studies, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Kubinger, Klaus D. – International Journal of Testing, 2005
In this article, we emphasize that the Rasch model is not only very useful for psychological test calibration but is also necessary if the number of solved items is to be used as an examinee's score. Simplified proof that the Rasch model implies specific objective parameter comparisons is given. Consequently, a model check per se is possible. For…
Descriptors: Psychometrics, Psychological Testing, Item Banks, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Hall, John D.; Howerton, D. Lynn; Bolin, Aaron U. – International Journal of Testing, 2005
The controversial practice of using unlicensed individuals to administered psychological tests has been questioned by some psychologists, professional organizations, state and provincial boards of psychology, state governments, departments of education, and third-party health care providers. This article provides an overview of the ethical, legal,…
Descriptors: Psychologists, Psychology, Psychological Testing, Ethics
Peer reviewed Peer reviewed
Direct linkDirect link
Cascallar, Alicia S.; Dorans, Neil J. – International Journal of Testing, 2005
This study compares two methods commonly used (concordance and prediction) to establish linkages between scores from tests of similar content given in different languages. Score linkages between the Verbal and Math sections of the SAT I and the corresponding sections of the Spanish-language admissions test, the Prueba de Aptitud Academica (PAA),…
Descriptors: Prediction, Scores, Comparative Analysis, Aptitude Tests
Pages: 1  |  ...  |  8  |  9  |  10  |  11  |  12  |  13  |  14  |  15  |  16  |  ...  |  19