Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 4 |
Descriptor
Source
| ETS Research Report Series | 2 |
| Behavioral Research and… | 1 |
| College Board | 1 |
| Language Testing | 1 |
| National Center for Education… | 1 |
| Partnership for Assessment of… | 1 |
| Pearson | 1 |
Author
| Alonzo, Julie | 1 |
| Ferrara, Steve | 1 |
| Futagi, Yoko | 1 |
| Gierl, Mark J. | 1 |
| Glas, Cees A. W. | 1 |
| Gonzalez, Magaly | 1 |
| Goodman, Joshua | 1 |
| Green, Bert F. | 1 |
| Hemat, Ramin | 1 |
| Kinsman, Amy | 1 |
| Kostin, Irene | 1 |
| More ▼ | |
Publication Type
| Numerical/Quantitative Data | 15 |
| Reports - Research | 11 |
| Journal Articles | 3 |
| Reports - Evaluative | 2 |
| Speeches/Meeting Papers | 2 |
| Reports - Descriptive | 1 |
| Tests/Questionnaires | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
| National Assessment of… | 2 |
| Armed Services Vocational… | 1 |
| Comprehensive Tests of Basic… | 1 |
| Graduate Record Examinations | 1 |
| SAT (College Admission Test) | 1 |
| Trends in International… | 1 |
What Works Clearinghouse Rating
Ferrara, Steve; Steedle, Jeffrey; Kinsman, Amy – Partnership for Assessment of Readiness for College and Careers, 2015
We report results from the following three analyses of PARCC [Partnership for Assessment of Readiness for College and Careers] cognitive complexity measures, based on 2014 field test item and task development and field test data. We conducted classification and regression tree analyses using 2014 PARCC field test data to do the following: (1)…
Descriptors: Cognitive Processes, Difficulty Level, Test Items, Mathematics Tests
National Center for Education Statistics, 2013
The 2011 NAEP-TIMSS linking study conducted by the National Center for Education Statistics (NCES) was designed to predict Trends in International Mathematics and Science Study (TIMSS) scores for the U.S. states that participated in 2011 National Assessment of Educational Progress (NAEP) mathematics and science assessment of eighth-grade students.…
Descriptors: Grade 8, Research Methodology, Research Design, Trend Analysis
Alonzo, Julie; Gonzalez, Magaly; Tindal, Gerald – Behavioral Research and Teaching, 2013
In this study, we describe two studies used to select appropriate assessments to measure phonemic awareness, alphabetic principle, and fluency in the Spanish language for students receiving literacy instruction in Spanish. We first describe two studies in which we use linear regression and correlations to examine the appropriateness of different…
Descriptors: Curriculum Based Assessment, Spanish, Phonemic Awareness, Alphabets
Meyers, Jason L.; Murphy, Stephen; Goodman, Joshua; Turhan, Ahmet – Pearson, 2012
Operational testing programs employing item response theory (IRT) applications benefit from of the property of item parameter invariance whereby item parameter estimates obtained from one sample can be applied to other samples (when the underlying assumptions are satisfied). In theory, this feature allows for applications such as computer-adaptive…
Descriptors: Equated Scores, Test Items, Test Format, Item Response Theory
PDF pending restorationGreen, Bert F. – 2002
Maximum likelihood and Bayesian estimates of proficiency, typically used in adaptive testing, use item weights that depend on test taker proficiency to estimate test taker proficiency. In this study, several methods were explored through computer simulation using fixed item weights, which depend mainly on the items difficulty. The simpler scores…
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Testing, Computer Simulation
Sheehan, Kathleen M.; Kostin, Irene; Futagi, Yoko; Hemat, Ramin; Zuckerman, Daniel – ETS Research Report Series, 2006
This paper describes the development, implementation, and evaluation of an automated system for predicting the acceptability status of candidate reading-comprehension stimuli extracted from a database of journal and magazine articles. The system uses a combination of classification and regression techniques to predict the probability that a given…
Descriptors: Automation, Prediction, Reading Comprehension, Classification
de la Torre, Jimmy; Patz, Richard J. – 2001
This paper seeks to extend the application of Markov chain Monte Carlo (MCMC) methods in item response theory (IRT) to include the estimation of equating relationships along with the estimation of test item parameters. A method is proposed that incorporates estimation of the equating relationship in the item calibration phase. Item parameters from…
Descriptors: Achievement Tests, Bayesian Statistics, Equated Scores, Estimation (Mathematics)
Krass, Iosif A.; Thomasson, Gary L. – 1999
New items are being calibrated for the next generation of the computerized adaptive (CAT) version of the Armed Services Vocational Aptitude Battery (ASVAB) (Forms 5 and 6). The requirements that the items be "good" three-parameter logistic (3-PL) model items and typically "like" items in the previous CAT-ASVAB tests have…
Descriptors: Adaptive Testing, Algorithms, Computer Assisted Testing, Nonparametric Statistics
Glas, Cees A. W.; Vos, Hans J. – 1998
A version of sequential mastery testing is studied in which response behavior is modeled by an item response theory (IRT) model. First, a general theoretical framework is sketched that is based on a combination of Bayesian sequential decision theory and item response theory. A discussion follows on how IRT based sequential mastery testing can be…
Descriptors: Adaptive Testing, Bayesian Statistics, Item Response Theory, Mastery Tests
Zhang, Jinming – ETS Research Report Series, 2005
Lord's bias function and the weighted likelihood estimation method are effective in reducing the bias of the maximum likelihood estimate of an examinee's ability under the assumption that the true item parameters are known. This paper presents simulation studies to determine the effectiveness of these two methods in reducing the bias when the item…
Descriptors: Statistical Bias, Maximum Likelihood Statistics, Computation, Ability
Samejima, Fumiko – 1984
In order to evaluate our methods and approaches of estimating the operating characteristics of discrete item responses, it is necessary to try other comparable methods on similar sets of data. LOGIST 5 was taken up for this reason, and was tried upon the hypothetical test items, which follow the normal ogive model and were used frequently in…
Descriptors: Computer Simulation, Computer Software, Estimation (Mathematics), Item Analysis
Peer reviewedReynolds, Trudy; And Others – Language Testing, 1994
Presents a study conducted to provide a comparative analysis of five item analysis indices using both IRT and non-IRT indices to describe the characteristics of flagged items and to investigate the appropriateness of logistic regression as an item analysis technique for further studies. The performance of five item analysis indices was examined.…
Descriptors: College Students, Comparative Analysis, English (Second Language), Item Analysis
Education Commission of the States, Denver, CO. National Assessment of Educational Progress. – 1977
The National Assessment of Educational Progress (NAEP) administered the selected supplemental mathematics exercises to 13-year-old students during October and November 1975 and to 17-year-old students during March and April 1976. This assessment represents a specially modified supplement to 1972-73 full-scale mathematics assessment and was…
Descriptors: Computation, Definitions, Educational Assessment, Elementary Secondary Education
Gierl, Mark J.; Tan, Xuan; Wang, Changjiang – College Board, 2005
The results of this study conclude that there is a multidimensional basis for test score inferences on the mathematics and critical reading sections of the SAT. Results from the exploratory analyses indicate that the data are multidimensional, as mathematics displayed two dimensions and critical reading displayed three dimensions. The correlations…
Descriptors: College Entrance Examinations, Standardized Tests, Scores, Inferences
Rock, Donald A.; And Others – 1995
This report documents the development and validation of the National Education Longitudinal Study of 1988 (NELS:88) cognitive test battery. The cognitive test battery assesses growth between grades 8 and 12 in the content areas of reading comprehension, mathematics, science, and history/citizenship/geography. The battery was part of the NELS:88…
Descriptors: Academic Achievement, Achievement Gains, Bayesian Statistics, Citizenship Education

Direct link
