NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 301 to 315 of 636 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Cui, Zhongmin; Kolen, Michael J. – Applied Psychological Measurement, 2008
This article considers two methods of estimating standard errors of equipercentile equating: the parametric bootstrap method and the nonparametric bootstrap method. Using a simulation study, these two methods are compared under three sample sizes (300, 1,000, and 3,000), for two test content areas (the Iowa Tests of Basic Skills Maps and Diagrams…
Descriptors: Test Length, Test Content, Simulation, Computation
Peer reviewed Peer reviewed
Alsawalmeh, Yousef M.; Feldt, Leonard S. – Educational and Psychological Measurement, 1999
Develops a statistical test for the hypothesis that alpha'(1) =alpha'(2) when alpha'(1) is the Spearman-Brown extrapolated value of Cronbach's alpha reliability for test 1 and alpha'(2) is the unadjusted coefficient for test 2. The test is shown to exercise tight control of Type I error. (Author/SLD)
Descriptors: Reliability, Test Length
Peer reviewed Peer reviewed
Direct linkDirect link
Weissman, Alexander – Educational and Psychological Measurement, 2007
A general approach for item selection in adaptive multiple-category classification tests is provided. The approach uses mutual information (MI), a special case of the Kullback-Leibler distance, or relative entropy. MI works efficiently with the sequential probability ratio test and alleviates the difficulties encountered with using other local-…
Descriptors: Scientific Concepts, Probability, Test Length, Item Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Lee, Yi-Hsuan; Zhang, Jinming – ETS Research Report Series, 2008
The method of maximum-likelihood is typically applied to item response theory (IRT) models when the ability parameter is estimated while conditioning on the true item parameters. In practice, the item parameters are unknown and need to be estimated first from a calibration sample. Lewis (1985) and Zhang and Lu (2007) proposed the expected response…
Descriptors: Item Response Theory, Comparative Analysis, Computation, Ability
Bentley-Williams, Robyn; Forbes, Anne – Australian Association for Research in Education (NJ1), 2012
This investigation examined the course experiences of Bachelor of Education Primary students across each year of the course. The aims of the study were to identify gaps in what we know about our students; to identify relevant domains in student experiences and to assist with course improvements. A reflective inquiry paradigm was adopted for…
Descriptors: Foreign Countries, Bachelors Degrees, Preservice Teachers, Student Teacher Attitudes
Wei, Youhua – ProQuest LLC, 2008
Scale linking is the process of developing the connection between scales of two or more sets of parameter estimates obtained from separate test calibrations. It is the prerequisite for many applications of IRT, such as test equating and differential item functioning analysis. Unidimensional scale linking methods have been studied and applied…
Descriptors: Test Length, Test Items, Sample Size, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Woods, Carol M. – Applied Psychological Measurement, 2008
In Ramsay-curve item response theory (RC-IRT), the latent variable distribution is estimated simultaneously with the item parameters of a unidimensional item response model using marginal maximum likelihood estimation. This study evaluates RC-IRT for the three-parameter logistic (3PL) model with comparisons to the normal model and to the empirical…
Descriptors: Test Length, Computation, Item Response Theory, Maximum Likelihood Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Glas, Cees A. W.; Pimentel, Jonald L. – Educational and Psychological Measurement, 2008
In tests with time limits, items at the end are often not reached. Usually, the pattern of missing responses depends on the ability level of the respondents; therefore, missing data are not ignorable in statistical inference. This study models data using a combination of two item response theory (IRT) models: one for the observed response data and…
Descriptors: Intelligence Tests, Statistical Inference, Item Response Theory, Modeling (Psychology)
Wu, Margaret – OECD Publishing (NJ1), 2010
This paper makes an in-depth comparison of the PISA (OECD) and TIMSS (IEA) mathematics assessments conducted in 2003. First, a comparison of survey methodologies is presented, followed by an examination of the mathematics frameworks in the two studies. The methodologies and the frameworks in the two studies form the basis for providing…
Descriptors: Mathematics Achievement, Foreign Countries, Gender Differences, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Wells, Craig S.; Bolt, Daniel M. – Applied Measurement in Education, 2008
Tests of model misfit are often performed to validate the use of a particular model in item response theory. Douglas and Cohen (2001) introduced a general nonparametric approach for detecting misfit under the two-parameter logistic model. However, the statistical properties of their approach, and empirical comparisons to other methods, have not…
Descriptors: Test Length, Test Items, Monte Carlo Methods, Nonparametric Statistics
Peer reviewed Peer reviewed
Edwards, Roger H. – Educational and Psychological Measurement, 1981
Under certain conditions, a validity Coefficient of Effective Length (CEL) can produce highly misleading results. A modified coefficent is suggested for use when empirical studies indicate that underlying assumptions have been violated. (Author/BW)
Descriptors: Mathematical Formulas, Test Construction, Test Length
Peer reviewed Peer reviewed
Direct linkDirect link
Bridgeman, Brent; Trapani, Catherine; Curley, Edward – Journal of Educational Measurement, 2004
The impact of allowing more time for each question on the SAT I: Reasoning Test scores was estimated by embedding sections with a reduced number of questions into the standard 30-minute equating section of two national test administrations. Thus, for example, questions were deleted from a verbal section that contained 35 questions to produce forms…
Descriptors: College Entrance Examinations, Test Length, Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ricker, Kathryn L.; von Davier, Alina A. – ETS Research Report Series, 2007
This study explored the effects of external anchor test length on final equating results of several equating methods, including equipercentile (frequency estimation), chained equipercentile, kernel equating (KE) poststratification PSE with optimal bandwidths, and KE PSE linear (large bandwidths) when using the nonequivalent groups anchor test…
Descriptors: Equated Scores, Test Items, Statistical Analysis, Test Length
Peer reviewed Peer reviewed
Direct linkDirect link
Woods, Carol M. – Applied Psychological Measurement, 2007
Ramsay curve item response theory (RC-IRT) was recently developed to detect and correct for nonnormal latent variables when unidimensional IRT models are fitted to data using maximum marginal likelihood estimation. The purpose of this research is to evaluate the performance of RC-IRT for Likert-type item responses with varying test lengths, sample…
Descriptors: Test Length, Item Response Theory, Sample Size, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Frazier, Thomas W.; Youngstrom, Eric A. – Intelligence, 2007
A historical increase in the number of factors purportedly measured by commercial tests of cognitive ability may result from four distinct pressures including: increasingly complex models of intelligence, test publishers' desires to provide clinically useful assessment instruments with greater interpretive value, test publishers' desires to…
Descriptors: Evaluation Criteria, Factor Structure, Cognitive Ability, Intelligence Tests
Pages: 1  |  ...  |  17  |  18  |  19  |  20  |  21  |  22  |  23  |  24  |  25  |  ...  |  43