NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
Job Training Partnership Act…1
What Works Clearinghouse Rating
Showing 16 to 30 of 91 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Garaschuk, Kseniya M.; Cytrynbaum, Eric N. – PRIMUS, 2019
Active learning techniques, such as peer instruction and group work, have been gaining a lot of traction in universities. Taking a natural next step in re-evaluating current practices, many institutions recently started experimenting student-centred group exams. In order to assess the feasibility and effectiveness of collaborative assessments, we…
Descriptors: Instructional Effectiveness, Mathematics Instruction, Group Testing, Group Activities
Peer reviewed Peer reviewed
Direct linkDirect link
Isbell, Dan; Winke, Paula – Language Testing, 2019
The American Council on the Teaching of Foreign Languages (ACTFL) oral proficiency interview -- computer (OPIc) testing system represents an ambitious effort in language assessment: Assessing oral proficiency in over a dozen languages, on the same scale, from virtually anywhere at any time. Especially for users in contexts where multiple foreign…
Descriptors: Oral Language, Language Tests, Language Proficiency, Second Language Learning
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Bae, Minryoung; Lee, Byungmin – English Teaching, 2018
This study examines the effects of text length and question type on Korean EFL readers' reading comprehension of the fill-in-the-blank items in Korean CSAT. A total of 100 Korean EFL college students participated in the study. After divided into three different proficiency groups, the participants took a reading comprehension test which consisted…
Descriptors: Test Items, Language Tests, Second Language Learning, Second Language Instruction
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kabasakal, Kübra Atalay; Kelecioglu, Hülya – Educational Sciences: Theory and Practice, 2015
This study examines the effect of differential item functioning (DIF) items on test equating through multilevel item response models (MIRMs) and traditional IRMs. The performances of three different equating models were investigated under 24 different simulation conditions, and the variables whose effects were examined included sample size, test…
Descriptors: Test Bias, Equated Scores, Item Response Theory, Simulation
Wu, Yi-Fang – ProQuest LLC, 2015
Item response theory (IRT) uses a family of statistical models for estimating stable characteristics of items and examinees and defining how these characteristics interact in describing item and test performance. With a focus on the three-parameter logistic IRT (Birnbaum, 1968; Lord, 1980) model, the current study examines the accuracy and…
Descriptors: Item Response Theory, Test Items, Accuracy, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Chon, Kyong Hee; Lee, Won-Chan; Ansley, Timothy N. – Applied Measurement in Education, 2013
Empirical information regarding performance of model-fit procedures has been a persistent need in measurement practice. Statistical procedures for evaluating item fit were applied to real test examples that consist of both dichotomously and polytomously scored items. The item fit statistics used in this study included the PARSCALE's G[squared],…
Descriptors: Test Format, Test Items, Item Analysis, Goodness of Fit
Peer reviewed Peer reviewed
Direct linkDirect link
Socha, Alan; DeMars, Christine E. – Educational and Psychological Measurement, 2013
Modeling multidimensional test data with a unidimensional model can result in serious statistical errors, such as bias in item parameter estimates. Many methods exist for assessing the dimensionality of a test. The current study focused on DIMTEST. Using simulated data, the effects of sample size splitting for use with the ATFIND procedure for…
Descriptors: Sample Size, Test Length, Correlation, Test Format
Wang, Wei – ProQuest LLC, 2013
Mixed-format tests containing both multiple-choice (MC) items and constructed-response (CR) items are now widely used in many testing programs. Mixed-format tests often are considered to be superior to tests containing only MC items although the use of multiple item formats leads to measurement challenges in the context of equating conducted under…
Descriptors: Equated Scores, Test Format, Test Items, Test Length
Sunnassee, Devdass – ProQuest LLC, 2011
Small sample equating remains a largely unexplored area of research. This study attempts to fill in some of the research gaps via a large-scale, IRT-based simulation study that evaluates the performance of seven small-sample equating methods under various test characteristic and sampling conditions. The equating methods considered are typically…
Descriptors: Test Length, Test Format, Sample Size, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Earl, Kerry – Distance Education, 2013
Tertiary educators (and their institutions) are turning to technology to help meet increased demands in a changing environment. Assessment is one area where such moves are being made. This paper reminds us that assessment is more than a summative check on student knowledge and skills, it is an experience and part of the communication, and…
Descriptors: Online Courses, Student Attitudes, Assignments, Test Format
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Lin, Chuan-Ju – Journal of Technology, Learning, and Assessment, 2010
Assembling equivalent test forms with minimal test overlap across forms is important in ensuring test security. Chen and Lei (2009) suggested a exposure control technique to control test overlap-ordered item pooling on the fly based on the essence that test overlap rate--ordered item pooling for the first t examinees is a function of test overlap…
Descriptors: Test Length, Test Format, Evaluation Criteria, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Camilli, Gregory – Educational Research and Evaluation, 2013
In the attempt to identify or prevent unfair tests, both quantitative analyses and logical evaluation are often used. For the most part, fairness evaluation is a pragmatic attempt at determining whether procedural or substantive due process has been accorded to either a group of test takers or an individual. In both the individual and comparative…
Descriptors: Alternative Assessment, Test Bias, Test Content, Test Format
Peer reviewed Peer reviewed
Netemeyer, Richard G.; Williamson, Donald A.; Burton, Scot; Biswas, Dipayan; Jindal, Supriya; Landreth, Stacy; Mills, Gregory; Primeaux, Sonya – Educational and Psychological Measurement, 2002
Derived shortened versions of the Automatic Thoughts Questionnaire (ATQ) (S. Hollon and P. Kendall, 1980) using samples of 434 and 419 adults. Cross-validation with samples of 163 and 91 adults showed support for the shortened versions. Overall, results suggest that these short forms are useful in measuring cognitions associated with depression.…
Descriptors: Adults, Depression (Psychology), Psychometrics, Test Format
Pommerich, Mary – Journal of Technology, Learning, and Assessment, 2007
Computer administered tests are becoming increasingly prevalent as computer technology becomes more readily available on a large scale. For testing programs that utilize both computer and paper administrations, mode effects are problematic in that they can result in examinee scores that are artificially inflated or deflated. As such, researchers…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Format, Scores
Peer reviewed Peer reviewed
Silverstein, A. B. – Journal of Clinical Psychology, 1985
The findings of research on short forms of the Wechsler Adult Intelligence Scales-Revised are used to illustrate points about three criteria for evaluating the usefulness of a short form. Results indicate there is little justification for regarding the three criteria as criteria. (Author/BL)
Descriptors: Correlation, Evaluation Criteria, Test Format, Test Interpretation
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7