NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 6,211 to 6,225 of 9,530 results Save | Export
Peer reviewed Peer reviewed
Huynh, Huynh; Saunders, Joseph C. – Journal of Educational Measurement, 1980
Single administration (beta-binomial) estimates for the raw agreement index p and the corrected-for-chance kappa index in mastery testing are compared with those based on two test administrations in terms of estimation bias and sampling variability. Bias is about 2.5 percent for p and 10 percent for kappa. (Author/RL)
Descriptors: Comparative Analysis, Error of Measurement, Mastery Tests, Mathematical Models
Peer reviewed Peer reviewed
Jaeger, Richard M. – Journal of Educational Measurement, 1981
Five indices are discussed that should logically discriminate between situations in which: (1) the linear equating method (LEM) adequately adjusts for difference between score distributions of two approximately parallel test forms; or (2) a method more complex than the linear equating method is needed. (RL)
Descriptors: College Entrance Examinations, Comparative Analysis, Difficulty Level, Equated Scores
Peer reviewed Peer reviewed
Hoste, R. – British Journal of Educational Psychology, 1981
In this paper, a proposal is made by which a content validity coefficient can be calculated. An example of the use of the coefficient is given, demonstrating that different question combinations in a CSE biology examination in which a choice of questions was given gave different levels of content validity. (Author)
Descriptors: Achievement Tests, Biology, Content Analysis, Item Sampling
Peer reviewed Peer reviewed
Warren, Gordon – Journal of Research in Science Teaching, 1979
The paper documents experiments designed to compare essay and multiple-choice tests as means of testing science learning. Results indicate it is easier to score high on multiple-choice tests; some students substitute quantity on essay tests; and essay tests reveal weaknesses hidden by multiple-choice tests. (RE)
Descriptors: Educational Research, Educational Theories, Evaluation Methods, Instructional Materials
Peer reviewed Peer reviewed
Knifong, J. Dan – Journal for Research in Mathematics Education, 1980
The computational difficulty of the word problem sections of eight standardized achievement tests was analyzed with respect to the variety of computational procedures and the number of digits per whole number computation. Analysis reveals considerable variation among the current tests in terms of computational procedure and difficulty. (Author/MK)
Descriptors: Computation, Difficulty Level, Educational Research, Elementary Education
Peer reviewed Peer reviewed
Holliday, William G.; Partridge, Louise A. – Journal of Research in Science Teaching, 1979
Investigates two evaluative hypotheses related to test item sequence and the performance of the students who take the tests. (SA)
Descriptors: Educational Research, Elementary Education, Evaluation, Measurement
Peer reviewed Peer reviewed
Johanson, George A. – Evaluation Practice, 1997
A discussion of differential item functioning (DIF) in the context of attitude assessment is followed by examples involving the detection of DIF on an attitudes-toward-science scale completed by 1,550 elementary school students and the finding of no DIF in a workshop evaluation completed by 1,682 adults. (SLD)
Descriptors: Adults, Attitude Measures, Attitudes, Elementary Education
Peer reviewed Peer reviewed
Zenisky, April L.; Sireci, Stephen G. – Applied Measurement in Education, 2002
Reviews and illustrates some of the current technological developments in computer-based testing, focusing on novel item formats and automated scoring methodologies. The review shows a number of innovations being researched and implemented. (SLD)
Descriptors: Educational Innovation, Educational Technology, Elementary Secondary Education, Large Scale Assessment
Peer reviewed Peer reviewed
Ban, Jae-Chun; Hanson, Bradley A.; Yi, Qing; Harris, Deborah J. – Journal of Educational Measurement, 2002
Compared three online pretest calibration scaling methods through simulation: (1) marginal maximum likelihood with one expectation maximization (EM) cycle (OEM) method; (2) marginal maximum likelihood with multiple EM cycles (MEM); and (3) M. Stocking's method B. MEM produced the smallest average total error in parameter estimation; OEM yielded…
Descriptors: Computer Assisted Testing, Error of Measurement, Maximum Likelihood Statistics, Online Systems
Peer reviewed Peer reviewed
Bline, Dennis; Lowe, Dana R.; Meixner, Wilda F.; Nouri, Hossein – Journal of Business Communication, 2003
Presents the results of an investigation about the effect of question order randomization on the psychometric properties of two frequently used oral and written apprehension instruments: McCroskey's oral communication apprehension scale and Daly and Miller's writing apprehension scale. Shows that the measurement properties of these instruments…
Descriptors: Communication Apprehension, Communication Research, Higher Education, Questionnaires
Peer reviewed Peer reviewed
Hynan, Linda S.; Foster, Barbara M. – Teaching of Psychology, 1997
Describes a project used in a sophomore-level psychological testing and measurement course. Students worked through the different phases of developing a test focused on item writing, reliability, and validity. Responses from both students and instructors have been consistently positive. (MJP)
Descriptors: Higher Education, Item Analysis, Item Response Theory, Psychological Testing
Peer reviewed Peer reviewed
Berger, Martijn P. F.; Veerkamp, Wim J. J. – Journal of Educational and Behavioral Statistics, 1997
Some alternative criteria for item selection in adaptive testing are proposed that take into account uncertainty in the ability estimates. A simulation study shows that the likelihood weighted information criterion is a good alternative to the maximum information criterion. Another good alternative uses a Bayesian expected a posteriori estimator.…
Descriptors: Ability, Adaptive Testing, Bayesian Statistics, Computer Assisted Testing
Peer reviewed Peer reviewed
Chang, Hua-Hua; Ying, Zhiliang – Applied Psychological Measurement, 1996
An item selection procedure for computerized adaptive testing based on average global information is proposed. Results from simulation studies comparing the approach with the usual maximum item information item selection indicate that the new method leads to improvement in terms of bias and mean squared error reduction under many circumstances.…
Descriptors: Adaptive Testing, Computer Assisted Testing, Error of Measurement, Item Response Theory
Peer reviewed Peer reviewed
Jodoin, Michael G. – Journal of Educational Measurement, 2003
Analyzed examinee responses to conventional (multiple-choice) and innovative item formats in a computer-based testing program for item response theory (IRT) information with the three parameter and graded response models. Results for more than 3,000 adult examines for 2 tests show that the innovative item types in this study provided more…
Descriptors: Ability, Adults, Computer Assisted Testing, Item Response Theory
Barnette, J. Jackson – Research in the Schools, 2001
Studied the primacy effect (tendency to select items closer to the left side of the response scale) in Likert scales worded from "Strongly Disagree" to "Strongly Agree" and in the opposite direction. Findings for 386 high school and college students show no primacy effect, although negatively worded stems had an effect on Cronbach's alpha. (SLD)
Descriptors: College Students, High School Students, High Schools, Higher Education
Pages: 1  |  ...  |  411  |  412  |  413  |  414  |  415  |  416  |  417  |  418  |  419  |  ...  |  636