Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 1 |
| Since 2017 (last 10 years) | 5 |
| Since 2007 (last 20 years) | 15 |
Descriptor
| Computer Assisted Testing | 15 |
| Construct Validity | 15 |
| Scores | 15 |
| Correlation | 6 |
| Factor Analysis | 6 |
| Second Language Learning | 6 |
| English (Second Language) | 5 |
| Language Tests | 5 |
| Scoring | 5 |
| Statistical Analysis | 5 |
| Writing Tests | 5 |
| More ▼ | |
Source
Author
| Attali, Yigal | 2 |
| Sinharay, Sandip | 2 |
| Wilson, Joshua | 2 |
| Aydin, Belgin | 1 |
| Boyne, James | 1 |
| Chen, Dandan | 1 |
| Clariana, Roy B. | 1 |
| Demir, Selcuk Besir | 1 |
| Deng, Hui | 1 |
| Dermo, John | 1 |
| Ercan, Recep | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 14 |
| Reports - Research | 12 |
| Reports - Descriptive | 2 |
| Tests/Questionnaires | 2 |
| Dissertations/Theses -… | 1 |
Education Level
| Higher Education | 7 |
| Postsecondary Education | 5 |
| Secondary Education | 3 |
| Elementary Education | 2 |
| Elementary Secondary Education | 1 |
| Grade 7 | 1 |
| Grade 8 | 1 |
| High Schools | 1 |
| Junior High Schools | 1 |
| Middle Schools | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
| Test of English as a Foreign… | 2 |
| Graduate Record Examinations | 1 |
| International English… | 1 |
| SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Myers, Matthew C.; Wilson, Joshua – International Journal of Artificial Intelligence in Education, 2023
This study evaluated the construct validity of six scoring traits of an automated writing evaluation (AWE) system called "MI Write." Persuasive essays (N = 100) written by students in grades 7 and 8 were randomized at the sentence-level using a script written with Python's NLTK module. Each persuasive essay was randomized 30 times (n =…
Descriptors: Construct Validity, Automation, Writing Evaluation, Algorithms
Isler, Cemre; Aydin, Belgin – International Journal of Assessment Tools in Education, 2021
This study is about the development and validation process of the Computerized Oral Proficiency Test of English as a Foreign Language (COPTEFL). The test aims at assessing the speaking proficiency levels of students in Anadolu University School of Foreign Languages (AUSFL). For this purpose, three monologic tasks were developed based on the Global…
Descriptors: Test Construction, Construct Validity, Interrater Reliability, Scores
Wilson, Joshua; Chen, Dandan; Sandbank, Micheal P.; Hebert, Michael – Journal of Educational Psychology, 2019
The present study examined issues pertaining to the reliability of writing assessment in the elementary grades, and among samples of struggling and nonstruggling writers. The present study also extended nascent research on the reliability and the practical applications of automated essay scoring (AES) systems in Response to Intervention frameworks…
Descriptors: Computer Assisted Testing, Automation, Scores, Writing Tests
Jin, Yan; Yan, Ming – Language Assessment Quarterly, 2017
One major threat to validity in high-stakes testing is construct-irrelevant variance. In this study we explored whether the transition from a paper-and-pencil to a computer-based test mode in a high-stakes test in China, the College English Test, has brought about variance irrelevant to the construct being assessed in this test. Analyses of the…
Descriptors: Writing Tests, Computer Assisted Testing, Computer Literacy, Construct Validity
Wei, Wei; Zheng, Ying – Computer Assisted Language Learning, 2017
This research provided a comprehensive evaluation and validation of the listening section of a newly introduced computerised test, Pearson Test of English Academic (PTE Academic). PTE Academic contains 11 item types assessing academic listening skills either alone or in combination with other skills. First, task analysis helped identify skills…
Descriptors: Listening Comprehension Tests, Computer Assisted Testing, Language Tests, Construct Validity
Kebble, Paul Graham – The EUROCALL Review, 2016
The C-Test as a tool for assessing language competence has been in existence for nearly 40 years, having been designed by Professors Klein-Braley and Raatz for implementation in German and English. Much research has been conducted over the ensuing years, particularly in regards to reliability and construct validity, for which it is reported to…
Descriptors: Language Tests, Computer Software, Test Construction, Test Reliability
Ercan, Recep; Yaman, Tugba; Demir, Selcuk Besir – Journal of Education and Training Studies, 2015
The objective of this study is to develop a valid and reliable attitude scale having quality psychometric features that can measure secondary school students' attitudes towards human rights. The study group of the research is comprised by 710 6th, 7th and 8th grade students who study at 4 secondary schools in the centre of Sivas. The study group…
Descriptors: Civil Rights, Attitude Measures, Factor Analysis, Construct Validity
Attali, Yigal; Sinharay, Sandip – ETS Research Report Series, 2015
The "e-rater"® automated essay scoring system is used operationally in the scoring of "TOEFL iBT"® independent and integrated tasks. In this study we explored the psychometric added value of reporting four trait scores for each of these two tasks, beyond the total e-rater score.The four trait scores are word choice, grammatical…
Descriptors: Writing Tests, Scores, Language Tests, English (Second Language)
Dermo, John; Boyne, James – Practitioner Research in Higher Education, 2014
We describe a study conducted during 2009-12 into innovative assessment practice, evaluating an assessed coursework task on a final year Medical Genetics module for Biomedical Science undergraduates. An authentic e-assessment coursework task was developed, integrating objectively marked online questions with an online DNA sequence analysis tool…
Descriptors: Biomedicine, Medical Education, Computer Assisted Testing, Courseware
Attali, Yigal; Sinharay, Sandip – ETS Research Report Series, 2015
The "e-rater"® automated essay scoring system is used operationally in the scoring of the argument and issue tasks that form the Analytical Writing measure of the "GRE"® General Test. For each of these tasks, this study explored the value added of reporting 4 trait scores for each of these 2 tasks over the total e-rater score.…
Descriptors: Scores, Computer Assisted Testing, Computer Software, Grammar
Freund, Philipp Alexander; Holling, Heinz – Intelligence, 2011
The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of…
Descriptors: Intelligence, Test Results, Individual Testing, Construct Validity
Sawaki, Yasuyo; Stricker, Lawrence J.; Oranje, Andreas H. – Language Testing, 2009
This construct validation study investigated the factor structure of the Test of English as a Foreign Language[TM] Internet-based test (TOEFL[R] iBT). An item-level confirmatory factor analysis was conducted for a test form completed by participants in a field study. A higher-order factor model was identified, with a higher-order general factor…
Descriptors: Speech Communication, Construct Validity, Factor Structure, Factor Analysis
Itomitsu, Masayuki – ProQuest LLC, 2009
This dissertation reports development and validation studies of a Web-based standardized test of Japanese as a foreign language (JFL), designed to measure learners' off-line grammatical and pragmatic knowledge in multiple-choice format. Targeting Japanese majors in the U.S. universities and colleges, the test is designed to explore possible…
Descriptors: Sentences, Speech Acts, Grammar, Second Language Learning
Kobrin, Jennifer L.; Deng, Hui; Shaw, Emily J. – Journal of Applied Testing Technology, 2007
This study was designed to address two frequent criticisms of the SAT essay--that essay length is the best predictor of scores, and that there is an advantage in using more "sophisticated" examples as opposed to personal experience. The study was based on 2,820 essays from the first three administrations of the new SAT. Each essay was…
Descriptors: Testing Programs, Computer Assisted Testing, Construct Validity, Writing Skills
Clariana, Roy B.; Wallace, Patricia – Journal of Educational Computing Research, 2007
This proof-of-concept investigation describes a computer-based approach for deriving the knowledge structure of individuals and of groups from their written essays, and considers the convergent criterion-related validity of the computer-based scores relative to human rater essay scores and multiple-choice test scores. After completing a…
Descriptors: Computer Assisted Testing, Multiple Choice Tests, Construct Validity, Cognitive Structures

Peer reviewed
Direct link
