Publication Date
In 2025 | 1 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 8 |
Since 2016 (last 10 years) | 10 |
Since 2006 (last 20 years) | 16 |
Descriptor
Source
Language Testing | 16 |
Author
Batty, Aaron Olaf | 2 |
Lee, Senyung | 2 |
McLean, Stuart | 2 |
Shin, Sun-Young | 2 |
Stewart, Jeffrey | 2 |
Culligan, Brent | 1 |
David, Gergely | 1 |
Dudley, Albert | 1 |
Eberharter, Kathrin | 1 |
Ebner, Viktoria S. | 1 |
Filipi, Anna | 1 |
More ▼ |
Publication Type
Journal Articles | 16 |
Reports - Research | 13 |
Reports - Evaluative | 2 |
Information Analyses | 1 |
Tests/Questionnaires | 1 |
Education Level
Higher Education | 11 |
Postsecondary Education | 9 |
Secondary Education | 4 |
Elementary Education | 1 |
Grade 8 | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Audience
Location
Japan | 5 |
Australia | 1 |
Austria | 1 |
Denmark | 1 |
Hungary | 1 |
Indiana | 1 |
Norway | 1 |
Saudi Arabia | 1 |
South Korea | 1 |
Sweden | 1 |
Turkey (Ankara) | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
Test of English for… | 2 |
Test of English as a Foreign… | 1 |
What Works Clearinghouse Rating
Xueliang Chen; Vahid Aryadoust; Wenxin Zhang – Language Testing, 2025
The growing diversity among test takers in second or foreign language (L2) assessments makes the importance of fairness front and center. This systematic review aimed to examine how fairness in L2 assessments was evaluated through differential item functioning (DIF) analysis. A total of 83 articles from 27 journals were included in a systematic…
Descriptors: Second Language Learning, Language Tests, Test Items, Item Analysis
Park, Yena; Lee, Senyung; Shin, Sun-Young – Language Testing, 2022
Despite consistent calls for authentic stimuli in listening tests for better construct representation, unscripted texts have been rarely adopted in high-stakes listening tests due to perceived inefficiency. This study details how a local academic listening test was developed using authentic unscripted audio-visual texts from the local target…
Descriptors: Listening Comprehension Tests, English for Academic Purposes, Test Construction, Foreign Students
Liao, Ray J. T. – Language Testing, 2023
Among the variety of selected response formats used in L2 reading assessment, multiple-choice (MC) is the most commonly adopted, primarily due to its efficiency and objectiveness. Given the impact of assessment results on teaching and learning, it is necessary to investigate the degree to which the MC format reliably measures learners' L2 reading…
Descriptors: Reading Tests, Language Tests, Second Language Learning, Second Language Instruction
Eberharter, Kathrin; Kormos, Judit; Guggenbichler, Elisa; Ebner, Viktoria S.; Suzuki, Shungo; Moser-Frötscher, Doris; Konrad, Eva; Kremmel, Benjamin – Language Testing, 2023
In online environments, listening involves being able to pause or replay the recording as needed. Previous research indicates that control over the listening input could improve the measurement accuracy of listening assessment. Self-pacing also supports the second language (L2) comprehension processes of test-takers with specific learning…
Descriptors: Literacy, Native Language, Second Language Learning, Second Language Instruction
Gyllstad, Henrik; McLean, Stuart; Stewart, Jeffrey – Language Testing, 2021
The last three decades have seen an increase of tests aimed at measuring an individual's vocabulary level or size. The target words used in these tests are typically sampled from word frequency lists, which are in turn based on language corpora. Conventionally, test developers sample items from frequency bands of 1000 words; different tests employ…
Descriptors: Vocabulary Development, Sample Size, Language Tests, Test Items
Nicklin, Christopher; Vitta, Joseph P. – Language Testing, 2022
Instrument measurement conducted with Rasch analysis is a common process in language assessment research. A recent systematic review of 215 studies involving Rasch analysis in language testing and applied linguistics research reported that 23 different software packages had been utilized. However, none of the analyses were conducted with one of…
Descriptors: Programming Languages, Vocabulary Development, Language Tests, Computer Software
Shin, Sun-Young; Lee, Senyung; Lidster, Ryan – Language Testing, 2021
In this study we investigated the potential for a shared-first-language (shared-L1) effect on second language (L2) listening test scores using differential item functioning (DIF) analyses. We did this in order to understand how accented speech may influence performance at the item level, while controlling for key variables including listening…
Descriptors: Listening Comprehension Tests, Language Tests, Native Language, Scores
Kaya, Elif; O'Grady, Stefan; Kalender, Ilker – Language Testing, 2022
Language proficiency testing serves an important function of classifying examinees into different categories of ability. However, misclassification is to some extent inevitable and may have important consequences for stakeholders. Recent research suggests that classification efficacy may be enhanced substantially using computerized adaptive…
Descriptors: Item Response Theory, Test Items, Language Tests, Classification
McLean, Stuart; Stewart, Jeffrey; Batty, Aaron Olaf – Language Testing, 2020
Vocabulary's relationship to reading proficiency is frequently cited as a justification for the assessment of L2 written receptive vocabulary knowledge. However, to date, there has been relatively little research regarding which modalities of vocabulary knowledge have the strongest correlations to reading proficiency, and observed differences have…
Descriptors: Prediction, Reading Tests, Language Proficiency, Test Items
Tengberg, Michael – Language Testing, 2017
Reading comprehension tests are often assumed to measure the same, or at least similar, constructs. Yet, reading is not a single but a multidimensional form of processing, which means that variations in terms of reading material and item design may emphasize one aspect of the construct at the cost of another. The educational systems in Denmark,…
Descriptors: Foreign Countries, National Competency Tests, Reading Tests, Comparative Analysis
Culligan, Brent – Language Testing, 2015
This study compared three common vocabulary test formats, the Yes/No test, the Vocabulary Knowledge Scale (VKS), and the Vocabulary Levels Test (VLT), as measures of vocabulary difficulty. Vocabulary difficulty was defined as the item difficulty estimated through Item Response Theory (IRT) analysis. Three tests were given to 165 Japanese students,…
Descriptors: Language Tests, Test Format, Comparative Analysis, Vocabulary
Batty, Aaron Olaf – Language Testing, 2015
The rise in the affordability of quality video production equipment has resulted in increased interest in video-mediated tests of foreign language listening comprehension. Although research on such tests has continued fairly steadily since the early 1980s, studies have relied on analyses of raw scores, despite the growing prevalence of item…
Descriptors: Listening Comprehension Tests, Comparative Analysis, Video Technology, Audio Equipment
Pae, Tae-Il – Language Testing, 2012
This study tracked gender differential item functioning (DIF) on the English subtest of the Korean College Scholastic Aptitude Test (KCSAT) over a nine-year period across three data points, using both the Mantel-Haenszel (MH) and item response theory likelihood ratio (IRT-LR) procedures. Further, the study identified two factors (i.e. reading…
Descriptors: Aptitude Tests, Academic Aptitude, Language Tests, Test Items
Filipi, Anna – Language Testing, 2012
The Assessment of Language Competence (ALC) certificates is an annual, international testing program developed by the Australian Council for Educational Research to test the listening and reading comprehension skills of lower to middle year levels of secondary school. The tests are developed for three levels in French, German, Italian and…
Descriptors: Listening Comprehension Tests, Item Response Theory, Statistical Analysis, Foreign Countries
David, Gergely – Language Testing, 2007
Some educational contexts almost mandate the application of multiple-choice (MC) testing techniques, even if they are deplored by many practitioners in the field. In such contexts especially, research into how well these types of item perform and how their performance may be characterised is both appropriate and desirable. The focus of this paper…
Descriptors: Student Evaluation, Grammar, Language Tests, Test Items
Previous Page | Next Page »
Pages: 1 | 2