EALTA Special Interest Group: Artificial Intelligence for Language Assessment

 

News

SIG Meeting

  • Date: 05.03.2025 | 13:00 to 16:00 (UK time)
  • Programme: To be published soon
  • Registration (free): Clic here.

Rationale

Recent years have seen an exponential increase in the use of Artificial Intelligence (AI) for language assessment. These uses include, but are not limited to, automated scoring enabled with natural language processing, AI driven test proctoring, and most recently, item writing using generative AI. Because the field of AI is developing very fast, its uses have not been widely researched and there is no legislation in place that would regulate those uses. In the field of language testing and assessment, although there have been some studies into the use of AI (e.g., Attali et al., 2022; Khademi, 2023; Mizumoto & Eguchi, 2023; O’Sullivan et al, 2023; Román, 2023; Su, Lin, & Lai, 2023), there is still a need for a better understanding of how AI works as well as for monitoring its uses. In parallel to this understanding, it is equally important to acknowledge and navigate the uncertainties that come with such progressive technological advancements. A number of academics, appreciating the extraordinary potential of AI-powered tools, have nevertheless expressed worries over their impact on academic integrity (Barrot, 2023) and the risk of psychological dependence (Alharbi, 2023). These concerns are far from insignificant and underscore the urgent need for an ongoing dialogue regarding the responsible, ethical, and thoughtful application of AI in the realm of language assessments.

Aims

  • Share professional expertise in the use of AI for language testing and assessment;
  • Work to improve the uses of AI for language testing and assessment systems in Europe;
  • Make the expertise in using AI for language testing and assessment readily available;
  • Engage with EU regulatory bodies to produce standard frameworks for the improvement in the uses of AI for language testing and assessment within Europe;
  • Provide a forum for sharing and solving problems related to using AI for language testing and assessment.

Focus

We expect the SIG to be a space for all interested in AI uses for language testing and assessment, but particularly for:

  • Test developers
  • Test administrators
  • Score users
  • Educators
  • Researchers

Organisation

The aim of the SIG is to provide a dedicated forum for discussing and sharing knowledge and practices in the use of AI in language testing and assessment. It seeks to enhance transparency in the utilization of AI for language assessment, explore present and future developments in the field, and foster meaningful discussions on this subject. In particular, the SIG aims to engage in:

  • Analysing existing policies and practices concerning the implementation of AI in language testing and assessment;
  • Establishing good practices for valid, meaningful, and fair uses of AI for language testing and assessment;
  • Collecting information from key stakeholders on their expectations, needs, attitudes, and concerns regarding the use of AI for language testing and assessment;
  • Disseminating knowledge and sharing best practices in the use of AI for language testing and assessment;
  • Encouraging research into the use of AI in language testing and assessment;
  • Promoting a constructive dialogue with European regulatory bodies regarding the use of AI for language testing and assessment;
  • Producing regulatory documents, including guidelines, to assist various stakeholders, such as language testing organisations, policy makers, and score users in implementing AI responsibly for language assessment;
  • Increasing assessment literacy of score users regarding the use of AI in language testing.

About us

Convenor

Dr Olena Rossi, independent language testing consultant and researcher. Olena’s interests within language testing lie in the area of item writing and, in particular, in using AI for item generation. Olena used to serve as the chief of ILTA’s Graduate Student Assembly, she was also a founding member and a co-convenor of the UKALTA’s Postgraduate Researcher Network.

Convenor

Dr Sha Liu, recently completed her PhD in Language Assessment from the University of Bristol. Her research is deeply embedded in the domain of technology-assisted language learning and assessment, with a strong focus on automated writing evaluation leveraging AI algorithms and investigating learner engagement through eye-tracking. She presently serves as the Communications Officer for the British Association for Applied Linguistics Testing, Evaluation, and Assessment Special Interest Group (BAAL TEASIG) Committee.

Content specialist

Dr. Nazlınur Göktürk is a language assessment specialist working for the Board of Education at the Republic of Türkiye Ministry of National Education. She obtained her PhD in Applied Linguistic and Technology from Iowa State University. Her research interests encompass L2 oral communication assessment, spoken dialog systems, and the use of AI for language test development. Within her role at the Ministry, she oversees various project activities aimed at aligning foreign language education with the CEFR CV.

Website manager

Mr. Josep Maria Montcada, Senior Assessment Expert, is currently in charge of language test design and language evaluation at the Ministry of Education of the Generalitat de Catalunya. He coordinates the development of high-stakes certification tests for the Official Language Schools (EOI) in Catalonia in 8 different languages and at levels B1, B2, C1 and C2. He trained in language assessment at Lancaster University and his main interests in this field are test validation and item writing.

Publication project lead

Dr Stefan O’Grady, associate lecturer in academic English and TESOL at the University of St Andrews. Stefan obtained his PhD from the Centre for English Language Learning and Assessment at the University of Bedfordshire. His research interests are in language test development and validation primarily for English medium university admissions and the potential for AI to impact on the assessment process.

Content specialist

Mr. Darren Perrett, Senior Assessment Services Manager, Cambridge University Press & Assessment. Darren is in his final year of his PhD with Leeds University, which is focused on the validation of Cambridge reading texts with NLP features and training an ML algorithm for CEFR classification of previously unseen texts. His main interests are centred around Automated Item Generation (AIG) for high-stakes assessment.

 

The following EALTA expert members act as AI SIG advisors:

  • Prof. Tony Green, University of Bedfordshire
  • Prof. Barry O’Sullivan, British Council
  • Dr Veronika Timpe-Laughlin, Educational Testing Service

Back to top.


Upcoming events

SIG Meeting

  • Date: 05.03.2025 | 13:00 to 16:00 (UK time)
  • Programme: To be published soon
  • Registration (free): Clic here.

Back to top.


Past Meetings

SIG Inaugural Meeting | EALTA Annual Conference Belfast 2024
6th June, 2 to 5:30 pm | Stranmillis University College (Belfast)
Programme

Back to top.


Resources

Guidelines | Fairness | Ethics
Australian Framework for Generative Artificial Intelligence (AI) in Schools The Ethical Framework for AI in Education | The University of Buckingham Ethical guidelines on the use of artificial intelligence (AI) and data in teaching and learning for educators | The European Commission Artificial intelligence and English language teaching: Preparing for the future | British Council
Artificial Intelligence and the Future of Teaching and Learning | US Department of Education ChatGPT and artificial intelligence in higher education: quick start guide | ETICO UNESCO’s International Institute for Education Planning

Guía sobre el uso de la inteligencia artificial en el ámbito educativo | Spanish Department of Education

 

Generative AI and Item Writing
Item Writing for Language Testing | Generative AI and Item Writing

 

Courses
Generative AI for beginners | Microsoft

Back to top.


Digital Library

Tags: listening, task and item generation

Tags: reading, task and item generation 

Tags: fairness, bias, differential item functioning, task and item generation

Tags: text generation, reading

Tags: task and item generation, test development

Tags: text generation, listening

Tags: AIG

Tags: writing, scoring, validity

Tags: scoring, speaking, young learners

Tags: criterion related evidence, predictive validity

Tags: test taker engagement, validation, test development 

Tags: writing, task and item generation

Tags: task design, listening 

Tags: scoring, writing

Tags: listening, task and item generation 

Tags: validity, validation

Tags: scoring

Tags: task design, speaking 

Tags: item generation, reading

Tags: feedback

Tags: item generation, reading

Tags: scoring

Tags: construct definition, scoring and rubric design, validity, fairness, equity, bias, copyright

Tags: validity

Tags: scoring, validity

Tags: vocabulary, task and item generation

Tags: writing, scoring

Back to top.


Media

Current applications of Artificial Intelligence in Language Assessment
Voss, E. (2024, April 8)

Tags: AI, current applications, future directions

The role of AI in educational assessment: What every assessment team should know for 2024.
Scharaschkin, A. (2024, February 14). Webinar delivered by AQA Global Assessment Services.

Tags: scoring, test development, feedback, bias, fairness

Responsible AI standards in assessment.
Burstein, J. (2024, January 19). Webinar delivered as part of the Duolingo webinar series.

Tags: test development, fairness, bias

Automated scoring of writing: Comparing deep learning and feature-based approaches.
Wei, J., van Moere, A., & Lattanzio, S. (2023, October 6). Webinar delivered as part of the Duolingo webinar series.

Tags: scoring

Using AI for test item generation: Opportunities and challenges.
Rossi, O. (2023, May 30). Webinar delivered as part of the EALTA Webinar series.

Tags: test development, item generation

A new paradigm for test development.
Attali, Y., LaFlair, G., & Runge, A. (2023, March 31). Webinar delivered as part of the Duolingo webinar series.

Tags: listening / reading, task and item generation

Generative AI for test development.
Davier, A. (2023, February 27). Talk given for the Department of Education, University of Oxford.

Tags: test development, item generation

Artificial Intelligence, Neuroscience, and Validity | Interface 1
Aryadoust, V. (2021, December 11)

Tags: test validity

Back to top.


Contact

ealtasigai@gmail.com

Back to top.