Generating Contexts for ESP Vocabulary Exercises with LLMs

Authors

  • Iglika Nikolova-Stoupak CENTAL, Universit´e catholique de Louvain
  • Serge Bibauw GIRSEF, Universit´e catholique de Louvain
  • Amandine Dumont ILV, Universit´e catholique de Louvain
  • Françoise Stas ILV, Universit´e catholique de Louvain
  • Patrick Watrin CENTAL, Universit´e catholique de Louvain
  • Thomas François CENTAL, Universit´e catholique de Louvain

DOI:

https://doi.org/10.3384/ecp211012

Keywords:

large language models, vocabulary, item generation, contextualized exercises, English for specific purposes

Abstract

The current paper addresses the need for language students and teachers to have access to a large number of pedagogically sound contexts for vocabulary acquisition and testing. We investigate the automatic derivation of contexts for a vocabulary list of English for Specific Purposes (ESP). The contexts are generated by contemporary Large Language Models (namely, Mistral-7B-Instruct and Gemini 1.0 Pro) in zero-shot and few-shot settings, or retrieved from a web-crawled repository of domain-relevant websites. The resulting contexts are compared to a professionally crafted reference corpus based on their textual characteristics (length, morphosyntactic, lexico-semantic, and discourse-related). In addition, we annotated the automatically derived contexts regarding their direct applicability, comprehensibility, and domain relevance. The 'Gemini, zero-shot' contexts are rated most highly by human annotators in terms of pedagogical usability, while the 'Mistral, few-shot' contexts are globally closest to the reference based on textual characteristics.

Downloads

Published

2024-10-15