Abstract
With the rise of generative AI models, such as large language models (LLMs), in educational settings, there is a growing demand to ensure the quality of AI-generated multiple-choice questions (MCQs) used in higher education. Traditional quiz development methods fall short in addressing the unique challenges posed by AI-generated content, such as consistency, cognitive demand, and question uniqueness. This paper presents the QUEST framework, a structured approach designed specifically to evaluate the quality of LLM-generated MCQs across five dimensions: Quality, Uniqueness, Effort, Structure, and Transparency. Following an iterative research process, AI-generated questions were assessed and refined using QUEST, revealing that the framework effectively improves question clarity, relevance, and educational value. The findings suggest that QUEST is a viable tool for educators to maintain high-quality standards in AI-generated assessments, ensuring these resources meet the pedagogical needs of diverse learners in higher education.
Original language | English |
---|---|
Title of host publication | New Media Pedagogy: Research Trends, Methodological Challenges, and Successful Implementations |
Publisher | Springer, Cham |
Pages | 293-303 |
ISBN (Electronic) | 978-3-031-95627-0 |
ISBN (Print) | 978-3-031-95626-3 |
DOIs | |
Publication status | Published - 29 Jun 2025 |
Event | 3rd International Conference, NMP 2024 - Kraków, Poland Duration: 28 Nov 2024 → 29 Nov 2024 |
Conference
Conference | 3rd International Conference, NMP 2024 |
---|---|
Country/Territory | Poland |
City | Kraków |
Period | 28/11/24 → 29/11/24 |
Fields of Expertise
- Information, Communication & Computing