TY - JOUR
T1 - Using generative AI for reading question creation based on PIRLS 2011 framework
AU - Wen, Zhipeng
AU - Chu, Samuel Kai Wah
N1 - Publisher Copyright:
© 2025 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group.
PY - 2025
Y1 - 2025
N2 - This study examined the use of ChatGPT 3.5 for reading questions generation based on four processes of comprehension of PIRLS 2011 assessment framework. Using an instrumental case study approach and a usability testing method, we employed an input story text to assess ChatGPT 3.5’s effectiveness. A total of twenty questions were generated and evaluated using specific criteria. We analyzed the content of the obtained questions and the quality of revised questions after further adjustment instructions. Findings reveal that ChatGPT 3.5 excels at generating factual questions, leveraging explicit details from text, and demonstrates improvement in interpret-and-integrate questions when detailed instructions are provided. However, it shows significant limitations with higher-order questions requiring inference and evaluation, where contextual accuracy and deeper comprehension are critical. Statistical analysis revealed significant differences between question types (p = 0.004). Thematic analysis highlighted recurring challenges, such as content misalignment, oversimplification, and difficulty processing complex or nuanced user requirements. Tips for users on potential issues and ethical concerns were also discussed. Despite limitations, ChatGPT 3.5 remains a valuable tool for educators and students to enhance question creation productivity. These findings contribute to understanding generative AI’s role in education and provide actionable insights for improving question generation efficiency.
AB - This study examined the use of ChatGPT 3.5 for reading questions generation based on four processes of comprehension of PIRLS 2011 assessment framework. Using an instrumental case study approach and a usability testing method, we employed an input story text to assess ChatGPT 3.5’s effectiveness. A total of twenty questions were generated and evaluated using specific criteria. We analyzed the content of the obtained questions and the quality of revised questions after further adjustment instructions. Findings reveal that ChatGPT 3.5 excels at generating factual questions, leveraging explicit details from text, and demonstrates improvement in interpret-and-integrate questions when detailed instructions are provided. However, it shows significant limitations with higher-order questions requiring inference and evaluation, where contextual accuracy and deeper comprehension are critical. Statistical analysis revealed significant differences between question types (p = 0.004). Thematic analysis highlighted recurring challenges, such as content misalignment, oversimplification, and difficulty processing complex or nuanced user requirements. Tips for users on potential issues and ethical concerns were also discussed. Despite limitations, ChatGPT 3.5 remains a valuable tool for educators and students to enhance question creation productivity. These findings contribute to understanding generative AI’s role in education and provide actionable insights for improving question generation efficiency.
KW - ChatGPT 3.5
KW - Information Technology
KW - Open & Distance Education and eLearning
KW - Study Skills
KW - Teachers & Teacher Education
KW - Teaching Assistants
KW - generative AI tools
KW - reading questions generation
KW - tips
UR - http://www.scopus.com/inward/record.url?scp=85216692839&partnerID=8YFLogxK
U2 - 10.1080/2331186X.2025.2458653
DO - 10.1080/2331186X.2025.2458653
M3 - Article
AN - SCOPUS:85216692839
VL - 12
JO - Cogent Education
JF - Cogent Education
IS - 1
M1 - 2458653
ER -