Search
On-Site Program Calendar
Browse By Day
Browse By Time
Browse By Person
Browse By Room
Browse By Unit
Browse By Session Type
Search Tips
Change Preferences / Time Zone
Sign In
Bluesky
Threads
X (Twitter)
YouTube
GPT is increasingly used in educational settings to create reading materials tailored to specific grade levels. However, the accuracy of GPT-generated readability scores has not yet been systematically evaluated. This study assesses GPT’s effectiveness in readability evaluation by comparing its grade-level predictions with traditional readability metrics across three genres: language arts, social studies, and science. The results revealed moderate to strong correlations between GPT-generated scores and traditional metrics, with the strongest alignment observed in science texts, followed by social studies, and the weakest in language arts. Additionally, zero-shot prompting performed comparably to few-shot prompting in estimating readability. These findings underscore GPT’s potential for automated readability assessment and offer practical implications for leveraging generative AI in educational contexts.