Search
On-Site Program Calendar
Browse By Day
Browse By Time
Browse By Person
Browse By Room
Browse By Unit
Browse By Session Type
Search Tips
Change Preferences / Time Zone
Sign In
Bluesky
Threads
X (Twitter)
YouTube
This study investigates the consistency and differences between human raters and ChatGPT in assessing Chinese EFL learners' expository writing. Eighty-two compositions were evaluated on language, content, and organization by humans and two ChatGPT sessions (ChatGPT1 and ChatGPT2). A repeated-measures two-way ANOVA and post-hoc t-tests analyzed rating patterns. Results show ChatGPT gave higher scores for content and organization, while humans rated language more favorably. No significant differences emerged between the two ChatGPT sessions, indicating strong internal consistency. The findings suggest AI can serve as a reliable complementary tool in writing assessment and support the development of hybrid evaluation models that combine human and AI strengths in second language writing.