Search
On-Site Program Calendar
Browse By Day
Browse By Time
Browse By Person
Browse By Room
Browse By Unit
Browse By Session Type
Search Tips
Change Preferences / Time Zone
Sign In
Bluesky
Threads
X (Twitter)
YouTube
This study develops and tests a Large Language Model-based assessment framework for analyzing written responses and identifying response uncertainties to support learning. We conducted two case studies with 837 middle and 90 elementary school students to apply and evaluate the framework. The AI models achieved an average overall accuracy of 86% and successfully detected uncertainty in student responses. The framework explains scoring decisions, flags uncertainty, and highlights weaknesses aligned with learning goals to improve usable knowledge. It is especially effective with vague or inconsistent responses, showing the potential of LLMs in classroom assessment. By providing AI-generated scoring reasons and uncertainty data, the framework enables more targeted, timely feedback to support teaching and learning across diverse student backgrounds and classroom contexts.