Search
On-Site Program Calendar
Browse By Day
Browse By Time
Browse By Person
Browse By Room
Browse By Unit
Browse By Session Type
Search Tips
Change Preferences / Time Zone
Sign In
Bluesky
Threads
X (Twitter)
YouTube
This study investigates the use of large language models (LLMs) in teacher professional development through a prompt-engineered framework in cognitive apprenticeship. Implemented in a 10-week asynchronous programming course for K–9 educators, the design simulates instructional phases—modeling, scaffolding, coaching, and reflection—via structured Custom GPT interactions. Design-based methods and qualitative analysis of hundreds of logs revealed (1) a 14% model error rate, (2) frequent student overreliance without constraints, (3) successful orchestration of instructional roles via prompt design, and (4) institutional feasibility and FERPA-compliant deployment. The framework preserves instructor authority as the primary evaluator, ensuring alignment with curricular goals. Findings contribute a tested blueprint for ethical, scalable, dialogic use of LLMs that retain instructor control while encouraging learner reflection.