Paper Summary
Share...

Direct link:

Using Large Language Model to Analyze Chemistry Undergraduate Students’ Self-Constructed Concept Maps

Fri, April 10, 7:45 to 9:15am PDT (7:45 to 9:15am PDT), Los Angeles Convention Center, Floor: Level Two, Poster Hall - Exhibit Hall A

Abstract

This study investigates using generative artificial intelligence (AI), specifically GPT‑4o, in assessing self-constructed concept maps in undergraduate science education. Drawing on Assimilation Theory and Generative Learning Theory, we examine whether AI can approximate expert human grading by analyzing the accuracy and completeness of concept maps created by 257 students. A reproducible pipeline was developed using both hard and soft voting strategies to aggregate GPT‑4o outputs. Results indicate that hard voting outperformed soft voting in accuracy, precision, and F1 score, with the highest accuracy reaching 96.8% under a categorical scoring rubric. These findings demonstrate the feasibility of using AI to assess complex student-generated visual representations and highlight the potential of AI-integrated assessments to support scalable and meaningful feedback in educational contexts.

Authors