Grade Guard: A Smart System for Short Answer Automated Grading
Grade Guard is an LLM-based framework for automated short answer grading that introduces an Indecisiveness Score to reflect grading uncertainty and enables human re-evaluation when confidence is low. The system uses fine-tuned temperature parameters and confidence-aware loss to improve grading accuracy compared to baseline LLM approaches.
The advent of large language models (LLMs) in the education sector has provided impetus to automate grading short answer questions. LLMs make evaluating short answers very efficient, thus addressing issues like staff shortage. However, in the task of Automated Short Answer Grading (ASAG), LLM responses are influenced by diverse perspectives in their training dataset, leading to inaccuracies in evaluating nuanced or partially correct answers. To address this challenge, we propose a novel framewor