LLMs are Biased Teachers: Evaluating LLM Bias in Personalized Education

Benchmark (Published & Automated) Relevance: 8/10 28 cited 2024 paper

This paper evaluates bias in LLMs acting as personalized tutors by measuring how models generate and select educational content differently for students with varying demographic characteristics (race, gender, disability, income, etc.). The study introduces two bias metrics (MAB and MDB) and applies them to 9 LLMs using over 17,000 educational explanations across multiple difficulty levels and subjects.

With the increasing adoption of large language models (LLMs) in education, concerns about inherent biases in these models have gained prominence. We evaluate LLMs for bias in the personalized educational setting, specifically focusing on the models' roles as"teachers."We reveal significant biases in how models generate and select educational content tailored to different demographic groups, including race, ethnicity, sex, gender, disability status, income, and national origin. We introduce and a

Study Type

Benchmark (Published & Automated)

Tool Types

AI Tutors 1-to-1 conversational tutoring systems.
Personalised Adaptive Learning Systems that adapt content and difficulty to individual learners.

Tags

large language model evaluation educationcomputer-science