Stanford Research Reveals AI Writing Feedback Exhibits Racial and Gender Bias
Key Takeaways
- ▸AI writing feedback systems exhibit measurable bias based on student race, gender, motivation, and ability status, with identical essays receiving different feedback depending on perceived student demographics
- ▸Black students receive more praise and encouragement, while Hispanic/English learner students face more grammar corrections—reflecting and potentially amplifying human teacher biases
- ▸Different feedback types (praise vs. critical suggestions) are distributed unequally, potentially steering different student groups toward different academic trajectories
Summary
Researchers at Stanford University conducted an analysis revealing that AI models used for automated writing feedback exhibit significant bias based on student demographics. The team submitted 600 middle school essays to four different AI models and resubmitted each essay 12 additional times with varying descriptions of the student's race, gender, motivation level, and disability status. The results showed consistent patterns of "positive feedback bias" and "feedback withholding bias"—essays attributed to Black students received more praise and encouragement, while those labeled as Hispanic or English learners received more grammar corrections. White students' essays prompted feedback focused on argument structure and evidence, female students received more affectionate language, and high-achieving students got critical suggestions while unmotivated students received upbeat encouragement.
The researchers attribute this bias to the AI models' training on vast amounts of human language, which embeds human biases including teachers' tendency to soften criticism for certain demographic groups. The unpublished paper, titled "Marked Pedagogies: Examining Linguistic Biases in Personalized Automated Writing Feedback," was nominated for best paper at the International Learning Analytics and Knowledge Conference in Norway, where it will be presented on April 30, 2026.
- AI models are picking up on and replicating human biases from their training data, raising concerns about fairness and equity as schools increasingly adopt AI-powered classroom tools
Editorial Opinion
This research exposes a critical blind spot in educational AI: while these tools are being rolled out to support students, they may actually be perpetuating systemic inequities at scale. The finding that identical work receives fundamentally different feedback based on perceived identity is deeply troubling—especially when these biases align with historical patterns of lower expectations for students of color. Schools must demand bias audits and transparency from AI vendors before deployment, and policymakers should treat this as a red flag warranting regulation before AI grading and feedback tools become entrenched in classrooms.



