AI tools alter student feedback based on race, gender, and ability

A Stanford study reveals AI tools alter writing feedback based on a student's race and gender. Discover what this means for equity in personalized learning.

Monday, April 27, 2026

Key Takeaways

  • A [Stanford University](https://www.stanford.edu) study found that artificial intelligence models give superficial praise and less structural criticism to essays attributed to Black students than those attributed to white students.
  • AI educational platforms show a "feedback withholding bias." Marginalized students receive only grammar corrections, while high-achieving or white students receive actionable advice on reasoning and argumentation.
  • While 73% of educational AI systems exhibit some form of bias, only 23% of school administrators conduct equity audits before using these tools in their classrooms.
  • AI teacher assistants generate more punitive behavior intervention plans for students with Black-coded names, while they offer supportive reinforcement for white-coded names.

As we previously reported, artificial intelligence is entering classrooms as a tool for personalized learning. New research shows these systems change their expectations and feedback based on a student's race, gender, and perceived ability.

What Happened

Researchers from Stanford University submitted 600 middle school essays to four different AI models. They asked the models to evaluate the same essays while changing the demographic description of the student writer. The results, detailed in the paper "Marked Pedagogies: Examining Linguistic Biases in Personalized Automated Writing Feedback", show shifts in how the AI responded.

When essays were attributed to Black students, the models provided more praise and encouragement. When the same essays were labeled as written by white students, the AI focused on argument structure and evidence.

The models stereotyped based on other identities. Essays labeled as written by Hispanic students or English language learners triggered basic corrections about grammar. Female students received more affectionate language, while students described as highly motivated were given direct, critical suggestions for improvement.

The Bigger Picture

Experts call this "feedback withholding bias". AI platforms offer positive reinforcement to marginalized groups while reserving high-level, actionable input for their peers. This creates an unequal learning environment. Studies show that precision feedback drives cognitive development. When AI models withhold this evaluation, they limit a student's academic growth.

This issue extends beyond writing. AI teacher assistants, including MagicSchool and Google Gemini, have generated more punitive behavior intervention plans for students with Black-coded names compared to white-coded names. Automated grading systems rate identical work from Black students 0.3 points lower than work from white students, while penalizing students who use African American Vernacular English.

Oversight is lagging. Industry data shows 73% of educational AI systems exhibit bias, yet only 23% of school administrators perform equity audits before purchasing these tools. Half of U.S. school districts have not provided any professional development training on AI implementation.

What This Means for Families

Marketing around educational AI claims it is a neutral tutor. These findings prove that algorithms are not immune to prejudice; they inherit and amplify the biases present in their training data.

If a student relies on AI-generated evaluations, they may receive a false sense of security. Superficial praise without structural critique means a student is not receiving the instruction required to advance their critical thinking. Families cannot assume that an AI tutor pushes every student toward the same academic standard.

What You Can Do

  • Look past the praise: When reviewing AI-generated feedback, check if the suggestions are actionable. If the feedback is entirely positive but lacks specific advice on how to improve the core argument or logic, ask a human teacher to review the assignment.
  • Question the implementation: Ask your local school board or principal what specific tools they use and whether those programs have undergone independent equity audits.
  • Monitor the data inputs: If a school requires students to use an AI platform, investigate whether the tool requires students to input demographic data or fill out profiles that could trigger biased responses.
Share: