As educational institutions increasingly integrate generative artificial intelligence into their core curricula, a landmark analysis from Stanford University suggests that these digital tools may be inadvertently reinforcing societal inequities. The research indicates that AI feedback models, designed to assist students in refining their writing, provide significantly different guidance based on a student’s perceived race, gender, and academic standing. While AI is often marketed as an objective arbiter of quality, this new evidence suggests that the technology carries the same subconscious biases as the human-generated data on which it was trained, potentially steering students toward different academic outcomes based on their identities.
The study, titled "Marked Pedagogies: Examining Linguistic Biases in Personalized Automated Writing Feedback," was conducted by researchers Mei Tan, Lena Phalen, and Dorottya Demszky of the Stanford Graduate School of Education. By isolating the variables of student identity while keeping the quality of writing constant, the researchers uncovered a systematic pattern of "feedback withholding" and "positive feedback bias" that could hinder the long-term development of marginalized students.
Experimental Methodology and the Identity Variable
To investigate the presence of bias, the Stanford team utilized a dataset of 600 argumentative essays written by middle school students. These essays focused on two specific prompts: whether schools should mandate community service and the validity of a scientific theory regarding a specific hill formation on Mars. These topics were selected from a collection of student writing frequently used in educational research to ensure a representative sample of adolescent writing styles and capabilities.
The researchers submitted these essays to four prominent large language models (LLMs) commonly used in educational software. The experiment’s core innovation lay in its iterative approach: each individual essay was submitted to the AI models 12 separate times. In each instance, the researchers provided a different descriptive "persona" for the student author. These personas identified the writer across several categories, including race (Black, white, Hispanic, or Asian), gender (male or female), motivation level (highly motivated or unmotivated), and the presence of learning disabilities.
By keeping the text of the essay identical but changing the student’s profile, the researchers were able to isolate the AI’s reaction to the student’s identity. The results revealed that the AI models did not treat the writing as a static object of analysis; instead, they adjusted their tone, expectations, and pedagogical strategies based on the labels provided.

Findings: Racial and Ethnic Divergence in Feedback
The most striking disparities emerged in how AI models addressed students of different racial backgrounds. When an essay was attributed to a Black student, the AI models were significantly more likely to provide praise and encouragement rather than constructive criticism. The feedback often focused on the student’s "voice" or "personal power," using phrases such as, "Your personal story is powerful! Adding more about how your experiences can connect with others could make this even stronger."
While such comments appear supportive, they often lacked the technical rigor found in the feedback given to white students. When the exact same essays were labeled as being written by white students, the AI shifted its focus toward the mechanics of argumentation. The feedback for white students emphasized structure, the quality of evidence, and the clarity of logic—the specific types of "growth-oriented" critiques that push a writer to improve their intellectual output.
In contrast, essays attributed to Hispanic students or English Language Learners (ELL) were frequently met with a disproportionate focus on grammar and "proper" English usage. Rather than engaging with the student’s ideas or the strength of their argument, the AI models defaulted to a corrective mode, prioritizing linguistic conformity over intellectual development. This suggests a "deficit-based" approach to minority students, where the AI prioritizes fixing perceived flaws rather than fostering advanced rhetorical skills.
Gendered Language and Motivation Biases
The study also identified clear gender-based patterns in AI communication. Female students were consistently addressed with more affectionate language and a higher frequency of first-person pronouns, such as "I love your confidence!" or "I can see how much you care about this topic." This "softer" tone, while seemingly friendly, contrasts with the more direct and professional tone typically directed at male personas.
Furthermore, the AI’s "pedagogical" approach shifted based on the student’s described motivation level. Students labeled as "unmotivated" received upbeat, highly encouraging feedback designed to boost confidence. Conversely, students described as "high-achieving" or "highly motivated" were met with direct, critical suggestions aimed at refining their work for a higher standard.
The researchers characterized this as a form of "feedback withholding." By shielding certain groups—such as Black students or those perceived as unmotivated—from rigorous criticism, the AI may be inadvertently lowering the bar for their achievement. This phenomenon mirrors the "soft bigotry of low expectations" sometimes observed in human classrooms, where teachers avoid harsh critiques of marginalized students to avoid appearing biased or discouraging, ultimately resulting in those students receiving less rigorous instruction.

Chronology of AI Integration and Research
The Stanford study arrives at a critical juncture in the history of educational technology. The timeline of this issue traces back to the late 2022 release of ChatGPT, which sparked a gold rush in the "EdTech" sector.
- November 2022: OpenAI releases ChatGPT, leading to immediate concerns regarding plagiarism and the future of the essay as an assessment tool.
- Early 2023: Major educational platforms begin integrating LLMs to provide "instant feedback" to students, promising to reduce teacher workloads.
- Late 2023 – Early 2024: Research begins to emerge regarding general algorithmic bias in AI, specifically in hiring and healthcare. Stanford researchers begin designing a study to see if these biases translate to the K-12 classroom.
- April 2026: The Stanford study is finalized and nominated for "Best Paper" at the 16th International Learning Analytics and Knowledge Conference in Norway.
- April 30, 2026: The findings are officially presented to the global academic community, prompting calls for stricter regulation of AI in schools.
Supporting Data: The Top 20 Words
To quantify these biases, the researchers analyzed the top 20 statistically significant words used by AI models for different demographics. The data revealed that:
- Black Students: Received words like "powerful," "leadership," "connect," and "encourage."
- White Students: Received words like "structure," "evidence," "clarity," "refine," and "strengthen."
- Female Students: Received words like "love," "confidence," "wonderful," and "personal."
- Hispanic/ELL Students: Received words like "grammar," "proper," "correct," and "standard."
These word choices indicate that AI is not just evaluating writing; it is performing a "marked pedagogy," where the identity of the student dictates the instruction style.
Expert Reactions and Official Responses
The educational community has reacted to the Stanford findings with a mixture of concern and caution. Mei Tan, the lead author of the study, emphasized that the AI is not creating these biases in a vacuum. "They are picking up on the biases that humans exhibit," Tan stated. Because LLMs are trained on vast repositories of human text—including historical literature, news, and internet forums—they absorb and replicate the societal stereotypes embedded in that language.
Tanya Baker, Executive Director of the National Writing Project, expressed deep concern regarding the long-term impact on student growth. Following a presentation of the study, Baker noted that if Black and Hispanic students are consistently given praise instead of critique, they are essentially being "denied the push to learn." She argued that while encouragement is vital, it cannot replace the rigorous feedback necessary to master complex writing.
Advocates for "culturally responsive teaching" have pointed out a difficult paradox. While acknowledging a student’s identity is generally considered a best practice in education, the Stanford study shows that when an AI does it, the result is often stereotyping rather than genuine support.

Broader Impact and Implications for the Future
The implications of this research extend far beyond the English classroom. As AI becomes embedded in school databases, the risk of "invisible bias" grows. Many modern learning management systems (LMS) already store data on a student’s race, socioeconomic status, and prior test scores. If an AI feedback tool is granted access to this data to "personalize" instruction, it may automatically apply these biased pedagogical filters without the teacher or student ever realizing it.
Furthermore, the study challenges the primary selling point of AI in education: efficiency. Many districts have adopted AI feedback tools because they provide instantaneous responses, allowing students to revise their work in real-time. However, if these responses are fundamentally biased, their speed becomes a liability rather than an asset.
The Stanford researchers suggest a "human-in-the-loop" approach, where teachers review AI-generated feedback before it reaches the student. However, this solution undermines the time-saving benefits that make AI attractive to overworked educators.
As the 16th International Learning Analytics and Knowledge Conference concludes, the consensus among experts is that the "neutrality" of AI is a myth. For AI to become a truly equitable tool in the classroom, developers must find ways to "de-bias" the training data or implement strict pedagogical guardrails that ensure every student—regardless of their background—is pushed to achieve the same high standards of excellence. Without such interventions, the "digital divide" may evolve from a lack of access to technology into a lack of access to rigorous, unbiased instruction.









Leave a Reply