Event Information
Can AI Level the Playing Field?
Introduction & Background:
I'll start by introducing my research question: Can AI help us build a more equitable society? I’ll briefly explain what large language models (LLMs) are, share what past research has shown about their potential to reinforce bias, and highlight why it's important to examine whether today’s AI systems are advancing or undermining fairness.
Methodology:
I’ll walk through how I designed two experiments using 20 fictional student profiles with identical academic transcripts, varying only race and gender, and analyzed over 300 responses from four major LLMs to assess how they responded to socio-demographic differences.
Key Results:
I’ll discuss my key findings showing that newer models exhibited a marked shift toward fairness, and in some cases, even signs of overcorrection.
a. Newer LLMs showed statistically significant reductions in gender bias and improved output consistency compared to older models.
b. African-American female students were, on average, recommended higher-quality community colleges than their white peers, suggesting a possible overcorrectionーintentional or otherwiseーthrough interventions aimed at addressing historical inequalities.
c. Gender bias in career recommendations (measured via weighted salary) was significant in older models like ChatGPT-3.5 but was largely eliminated in newer LLMs.
d. Across both experiments, newer models appeared more fair and consistent.
Conclusion:
In the final part of my presentation, I’ll reflect on how my results show signs of progress toward fairness in AI, and what that could mean for using AI to promote equity, especially in education policy. I’ll also highlight the importance of continued experimentation as models evolve, and discuss the ethical implications of bias correction—how we guide AI to be fair without introducing new problems.
Q&A:
Finally, I’ll open the floor for discussion and questions from the audience.
Guo, Y., Guo, M., Su, J., Yang, Z., Zhu, M., Li, H., Qiu, M., & Liu, S. S. (2024, November 16). Bias in large language models: origin, evaluation, and mitigation. arXiv.org. https://arxiv.org/abs/2411.10915?utm_source
Co-Intelligence: Living and Working with AI by Ethan Mollick