Identifying, Measuring, and Addressing Algorithmic Bias in AI Admission Systems for Graduate Education
Files
TR Number
Date
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Abstract
The number of graduate students has been increasing rapidly to meet industry demands, with over 200% increase in competitive fields like computer science (CS) in the past decade. Several universities have adopted AI in their admissions processes for various tasks such as evaluating transcripts, extracting important information from essays, and scoring applications. While AI can greatly increase the efficiency of processing a large volume of applications, it is prone to data and algorithmic bias, which can lead to unfair outcomes for underprivileged subgroups among applicants. Recent changes in legislation such as the ban of affirmative action by the U.S. Supreme Court make it increasingly relevant to study the demographic composition of admitted students and ensure that we develop fair machine learning systems. We present a comprehensive two-phase methodology for detecting and mitigating algorithmic bias. Through analysis of graduate admissions data of the Computer Science department of a large R1 university, we found significant post-ban demographic shifts, including decreased applications from underrepresented groups and a 66% increase in applicants declining to report race post the affirmative action ban. Our preemptive bias detection phase includes exploratory data analysis, clustering, and subgroup discovery to identify both independent and intersectional sources of bias, revealing significant disparities based on citizenship status, gender and race. We then developed and evaluated a neural network model using fairness metrics, discovering substantial bias amplification for gender and citizenship status. Our fairness evaluation and bias correction phase demonstrated that preprocessing and postprocessing mitigation techniques could significantly improve fairness metrics, though with varying effectiveness across different protected attributes. SHAP analysis confirmed that while academic metrics like GPA remained the strongest predictors, demographic features substantially influenced model decisions even after mitigation. This work provides a systematic framework for institutions seeking to implement fair AI admissions systems while navigating new legal constraints on affirmative action, emphasizing the importance of proactive bias detection and mitigation to maintain diversity in higher education.