top of page
Search

Objectivity In, Bias Out: How We Taught an AI to Read College Applications


By:

Mihir Gupta and Professor Vidyanand Choudhary


Overview

The college admissions process has long struggled with the challenge of balancing human judgment and fairness. This article explores how a team at the University of California, Irvine, developed an AI-driven, evidence-based admissions evaluation system that brings objectivity, transparency, and scalability to the early stages of application review. Using a Large Language Model (LLM) and a carefully designed deterministic rubric, the system demonstrates how generative AI can reduce bias, ensure consistency, and augment human decision-making — not replace it. The work showcases a powerful proof-of-concept for how educational institutions can ethically and effectively integrate AI into high-stakes decision-making.


The Challenge: From “Magic” to Mechanics

The college admissions process is one of the most consequential and subjective experiences in a young person’s life. Behind every decision are dedicated admissions officers tasked with building a diverse, talented class — all while managing thousands of applications. The key question we sought to answer:


Can AI help ensure every applicant receives a fair, consistent, and evidence-based evaluation?


Large Language Models (LLMs) like GPT-4 are powerful but can be unpredictable. For admissions, we couldn’t rely on a “black box” model. We needed a system that produced the same answer, for the same reasons, every time — turning creative AI into a deterministic analytical tool.


The Three Pillars of the System

1. The Foundation: Real Data for Real Problems

We began by constructing a dataset of authentic student profiles sourced from publicly available forums on Reddit. These profiles provided the messy, unstructured data that mirrors real-world student narratives — ideal for testing robustness and interpretability. This data formed the bedrock for both training and validating our system.


2. The Rulebook: Building a “Mechanically-Strict” Rubric

At the heart of our project lies a 21-criterion evaluation rubric, covering dimensions like academic rigor, extracurricular impact, and personal expression.Creating this rubric required over 20 iterations and careful testing using OpenAI’s o4-mini model as an adversarial reviewer. Each iteration revealed ambiguities, loopholes, and potential biases, which we systematically refined.


The finalized rubric enforces three non-negotiable principles:

  • Evidence Mandate: Every awarded point must be justified by a direct, verbatim quote from the application — no paraphrasing.

  • Sequential Evaluation: Each criterion is scored independently to prevent halo effects.

  • Hybrid Computation: The LLM handles text interpretation, while a Python script performs all calculations for perfect numerical accuracy.

This structure ensures that the evaluation is auditable, repeatable, and bias-resistant.


3. The Engine: Automating the Review Pipeline

The rubric was integrated into a Python-driven automation engine that connects to the GPT-4o API, enabling the concurrent evaluation of hundreds of student profiles. The system’s hybrid design provides both transparency and scalability, with every decision traceable to textual evidence.


Results: Stable, Sensible, and Transparent

Testing produced impressive results:

  • Over 140 repeated evaluations of the same profile yielded a standard deviation of σ ≈ 0.011, proving remarkable stability.

  • “Sensitivity tests” — minor modifications to applications — produced logically consistent score changes, confirming the model was reading, not memorizing.

  • When compared with human expert tags (e.g., “good” vs. “excellent”), the AI’s scoring aligned closely with human judgment, further validating its reliability.

Minor inconsistencies, such as score flickering in ambiguous sections, were logged and addressed, demonstrating the value of transparency in iterative AI system design.


Implications for the Future of Admissions

This initiative is not about replacing human reviewers, but empowering them.By handling the mechanical, evidence-based evaluation, the AI system frees admissions officers to focus on holistic, context-rich aspects of applications — the personal stories and potential that define each student.

In the future, similar systems could be used to:

  • Improve fairness and reduce unconscious bias in admissions.

  • Offer auditable, explainable metrics for institutional accountability.

  • Scale objective evaluation processes across universities globally.

By pairing the tireless objectivity of AI with the nuanced wisdom of human judgment, we move toward an admissions process that is both efficient and equitable.


Call to Action

Through our GradeBot platform (https://gradebot.ai), we are pioneering the application of Generative AI for education, evaluation, and decision support.

If you’re an educator, administrator, or institution interested in building fair, explainable, and auditable AI systems, visit GradeBot.ai to explore our tools for automated grading, rubric-based evaluation, and ethical AI deployment in education.

Join us in shaping a future where AI enhances human judgment — not replaces it.


References (2024–2025)

Improve Quality in Generative AI solutions”, AI Governance Conference, Washington DC, Dec 2024,

 
 
 

Comments


  • Facebook
  • LinkedIn
  • YouTube

© 2025 Applied AI Institute. All rights reserved.

bottom of page