Computer Science Department , PhD Dissertation Defense, Mallak Alkhathlan " Group Fairness of Ranking-Based Decisions: Do Metrics Reflect Human Perception?"

Tuesday, February 18, 2025
12:00 pm to 1:00 pm

Mallak Alkhathlan

 Ph.D. Candidate

WPI – Computer Science Department 

 

Tuesday, February 18, 2025

Time: 12:00PM - 1:00PM

Location: Gordon Library Conference Room

  

Dissertation Committee:

Dr. Elke Rundensteiner, Professor, WPI (Advisor)

Dr. Lane Harrison, Associate Professor, WPI (Co-Advisor)

Dr. Fabricio Murai, Assistant Professor, WPI

Dr. Mahsan Nourani, Assistant Professor, Northeastern University

 As automated decision-making systems increasingly influence critical aspects of daily life, concerns have grown over their potential to perpetuate biases and disproportionately harm marginalized groups. Ranking-based decision-making, widely used in hiring, university admissions, and bail decisions, can embed disparities due to reliance on historical data and predefined criteria. While fairness metrics aim to mitigate bias, a fundamental challenge remains: do these metrics align with human perceptions of fairness? This dissertation investigates this alignment through HCI studies on group fairness, ultimately guiding the design of ranking systems that are not only mathematically fair but also widely accepted.  

 This dissertation is structured into three parts. First, we study how ranking fairness metrics align with people’s subjective perceptions of fairness. 480 participants evaluated a scholarship distribution across 12 conditions, varying by group sensitivity, size, and performance scores. Results show a strong reliance on explicit score values, with group sensitivity influencing perceptions, particularly in unbalanced groups. Fairness perceptions were more consistent when candidates had similar grades, regardless of group sensitivity or size. Second, using JND, the minimum detectable difference between stimuli, and 2AFC methods, we address the limitations of Likert-scale studies in assessing fairness perceptions. With 224 participants and 170,000 judgments, we find that longer lists lead to more consistent and precise fairness judgments, revealing cases where fairness metrics diverge from human intuition. Third, a thematic analysis of 6,000 crowdsourced responses revealed concerns about merit-based discontent and transparency in socio-economic adjustments. Some participants relied on cognitive strategies and visual cues to assess fairness. These findings highlight the challenges of aligning fairness metrics with people’s intuition and the need for fairness-aware ranking methods that balance quantitative metrics with qualitative perceptions.

Audience(s)

Department(s):

Computer Science