Home /  Workshop /  Schedules /  Estimating and Controlling for Fairness via Sensitive Attribute Predictors

Estimating and Controlling for Fairness via Sensitive Attribute Predictors

Introductory Workshop: Algorithms, Fairness, and Equity August 28, 2023 - September 01, 2023

August 30, 2023 (03:00 PM PDT - 04:00 PM PDT)
Speaker(s): Jeremias Sulam (Johns Hopkins University)
Location: SLMath: Eisenbud Auditorium, Online/Virtual
Tags/Keywords
  • Fairness

  • Proxy Attributes

  • Missing Data

  • computer science

  • Artificial Intelligence

Primary Mathematics Subject Classification No Primary AMS MSC
Secondary Mathematics Subject Classification No Secondary AMS MSC
Video

Estimating and Controlling for Fairness via Sensitive Attribute Predictors

Abstract

As the use of machine learning models in real world high-stakes decision settings continues to grow, it is highly important that we are able to audit and control for any potential fairness violations these models may exhibit towards certain groups. To do so, one naturally requires access to sensitive attributes, such as demographics, gender, or other potentially sensitive features that determine group membership. Unfortunately, in many settings, this information is often unavailable. In this talk, I will present recent work centering on the well known equalized odds (EOD) definition of fairness. In a setting without sensitive attributes, we first provide tight and computable upper bounds for the EOD violation of a predictor, precisely reflect the worst possible EOD violation. Second, we demonstrate how one can provably control the worst-case EOD by a new post-processing correction method. Our results characterize when directly controlling for EOD with respect to the predicted sensitive attributes is -- and when is not -- optimal when it comes to controlling worst-case EOD. Our results hold under assumptions that are milder than previous works, and we illustrate these results with experiments on synthetic and real datasets. Time permitting, I will also present recent results on interpretability of machine learning models, linking common notions of feature importance to well-understood and traditional statistical tests.

Supplements No Notes/Supplements Uploaded
Video/Audio Files

Estimating and Controlling for Fairness via Sensitive Attribute Predictors

Troubles with video?

Please report video problems to itsupport@slmath.org.

See more of our Streaming videos on our main VMath Videos page.