Adversarial Attacks on Fairness of Graph Neural Networks
Binchi Zhang, Yushun Dong, et al.
ICLR 2024
Machine learning (ML) models often perform differently under distribution shifts, in terms of utility, fairness, and other dimensions. We propose the Adversarial Auditor for measuring the utility and fairness performance of ML models under compound shifts of outcome and protected attributes. We use Multi-Objective Bayesian Optimization (MOBO) to account for multiple metrics and identify shifts where model performance is extreme, both good and bad. Using two case studies, we show that MOBO performed better than random and grid-based approaches in identifying scenarios by adversarially optimizing objectives, highlighting the value of such an auditor for developing fair, accurate and shift-robust models.
Binchi Zhang, Yushun Dong, et al.
ICLR 2024
Natalia Martinez Gil, Kanthi Sarpatwar, et al.
NeurIPS 2023
Ching-Yao chuang, Youssef Mroueh
ICLR 2021
Barbara A. Han, Subhabrata Majumdar, et al.
Epidemics