Algorithms (Mar 2021)

Local Data Debiasing for Fairness Based on Generative Adversarial Training

  • Ulrich Aïvodji,
  • François Bidet,
  • Sébastien Gambs,
  • Rosin Claude Ngueveu,
  • Alain Tapp

DOI
https://doi.org/10.3390/a14030087
Journal volume & issue
Vol. 14, no. 3
p. 87

Abstract

Read online

The widespread use of automated decision processes in many areas of our society raises serious ethical issues with respect to the fairness of the process and the possible resulting discrimination. To solve this issue, we propose a novel adversarial training approach called GANSan for learning a sanitizer whose objective is to prevent the possibility of any discrimination (i.e., direct and indirect) based on a sensitive attribute by removing the attribute itself as well as the existing correlations with the remaining attributes. Our method GANSan is partially inspired by the powerful framework of generative adversarial networks (in particular Cycle-GANs), which offers a flexible way to learn a distribution empirically or to translate between two different distributions. In contrast to prior work, one of the strengths of our approach is that the sanitization is performed in the same space as the original data by only modifying the other attributes as little as possible, thus preserving the interpretability of the sanitized data. Consequently, once the sanitizer is trained, it can be applied to new data locally by an individual on their profile before releasing it. Finally, experiments on real datasets demonstrate the effectiveness of the approach as well as the achievable trade-off between fairness and utility.

Keywords