Learning Fair Representations through Uniformly Distributed Sensitive Attributes

被引:1
|
作者
Kenfack, Patrik Joslin [1 ]
Rivera, Adin Ramirez [3 ]
Khan, Adil Mehmood [1 ,4 ]
Mazzara, Manuel [2 ]
机构
[1] Innopolis Univ, Machine Learning & Knowledge Representat Lab, Innopolis, Russia
[2] Innopolis Univ, Inst Software Dev & Engn, Innopolis 420500, Russia
[3] Univ Oslo, Dept Informat, Digital Signal Proc & Image Anal DSB Grp, N-0373 Oslo, Norway
[4] Univ Hull, Sch Comp Sci, Kingston Upon Hull HU6 7RX, England
关键词
Fairness; Fair representation; Bias; Decision making;
D O I
10.1109/SaTML54575.2023.00014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine Learning (ML) models trained on biased data can reproduce and even amplify these biases. Since such models are deployed to make decisions that can affect people's lives, ensuring their fairness is critical. One approach to mitigate possible unfairness of ML models is to map the input data into a less-biased new space by means of training the model on fair representations. Several methods based on adversarial learning have been proposed to learn fair representation by fooling an adversary in predicting the sensitive attribute (e.g., gender or race). However, adversarial-based learning can be too difficult to optimize in practice; besides, it penalizes the utility of the representation. Hence, in this research effort we train bias-free representations from the input data by inducing a uniform distribution over the sensitive attributes in the latent space. In particular, we propose a probabilistic framework that learns these representations by enforcing the correct reconstruction of the original data, plus the prediction of the attributes of interest while eliminating the possibility of predicting the sensitive ones. Our method leverages the inability of Deep Neural Networks (DNNs) to generalize when trained on a noisy label space to regularize the latent space. We use a network head that predicts a noisy version of the sensitive attributes in order to increase the uncertainty of their predictions at test time. Our experiments in two datasets demonstrated that the proposed model significantly improves fairness while maintaining the prediction accuracy of downstream tasks.
引用
收藏
页码:58 / 67
页数:10
相关论文
共 50 条
  • [1] Learning fair models without sensitive attributes: A generative approach
    Zhu, Huaisheng
    Dai, Enyan
    Liu, Hui
    Wang, Suhang
    NEUROCOMPUTING, 2023, 561
  • [2] Fair Supervised Learning with A Simple Random Sampler of Sensitive Attributes
    Sohn, Jinwon
    Song, Qifan
    Lin, Guang
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [3] Learning Smooth and Fair Representations
    Gitiaux, Xavier
    Rangwala, Huzefa
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130 : 253 - +
  • [4] Learning fair models and representations
    Oneto, Luca
    INTELLIGENZA ARTIFICIALE, 2020, 14 (01) : 151 - 178
  • [5] Learning Controllable Fair Representations
    Song, Jiaming
    Kalluri, Pratyusha
    Grover, Aditya
    Zhao, Shengjia
    Ermon, Stefano
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [6] A Sequentially Fair Mechanism for Multiple Sensitive Attributes
    Hu, Francois
    Ratz, Philipp
    Charpentier, Arthur
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 11, 2024, : 12502 - 12510
  • [7] The Crucial Role of Sensitive Attributes in Fair Classification
    Haeri, Maryam Amir
    Zweig, Katharina Anna
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 2993 - 3002
  • [8] On Fairness of Medical Image Classification with Multiple Sensitive Attributes via Learning Orthogonal Representations
    Deng, Wenlong
    Zhong, Yuan
    Dou, Qi
    Li, Xiaoxiao
    INFORMATION PROCESSING IN MEDICAL IMAGING, IPMI 2023, 2023, 13939 : 158 - 169
  • [9] Shielded Representations: Protecting Sensitive Attributes Through Iterative Gradient-Based Projection
    Iskander, Shadi
    Radinsky, Kira
    Belinkov, Yonatan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5961 - 5977
  • [10] Inherent Tradeoffs in Learning Fair Representations
    Zhao, Han
    Gordon, Geoffrey J.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32