Understanding and mitigating the impact of race with adversarial autoencoders

被引:0
|
作者
Sarullo, Kathryn [1 ]
Swamidass, S. Joshua [2 ]
机构
[1] Washington Univ, Dept Comp Sci, St. Louis, MO 63130 USA
[2] Washington Univ, Dept Pathol & Immunol, Sch Med St. Louis, St. Louis, MO USA
来源
COMMUNICATIONS MEDICINE | 2024年 / 4卷 / 01期
关键词
D O I
10.1038/s43856-024-00627-3
中图分类号
R-3 [医学研究方法]; R3 [基础医学];
学科分类号
1001 ;
摘要
BackgroundArtificial intelligence carries the risk of exacerbating some of our most challenging societal problems, but it also has the potential of mitigating and addressing these problems. The confounding effects of race on machine learning is an ongoing subject of research. This study aims to mitigate the impact of race on data-derived models, using an adversarial variational autoencoder (AVAE). In this study, race is a self-reported feature. Race is often excluded as an input variable, however, due to the high correlation between race and several other variables, race is still implicitly encoded in the data.MethodsWe propose building a model that (1) learns a low dimensionality latent spaces, (2) employs an adversarial training procedure that ensure its latent space does not encode race, and (3) contains necessary information for reconstructing the data. We train the autoencoder to ensure the latent space does not indirectly encode race.ResultsIn this study, AVAE successfully removes information about race from the latent space (AUC ROC = 0.5). In contrast, latent spaces constructed using other approaches still allow the reconstruction of race with high fidelity. The AVAE's latent space does not encode race but conveys important information required to reconstruct the dataset. Furthermore, the AVAE's latent space does not predict variables related to race (R2 = 0.003), while a model that includes race does (R2 = 0.08).ConclusionsThough we constructed a race-independent latent space, any variable could be similarly controlled. We expect AVAEs are one of many approaches that will be required to effectively manage and understand bias in ML. Computer models used in healthcare can sometimes be biased based on race, leading to unfair outcomes. Our study focuses on understanding and reducing the impact of self-reported race in computer models that learn from data. We use a model called an Adversarial Variational Autoencoder (AVAE), which helps ensure that the models don't accidentally use race in their calculations. The AVAE technique creates a simplified version of the data, called a latent space, that leaves out race information but keeps other important details needed for accurate predictions. Our results show that this approach successfully removes race information from the models while still allowing them to work well. This method is one of many steps needed to address bias in computer learning and ensure fairer outcomes. Our findings highlight the importance of developing tools that can manage and understand bias, contributing to more equitable and trustworthy technology. Sarullo and Swamidass use an adversarial variational autoencoder (AVAE) to remove race information from computer models while retaining essential data for accurate predictions, effectively reducing bias. This approach highlights the importance of developing tools to manage bias, ensuring fairer and more trustworthy technology.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] WaveGuard: Understanding and Mitigating Audio Adversarial Examples
    Hussain, Shehzeen
    Neekhara, Paarth
    Dubnov, Shlomo
    McAuley, Julian
    Koushanfar, Farinaz
    PROCEEDINGS OF THE 30TH USENIX SECURITY SYMPOSIUM, 2021, : 2273 - 2290
  • [2] Robustness of Autoencoders for Anomaly Detection Under Adversarial Impact
    Goodge, Adam
    Hooi, Bryan
    Ng, See Kiong
    Ng, Wee Siong
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 1244 - 1250
  • [3] MITIGATING ADVERSARIAL ATTACKS ON MEDICAL IMAGE UNDERSTANDING SYSTEMS
    Paul, Rahul
    Schabath, Matthew
    Gillies, Robert
    Hall, Lawrence
    Goldgof, Dmitry
    2020 IEEE 17TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2020), 2020, : 1517 - 1521
  • [4] Denoising Adversarial Autoencoders
    Creswell, Antonia
    Bharath, Anil Anthony
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (04) : 968 - 984
  • [5] Towards Understanding and Mitigating Audio Adversarial Examples for Speaker Recognition
    Chen, Guangke
    Zhao, Zhe
    Song, Fu
    Chen, Sen
    Fan, Lingling
    Wang, Feng
    Wang, Jiashui
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2023, 20 (05) : 3970 - 3987
  • [6] Mitigating the impact of adversarial attacks in very deep networks
    Hassanin, Mohammed
    Radwan, Ibrahim
    Moustafa, Nour
    Tahtali, Murat
    Kumar, Neeraj
    APPLIED SOFT COMPUTING, 2021, 105 (105)
  • [7] Dual Adversarial Autoencoders for Clustering
    Ge, Pengfei
    Ren, Chuan-Xian
    Dai, Dao-Qing
    Feng, Jiashi
    Yan, Shuicheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (04) : 1417 - 1424
  • [8] Learning Priors for Adversarial Autoencoders
    Wang, Hui-Po
    Ko, Wei-Jan
    Peng, Wen-Hsiao
    2018 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2018, : 1388 - 1396
  • [9] Learning priors for adversarial autoencoders
    Wang, Hui-Po
    Peng, Wen-Hsiao
    Ko, Wei-Jan
    APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING, 2020, 9
  • [10] Understanding and mitigating impact of immunogenicity on pharmacokinetic assays
    White, Joleen T.
    Golob, Michaela
    Sailstad, Jeffrey
    BIOANALYSIS, 2011, 3 (16) : 1799 - 1803