Lack of Transparency and Potential Bias in Artificial Intelligence Data Sets and Algorithms A Scoping Review

被引:131
|
作者
Daneshjou, Roxana [1 ,2 ]
Smith, Mary P. [3 ]
Sun, Mary D. [4 ]
Rotemberg, Veronica [5 ]
Zou, James [6 ,7 ,8 ]
机构
[1] Stanford Sch Med, Stanford Dept Dermatol, 450 Broadway, Redwood City, CA 94061 USA
[2] Stanford Sch Med, Stanford Dept Biomed Data Sci, Stanford, CA 94305 USA
[3] Mem Sloan Kettering Canc Ctr, Dept Med, 1275 York Ave, New York, NY 10021 USA
[4] Icahn Sch Med Mt Sinai, New York, NY 10029 USA
[5] Mem Sloan Kettering Canc Ctr, Dermatol Serv, 1275 York Ave, New York, NY 10021 USA
[6] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[7] Stanford Univ, Dept Biomed Data Sci, Stanford, CA 94305 USA
[8] Chan Zuckerberg Biohub, San Francisco, CA USA
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
CONVOLUTIONAL NEURAL-NETWORK; SKIN-CANCER; IMAGE CLASSIFICATION; DERMATOLOGISTS; MELANOMA; PERFORMANCE; DIAGNOSIS; TIME; ACCURACY; SUPERIOR;
D O I
10.1001/jamadermatol.2021.3129
中图分类号
R75 [皮肤病学与性病学];
学科分类号
100206 ;
摘要
IMPORTANCE Clinical artificial intelligence (AI) algorithms have the potential to improve clinical care, but fair, generalizable algorithms depend on the clinical data on which they are trained and tested. OBJECTIVE To assess whether data sets used for training diagnostic AI algorithms addressing skin disease are adequately described and to identify potential sources of bias in these data sets. DATA SOURCES In this scoping review, PubMed was used to search for peer-reviewed research articles published between January 1, 2015, and November 1, 2020, with the following paired search terms: deep learning and dermatology, artificial intelligence and dermatology, deep learning and dermatologist, and artificial intelligence and dermatologist. STUDY SELECTION Studies that developed or tested an existing deep learning algorithm for triage, diagnosis, or monitoring using clinical or dermoscopic images of skin disease were selected, and the articles were independently reviewed by 2 investigators to verify that they met selection criteria. CONSENSUS PROCESS Data set audit criteria were determined by consensus of all authors after reviewing existing literature to highlight data set transparency and sources of bias. RESULTS A total of 70 unique studies were included. Among these studies, 1 065 291 images were used to develop or test AI algorithms, of which only 257 372 (24.2%) were publicly available. Only 14 studies (20.0%) included descriptions of patient ethnicity or race in at least 1 data set used. Only 7 studies (10.0%) included any information about skin tone in at least 1 data set used. Thirty-six of the 56 studies developing new AI algorithms for cutaneous malignant neoplasms (64.3%) met the gold standard criteria for disease labeling. Public data sets were cited more often than private data sets, suggesting that public data sets contribute more to new development and benchmarks. CONCLUSIONS AND RELEVANCE This scoping review identified 3 issues in data sets that are used to develop and test clinical AI algorithms for skin disease that should be addressed before clinical translation: (1) sparsity of data set characterization and lack of transparency, (2) nonstandard and unverified disease labels, and (3) inability to fully assess patient diversity used for algorithm development and testing.
引用
收藏
页码:1362 / 1369
页数:8
相关论文
共 50 条
  • [1] Transparency in Artificial Intelligence Reporting in Ophthalmology-A Scoping Review
    Chen, Dinah
    Geevarghese, Alexi
    Lee, Samuel
    Plovnick, Caitlin
    Elgin, Cansu
    Zhou, Raymond
    Oermann, Eric
    Aphinyonaphongs, Yindalon
    Al-Aswad, Lama A.
    OPHTHALMOLOGY SCIENCE, 2024, 4 (04):
  • [2] Age-related bias and artificial intelligence: a scoping review
    Charlene H. Chu
    Simon Donato-Woodger
    Shehroz S. Khan
    Rune Nyrup
    Kathleen Leslie
    Alexandra Lyn
    Tianyu Shi
    Andria Bianchi
    Samira Abbasgholizadeh Rahimi
    Amanda Grenier
    Humanities and Social Sciences Communications, 10
  • [3] Age-related bias and artificial intelligence: a scoping review
    Chu, Charlene H.
    Donato-Woodger, Simon
    Khan, Shehroz S.
    Nyrup, Rune
    Leslie, Kathleen
    Lyn, Alexandra
    Shi, Tianyu
    Bianchi, Andria
    Rahimi, Samira Abbasgholizadeh
    Grenier, Amanda
    HUMANITIES & SOCIAL SCIENCES COMMUNICATIONS, 2023, 10 (01):
  • [4] Openness and Transparency in the Evaluation of Bias in Artificial Intelligence
    Larson, David B.
    RADIOLOGY, 2023, 306 (02)
  • [5] Transparency and Representation in Clinical Research Utilizing Artificial Intelligence in Oncology: A Scoping Review
    D'Amiano, Anjali J.
    Cheunkarndee, Tia
    Azoba, Chinenye
    Chen, Krista Y.
    Mak, Raymond H.
    Perni, Subha
    CANCER MEDICINE, 2025, 14 (05):
  • [6] Transparency and representation in clinical research utilizing artificial intelligence in oncology: A scoping review
    D'Amiano, Anjali
    Cheunkarndee, Tia
    Azoba, Chinenye C.
    Bitterman, Danielle Sara
    Mak, Raymond H.
    Perni, Subha
    JOURNAL OF CLINICAL ONCOLOGY, 2022, 40 (28) : 308 - 308
  • [7] Consumer bias against evaluations received by artificial intelligence: the mediation effect of lack of transparency anxiety
    Lopez, Alberto
    Garza, Ricardo
    JOURNAL OF RESEARCH IN INTERACTIVE MARKETING, 2023, 17 (06) : 831 - 847
  • [8] Artificial Intelligence Algorithms for Expert Identification in Medical Domains: A Scoping Review
    Borna, Sahar
    Barry, Barbara A.
    Makarova, Svetlana
    Parte, Yogesh
    Haider, Clifton R.
    Sehgal, Ajai
    Leibovich, Bradley C.
    Forte, Antonio Jorge
    EUROPEAN JOURNAL OF INVESTIGATION IN HEALTH PSYCHOLOGY AND EDUCATION, 2024, 14 (05) : 1182 - 1196
  • [9] The potential of artificial intelligence to improve patient safety: a scoping review
    Bates, David W.
    Levine, David
    Syrowatka, Ania
    Kuznetsova, Masha
    Craig, Kelly Jean Thomas
    Rui, Angela
    Jackson, Gretchen Purcell
    Rhee, Kyu
    NPJ DIGITAL MEDICINE, 2021, 4 (01)
  • [10] The potential of artificial intelligence to improve patient safety: a scoping review
    David W. Bates
    David Levine
    Ania Syrowatka
    Masha Kuznetsova
    Kelly Jean Thomas Craig
    Angela Rui
    Gretchen Purcell Jackson
    Kyu Rhee
    npj Digital Medicine, 4