On Training Sample Memorization: Lessons from Benchmarking Generative Modeling with a Large-scale Competition

被引:16
|
作者
Bai, Ching-Yuan [1 ]
Lin, Hsuan-Tien [1 ]
Raffel, Colin [2 ]
Kan, Wendy Chih-wen [2 ]
机构
[1] Natl Taiwan Univ, Comp Sci & Informat Engn, Taipei, Taiwan
[2] Google, Mountain View, CA 94043 USA
关键词
benchmark; competition; neural networks; generative models; memorization; datasets; computer vision;
D O I
10.1145/3447548.3467198
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many recent developments on generative models for natural images have relied on heuristically-motivated metrics that can be easily gamed by memorizing a small sample from the true distribution or training a model directly to improve the metric. In this work, we critically evaluate the gameability of these metrics by designing and deploying a generative modeling competition. Our competition received over 11000 submitted models. The competitiveness between participants allowed us to investigate both intentional and unintentional memorization in generative modeling. To detect intentional memorization, we propose the "Memorization-Informed Frechet Inception Distance" (MiFID) as a new memorization-aware metric and design benchmark procedures to ensure that winning submissions made genuine improvements in perceptual quality. Furthermore, we manually inspect the code for the 1000 top-performing models to understand and label different forms of memorization. Our analysis reveals that unintentional memorization is a serious and common issue in popular generative models. The generated images and our memorization labels of those models as well as code to compute MiFID are released to facilitate future studies on benchmarking generative models.
引用
收藏
页码:2534 / 2542
页数:9
相关论文
共 50 条
  • [41] Lessons Learned: Large-Scale Perfused Cadaver Training in Three Different Curricular Environments
    Koo, Alex Y.
    Rodgers, David K.
    Hohman, Marc H.
    Muise, Jason R.
    Couperus, Kyle S.
    Phelps, Jillian F.
    MILITARY MEDICINE, 2024, 189 (9-10) : e1871 - e1878
  • [42] Consistency checking in an infrastructure for large-scale generative programming
    Rauschmayer, A
    Knapp, A
    Wirsing, M
    19TH INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, PROCEEDINGS, 2004, : 238 - 247
  • [43] GENERATIVE MODELS FOR LARGE-SCALE SIMULATIONS OF CONNECTOME DEVELOPMENT
    Brooks, Skylar J.
    Stamoulis, Catherine
    2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [44] Will Large-scale Generative Models Corrupt Future Datasets?
    Hataya, Ryuichiro
    Bao, Han
    Arai, Hiromi
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 20498 - 20508
  • [45] Generative models and abstractions for large-scale neuroanatomy datasets
    Rolnick, David
    Dyer, Eva L.
    CURRENT OPINION IN NEUROBIOLOGY, 2019, 55 : 112 - 120
  • [46] Generative pretraining from large-scale transcriptomes for single-cell deciphering
    Shen, Hongru
    Liu, Jilei
    Hu, Jiani
    Shen, Xilin
    Zhang, Chao
    Wu, Dan
    Feng, Mengyao
    Yang, Meng
    Li, Yang
    Yang, Yichen
    Wang, Wei
    Zhang, Qiang
    Yang, Jilong
    Chen, Kexin
    Li, Xiangchun
    ISCIENCE, 2023, 26 (05)
  • [47] Competitive Benchmarking: Lessons Learned from the Trading Agent Competition
    Ketter, Wolfgang
    Symeonidis, Andreas L.
    AI MAGAZINE, 2012, 33 (02) : 103 - 107
  • [48] THE DESIGN OF LARGE-SCALE TRAINING GAMES
    HARTLEY, DA
    RITCHIE, GN
    FITZSIMONS, EA
    SIMULATION & GAMING, 1981, 12 (02) : 141 - 152
  • [49] Evaluating large-scale training simulations
    Simpson, H
    Oser, RL
    MILITARY PSYCHOLOGY, 2003, 15 (01) : 25 - 40
  • [50] Large-scale disasters: Leadership and management lessons
    Department of Civil and Environmental Engineering, Colorado State University, Fort Collins, United States
    Leadersh. Manage. Eng., 3 (97-100):