Learning to Unlearn: Instance-Wise Unlearning for Pre-trained Classifiers

被引:0
|
作者
Cha, Sungmin [1 ]
Cho, Sungjun [2 ]
Hwang, Dasol [2 ]
Lee, Honglak [2 ]
Moon, Taesup [3 ]
Lee, Moontae [2 ,4 ]
机构
[1] New York Univ, New York, NY USA
[2] LG AI Res, Seoul, South Korea
[3] Seoul Natl Univ, INMC, ASRI, Seoul, South Korea
[4] Univ Illinois, Chicago, IL USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since the recent advent of regulations for data protection (e.g., the General Data Protection Regulation), there has been increasing demand in deleting information learned from sensitive data in pre-trained models without retraining from scratch. The inherent vulnerability of neural networks towards adversarial attacks and unfairness also calls for a robust method to remove or correct information in an instancewise fashion, while retaining the predictive performance across remaining data. To this end, we consider instance-wise unlearning, of which the goal is to delete information on a set of instances from a pre-trained model, by either misclassifying each instance away from its original prediction or relabeling the instance to a different label. We also propose two methods that reduce forgetting on the remaining data: 1) utilizing adversarial examples to overcome forgetting at the representation-level and 2) leveraging weight importance metrics to pinpoint network parameters guilty of propagating unwanted information. Both methods only require the pretrained model and data instances to forget, allowing painless application to real-life settings where the entire training set is unavailable. Through extensive experimentation on various image classification benchmarks, we show that our approach effectively preserves knowledge of remaining data while unlearning given instances in both single-task and continual unlearning scenarios.
引用
收藏
页码:11186 / 11194
页数:9
相关论文
共 50 条
  • [21] Target to Source Coordinate-Wise Adaptation of Pre-trained Models
    Zhang, Luxin
    Germain, Pascal
    Kessaci, Yacine
    Biernacki, Christophe
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2020, PT I, 2021, 12457 : 378 - 394
  • [22] Self-Supervised Video Representation Learning Using Improved Instance-Wise Contrastive Learning and Deep Clustering
    Zhu, Yisheng
    Shuai, Hui
    Liu, Guangcan
    Liu, Qingshan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) : 6741 - 6752
  • [23] Towards Inadequately Pre-trained Models in Transfer Learning
    Deng, Andong
    Li, Xingjian
    Hu, Di
    Wang, Tianyang
    Xiong, Haoyi
    Xu, Cheng-Zhong
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 19340 - 19351
  • [24] PTMA: Pre-trained Model Adaptation for Transfer Learning
    Li, Xiao
    Yan, Junkai
    Jiang, Jianjian
    Zheng, Wei-Shi
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, KSEM 2024, 2024, 14884 : 176 - 188
  • [25] Transfer learning with pre-trained conditional generative models
    Yamaguchi, Shin'ya
    Kanai, Sekitoshi
    Kumagai, Atsutoshi
    Chijiwa, Daiki
    Kashima, Hisashi
    MACHINE LEARNING, 2025, 114 (04)
  • [26] Instance-wise Hard Negative Example Generation for Contrastive Learning in Unpaired Image-to-Image Translation
    Wang, Weilun
    Zhou, Wengang
    Bao, Jianmin
    Chen, Dong
    Li, Houqiang
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 14000 - 14009
  • [27] Instance-aware Dynamic Prompt Tuning for Pre-trained Point Cloud Models
    Zha, Yaohua
    Wang, Jinpeng
    Dai, Tao
    Bin Chen
    Wang, Zhi
    Xia, Shu-Tao
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 14115 - 14124
  • [28] CUE: An Uncertainty Interpretation Framework for Text Classifiers Built on Pre-Trained Language Models
    Li, Jiazheng
    Sun, Zhaoyue
    Liang, Bin
    Gui, Lin
    He, Yulan
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1253 - 1262
  • [29] Pre-trained Classifiers with One Shot Similarity for Context Aware Face Verification and Identification
    Sharma, Monika
    Hebbalaguppe, Ramya
    Vig, Lovekesh
    2017 IEEE INTERNATIONAL CONFERENCE ON IDENTITY, SECURITY AND BEHAVIOR ANALYSIS (ISBA), 2017,
  • [30] Federated Learning from Pre-Trained Models: A Contrastive Learning Approach
    Tan, Yue
    Long, Guodong
    Ma, Jie
    Liu, Lu
    Zhou, Tianyi
    Jiang, Jing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,