Learning to Unlearn: Instance-Wise Unlearning for Pre-trained Classifiers

被引:0
|
作者
Cha, Sungmin [1 ]
Cho, Sungjun [2 ]
Hwang, Dasol [2 ]
Lee, Honglak [2 ]
Moon, Taesup [3 ]
Lee, Moontae [2 ,4 ]
机构
[1] New York Univ, New York, NY USA
[2] LG AI Res, Seoul, South Korea
[3] Seoul Natl Univ, INMC, ASRI, Seoul, South Korea
[4] Univ Illinois, Chicago, IL USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since the recent advent of regulations for data protection (e.g., the General Data Protection Regulation), there has been increasing demand in deleting information learned from sensitive data in pre-trained models without retraining from scratch. The inherent vulnerability of neural networks towards adversarial attacks and unfairness also calls for a robust method to remove or correct information in an instancewise fashion, while retaining the predictive performance across remaining data. To this end, we consider instance-wise unlearning, of which the goal is to delete information on a set of instances from a pre-trained model, by either misclassifying each instance away from its original prediction or relabeling the instance to a different label. We also propose two methods that reduce forgetting on the remaining data: 1) utilizing adversarial examples to overcome forgetting at the representation-level and 2) leveraging weight importance metrics to pinpoint network parameters guilty of propagating unwanted information. Both methods only require the pretrained model and data instances to forget, allowing painless application to real-life settings where the entire training set is unavailable. Through extensive experimentation on various image classification benchmarks, we show that our approach effectively preserves knowledge of remaining data while unlearning given instances in both single-task and continual unlearning scenarios.
引用
收藏
页码:11186 / 11194
页数:9
相关论文
共 50 条
  • [41] Learning to Remove: Towards Isotropic Pre-trained BERT Embedding
    Liang, Yuxin
    Cao, Rui
    Zheng, Jie
    Ren, Jie
    Gao, Ling
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2021, PT V, 2021, 12895 : 448 - 459
  • [42] MODEL SPIDER: Learning to Rank Pre-Trained Models Efficiently
    Zhang, Yi-Kai
    Huang, Ting-Ji
    Ding, Yao-Xiang
    Zhan, De-Chuan
    Ye, Han-Jia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [43] Pre-Trained Image Encoder for Generalizable Visual Reinforcement Learning
    Yuan, Zhecheng
    Xue, Zhengrong
    Yuan, Bo
    Wang, Xueqian
    Wu, Yi
    Gao, Yang
    Xu, Huazhe
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [44] Do Pre-trained Models Benefit Equally in Continual Learning?
    Lee, Kuan-Ying
    Zhong, Yuanyi
    Wang, Yu-Xiong
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 6474 - 6482
  • [45] Transfer Learning from Pre-trained BERT for Pronoun Resolution
    Bao, Xingce
    Qiao, Qianqian
    GENDER BIAS IN NATURAL LANGUAGE PROCESSING (GEBNLP 2019), 2019, : 82 - 88
  • [46] TransTailor: Pruning the Pre-trained Model for Improved Transfer Learning
    Liu, Bingyan
    Cai, Yifeng
    Guo, Yao
    Chen, Xiangqun
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8627 - 8634
  • [47] Class-Incremental Learning with Strong Pre-trained Models
    Wu, Tz-Ying
    Swaminathan, Gurumurthy
    Li, Zhizhong
    Ravichandran, Avinash
    Vasconcelos, Nuno
    Bhotika, Rahul
    Soatto, Stefano
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9591 - 9600
  • [48] Pre-trained Online Contrastive Learning for Insurance Fraud Detection
    Zhang, Rui
    Cheng, Dawei
    Yang, Jie
    Ouyang, Yi
    Wu, Xian
    Zheng, Yefeng
    Jiang, Changjun
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 20, 2024, : 22511 - 22519
  • [49] High fidelity deep learning-based MRI reconstruction with instance-wise discriminative feature matching loss
    Wang, Ke
    Tamir, Jonathan, I
    De Goyeneche, Alfredo
    Wollner, Uri
    Brada, Rafi
    Yu, Stella X.
    Lustig, Michael
    MAGNETIC RESONANCE IN MEDICINE, 2022, 88 (01) : 476 - 491
  • [50] Pre-trained combustion model and transfer learning in thermoacoustic instability
    Qin, Ziyu
    Wang, Xinyao
    Han, Xiao
    Lin, Yuzhen
    Zhou, Yuchen
    PHYSICS OF FLUIDS, 2023, 35 (03)