共 50 条
- [31] AD-KD: Attribution-Driven Knowledge Distillation for Language Model Compression PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 8449 - 8465
- [32] HRKD: Hierarchical Relational Knowledge Distillation for Cross-domain Language Model Compression 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3126 - 3136
- [33] Knowledge Distillation Based on Pruned Model BLOCKCHAIN AND TRUSTWORTHY SYSTEMS, BLOCKSYS 2019, 2020, 1156 : 598 - 603
- [34] On-Demand Deep Model Compression for Mobile Devices: A Usage-Driven Model Selection Framework MOBISYS'18: PROCEEDINGS OF THE 16TH ACM INTERNATIONAL CONFERENCE ON MOBILE SYSTEMS, APPLICATIONS, AND SERVICES, 2018, : 389 - 400
- [37] Data-Free Ensemble Knowledge Distillation for Privacy-conscious Multimedia Model Compression PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 1803 - 1811
- [39] Efficient Knowledge Distillation from Model Checkpoints ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
- [40] Contrastive Distillation on Intermediate Representations for Language Model Compression PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 498 - 508