共 50 条
- [41] AD-KD: Attribution-Driven Knowledge Distillation for Language Model Compression PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 8449 - 8465
- [42] HRKD: Hierarchical Relational Knowledge Distillation for Cross-domain Language Model Compression 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3126 - 3136
- [44] Knowledge Distillation via Information Matching NEURAL INFORMATION PROCESSING, ICONIP 2023, PT IV, 2024, 14450 : 405 - 417
- [46] Lifelong Compression Mixture Model via Knowledge Relationship Graph THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10900 - 10908
- [47] Knowledge Distillation for Sequence Model 19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3703 - 3707
- [48] Combining Weight Pruning and Knowledge Distillation For CNN Compression 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3185 - 3192
- [49] Few Sample Knowledge Distillation for Efficient Network Compression 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, : 14627 - 14635
- [50] Differentially Private Model Compression ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,