共 50 条
- [21] Generative Inference of Large Language Models in Edge Computing: An Energy Efficient Approach 20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024, 2024, : 244 - 249
- [22] Steering Large Language Models for Machine Translation with Finetuning and In-Context Learning FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 11127 - 11148
- [25] Inference to the Best Explanation in Large Language Models PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 217 - 235
- [26] Assessing Inference Time in Large Language Models SYSTEM DEPENDABILITY-THEORY AND APPLICATIONS, DEPCOS-RELCOMEX 2024, 2024, 1026 : 296 - 305
- [27] InfiniGen: Efficient Generative Inference of Large Language Models with Dynamic KV Cache Management PROCEEDINGS OF THE 18TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, OSDI 2024, 2024, : 155 - 172
- [28] Large Language Models Need Symbolic AI NEURAL-SYMBOLIC LEARNING AND REASONING 2023, NESY 2023, 2023,
- [29] Debiasing Large Language Models with Structured Knowledge FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 10274 - 10287
- [30] PuMer: Pruning and Merging Tokens for Efficient Vision Language Models PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 12890 - 12903