共 50 条
- [1] Large Language Models (LLMs) Inference Offloading and Resource Allocation in Cloud-Edge Networks: An Active Inference Approach 2023 IEEE 98TH VEHICULAR TECHNOLOGY CONFERENCE, VTC2023-FALL, 2023,
- [2] Incentive-driven Computation Offloading and Resource Allocation in Mobile Cloud-Edge Computing 2022 IEEE 42ND INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS WORKSHOPS (ICDCSW), 2022, : 157 - 162
- [3] Generative Inference of Large Language Models in Edge Computing: An Energy Efficient Approach 20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024, 2024, : 244 - 249
- [7] Adaptive Data Sharing and Computation Offloading in Cloud-Edge Computing with Resource Constraints 2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 2842 - 2849
- [8] Energy-Efficient Cloud-Edge Collaborative Computing: Joint Task Offloading, Resource Allocation, and Service Caching ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT V, ICIC 2024, 2024, 14879 : 285 - 296