共 50 条
- [1] Accelerating DNN Inference with GraphBLAS and the GPU 2019 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2019,
- [2] Virtual PIM: Resource-aware Dynamic DPU Allocation and Workload Scheduling Framework for Multi-DPU PIM Architecture 2023 32ND INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES, PACT, 2023, : 112 - 123
- [4] Accelerating DNN Inference by Edge-Cloud Collaboration 2021 IEEE INTERNATIONAL PERFORMANCE, COMPUTING, AND COMMUNICATIONS CONFERENCE (IPCCC), 2021,
- [5] Irina: Accelerating DNN Inference with Efficient Online Scheduling PROCEEDINGS OF 2020 4TH ASIA-PACIFIC WORKSHOP ON NETWORKING, APNET 2020, 2020, : 36 - 43
- [6] Multi-exit DNN inference acceleration for intelligent terminal with heterogeneous processors SUSTAINABLE COMPUTING-INFORMATICS & SYSTEMS, 2023, 40
- [7] HarmonyBatch: Batching multi-SLO DNN Inference with Heterogeneous Serverless Functions 2024 IEEE/ACM 32ND INTERNATIONAL SYMPOSIUM ON QUALITY OF SERVICE, IWQOS, 2024,
- [8] EdgeSP: Scalable Multi-device Parallel DNN Inference on Heterogeneous Edge Clusters ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2021, PT II, 2022, 13156 : 317 - 333