共 7 条
- [1] CoFB: latency-constrained co-scheduling of flows and batches for deep learning inference service on the CPU–GPU system The Journal of Supercomputing, 2023, 79 : 14172 - 14199
- [3] Orchestrated Co-scheduling, Resource Partitioning, and Power Capping on CPU-GPU Heterogeneous Systems via Machine Learning ARCHITECTURE OF COMPUTING SYSTEMS, ARCS 2022, 2022, 13642 : 51 - 67
- [4] Demystifying the TensorFlow Eager Execution of Deep Learning Inference on a CPU-GPU Tandem 2022 25TH EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD), 2022, : 446 - 455
- [5] The Best of Many Worlds: Scheduling Machine Learning Inference on CPU-GPU Integrated Architectures 2022 IEEE 36TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW 2022), 2022, : 55 - 64
- [6] Fault-tolerant deep learning inference on CPU-GPU integrated edge devices with TEEs FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 161 : 404 - 414