Effective Cross-Region Courier-Displacement for Instant Delivery via Reinforcement Learning

被引:7
|
作者
Hu, Shijie [1 ]
Guo, Baoshen [1 ]
Wang, Shuai [1 ]
Zhou, Xiaolei [1 ,2 ]
机构
[1] Southeast Univ, Sch Comp Sci & Technol, Nanjing, Peoples R China
[2] Natl Univ Def Technol, Res Inst 63, Zunyi, Guizhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Courier displacement; Reinforcement learning; Instant delivery;
D O I
10.1007/978-3-030-85928-2_23
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the rapid development of mobile phones and the Internet of Things, instant delivery services (e.g., UberEats and MeiTuan) have become a popular choice for people to order foods, fruits, and other groceries online, especially after the impact of COVID-19. In instant delivery services, it is important to dispatch massive orders to limited couriers, especially in rush hours. To meet this need, an efficient courier displacement mechanism not only can balance the demand (picking up orders) and supply (couriers' capacity) but also improve the efficiency of order delivery by reducing idle displacing time. Existing studies on fleet management of rider-sharing or bike rebalancing cannot apply to courier displacement problems in instant delivery due to unique practical factors of instant delivery including region difference and strict delivery time constraints. In this work, we propose an efficient cross-region courier displacement method Courier Displacement Reinforcement Learning (short for CDRL), based on multi-agent actor-critic, considering the dynamic demand and supply at the region level and strict time constraints. Specifically, the multi-agent actor-critic reinforcement learning-based courier displacement framework utilizes a policy network to generate displacement decisions considering multiple practical factors and designs a value network to evaluate decisions of the policy network. One month of real-world order records data-set of Shanghai collecting from Eleme (i.e., one of the biggest instant delivery services in China) are utilized in the evaluation and the results show that our method offering up to 36% increase in courier displacement performance and reduce idle ride time by 17%.
引用
收藏
页码:288 / 300
页数:13
相关论文
共 47 条
  • [21] Heterogeneous Attentions for Solving Pickup and Delivery Problem via Deep Reinforcement Learning
    Li, Jingwen
    Xin, Liang
    Cao, Zhiguang
    Lim, Andrew
    Song, Wen
    Zhang, Jie
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (03) : 2306 - 2315
  • [22] Root Cause Attribution of Delivery Risks via Causal Discovery with Reinforcement Learning
    Bo, Shi
    Xiao, Minheng
    ALGORITHMS, 2024, 17 (11)
  • [23] Effective Automated Feature Derivation via Reinforcement Learning for Microcredit Default Prediction
    Song, Mengnan
    Wang, Jiasong
    Zhang, Tongtong
    Zhang, Guoguang
    Zhang, Ruijun
    Su, Suisui
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [24] Value Enhancement of Reinforcement Learning via Efficient and Robust Trust Region Optimization
    Shi, Chengchun
    Qi, Zhengling
    Wang, Jianing
    Zhou, Fan
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2024, 119 (547) : 2011 - 2025
  • [25] Cross-data Automatic Feature Engineering via Meta-learning and Reinforcement Learning
    Zhang, Jianyu
    Hao, Jianye
    Fogelman-Soulie, Francoise
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2020, PT I, 2020, 12084 : 818 - 829
  • [26] Order dispatching for an ultra-fast delivery service via deep reinforcement learning
    Eray Mert Kavuk
    Ayse Tosun
    Mucahit Cevik
    Aysun Bozanta
    Sibel B. Sonuç
    Mehmetcan Tutuncu
    Bilgin Kosucu
    Ayse Basar
    Applied Intelligence, 2022, 52 : 4274 - 4299
  • [27] Order dispatching for an ultra-fast delivery service via deep reinforcement learning
    Kavuk, Eray Mert
    Tosun, Ayse
    Cevik, Mucahit
    Bozanta, Aysun
    Sonuc, Sibel B.
    Tutuncu, Mehmetcan
    Kosucu, Bilgin
    Basar, Ayse
    APPLIED INTELLIGENCE, 2022, 52 (04) : 4274 - 4299
  • [28] Effective, Platform-Independent GUI Testing via Image Embedding and Reinforcement Learning
    Yu, Shengcheng
    Fang, Chunrong
    Li, Xin
    Ling, Yuchen
    Chen, Zhenyu
    Su, Zhendong
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (07)
  • [29] AUTOMATIC DATA AUGMENTATION VIA DEEP REINFORCEMENT LEARNING FOR EFFECTIVE KIDNEY TUMOR SEGMENTATION
    Qin, Tiexin
    Wang, Ziyuan
    He, Kelei
    Shi, Yinghuan
    Gao, Yang
    Shen, Dinggang
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 1419 - 1423
  • [30] Multi-Task Reinforcement Learning in Reproducing Kernel Hilbert Spaces via Cross-Learning
    Cervino, Juan
    Bazerque, Juan Andres
    Calvo-Fullana, Miguel
    Ribeiro, Alejandro
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 5947 - 5962