Model-Based Offline Reinforcement Learning for Autonomous Delivery of Guidewire

被引:0
|
作者
Li, Hao [1 ]
Zhou, Xiao-Hu [1 ]
Xie, Xiao-Liang [1 ]
Liu, Shi-Qi [1 ]
Feng, Zhen-Qiu [1 ]
Gui, Mei-Jiang [1 ]
Xiang, Tian-Yu [1 ]
Huang, De-Xing [1 ]
Hou, Zeng-Guang [1 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence, Beijing 100190, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Data models; Training; Arteries; Reinforcement learning; Instruments; Catheters; Predictive models; Offline reinforcement learning; deep neural network; vascular robotic system; robot assisted intervention; PERCUTANEOUS CORONARY INTERVENTION;
D O I
10.1109/TMRB.2024.3407349
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Guidewire delivery is a fundamental procedure in percutaneous coronary intervention. The inherent flexibility of the guidewire poses challenges in precise control, necessitating long-term training and substantial expertise. In response, this paper proposes a novel offline reinforcement learning (RL) algorithm, Conservative Offline Reinforcement Learning with Variational Environment Model (CORVE), for autonomous delivery of guidewire. CORVE first uses offline data to train an environment model and then optimizes the policy with both offline and model-generated data. The proposed method shares an encoder between the environmental model, policy, and Q-function, mitigating the common sample inefficiency in image-based RL. Besides, CORVE utilizes model prediction errors to forecast wrong deliveries in inference, which is an attribute absent in existing methods. The experimental results show that CORVE obtains superior performance in guidewire deliveries, achieving notably higher success rates and smoother movements than existing methods. These findings suggest that CORVE holds significant potential for enhancing the autonomy of vascular robotic systems in clinical settings.
引用
收藏
页码:1054 / 1062
页数:9
相关论文
共 50 条
  • [31] Towards Accurate And Robust Dynamics and Reward Modeling for Model-Based Offline Inverse Reinforcement Learning
    Zhang, Gengyu
    Yan, Yan
    2024 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS 2024, 2024, : 611 - 618
  • [32] Model-based Reinforcement Learning: A Survey
    Moerland, Thomas M.
    Broekens, Joost
    Plaat, Aske
    Jonker, Catholijn M.
    FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2023, 16 (01): : 1 - 118
  • [33] A survey on model-based reinforcement learning
    Fan-Ming LUO
    Tian XU
    Hang LAI
    Xiong-Hui CHEN
    Weinan ZHANG
    Yang YU
    Science China(Information Sciences), 2024, 67 (02) : 59 - 84
  • [34] Nonparametric model-based reinforcement learning
    Atkeson, CG
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 10, 1998, 10 : 1008 - 1014
  • [35] The ubiquity of model-based reinforcement learning
    Doll, Bradley B.
    Simon, Dylan A.
    Daw, Nathaniel D.
    CURRENT OPINION IN NEUROBIOLOGY, 2012, 22 (06) : 1075 - 1081
  • [36] Multiple model-based reinforcement learning
    Doya, K
    Samejima, K
    Katagiri, K
    Kawato, M
    NEURAL COMPUTATION, 2002, 14 (06) : 1347 - 1369
  • [37] A survey on model-based reinforcement learning
    Luo, Fan-Ming
    Xu, Tian
    Lai, Hang
    Chen, Xiong-Hui
    Zhang, Weinan
    Yu, Yang
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (02)
  • [38] Offline Reinforcement Learning Based on Prioritized Sampling Model
    Gu Y.
    Cheng Y.-H.
    Wang X.-S.
    Zidonghua Xuebao/Acta Automatica Sinica, 2024, 50 (01): : 143 - 153
  • [39] Scaling model-based average-reward reinforcement learning for product delivery
    Proper, Scott
    Tadepalli, Prasad
    MACHINE LEARNING: ECML 2006, PROCEEDINGS, 2006, 4212 : 735 - 742
  • [40] Uncertainty-Aware Model-Based Reinforcement Learning: Methodology and Application in Autonomous Driving
    Wu, Jingda
    Huang, Zhiyu
    Lv, Chen
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (01): : 194 - 203