Model-Based Offline Reinforcement Learning for Autonomous Delivery of Guidewire

被引:0
|
作者
Li, Hao [1 ]
Zhou, Xiao-Hu [1 ]
Xie, Xiao-Liang [1 ]
Liu, Shi-Qi [1 ]
Feng, Zhen-Qiu [1 ]
Gui, Mei-Jiang [1 ]
Xiang, Tian-Yu [1 ]
Huang, De-Xing [1 ]
Hou, Zeng-Guang [1 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence, Beijing 100190, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Data models; Training; Arteries; Reinforcement learning; Instruments; Catheters; Predictive models; Offline reinforcement learning; deep neural network; vascular robotic system; robot assisted intervention; PERCUTANEOUS CORONARY INTERVENTION;
D O I
10.1109/TMRB.2024.3407349
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Guidewire delivery is a fundamental procedure in percutaneous coronary intervention. The inherent flexibility of the guidewire poses challenges in precise control, necessitating long-term training and substantial expertise. In response, this paper proposes a novel offline reinforcement learning (RL) algorithm, Conservative Offline Reinforcement Learning with Variational Environment Model (CORVE), for autonomous delivery of guidewire. CORVE first uses offline data to train an environment model and then optimizes the policy with both offline and model-generated data. The proposed method shares an encoder between the environmental model, policy, and Q-function, mitigating the common sample inefficiency in image-based RL. Besides, CORVE utilizes model prediction errors to forecast wrong deliveries in inference, which is an attribute absent in existing methods. The experimental results show that CORVE obtains superior performance in guidewire deliveries, achieving notably higher success rates and smoother movements than existing methods. These findings suggest that CORVE holds significant potential for enhancing the autonomy of vascular robotic systems in clinical settings.
引用
收藏
页码:1054 / 1062
页数:9
相关论文
共 50 条
  • [1] MOReL: Model-Based Offline Reinforcement Learning
    Kidambi, Rahul
    Rajeswaran, Aravind
    Netrapalli, Praneeth
    Joachims, Thorsten
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [2] Offline Reinforcement Learning with Reverse Model-based Imagination
    Wang, Jianhao
    Li, Wenzhe
    Jiang, Haozhe
    Zhu, Guangxiang
    Li, Siyuan
    Zhang, Chongjie
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [3] Offline Model-Based Reinforcement Learning for Tokamak Control
    Char, Ian
    Abbate, Joseph
    Bardoczi, Laszlo
    Boyer, Mark D.
    Chung, Youngseog
    Conlin, Rory
    Erickson, Keith
    Mehta, Viraj
    Richner, Nathan
    Kolemen, Egemen
    Schneider, Jeff
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [4] Model-Based Offline Reinforcement Learning with Local Misspecification
    Dong, Kefan
    Flet-Berliac, Yannis
    Nie, Allen
    Brunskill, Emma
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 7423 - 7431
  • [5] Weighted model estimation for offline model-based reinforcement learning
    Hishinuma, Toru
    Senda, Kei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [6] Bayesian Model-Based Offline Reinforcement Learning for Product Allocation
    Jenkins, Porter
    Wei, Hua
    Jenkins, J. Stockton
    Li, Zhenhui
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 12531 - 12537
  • [7] SETTLING THE SAMPLE COMPLEXITY OF MODEL-BASED OFFLINE REINFORCEMENT LEARNING
    Li, Gen
    Shi, Laixi
    Chen, Yuxin
    Chi, Yuejie
    Wei, Yuting
    ANNALS OF STATISTICS, 2024, 52 (01): : 233 - 260
  • [8] Model-based offline reinforcement learning for sustainable fishery management
    Ju, Jun
    Kurniawati, Hanna
    Kroese, Dirk
    Ye, Nan
    EXPERT SYSTEMS, 2025, 42 (01)
  • [9] OCEAN-MBRL: Offline Conservative Exploration for Model-Based Offline Reinforcement Learning
    Wu, Fan
    Zhang, Rui
    Yi, Qi
    Gao, Yunkai
    Guo, Jiaming
    Peng, Shaohui
    Lan, Siming
    Han, Husheng
    Pan, Yansong
    Yuan, Kaizhao
    Jin, Pengwei
    Chen, Ruizhi
    Chen, Yunji
    Li, Ling
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 15897 - 15905
  • [10] Comparing Model-free and Model-based Algorithms for Offline Reinforcement Learning
    Swazinna, Phillip
    Udluft, Steffen
    Hein, Daniel
    Runkler, Thomas
    IFAC PAPERSONLINE, 2022, 55 (15): : 19 - 26