Graph-Based Spatial Reasoning for Tracking Landmarks in Dynamic Laparoscopic Environments

被引:0
|
作者
Zhang, Jie [1 ]
Wang, Yiwei [1 ,2 ]
Zhou, Song [1 ]
Zhao, Huan [1 ]
Wan, Chidan [3 ]
Cai, Xiong [3 ]
Ding, Han [1 ]
机构
[1] Huazhong Univ Sci & Technol, State Key Lab Intelligent Mfg Equipment & Technol, Wuhan 430074, Peoples R China
[2] Huazhong Univ Sci & Technol, Sch Med Equipment Sci & Engn, Wuhan 430074, Peoples R China
[3] Huazhong Univ Sci & Technol, Union Hosp, Tongji Med Coll, Dept Hepatobiliary Surg, Wuhan 430022, Peoples R China
来源
基金
美国国家科学基金会;
关键词
Surgery; Feature extraction; Task analysis; Image reconstruction; Anatomy; Visualization; Laparoscopes; Computer vision for medical robotics; deep learning for visual perception; surgical robotics: laparoscopy;
D O I
10.1109/LRA.2024.3445654
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Accurate anatomical landmark tracking is crucial yet challenging in laparoscopic surgery due to the changing appearance of landmarks during dynamic tool-anatomy interactions and visual domain shifts between cases. Unlike appearance-based detection methods, this work proposes a novel graph-based approach to reconstruct the entire target landmark area by explicitly modeling the evolving spatial relations over time among scenario entities, including observable regions, surgical tools, and landmarks. Considering tool-anatomy interactions, we present the Tool-Anatomy Interaction Graph (TAI-G), a spatio-temporal graph that captures spatial dependencies among entities, attribute interactions within entities, and temporal dependencies of spatial relations. To mitigate domain shifts, geometric segmentation features are designated as node attributes, representing domain-invariant image information in the graph space. Message passing with attention helps propagate information across TAI-G, enhancing robust tracking by reconstructing landmark data. Evaluated on laparoscopic cholecystectomy, our framework demonstrates effective handling of complex tool-anatomy interactions and visual domain gaps to accurately track landmarks, showing promise in enhancing the stability and reliability of intricate surgical tasks.
引用
收藏
页码:8459 / 8466
页数:8
相关论文
共 50 条
  • [1] Incorporating Moving Landmarks within 2D Graph-Based SLAM for Dynamic Environments
    Aerts, Peter
    Slaets, Peter
    Demeester, Eric
    2021 6TH INTERNATIONAL CONFERENCE ON MECHANICAL ENGINEERING AND ROBOTICS RESEARCH (ICMERR), 2021, : 1 - 7
  • [2] Dynamic graph-based search in unknown environments
    Haynes, Paul S.
    Alboul, Lyuba
    Penders, Jacques
    JOURNAL OF DISCRETE ALGORITHMS, 2012, 12 : 2 - 13
  • [3] Graph-based inductive reasoning
    Boumans, Marcel
    STUDIES IN HISTORY AND PHILOSOPHY OF SCIENCE, 2016, 59 : 1 - 10
  • [4] Graph-Based Global Reasoning Networks
    Chen, Yunpeng
    Rohrbach, Marcus
    Yan, Zhicheng
    Yan, Shuicheng
    Feng, Jiashi
    Kalantidis, Yannis
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 433 - 442
  • [5] Multiple processes in graph-based reasoning
    Peebles, D
    Cheng, PCH
    Shadbolt, N
    PROCEEDINGS OF THE TWENTY FIRST ANNUAL CONFERENCE OF THE COGNITIVE SCIENCE SOCIETY, 1999, : 531 - 536
  • [6] STRATEGIES FOR MIXED REASONING WITH A GRAPH-BASED REASONING ENGINE
    GARNER, BJ
    TSUI, E
    KNOWLEDGE-BASED SYSTEMS, 1991, 4 (02) : 75 - 81
  • [7] GRAPH-BASED KNOWLEDGE REPRESENTATION AND REASONING
    Chein, M.
    ICEIS 2010: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON ENTERPRISE INFORMATION SYSTEMS, VOL 1: DATABASES AND INFORMATION SYSTEMS INTEGRATION, 2010, : IS17 - IS21
  • [8] GRAPH-BASED KINSHIP REASONING NETWORK
    Li, Wanhua
    Zhang, Yingqiang
    Lv, Kangchen
    Lu, Jiwen
    Feng, Jianjiang
    Zhou, Jie
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,
  • [9] Graph-based object tracking
    Gomila, C
    Meyer, F
    2003 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOL 2, PROCEEDINGS, 2003, : 41 - 44
  • [10] Online Graph-Based Tracking
    Nam, Hyeonseob
    Hong, Seunghoon
    Han, Bohyung
    COMPUTER VISION - ECCV 2014, PT V, 2014, 8693 : 112 - 126