Object and anatomical feature recognition in surgical video images based on a convolutional neural network

被引:18
|
作者
Bamba, Yoshiko [1 ]
Ogawa, Shimpei [1 ]
Itabashi, Michio [1 ]
Shindo, Hironari [2 ]
Kameoka, Shingo [3 ]
Okamoto, Takahiro [4 ]
Yamamoto, Masakazu [1 ]
机构
[1] Tokyo Womens Med Univ, Inst Gastroenterol, Dept Surg, Shinjuku Ku, 8-1 Kawadacho, Tokyo 1628666, Japan
[2] Otsuki Municipal Cent Hosp, Yamanashi, Japan
[3] Ushiku Aiwa Hosp, Ibaraki, Japan
[4] Tokyo Womens Med Univ, Dept Breast Endocrinol Surg, Tokyo, Japan
关键词
Image-guided navigation technology; Surgical education; Convolutional neural network; Computer vision; Object detection; GASTRIC-CANCER; SURGERY;
D O I
10.1007/s11548-021-02434-w
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Purpose Artificial intelligence-enabled techniques can process large amounts of surgical data and may be utilized for clinical decision support to recognize or forecast adverse events in an actual intraoperative scenario. To develop an image-guided navigation technology that will help in surgical education, we explored the performance of a convolutional neural network (CNN)-based computer vision system in detecting intraoperative objects. Methods The surgical videos used for annotation were recorded during surgeries conducted in the Department of Surgery of Tokyo Women's Medical University from 2019 to 2020. Abdominal endoscopic images were cut out from manually captured surgical videos. An open-source programming framework for CNN was used to design a model that could recognize and segment objects in real time through IBM Visual Insights. The model was used to detect the GI tract, blood, vessels, uterus, forceps, ports, gauze and clips in the surgical images. Results The accuracy, precision and recall of the model were 83%, 80% and 92%, respectively. The mean average precision (mAP), the calculated mean of the precision for each object, was 91%. Among surgical tools, the highest recall and precision of 96.3% and 97.9%, respectively, were achieved for forceps. Among the anatomical structures, the highest recall and precision of 92.9% and 91.3%, respectively, were achieved for the GI tract. Conclusion The proposed model could detect objects in operative images with high accuracy, highlighting the possibility of using AI-based object recognition techniques for intraoperative navigation. Real-time object recognition will play a major role in navigation surgery and surgical education.
引用
收藏
页码:2045 / 2054
页数:10
相关论文
共 50 条
  • [41] Structural Damage Recognition Based on Filtered Feature Selection and Convolutional Neural Network
    Jin, Zihan
    Teng, Shuai
    Zhang, Jiqiao
    Chen, Gongfa
    Cui, Fangsen
    INTERNATIONAL JOURNAL OF STRUCTURAL STABILITY AND DYNAMICS, 2022, 22 (12)
  • [42] Feature Extraction and Recognition of Human Physiological Signals Based on the Convolutional Neural Network
    Hurr, Chansol
    Li, Caiyan
    Li, Heng
    MOBILE INFORMATION SYSTEMS, 2022, 2022
  • [43] Facial Expression Recognition Based on Local Feature Fusion of Convolutional Neural Network
    Yao Lisha
    Xu Guoming
    Zhao Feng
    LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (04)
  • [44] Manufacturing Feature Recognition With a Sparse Voxel-Based Convolutional Neural Network
    Vatandoust, Farzad
    Yan, Xiaoliang
    Rosen, David
    Melkote, Shreyes N.
    JOURNAL OF COMPUTING AND INFORMATION SCIENCE IN ENGINEERING, 2025, 25 (03)
  • [45] Human Activity Recognition Based On Video Summarization And Deep Convolutional Neural Network
    Kushwaha, Arati
    Khare, Manish
    Bommisetty, Reddy Mounika
    Khare, Ashish
    COMPUTER JOURNAL, 2024,
  • [46] Video fire recognition based on multi-channel convolutional neural network
    Zhong, Chen
    Shao, Yu
    Ding, Hongjun
    Wang, Ke
    2020 3RD INTERNATIONAL CONFERENCE ON COMPUTER INFORMATION SCIENCE AND APPLICATION TECHNOLOGY (CISAT) 2020, 2020, 1634
  • [47] Human Activity Recognition Based On Video Summarization And Deep Convolutional Neural Network
    Kushwaha, Arati
    Khare, Manish
    Bommisetty, Reddy Mounika
    Khare, Ashish
    Computer Journal, 1600, 67 (08): : 2601 - 2609
  • [48] Object Tracking in the Video Stream by Means of a Convolutional Neural Network
    Zolotukhin, Yu N.
    Kotov, K. Yu
    Nesterov, A. A.
    Semenyuk, E. D.
    OPTOELECTRONICS INSTRUMENTATION AND DATA PROCESSING, 2020, 56 (06) : 642 - 648
  • [49] 3D Convolutional Neural Network based on memristor for video recognition
    Liu, Jiaqi
    Li, Zhenghao
    Tang, Yongliang
    Hu, Wei
    Wu, Jun
    PATTERN RECOGNITION LETTERS, 2020, 130 (130) : 116 - 124
  • [50] Object Tracking in the Video Stream by Means of a Convolutional Neural Network
    Yu. N. Zolotukhin
    K. Yu. Kotov
    A. A. Nesterov
    E. D. Semenyuk
    Optoelectronics, Instrumentation and Data Processing, 2020, 56 : 642 - 648