Knowledge Distillation in Histology Landscape by Multi-Layer Features Supervision

被引:6
|
作者
Javed, Sajid [1 ]
Mahmood, Arif [2 ]
Qaiser, Talha [3 ]
Werghi, Naoufel [1 ]
机构
[1] Khalifa Univ Sci & Technol, Dept Elect Engn & Comp Sci, Abu Dhabi, U Arab Emirates
[2] Informat Technol Univ, Dept Comp Sci, Lahore 54000, Pakistan
[3] Univ Warwick, Dept Comp Sci, Coventry CV4 7AL, England
关键词
Cancer; Training; Knowledge engineering; Histopathology; Task analysis; Predictive models; Neural networks; Knowledge distillation; features distillation; histology image classification; tissue phenotyping; FRAMEWORK; NETWORK;
D O I
10.1109/JBHI.2023.3237749
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Automatic tissue classification is a fundamental task in computational pathology for profiling tumor micro-environments. Deep learning has advanced tissue classification performance at the cost of significant computational power. Shallow networks have also been end-to-end trained using direct supervision however their performance degrades because of the lack of capturing robust tissue heterogeneity. Knowledge distillation has recently been employed to improve the performance of the shallow networks used as student networks by using additional supervision from deep neural networks used as teacher networks. In the current work, we propose a novel knowledge distillation algorithm to improve the performance of shallow networks for tissue phenotyping in histology images. For this purpose, we propose multi-layer feature distillation such that a single layer in the student network gets supervision from multiple teacher layers. In the proposed algorithm, the size of the feature map of two layers is matched by using a learnable multi-layer perceptron. The distance between the feature maps of the two layers is then minimized during the training of the student network. The overall objective function is computed by summation of the loss over multiple layers combination weighted with a learnable attention-based parameter. The proposed algorithm is named as Knowledge Distillation for Tissue Phenotyping (KDTP). Experiments are performed on five different publicly available histology image classification datasets using several teacher-student network combinations within the KDTP algorithm. Our results demonstrate a significant performance increase in the student networks by using the proposed KDTP algorithm compared to direct supervision-based training methods.
引用
收藏
页码:2037 / 2046
页数:10
相关论文
共 50 条
  • [1] Knowledge distillation based on multi-layer fusion features
    Tan, Shengyuan
    Guo, Rongzuo
    Tang, Jialiang
    Jiang, Ning
    Zou, Junying
    PLOS ONE, 2023, 18 (08):
  • [2] MulKD: Multi-layer Knowledge Distillation via collaborative learning
    Guermazi, Emna
    Mdhaffar, Afef
    Jmaiel, Mohamed
    Freisleben, Bernd
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [3] Medical image segmentation based on multi-layer features and spatial information distillation
    Zheng Y.
    Hao P.
    Wu D.
    Bai C.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2022, 48 (08): : 1409 - 1417
  • [4] Coherent multi-layer landscape synthesis
    Argudo, Oscar
    Andujar, Carlos
    Chica, Antonio
    Guerin, Eric
    Digne, Julie
    Peytavie, Adrien
    Galin, Eric
    VISUAL COMPUTER, 2017, 33 (6-8): : 1005 - 1015
  • [5] Coherent multi-layer landscape synthesis
    Oscar Argudo
    Carlos Andujar
    Antonio Chica
    Eric Guérin
    Julie Digne
    Adrien Peytavie
    Eric Galin
    The Visual Computer, 2017, 33 : 1005 - 1015
  • [6] Attention model with multi-layer supervision for text Classification
    Yue, Chunyi
    Cao, Hanqiang
    Xu, Guoping
    Dong, Youli
    2020 5TH INTERNATIONAL CONFERENCE ON MATHEMATICS AND ARTIFICIAL INTELLIGENCE (ICMAI 2020), 2020, : 103 - 109
  • [7] Knowledge Distillation with Deep Supervision
    Luo, Shiya
    Chen, Defang
    Wang, Can
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [8] Assessing the importance of features for multi-layer perceptrons
    Egmont-Petersen, M
    Talmon, JL
    Hasman, A
    Ambergen, AW
    NEURAL NETWORKS, 1998, 11 (04) : 623 - 635
  • [9] Performance Characterization of Supervision on Knowledge Distillation
    Chism, Joshua
    Lee, Byeong Kil
    PROCEEDINGS OF THE FUTURE TECHNOLOGIES CONFERENCE (FTC) 2021, VOL 2, 2022, 359 : 40 - 55
  • [10] Membrane distillation using surface modified multi-layer porous ceramics
    Cong, Shuo
    Liu, Xiaohua
    Guo, Fei
    INTERNATIONAL JOURNAL OF HEAT AND MASS TRANSFER, 2019, 129 : 764 - 772