共 50 条
Asymmetrical Contrastive Learning Network via Knowledge Distillation for No-Service Rail Surface Defect Detection
被引:1
|作者:
Zhou, Wujie
[1
,2
]
Sun, Xinyu
[1
]
Qian, Xiaohong
[1
]
Fang, Meixin
[3
]
机构:
[1] Zhejiang Univ Sci & Technol, Sch Informat & Elect Engn, Hangzhou 310023, Peoples R China
[2] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 308232, Singapore
[3] Zhejiang Univ, Sch Med, Hangzhou 310003, Peoples R China
基金:
中国国家自然科学基金;
关键词:
Biological system modeling;
Contrastive learning;
Feature extraction;
Adaptation models;
Rails;
Computational modeling;
Neural networks;
Defect detection;
Decoding;
Convolution;
graph mapping distillation;
knowledge distillation (KD);
rail surface defect detection (SDD);
SALIENT OBJECT DETECTION;
REFINEMENT;
FUSION;
D O I:
10.1109/TNNLS.2024.3479453
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Owing to extensive research on deep learning, significant progress has recently been made in trackless surface defect detection (SDD). Nevertheless, existing algorithms face two main challenges. First, while depth features contain rich spatial structure features, most models only accept red- green-blue (RGB) features as input, which severely constrains performance. Thus, this study proposes a dual-stream teacher model termed the asymmetrical contrastive learning network (ACLNet-T), which extracts both RGB and depth features to achieve high performance. Second, the introduction of the dual-stream model facilitates an exponential increase in the number of parameters. As a solution, we designed a single-stream student model (ACLNet-S) that extracted RGB features. We leveraged a contrastive distillation loss via knowledge distillation (KD) techniques to transfer rich multimodal features from the ACLNet-T to the ACLNet-S pixel by pixel and channel by channel. Furthermore, to compensate for the lack of contrastive distillation loss that focuses exclusively on local features, we employed multiscale graph mapping to establish long-range dependencies and transfer global features to the ACLNet-S through multiscale graph mapping distillation loss. Finally, an attentional distillation loss based on the adaptive attention decoder (AAD) was designed to further improve the performance of the ACLNet-S. Consequently, we obtained the ACLNet-S*, which achieved performance similar to that of ACLNet-T, despite having a nearly eightfold parameter count gap. Through comprehensive experimentation using the industrial RGB-D dataset NEU RSDDS-AUG, the ACLNet-S* (ACLNet-S with KD) was confirmed to outperform 16 state-of-the-art methods. Moreover, to showcase the generalization capacity of ACLNet-S*, the proposed network was evaluated on three additional public datasets, and ACLNet-S* achieved comparable results.
引用
收藏
页数:14
相关论文