GDALR: Global Dual Attention and Local Representations in transformer for surface defect detection

被引:9
|
作者
Zhou, Xin [1 ]
Zhou, Shihua [1 ]
Zhang, Yongchao [1 ]
Ren, Zhaohui [1 ]
Jiang, Zeyu [1 ]
Luo, Hengfa [1 ]
机构
[1] Northeastern Univ, Sch Mech Engn & Automat, Wenhua Rd, Shenyang 110819, Liaoning, Peoples R China
关键词
Surface defect detection; Semantic segmentation; Vision transformer; Dual-attention; Local transformer;
D O I
10.1016/j.measurement.2024.114398
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Automated surface detection has gradually emerged as a promising and crucial inspection method in the industrial sector, greatly enhancing production quality and efficiency. However, current semantic network models based on Vision Transformers are primarily trained on natural images, which exhibit complex object textures and backgrounds. Additionally, pure Vision Transformers lack the ability to capture local representations, making it challenging to directly apply existing semantic segmentation models to industrial production scenarios. In this paper, we propose a novel transformer segmentation model specifically designed for surface defect detection in industrial settings. Firstly, we employ a Dual -Attention Transformer (DAT) as the backbone of our model. This backbone replaces the generic 2D convolution block with a new self -attention block in the Spatial Reduction Attention module (SRA), enabling the establishment of a global view for each layer. Secondly, we enhance the collection of local information during decoding by initializing the relative position between query and key pixels. Finally, to strengthen the salient defect structure, we utilize Pixel Shuffle to rearrange the Ground Truth (GT) in order to guide the feature maps at each scale. Extensive experiments are conducted on three publicly industrial datasets, and evaluation results describe the outstanding performance of our network in surface defect detection.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Steel surface defect detection based on sparse global attention transformer
    Li, Yinghao
    Han, Zhiyong
    Wang, Wenmeng
    Xu, Heping
    Wei, Yongpeng
    Zai, Guangjun
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (04)
  • [2] LGAFormer: transformer with local and global attention for action detection
    Zhang, Haiping
    Zhou, Fuxing
    Wang, Dongjing
    Zhang, Xinhao
    Yu, Dongjin
    Guan, Liming
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (12): : 17952 - 17979
  • [3] Global and Local Attention-Based Transformer for Hyperspectral Image Change Detection
    Wang, Ziyi
    Gao, Feng
    Dong, Junyu
    Du, Qian
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2025, 22
  • [4] Defect transformer: An efficient hybrid transformer architecture for surface defect detection
    Wang, Junpu
    Xu, Guili
    Yan, Fuju
    Wang, Jinjin
    Wang, Zhengsheng
    MEASUREMENT, 2023, 211
  • [5] Road Surface Defect Detection Based on Partial Convolution and Global Attention
    Xie, Xiaoneng
    INTERNATIONAL JOURNAL OF PAVEMENT RESEARCH AND TECHNOLOGY, 2024,
  • [6] Dual-image differential transformer for periodic lace surface defect detection
    Jiao, Wenhua
    Cai, Xiaoyi
    Tian, Yuyu
    Zuo, Mingcheng
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 148
  • [7] LGGFormer: A dual-branch local-guided global self-attention network for surface defect segmentation
    Zhang, Gaowei
    Lu, Yang
    Jiang, Xiaoheng
    Jin, Shaohui
    Li, Shupan
    Xu, Mingliang
    ADVANCED ENGINEERING INFORMATICS, 2025, 64
  • [8] Dual Attention-Based Industrial Surface Defect Detection with Consistency Loss
    Li, Xuyang
    Zheng, Yu
    Chen, Bei
    Zheng, Enrang
    SENSORS, 2022, 22 (14)
  • [9] Fully Convolutional Transformer with Local-Global Attention
    Lee, Sihaeng
    Yi, Eojindl
    Lee, Janghyeon
    Yoo, Jinsu
    Lee, Honglak
    Kim, Seung Hwan
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 552 - 559
  • [10] GRA-Net: Global receptive attention network for surface defect detection
    Xiao, Meng
    Yang, Bo
    Wang, Shilong
    Mo, Fan
    He, Yan
    Gao, Yifan
    KNOWLEDGE-BASED SYSTEMS, 2023, 280