A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images

被引:0
|
作者
Wu, Yingbin [1 ,2 ]
Zhao, Peng [1 ]
Wang, Fubo [1 ]
Zhou, Mingquan [1 ,3 ]
Geng, Shengling [1 ,3 ]
Zhang, Dan [1 ,3 ]
机构
[1] Qinghai Normal Univ, Sch Comp Sci, Xining 810016, Peoples R China
[2] Yuncheng Univ, Sch Math & Informat Technol, Yuncheng 044000, Peoples R China
[3] State Key Lab Tibetan Intelligent Informat Proc &, Xining 810016, Peoples R China
基金
中国国家自然科学基金;
关键词
building segmentation; feature fusion; prior-guided information; dual branch network; parallel dilated convolution; EXTRACTION;
D O I
10.3390/buildings14072006
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
The domain of remote sensing image processing has witnessed remarkable advancements in recent years, with deep convolutional neural networks (CNNs) establishing themselves as a prominent approach for building segmentation. Despite the progress, traditional CNNs, which rely on convolution and pooling for feature extraction during the encoding phase, often fail to precisely delineate global pixel interactions, potentially leading to the loss of vital semantic details. Moreover, conventional CNN-based segmentation models frequently neglect the nuanced semantic differences between shallow and deep features during the decoding phase, which can result in subpar feature integration through rudimentary addition or concatenation techniques. Additionally, the unique boundary characteristics of buildings in remote sensing images, which offer a rich vein of prior information, have not been fully harnessed by traditional CNNs. This paper introduces an innovative approach to building segmentation in remote sensing images through a prior-guided dual branch multi-feature fusion network (PDBMFN). The network is composed of a prior-guided branch network (PBN) in the encoding process, a parallel dilated convolution module (PDCM) designed to incorporate prior information, and a multi-feature aggregation module (MAM) in the decoding process. The PBN leverages prior region and edge information derived from superpixels and edge maps to enhance edge detection accuracy during the encoding phase. The PDCM integrates features from both branches and applies dilated convolution across various scales to expand the receptive field and capture a more comprehensive semantic context. During the decoding phase, the MAM utilizes deep semantic information to direct the fusion of features, thereby optimizing segmentation efficacy. Through a sequence of aggregations, the MAM gradually merges deep and shallow semantic information, culminating in a more enriched and holistic feature representation. Extensive experiments are conducted across diverse datasets, such as WHU, Inria Aerial, and Massachusetts, revealing that PDBMFN outperforms other sophisticated methods in terms of segmentation accuracy. In the key segmentation metrics, including mIoU, precision, recall, and F1 score, PDBMFN shows a marked superiority over contemporary techniques. The ablation studies further substantiate the performance improvements conferred by the PBN's prior information guidance and the efficacy of the PDCM and MAM modules.
引用
收藏
页数:22
相关论文
共 50 条
  • [31] The Application of Multi-feature Fusion in Remote Sensing Image Recognition
    Liu, Jinmei
    Chen, Jinghua
    MATERIALS SCIENCE, CIVIL ENGINEERING AND ARCHITECTURE SCIENCE, MECHANICAL ENGINEERING AND MANUFACTURING TECHNOLOGY, PTS 1 AND 2, 2014, 488-489 : 1079 - +
  • [32] A Multi-Level Feature Fusion Network for Remote Sensing Image Segmentation
    Dong, Sijun
    Chen, Zhengchao
    SENSORS, 2021, 21 (04) : 1 - 18
  • [33] Prior-guided attention fusion transformer for multi-lesion segmentation of diabetic retinopathy
    Xu, Chenfangqian
    Guo, Xiaoxin
    Yang, Guangqi
    Cui, Yihao
    Su, Longchen
    Dong, Hongliang
    Hu, Xiaoying
    Che, Songtian
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [34] A Frequency Domain Feature-Guided Network for Semantic Segmentation of Remote Sensing Images
    Li, Xin
    Xu, Feng
    Gao, Hongmin
    Liu, Fan
    Lyu, Xin
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1369 - 1373
  • [35] A novel network for semantic segmentation of landslide areas in remote sensing images with multi-branch and multi-scale fusion
    Wang, Kai
    He, Daojie
    Sun, Qingqiang
    Yi, Lizhi
    Yuan, Xiaofeng
    Wang, Yalin
    APPLIED SOFT COMPUTING, 2024, 158
  • [36] Multi-View Feature Fusion and Rich Information Refinement Network for Semantic Segmentation of Remote Sensing Images
    Liu, Jiang
    Cheng, Shuli
    Du, Anyu
    REMOTE SENSING, 2024, 16 (17)
  • [37] Dense feature pyramid fusion deep network for building segmentation in remote sensing image
    Tian Qinglin
    Zhao Yingjun
    Qin Kai
    Li Yao
    Chen Xuejiao
    SEVENTH SYMPOSIUM ON NOVEL PHOTOELECTRONIC DETECTION TECHNOLOGY AND APPLICATIONS, 2021, 11763
  • [38] Multiscale feature fusion network for automatic port segmentation from remote sensing images
    Ju, Haoran
    Bi, Fukun
    Bian, Mingming
    Shi, Yinni
    JOURNAL OF APPLIED REMOTE SENSING, 2022, 16 (04)
  • [39] A dual-branch multi-feature deep fusion network framework for hyperspectral image classification
    Liu, Linfeng
    Zhang, Chengcai
    Luo, Weiran
    GEOCARTO INTERNATIONAL, 2022, 37 (27) : 18692 - 18715
  • [40] Based on Multi-Feature Information Attention Fusion for Multi-Modal Remote Sensing Image Semantic Segmentation
    Zhang, Chongyu
    2021 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATION (IEEE ICMA 2021), 2021, : 71 - 76