A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images

被引:0
|
作者
Wu, Yingbin [1 ,2 ]
Zhao, Peng [1 ]
Wang, Fubo [1 ]
Zhou, Mingquan [1 ,3 ]
Geng, Shengling [1 ,3 ]
Zhang, Dan [1 ,3 ]
机构
[1] Qinghai Normal Univ, Sch Comp Sci, Xining 810016, Peoples R China
[2] Yuncheng Univ, Sch Math & Informat Technol, Yuncheng 044000, Peoples R China
[3] State Key Lab Tibetan Intelligent Informat Proc &, Xining 810016, Peoples R China
基金
中国国家自然科学基金;
关键词
building segmentation; feature fusion; prior-guided information; dual branch network; parallel dilated convolution; EXTRACTION;
D O I
10.3390/buildings14072006
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
The domain of remote sensing image processing has witnessed remarkable advancements in recent years, with deep convolutional neural networks (CNNs) establishing themselves as a prominent approach for building segmentation. Despite the progress, traditional CNNs, which rely on convolution and pooling for feature extraction during the encoding phase, often fail to precisely delineate global pixel interactions, potentially leading to the loss of vital semantic details. Moreover, conventional CNN-based segmentation models frequently neglect the nuanced semantic differences between shallow and deep features during the decoding phase, which can result in subpar feature integration through rudimentary addition or concatenation techniques. Additionally, the unique boundary characteristics of buildings in remote sensing images, which offer a rich vein of prior information, have not been fully harnessed by traditional CNNs. This paper introduces an innovative approach to building segmentation in remote sensing images through a prior-guided dual branch multi-feature fusion network (PDBMFN). The network is composed of a prior-guided branch network (PBN) in the encoding process, a parallel dilated convolution module (PDCM) designed to incorporate prior information, and a multi-feature aggregation module (MAM) in the decoding process. The PBN leverages prior region and edge information derived from superpixels and edge maps to enhance edge detection accuracy during the encoding phase. The PDCM integrates features from both branches and applies dilated convolution across various scales to expand the receptive field and capture a more comprehensive semantic context. During the decoding phase, the MAM utilizes deep semantic information to direct the fusion of features, thereby optimizing segmentation efficacy. Through a sequence of aggregations, the MAM gradually merges deep and shallow semantic information, culminating in a more enriched and holistic feature representation. Extensive experiments are conducted across diverse datasets, such as WHU, Inria Aerial, and Massachusetts, revealing that PDBMFN outperforms other sophisticated methods in terms of segmentation accuracy. In the key segmentation metrics, including mIoU, precision, recall, and F1 score, PDBMFN shows a marked superiority over contemporary techniques. The ablation studies further substantiate the performance improvements conferred by the PBN's prior information guidance and the efficacy of the PDCM and MAM modules.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Deep multi-feature fusion network for remote sensing images
    Xiong, Wei
    Xiong, Zhenyu
    Cui, Yaqi
    Lv, Yafei
    REMOTE SENSING LETTERS, 2020, 11 (06) : 563 - 571
  • [2] A Multi-Branch Feature Fusion Network for Building Detection in Remote Sensing Images
    Li, Chao
    Huang, Xinyu
    Tang, Jiechen
    Wang, Kai
    IEEE ACCESS, 2021, 9 (09): : 168511 - 168519
  • [3] Building Multi-Feature Fusion Refined Network for Building Extraction from High-Resolution Remote Sensing Images
    Ran, Shuhao
    Gao, Xianjun
    Yang, Yuanwei
    Li, Shaohua
    Zhang, Guangbin
    Wang, Ping
    REMOTE SENSING, 2021, 13 (14)
  • [4] A classification method of building structures based on multi-feature fusion of UAV remote sensing images
    Haoguo Du
    Yanbo Cao
    Fanghao Zhang
    Jiangli Lv
    Shurong Deng
    Yongkun Lu
    Shifang He
    Yuanshuo Zhang
    Qinkun Yu
    EarthquakeResearchAdvances, 2021, 1 (04) : 38 - 47
  • [5] Prior-Guided Fuzzy-Aware Multibranch Network for Remote Sensing Image Segmentation
    Wei, Guangyi
    Xu, Jindong
    Chong, Qianpeng
    Huang, Jianjun
    Xing, Haihua
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 1
  • [6] Dual-Path Feature Fusion Network for Semantic Segmentation of Remote Sensing Images
    Li, Boyang
    Zhang, Yu
    Zhang, Youmei
    Li, Bin
    Li, Zhenhao
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [7] Semantic Segmentation of Remote Sensing Images by Interactive Representation Refinement and Geometric Prior-Guided Inference
    Li, Xin
    Xu, Feng
    Liu, Fan
    Tong, Yao
    Lyu, Xin
    Zhou, Jun
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 18
  • [8] Multi-feature Map Pyramid Fusion Deep Network for Semantic Segmentation on Remote Sensing Data
    Zhao Fei
    Zhang Wenkai
    Yan Zhiyuan
    Yu Hongfeng
    Diao Wenhui
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2019, 41 (10) : 2525 - 2531
  • [9] MDTrans: Multi-scale and dual-branch feature fusion network based on Swin Transformer for building extraction in remote sensing images
    Diao, Kuo
    Zhu, Jinlong
    Liu, Guangjie
    Li, Meng
    IET IMAGE PROCESSING, 2024, 18 (11) : 2930 - 2942
  • [10] A Multi-Feature Fusion and Attention Network for Multi-Scale Object Detection in Remote Sensing Images
    Cheng, Yong
    Wang, Wei
    Zhang, Wenjie
    Yang, Ling
    Wang, Jun
    Ni, Huan
    Guan, Tingzhao
    He, Jiaxin
    Gu, Yakang
    Tran, Ngoc Nguyen
    REMOTE SENSING, 2023, 15 (08)