DSMENet: A Road Segmentation Network Based on Dual-Branch Dynamic Snake Convolutional Encoding and Multi-modal Information Iterative Enhancement

被引:0
|
作者
Li, Zhiyang [1 ]
Pan, Xuran [1 ]
Yang, Shuhao [1 ]
Yang, Xinqi [1 ]
Xu, Kexing [1 ]
机构
[1] Tianjin Univ Sci & Technol, Coll Artificial Intelligence, Tianjin Econ & Technol Dev Area TEDA, 9 Dishisan Dajie, Tianjin 300457, Peoples R China
关键词
Remote Sensing Image; Multi-Modal; Road Segmentation; Snake Convolution; Information Enhancement;
D O I
10.1007/978-981-97-5615-5_14
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Road segmentation from remote sensing images plays an important role in basic map data processing and services. However, roads in remote sensing images are characterized by long and narrow spans, intricate topological structures and being easily obscured, making road segmentation a challenging task in the field of re-mote sensing image object segmentation. To improve the accuracy and connectivity of road segmentation, this paper proposes a method based on Dual-branch dynamic Snake convolutional encoding and Multi-modal information iterative Enhancement (DSMENet). The multi-modal data are first encoded separately by dual-branch dynamic snake convolution encoders to adaptively focus on slender and winding local structures, accurately capturing the features of tube-like roads; next, attention driven feature fusion of multi-modal features are performed at different stages of the encoders, which are then input into the decoder for spatial resolution restoration. Finally, a multi-modal information iterative enhancement module is embedded at the end of the network to fully exploit spatial detail features of original multi-modal data and enhance the features at the end of the de-coder, thereby improving the connectivity of road segmentation. Experimental evaluations on the BJRoad dataset demonstrate that (1) The dynamic snake convolution enables the model to focus on tube-like roads effectively, resulting in a significant reduction in false alarms and an improvement in road segmentation accuracy. (2) The multi-modal information iterative enhancement module can provide supplementary spatial detail information to the road segmentation results, mitigating the effects of shadow occlusions and enhancing the connectivity of road segmentation.
引用
收藏
页码:168 / 179
页数:12
相关论文
共 26 条
  • [21] DAG-Net: Dual-Branch Attention-Guided Network for Multi-Scale Information Fusion in Lung Nodule Segmentation
    Zhang, Bojie
    Zhu, Hongqing
    Wang, Ziying
    Luo, Lan
    Yu, Yang
    INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2024, 34 (06)
  • [22] Attention-based convolutional neural network with multi-modal temporal information fusion for motor imagery EEG decoding
    Ma X.
    Chen W.
    Pei Z.
    Zhang Y.
    Chen J.
    Computers in Biology and Medicine, 2024, 175
  • [23] Seeing helps hearing: A multi-modal dataset and a mamba-based dual branch parallel network for auditory attention decoding
    Fan, Cunhang
    Zhang, Hongyu
    Ni, Qinke
    Zhang, Jingjing
    Tao, Jianhua
    Zhou, Jian
    Yi, Jiangyan
    Lv, Zhao
    Wu, Xiaopei
    INFORMATION FUSION, 2025, 118
  • [24] Dual Encoder-Based Dynamic-Channel Graph Convolutional Network With Edge Enhancement for Retinal Vessel Segmentation
    Li, Yang
    Zhang, Yue
    Cui, Weigang
    Lei, Baiying
    Kuang, Xihe
    Zhang, Teng
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2022, 41 (08) : 1975 - 1989
  • [25] Image Enhancement Based on Dual-Branch Generative Adversarial Network Combining Spatial and Frequency Domain Information for Imbalanced Fault Diagnosis of Rolling Bearing
    Huang, Yuguang
    Wen, Bin
    Liao, Weiqing
    Shan, Yahui
    Fu, Wenlong
    Wang, Renming
    SYMMETRY-BASEL, 2024, 16 (05):
  • [26] Snow Detection in Gaofen-1 Multi-Spectral Images Based on Swin-Transformer and U-Shaped Dual-Branch Encoder Structure Network with Geographic Information
    Wu, Yue
    Shi, Chunxiang
    Shen, Runping
    Gu, Xiang
    Tie, Ruian
    Ge, Lingling
    Sun, Shuai
    REMOTE SENSING, 2024, 16 (17)