Real-Time Fusion Network for RGB-D Semantic Segmentation Incorporating Unexpected Obstacle Detection for Road-Driving Images

被引:109
|
作者
Sun, Lei [1 ]
Yang, Kailun [2 ]
Hu, Xinxin [1 ]
Hu, Weijian [1 ]
Wang, Kaiwei [3 ]
机构
[1] Zhejiang Univ, State Key Lab Modern Opt Instrumentat, Hangzhou 310027, Peoples R China
[2] Karlsruhe Inst Technol, Inst Anthropomat & Robot, D-76131 Karlsruhe, Germany
[3] Zhejiang Univ, Natl Opt Instrumentat Engn Technol Res Ctr, Hangzhou 310027, Peoples R China
关键词
Semantic scene understanding; RGB-D fusion; obstacle detection; autonomous driving;
D O I
10.1109/LRA.2020.3007457
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Semantic segmentation has made striking progress due to the success of deep convolutional neural networks. Considering the demands of autonomous driving, real-time semantic segmentation has become a research hotspot these years. However, few real-time RGB-D fusion semantic segmentation studies are carried out despite readily accessible depth information nowadays. In this letter, we propose a real-time fusion semantic segmentation network termed RFNet that effectively exploits complementary cross-modal information. Building on an efficient network architecture, RFNet is capable of running swiftly, which satisfies autonomous vehicles applications. Multi-dataset training is lever-aged to incorporate unexpected small obstacle detection, enriching the recognizable classes required to face unforeseen hazards in the real world. A comprehensive set of experiments demonstrates the effectiveness of our framework. On Cityscapes, Our method outperforms previous state-of-the-art semantic segmenters, with excellent accuracy and 22 Hz inference speed at the full 2048 x 1024 resolution, outperforming most existing RGB-D networks.
引用
收藏
页码:5558 / 5565
页数:8
相关论文
共 50 条
  • [1] MAFFNet: real-time multi-level attention feature fusion network with RGB-D semantic segmentation for autonomous driving
    Lv, Tongfei
    Zhang, Yu
    Luo, Lin
    Gao, Xiaorong
    APPLIED OPTICS, 2022, 61 (09) : 2219 - 2229
  • [2] In Defense of Pre-trained ImageNet Architectures for Real-time Semantic Segmentation of Road-driving Images
    Orsic, Marin
    Kreso, Ivan
    Bevandic, Petra
    Segvic, Sinisa
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 12599 - 12608
  • [3] EFINet: Efficient Feature Interaction Network for Real-Time RGB-D Semantic Segmentation
    Yang, Zhe
    Mu, Baozhong
    Wang, Mingxun
    Wang, Xin
    Xu, Jie
    Yang, Baolu
    Yang, Cheng
    Li, Hong
    Lv, Rongqi
    IEEE ACCESS, 2024, 12 : 151046 - 151062
  • [4] A Fusion Network for Semantic Segmentation Using RGB-D Data
    Yuan, Jiahui
    Zhang, Kun
    Xia, Yifan
    Qi, Lin
    Dong, Junyu
    NINTH INTERNATIONAL CONFERENCE ON GRAPHIC AND IMAGE PROCESSING (ICGIP 2017), 2018, 10615
  • [5] Cascaded Feature Network for Semantic Segmentation of RGB-D Images
    Lin, Di
    Chen, Guangyong
    Daniel Cohen-Or
    Heng, Pheng-Ann
    Huang, Hui
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1320 - 1328
  • [6] Cascaded Feature Network for Semantic Segmentation of RGB-D Images
    Lin, Di
    Chen, Guangyong
    Cohen-Or, Daniel
    Heng, Pheng-Ann
    Huang, Hui
    Proceedings of the IEEE International Conference on Computer Vision, 2017, 2017-October : 1320 - 1328
  • [7] SPNet: An RGB-D Sequence Progressive Network for Road Semantic Segmentation
    Zhou, Zhi
    Zhang, Yuhang
    Hua, Guoguang
    Long, Ruijing
    Tian, Shishun
    Zou, Wenbin
    2023 IEEE 25TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING, MMSP, 2023,
  • [8] Attention-based fusion network for RGB-D semantic segmentation
    Zhong, Li
    Guo, Chi
    Zhan, Jiao
    Deng, JingYi
    NEUROCOMPUTING, 2024, 608
  • [9] DDNet: Depth Dominant Network for Semantic Segmentation of RGB-D Images
    Rong, Peizhi
    SENSORS, 2024, 24 (21)
  • [10] Zig-Zag Network for Semantic Segmentation of RGB-D Images
    Lin, Di
    Huang, Hui
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (10) : 2642 - 2655