Light4Mars: A lightweight transformer model for semantic segmentation on unstructured environment like Mars

被引:2
|
作者
Xiong, Yonggang [1 ]
Xiao, Xueming [1 ,4 ]
Yao, Meibao [2 ]
Cui, Hutao [3 ]
Fu, Yuegang [4 ]
机构
[1] Changchun Univ Sci & Technol, CVIR Lab, Changchun, Peoples R China
[2] Jilin Univ, Sch Artificial Intelligence, Minist Educ, Engn Res Ctr Knowledge Driven Human Machine Intell, Changchun, Peoples R China
[3] Harbin Inst Technol, Deep Space Explorat Res Ctr, Harbin, Peoples R China
[4] Changchun Univ Sci & Technol, Minist Educ, Key Lab Optoelect Measurement & Opt Informat Trans, Changchun, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantic segmentation; Unstructured environment; Mars; Lightweight model; SYSTEM-DESIGN; TIANWEN-1; DESCENT; ENTRY; GUIDANCE; NETWORK;
D O I
10.1016/j.isprsjprs.2024.06.008
中图分类号
P9 [自然地理学];
学科分类号
0705 ; 070501 ;
摘要
Auto-semantic segmentation is important for robots on unstructured and dynamic environments like planets where ambient conditions cannot be controlled and the scale is larger than that found indoors. Current learning- based methods have achieved breathtaking improvements on this topic. For onboard applications, however, all those methods still suffer from huge computational costs and are difficult to deploy on edge devices. In this paper, unlike previous transformer-based SOTA approaches that heavily relied on complex design, we proposed Light4Mars, a lightweight model with minimal computational complexity while maintaining high segmenting accuracy. We designed a lightweight squeeze window transformer module that focuses on window-scale feature extraction and is more effective in learning global and local contextual information. The aggregated local attention decoder is utilized to fuse semantic information at different scales, especially for unstructured scenes. Since there are few all-terrain datasets for semantic segmentation of unstructured scenes like Mars, we built a synthetic dataset SynMars-TW, referencing images collected by the ZhuRong rover on the Tianwen-1 mission and the Curiosity rover. Extensive experiments on SynMars-TW and the real Mars dataset, MarsScapes show that our approach achieves state-of-the-art performance with favorable computational simplicity. To the best of our knowledge, the proposed Light4Mars-T network is the first segmentation model for Mars image segmentation with parameters lower than 0.1M. . Code and datasets are available at https://github.com/CVIR-Lab/Light4Mars.
引用
收藏
页码:167 / 178
页数:12
相关论文
共 8 条
  • [1] A hybrid attention semantic segmentation network for unstructured terrain on Mars
    Liu, Haiqiang
    Yao, Meibao
    Xiao, Xueming
    Cui, Hutao
    ACTA ASTRONAUTICA, 2023, 204 : 492 - 499
  • [2] ELiFormer: A hierarchical Transformer based Model with Efficient Encoder and Lightweight Decoder for Semantic Segmentation
    Wu, Zixuan
    Zhou, Yue
    2024 2ND ASIA CONFERENCE ON COMPUTER VISION, IMAGE PROCESSING AND PATTERN RECOGNITION, CVIPPR 2024, 2024,
  • [3] Implementation and validation of the GEANT4/AtRIS code to model the radiation environment at Mars
    Guo, Jingnan
    Banjac, Saga
    Roestel, Lennart
    Terasa, Jan C.
    Herbst, Konstantin
    Heber, Bernd
    Wimmer-Schweingruber, Robert F.
    JOURNAL OF SPACE WEATHER AND SPACE CLIMATE, 2019, 9
  • [4] MARS: Model-agnostic Biased Object Removal without Additional Supervision for Weakly-Supervised Semantic Segmentation
    Jo, Sanghyun
    Yu, In-Jae
    Kim, Kyungsu
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 614 - 623
  • [5] Light-Deeplabv3+: a lightweight real-time semantic segmentation method for complex environment perception
    Ding, Peng
    Qian, Huaming
    JOURNAL OF REAL-TIME IMAGE PROCESSING, 2024, 21 (01)
  • [6] Light-Deeplabv3+: a lightweight real-time semantic segmentation method for complex environment perception
    Peng Ding
    Huaming Qian
    Journal of Real-Time Image Processing, 2024, 21
  • [7] Development of an Ensembled Meta-Deep Learning Model for Semantic Road-Scene Segmentation in an Unstructured Environment
    Sivanandham, Sangavi
    Gunaseelan, Dharani Bai
    APPLIED SCIENCES-BASEL, 2022, 12 (23):
  • [8] MarS3D: A Plug-and-Play Motion-Aware Model for Semantic Segmentation on Multi-Scan 3D Point Clouds
    Liu, Jiahui
    Chang, Chirui
    Liu, Jianhui
    Wu, Xiaoyang
    Ma, Lan
    Qi, Xiaojuan
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 9372 - 9381