Exploring the Application of Large-Scale Pre-Trained Models on Adverse Weather Removal

被引:2
|
作者
Tan, Zhentao [1 ]
Wu, Yue [1 ]
Liu, Qiankun [2 ,3 ]
Chu, Qi [2 ,3 ]
Lu, Le [1 ]
Ye, Jieping [1 ]
Yu, Nenghai [2 ,3 ]
机构
[1] Alibaba Grp, Hangzhou 310052, Peoples R China
[2] Univ Sci & Technol China USTC, Sch Cyber Sci & Technol, CAS Key Lab Electromagnet Space Informat, Hefei 230026, Peoples R China
[3] Univ Sci & Technol China USTC, Anhui Prov Key Lab Digital Secur, Hefei 230026, Peoples R China
关键词
Meteorology; Task analysis; Training; Semantics; Image restoration; Rain; Feature extraction; Adverse weather removal; image restoration; multi-modal pre-trained model; NETWORK;
D O I
10.1109/TIP.2024.3368961
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image restoration under adverse weather conditions (e.g., rain, snow, and haze) is a fundamental computer vision problem that has important implications for various downstream applications. Distinct from early methods that are specially designed for specific types of weather, recent works tend to simultaneously remove various adverse weather effects based on either spatial feature representation learning or semantic information embedding. Inspired by various successful applications incorporating large-scale pre-trained models (e.g., CLIP), in this paper, we explore their potential benefits for leveraging large-scale pre-trained models in this task based on both spatial feature representation learning and semantic information embedding aspects: 1) spatial feature representation learning, we design a Spatially Adaptive Residual (SAR) encoder to adaptively extract degraded areas. To facilitate training of this model, we propose a Soft Residual Distillation (CLIP-SRD) strategy to transfer spatial knowledge from CLIP between clean and adverse weather images; 2) semantic information embedding, we propose a CLIP Weather Prior (CWP) embedding module to enable the network to adaptively respond to different weather conditions. This module integrates the sample-specific weather priors extracted by the CLIP image encoder with the distribution-specific information (as learned by a set of parameters) and embeds these elements using a cross-attention mechanism. Extensive experiments demonstrate that our proposed method can achieve state-of-the-art performance under various and severe adverse weather conditions. The code will be made available.
引用
收藏
页码:1683 / 1698
页数:16
相关论文
共 50 条
  • [41] EFFICIENT UTILIZATION OF LARGE PRE-TRAINED MODELS FOR LOW RESOURCE ASR
    Vieting, Peter
    Luescher, Christoph
    Dierkes, Julian
    Schlueter, Ralf
    Ney, Hermann
    2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [42] The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter
    Jaiswal, Ajay
    Liu, Shiwei
    Chen, Tianlong
    Wang, Zhangyang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [43] Exploring Accurate and Generic Simile Knowledge from Pre-trained Language Models
    Zhou, Shuhan
    Ma, Longxuan
    Shao, Yanqiu
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2023, 2023, 14232 : 348 - 363
  • [44] Exploring the Potential of Pre-Trained Language Models of Code for Automated Program Repair
    Hao, Sichong
    Shi, Xianjun
    Liu, Hongwei
    ELECTRONICS, 2024, 13 (07)
  • [45] Synergizing Large Language Models and Pre-Trained Smaller Models for Conversational Intent Discovery
    Liang, Jinggui
    Liao, Lizi
    Fei, Hao
    Jiang, Jing
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 14133 - 14147
  • [46] From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Models to Pre-trained Machine Reader
    Xu, Weiwen
    Li, Xin
    Zhang, Wenxuan
    Zhou, Meng
    Lam, Wai
    Si, Luo
    Bing, Lidong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [47] Annotating Columns with Pre-trained Language Models
    Suhara, Yoshihiko
    Li, Jinfeng
    Li, Yuliang
    Zhang, Dan
    Demiralp, Cagatay
    Chen, Chen
    Tan, Wang-Chiew
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA (SIGMOD '22), 2022, : 1493 - 1503
  • [48] Clinical efficacy of pre-trained large language models through the lens of aphasia
    Cong, Yan
    Lacroix, Arianna N.
    Lee, Jiyeon
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [49] Editorial for Special Issue on Pre-trained Large Language Models for Information Processing
    Wang, Bin
    Kawahara, Tatsuya
    Li, Haizhou
    Meng, Helen
    Wu, Chung-Hsien
    APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING, 2024, 13 (02)
  • [50] Grounding Ontologies with Pre-Trained Large Language Models for Activity Based Intelligence
    Azim, Anee
    Clark, Leon
    Lau, Caleb
    Cobb, Miles
    Jenner, Kendall
    SIGNAL PROCESSING, SENSOR/INFORMATION FUSION, AND TARGET RECOGNITION XXXIII, 2024, 13057