Exploring the Application of Large-Scale Pre-Trained Models on Adverse Weather Removal

被引:2
|
作者
Tan, Zhentao [1 ]
Wu, Yue [1 ]
Liu, Qiankun [2 ,3 ]
Chu, Qi [2 ,3 ]
Lu, Le [1 ]
Ye, Jieping [1 ]
Yu, Nenghai [2 ,3 ]
机构
[1] Alibaba Grp, Hangzhou 310052, Peoples R China
[2] Univ Sci & Technol China USTC, Sch Cyber Sci & Technol, CAS Key Lab Electromagnet Space Informat, Hefei 230026, Peoples R China
[3] Univ Sci & Technol China USTC, Anhui Prov Key Lab Digital Secur, Hefei 230026, Peoples R China
关键词
Meteorology; Task analysis; Training; Semantics; Image restoration; Rain; Feature extraction; Adverse weather removal; image restoration; multi-modal pre-trained model; NETWORK;
D O I
10.1109/TIP.2024.3368961
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image restoration under adverse weather conditions (e.g., rain, snow, and haze) is a fundamental computer vision problem that has important implications for various downstream applications. Distinct from early methods that are specially designed for specific types of weather, recent works tend to simultaneously remove various adverse weather effects based on either spatial feature representation learning or semantic information embedding. Inspired by various successful applications incorporating large-scale pre-trained models (e.g., CLIP), in this paper, we explore their potential benefits for leveraging large-scale pre-trained models in this task based on both spatial feature representation learning and semantic information embedding aspects: 1) spatial feature representation learning, we design a Spatially Adaptive Residual (SAR) encoder to adaptively extract degraded areas. To facilitate training of this model, we propose a Soft Residual Distillation (CLIP-SRD) strategy to transfer spatial knowledge from CLIP between clean and adverse weather images; 2) semantic information embedding, we propose a CLIP Weather Prior (CWP) embedding module to enable the network to adaptively respond to different weather conditions. This module integrates the sample-specific weather priors extracted by the CLIP image encoder with the distribution-specific information (as learned by a set of parameters) and embeds these elements using a cross-attention mechanism. Extensive experiments demonstrate that our proposed method can achieve state-of-the-art performance under various and severe adverse weather conditions. The code will be made available.
引用
收藏
页码:1683 / 1698
页数:16
相关论文
共 50 条
  • [21] Exploring Pre-trained Language Models for Event Extraction and Generation
    Yang, Sen
    Feng, Dawei
    Qiao, Linbo
    Kan, Zhigang
    Li, Dongsheng
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5284 - 5294
  • [22] Exploring Strategies for Generalizable Commonsense Reasoning with Pre-trained Models
    Ma, Kaixin
    Ilievski, Filip
    Francis, Jonathan
    Ozaki, Satoru
    Nyberg, Eric
    Oltramari, Alessandro
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 5474 - 5483
  • [23] Exploring Pre-trained Language Models for Vocabulary Alignment in the UMLS
    Hao, Xubing
    Abeysinghe, Rashmie
    Shi, Jay
    Cui, Licong
    ARTIFICIAL INTELLIGENCE IN MEDICINE, PT I, AIME 2024, 2024, 14844 : 273 - 278
  • [24] y-Tuning: an efficient tuning paradigm for large-scale pre-trained models via label representation learning
    Liu, Yitao
    An, Chenxin
    Qiu, Xipeng
    FRONTIERS OF COMPUTER SCIENCE, 2024, 18 (04)
  • [25] Hollowed-Out Icon Colorization with Pre-trained Large-Scale Image Generation Model
    Miyauchi, Koki
    Orihara, Ryohei
    Sei, Yuichi
    Tahara, Yasuyuki
    Ohsuga, Akihiko
    Transactions of the Japanese Society for Artificial Intelligence, 2024, 39 (06)
  • [26] ON THE USE OF MODALITY-SPECIFIC LARGE-SCALE PRE-TRAINED ENCODERS FOR MULTIMODAL SENTIMENT ANALYSIS
    Ando, Atsushi
    Masumura, Ryo
    Takashima, Akihiko
    Suzuki, Satoshi
    Makishima, Naoki
    Suzuki, Keita
    Moriya, Takafumi
    Ashihara, Takanori
    Sato, Hiroshi
    2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, : 739 - 746
  • [27] Text Detoxification using Large Pre-trained Neural Models
    Dale, David
    Voronov, Anton
    Dementieva, Daryna
    Logacheva, Varvara
    Kozlova, Olga
    Semenov, Nikita
    Panchenko, Alexander
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 7979 - 7996
  • [28] Probing Toxic Content in Large Pre-Trained Language Models
    Ousidhoum, Nedjma
    Zhao, Xinran
    Fang, Tianqing
    Song, Yangqiu
    Yeung, Dit-Yan
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4262 - 4274
  • [29] TrafficBERT: Pre-trained model with large-scale data for long-range traffic flow forecasting
    Jin, KyoHoon
    Wi, JeongA
    Lee, EunJu
    Kang, ShinJin
    Kim, SooKyun
    Kim, YoungBin
    Expert Systems with Applications, 2021, 186
  • [30] TrafficBERT: Pre-trained model with large-scale data for long-range traffic flow forecasting
    Jin, KyoHoon
    Wi, JeongA
    Lee, EunJu
    Kang, ShinJin
    Kim, SooKyun
    Kim, YoungBin
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 186