CRS-Diff: Controllable Remote Sensing Image Generation With Diffusion Model

被引:3
|
作者
Tang, Datao [1 ,2 ]
Cao, Xiangyong [1 ,2 ]
Hou, Xingsong [3 ]
Jiang, Zhongyuan [4 ]
Liu, Junmin [5 ]
Meng, Deyu [2 ,5 ,6 ]
机构
[1] Xi An Jiao Tong Univ, Sch Comp Sci & Technol, Xian 710049, Peoples R China
[2] Xi An Jiao Tong Univ, Key Lab Intelligent Networks & Network Secur, Minist Educ, Xian 710049, Peoples R China
[3] Xi An Jiao Tong Univ, Sch Informat & Commun Engn, Xian 710049, Shaanxi, Peoples R China
[4] Xidian Univ, Sch Cyber Engn, Xian 710049, Shaanxi, Peoples R China
[5] Xi An Jiao Tong Univ, Sch Math & Stat, Xian 710049, Shaanxi, Peoples R China
[6] Macau Univ Scienceand Technol, Macao Inst Syst Engn, Taipa, Macao, Peoples R China
关键词
Diffusion models; Image synthesis; Image resolution; Text to image; Remote sensing; Training; Task analysis; Controllable generation; deep learning; diffusion model; remote sensing (RS) image;
D O I
10.1109/TGRS.2024.3453414
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
The emergence of generative models has revolutionized the field of remote sensing (RS) image generation. Despite generating high-quality images, existing methods are limited in relying mainly on text control conditions, and thus do not always generate images accurately and stably. In this article, we propose CRS-Diff, a new RS generative framework specifically tailored for RS image generation, leveraging the inherent advantages of diffusion models while integrating more advanced control mechanisms. Specifically, CRS-Diff can simultaneously support text-condition, metadata-condition, and image-condition control inputs, thus enabling more precise control to refine the generation process. To effectively integrate multiple condition control information, we introduce a new conditional control mechanism to achieve multiscale feature fusion (FF), thus enhancing the guiding effect of control conditions. To the best of our knowledge, CRS-Diff is the first multiple-condition controllable RS generative model. Experimental results in single-condition and multiple-condition cases have demonstrated the superior ability of our CRS-Diff to generate RS images both quantitatively and qualitatively compared with previous methods. Additionally, our CRS-Diff can serve as a data engine that generates high-quality training data for downstream tasks, e.g., road extraction. The code is available at https://github.com/Sonettoo/CRS-Diff.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Diffusion-based remote sensing image fusion for classification
    Jiang, Yuling
    Liu, Shujun
    Wang, Huajun
    APPLIED INTELLIGENCE, 2025, 55 (03)
  • [32] Remote sensing image enhancement using hazy image model
    Kaplan, N. H.
    OPTIK, 2018, 155 : 139 - 148
  • [33] Surface and buried landmine scene generation and validation using the digital Imaging and remote sensing image generation model
    Peterson, ED
    Brown, SD
    Hattenberger, TJ
    Schott, JR
    IMAGING SPECTROMETRY X, 2004, 5546 : 312 - 323
  • [34] MIGN: Multiscale Image Generation Network for Remote Sensing Image Semantic Segmentation
    Nie, Jie
    Wang, Chenglong
    Yu, Shusong
    Shi, Jinjin
    Lv, Xiaowei
    Wei, Zhiqiang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 5601 - 5613
  • [35] Exploring Models and Data for Remote Sensing Image Caption Generation
    Lu, Xiaoqiang
    Wang, Binqiang
    Zheng, Xiangtao
    Li, Xuelong
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2018, 56 (04): : 2183 - 2195
  • [36] Swin-Diff: a single defocus image deblurring network based on diffusion model
    Liang, Hanyan
    Chai, Shuyao
    Zhao, Xixuan
    Kan, Jiangming
    COMPLEX & INTELLIGENT SYSTEMS, 2025, 11 (03)
  • [37] Scene Attention Mechanism for Remote Sensing Image Caption Generation
    Wu, Shiqi
    Zhang, Xiangrong
    Wang, Xin
    Li, Chen
    Jiao, Licheng
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [38] CDRM: Controllable diffusion restoration model for realistic image deblurring
    Chen, Ziyi
    Cui, Guangmang
    Zhao, Jufeng
    Nie, Jiahao
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 275
  • [39] UniControl: A Unified Diffusion Model for Controllable Visual Generation In the Wild
    Qin, Can
    Zhang, Shu
    Yu, Ning
    Feng, Yihao
    Yang, Xinyi
    Zhou, Yingbo
    Wang, Huan
    Niebles, Juan Carlos
    Xiong, Caiming
    Savarese, Silvio
    Ermon, Stefano
    Fu, Yun
    Xu, Ran
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [40] Universal Fingerprint Generation: Controllable Diffusion Model With Multimodal Conditions
    Grosz, Steven A.
    Jain, Anil K.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (02) : 1028 - 1041