A Diffusion Model Translator for Efficient Image-to-Image Translation

被引:3
|
作者
Xia, Mengfei [1 ]
Zhou, Yu [1 ]
Yi, Ran [2 ]
Liu, Yong-Jin [1 ]
Wang, Wenping [3 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, MOE Key Lab Pervas Comp, Beijing 100084, Peoples R China
[2] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai 200240, Peoples R China
[3] Texas A&M Univ, Dept Comp Sci & Comp Engn, College Stn, TX 77840 USA
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Task analysis; Noise reduction; Diffusion models; Diffusion processes; Training; Computer science; Trajectory; image translation; deep learning; generative models;
D O I
10.1109/TPAMI.2024.3435448
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Applying diffusion models to image-to-image translation (I2I) has recently received increasing attention due to its practical applications. Previous attempts inject information from the source image into each denoising step for an iterative refinement, thus resulting in a time-consuming implementation. We propose an efficient method that equips a diffusion model with a lightweight translator, dubbed a Diffusion Model Translator (DMT), to accomplish I2I. Specifically, we first offer theoretical justification that in employing the pioneering DDPM work for the I2I task, it is both feasible and sufficient to transfer the distribution from one domain to another only at some intermediate step. We further observe that the translation performance highly depends on the chosen timestep for domain transfer, and therefore propose a practical strategy to automatically select an appropriate timestep for a given task. We evaluate our approach on a range of I2I applications, including image stylization, image colorization, segmentation to image, and sketch to image, to validate its efficacy and general utility. The comparisons show that our DMT surpasses existing methods in both quality and efficiency. Code is available at https://github.com/THU-LYJ-Lab/dmt.
引用
收藏
页码:10272 / 10283
页数:12
相关论文
共 50 条
  • [41] Visualization Techniques applied to Image-to-Image Translation
    Protas, Eglen
    Bratti, Jose
    Ribeiro, Pedro O. C. S.
    Drews-, Paulo, Jr.
    Botelho, Silvia
    2018 7TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS), 2018, : 242 - 247
  • [42] Panoptic-aware Image-to-Image Translation
    Zhang, Liyun
    Ratsamee, Photchara
    Wang, Bowen
    Luo, Zhaojie
    Uranishi, Yuki
    Higashida, Manabu
    Takemura, Haruo
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 259 - 268
  • [43] Avoiding Shortcuts in Unpaired Image-to-Image Translation
    Fontanini, Tomaso
    Botti, Filippo
    Bertozzi, Massimo
    Prati, Andrea
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2022, PT I, 2022, 13231 : 463 - 475
  • [44] Image-to-image translation for wavefront and PSF estimation
    Smith, Jeffrey
    Cranney, Jesse
    Gretton, Charles
    Gratadour, Damien
    ADAPTIVE OPTICS SYSTEMS VIII, 2022, 12185
  • [45] Consistent Embedded GAN for Image-to-Image Translation
    Xiong, Feng
    Wang, Qianqian
    Gao, Quanxue
    IEEE ACCESS, 2019, 7 : 126651 - 126661
  • [46] Unaligned Image-to-Image Translation by Learning to Reweight
    Xie, Shaoan
    Gong, Mingming
    Xu, Yanwu
    Zhang, Kun
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 14154 - 14164
  • [47] Semantic Example Guided Image-to-Image Translation
    Huang, Jialu
    Liao, Jing
    Kwong, Sam
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 : 1654 - 1665
  • [48] Robotic Instrument Segmentation With Image-to-Image Translation
    Colleoni, Emanuele
    Stoyanov, Danail
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 935 - 942
  • [49] Contrastive learning for unsupervised image-to-image translation
    Lee, Hanbit
    Seol, Jinseok
    Lee, Sang-goo
    Park, Jaehui
    Shim, Junho
    APPLIED SOFT COMPUTING, 2024, 151
  • [50] Asymmetric GAN for Unpaired Image-to-Image Translation
    Li, Yu
    Tang, Sheng
    Zhang, Rui
    Zhang, Yongdong
    Li, Jintao
    Yan, Shuicheng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (12) : 5881 - 5896