GazeFusion: Saliency-Guided Image Generation

被引:0
|
作者
Zhang, Yunxiang [1 ]
Wu, Nan [2 ]
Lin, Connor Z. [2 ]
Wetzstein, Gordon [2 ]
Sun, Qi [1 ]
机构
[1] NYU, Brooklyn, NY 11201 USA
[2] Stanford Univ, Stanford, CA USA
关键词
Human Visual Attention; Perceptual Computer Graphics; Controllable Image Generation; VISUAL-ATTENTION; ALLOCATION; MODEL;
D O I
10.1145/3694969
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Diffusion models offer unprecedented image generation power given just a text prompt. While emerging approaches for controlling diffusion models have enabled users to specify the desired spatial layouts of the generated content, they cannot predict or control where viewers will pay more attention due to the complexity of human vision. Recognizing the significance of attention-controllable image generation in practical applications, we present a saliency-guided framework to incorporate the data priors of human visual attention mechanisms into the generation process. Given a user-specified viewer attention distribution, our control module conditions a diffusion model to generate images that attract viewers' attention toward the desired regions. To assess the efficacy of our approach, we performed an eye-tracked user study and a large-scale model-based saliency analysis. The results evidence that both the cross-user eye gaze distributions and the saliency models' predictions align with the desired attention distributions. Lastly, we outline several applications, including interactive design of saliency guidance, attention suppression in unwanted regions, and adaptive generation for varied display/viewing conditions.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Saliency-Guided Image Translation
    Jiang, Lai
    Xu, Mai
    Wang, Xiaofei
    Sigal, Leonid
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16504 - 16513
  • [2] Saliency-guided image translation
    Jiang, Lai
    Dai, Ning
    Xu, Mai
    Deng, Xin
    Li, Shengxi
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2023, 49 (10): : 2689 - 2698
  • [3] SALIENCY-GUIDED IMAGE STYLE TRANSFER
    Liu, Xiuwen
    Liu, Zhi
    Zhou, Xiaofei
    Chen, Minyu
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, : 66 - 71
  • [4] Saliency-Guided Lighting
    Lee, Chang Ha
    Kim, Youngmin
    Varshney, Amitabh
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2009, E92D (02): : 369 - 373
  • [5] Unsupervised saliency-guided SAR image change detection
    Zheng, Yaoguo
    Jiao, Licheng
    Liu, Hongying
    Zhang, Xiangrong
    Hou, Biao
    Wang, Shuang
    PATTERN RECOGNITION, 2017, 61 : 309 - 326
  • [6] BE NATURAL: A SALIENCY-GUIDED DEEP FRAMEWORK FOR IMAGE QUALITY
    Hou, Weilong
    Gao, Xinbo
    2014 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2014,
  • [7] Saliency-Guided Deep Framework for Image Quality Assessment
    Hou, Weilong
    Gao, Xinbo
    IEEE MULTIMEDIA, 2015, 22 (02) : 46 - 55
  • [8] Saliency-Guided Video Deinterlacing
    Trocan, Maria
    Coudoux, Francois-Xavier
    COMPUTATIONAL COLLECTIVE INTELLIGENCE (ICCCI 2015), PT II, 2015, 9330 : 24 - 33
  • [9] Saliency-Guided Remote Sensing Image Super-Resolution
    Liu, Baodi
    Zhao, Lifei
    Li, Jiaoyue
    Zhao, Hengle
    Liu, Weifeng
    Li, Ye
    Wang, Yanjiang
    Chen, Honglong
    Cao, Weijia
    REMOTE SENSING, 2021, 13 (24)
  • [10] SGSR: A SALIENCY-GUIDED IMAGE SUPER-RESOLUTION NETWORK
    Kim, Dayeon
    Kim, Munchurl
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 980 - 984