GazeFusion: Saliency-Guided Image Generation

被引:0
|
作者
Zhang, Yunxiang [1 ]
Wu, Nan [2 ]
Lin, Connor Z. [2 ]
Wetzstein, Gordon [2 ]
Sun, Qi [1 ]
机构
[1] NYU, Brooklyn, NY 11201 USA
[2] Stanford Univ, Stanford, CA USA
关键词
Human Visual Attention; Perceptual Computer Graphics; Controllable Image Generation; VISUAL-ATTENTION; ALLOCATION; MODEL;
D O I
10.1145/3694969
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Diffusion models offer unprecedented image generation power given just a text prompt. While emerging approaches for controlling diffusion models have enabled users to specify the desired spatial layouts of the generated content, they cannot predict or control where viewers will pay more attention due to the complexity of human vision. Recognizing the significance of attention-controllable image generation in practical applications, we present a saliency-guided framework to incorporate the data priors of human visual attention mechanisms into the generation process. Given a user-specified viewer attention distribution, our control module conditions a diffusion model to generate images that attract viewers' attention toward the desired regions. To assess the efficacy of our approach, we performed an eye-tracked user study and a large-scale model-based saliency analysis. The results evidence that both the cross-user eye gaze distributions and the saliency models' predictions align with the desired attention distributions. Lastly, we outline several applications, including interactive design of saliency guidance, attention suppression in unwanted regions, and adaptive generation for varied display/viewing conditions.
引用
收藏
页数:19
相关论文
共 50 条
  • [41] Saliency-guided feature learning in a probabilistic categorization task
    Chenkov, N.
    Nelson, J-D
    PERCEPTION, 2010, 39 : 55 - 56
  • [42] Saliency-Guided No-Reference Omnidirectional Image Quality Assessment via Scene Content Perceiving
    Zhang, Youzhi
    Wan, Lifei
    Liu, Deyang
    Zhou, Xiaofei
    An, Ping
    Shan, Caifeng
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73
  • [43] SiSL-Net: Saliency-guided self-supervised learning network for image classification
    Liu, Kun
    Meng, Rui
    Li, Longteng
    Mao, Jingkun
    Chen, Haiyong
    NEUROCOMPUTING, 2022, 510 : 193 - 202
  • [44] No-reference stereoscopic image quality assessment based on saliency-guided binocular feature consolidation
    Xu, Xiaogang
    Zhao, Yang
    Ding, Yong
    ELECTRONICS LETTERS, 2017, 53 (22) : 1468 - 1469
  • [45] A saliency-guided street view image inpainting framework for efficient last-meters wayfinding
    Hu, Chuanbo
    Jia, Shan
    Zhang, Fan
    Li, Xin
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2023, 195 : 365 - 379
  • [46] HAZY REMOTE SENSING IMAGE RESTORATION BASED ON SALIENCY-GUIDED TRANSMISSION OPTIMIZATION AND TEXTURE BOOSTING
    Liu, Yanmeng
    Zhang, Libao
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 6109 - 6112
  • [47] Saliency-Guided Unsupervised Feature Learning for Scene Classification
    Zhang, Fan
    Du, Bo
    Zhang, Liangpei
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2015, 53 (04): : 2175 - 2184
  • [48] Saliency-Guided Quality Assessment of Screen Content Images
    Gu, Ke
    Wang, Shiqi
    Yang, Huan
    Lin, Weisi
    Zhai, Guangtao
    Yang, Xiaokang
    Zhang, Wenjun
    IEEE TRANSACTIONS ON MULTIMEDIA, 2016, 18 (06) : 1098 - 1110
  • [49] Saliency-Guided Transformer Network combined with Local Embedding for No-Reference Image Quality Assessment
    Zhu, Mengmeng
    Hou, Guanqun
    Chen, Xinjia
    Xie, Jiaxing
    Lu, Haixian
    Che, Jun
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 1953 - 1962
  • [50] SGDViT: Saliency-Guided Dynamic Vision Transformer for UAV Tracking
    Yao, Liangliang
    Fu, Changhong
    Li, Sihang
    Zheng, Guangze
    Ye, Junjie
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 3353 - 3359