GazeFusion: Saliency-Guided Image Generation

被引:0
|
作者
Zhang, Yunxiang [1 ]
Wu, Nan [2 ]
Lin, Connor Z. [2 ]
Wetzstein, Gordon [2 ]
Sun, Qi [1 ]
机构
[1] NYU, Brooklyn, NY 11201 USA
[2] Stanford Univ, Stanford, CA USA
关键词
Human Visual Attention; Perceptual Computer Graphics; Controllable Image Generation; VISUAL-ATTENTION; ALLOCATION; MODEL;
D O I
10.1145/3694969
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Diffusion models offer unprecedented image generation power given just a text prompt. While emerging approaches for controlling diffusion models have enabled users to specify the desired spatial layouts of the generated content, they cannot predict or control where viewers will pay more attention due to the complexity of human vision. Recognizing the significance of attention-controllable image generation in practical applications, we present a saliency-guided framework to incorporate the data priors of human visual attention mechanisms into the generation process. Given a user-specified viewer attention distribution, our control module conditions a diffusion model to generate images that attract viewers' attention toward the desired regions. To assess the efficacy of our approach, we performed an eye-tracked user study and a large-scale model-based saliency analysis. The results evidence that both the cross-user eye gaze distributions and the saliency models' predictions align with the desired attention distributions. Lastly, we outline several applications, including interactive design of saliency guidance, attention suppression in unwanted regions, and adaptive generation for varied display/viewing conditions.
引用
收藏
页数:19
相关论文
共 50 条
  • [21] Saliency-Guided Consistent Color Harmonization
    Baveye, Yoann
    Urban, Fabrice
    Chamaret, Christel
    Demoulin, Vincent
    Hellier, Pierre
    COMPUTATIONAL COLOR IMAGING, CCIW 2013, 2013, 7786 : 105 - 118
  • [22] UAV Image Haze Removal Based on Saliency-Guided Parallel Learning Mechanism
    Zheng, Ruohui
    Zhang, Libao
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2023, 20
  • [23] Partial-Duplicate Image Retrieval via Saliency-Guided Visual Matching
    Li, Liang
    Jiang, Shuqiang
    Zha, Zheng-Jun
    Wu, Zhipeng
    Huang, Qingming
    IEEE MULTIMEDIA, 2013, 20 (03) : 13 - 23
  • [24] Cluster-Based Saliency-Guided Content-Aware Image Retargeting
    Li-Wei Kang
    Ching-Yu Tseng
    Chao-Long Jheng
    Ming-Fang Weng
    Chao-Yung Hsu
    Journal of Electronic Science and Technology, 2017, 15 (02) : 141 - 146
  • [25] Multilevel saliency-guided self-supervised learning for image anomaly detection
    Qin, Jianjian
    Gu, Chunzhi
    Yu, Jun
    Zhang, Chao
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (8-9) : 6339 - 6351
  • [26] Saliency-Guided Nonsubsampled Shearlet Transform for Multisource Remote Sensing Image Fusion
    Li, Liangliang
    Ma, Hongbing
    SENSORS, 2021, 21 (05) : 1 - 14
  • [27] SageMix: Saliency-Guided Mixup for Point Clouds
    Lee, Sanghyeok
    Jeon, Minkyu
    Kim, Injae
    Xiong, Yunyang
    Kim, Hyunwoo J.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [28] Saliency-Guided Complexity Control for HEVC Decoding
    Yang, Ren
    Xu, Mai
    Wang, Zulin
    Duan, Yiping
    Tao, Xiaoming
    IEEE TRANSACTIONS ON BROADCASTING, 2018, 64 (04) : 865 - 882
  • [29] Saliency-Guided Color Transfer between Images
    Xia, Jiazhi
    ADVANCES IN VISUAL COMPUTING, ISVC 2013, PT I, 2013, 8033 : 468 - 475
  • [30] Saliency-guided Adaptive Seeding for Supervoxel Segmentation
    Gao, Ge
    Lauri, Mikko
    Zhang, Jianwei
    Frintrop, Simone
    2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 4938 - 4943