Nested Deformable Multi-head Attention for Facial Image Inpainting

被引:3
|
作者
Phutke, Shruti S. [1 ]
Murala, Subrahmanyam [1 ]
机构
[1] Indian Inst Technol Ropar, CVPR Lab, Ropar, Punjab, India
关键词
NETWORK;
D O I
10.1109/WACV56688.2023.00602
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Extracting adequate contextual information is an important aspect of any image inpainting method. To achieve this, ample image inpainting methods are available that aim to focus on large receptive fields. Recent advancements in the deep learning field with the introduction of transformers for image inpainting paved the way toward plausible results. Stacking multiple transformer blocks in a single layer causes the architecture to become computationally complex. In this context, we propose a novel lightweight architecture with a nested deformable attention-based transformer layer for feature fusion. The nested attention helps the network to focus on long-term dependencies from encoder and decoder features. Also, multi-head attention consisting of a deformable convolution is proposed to delve into the diverse receptive fields. With the advantage of nested and deformable attention, we propose a lightweight architecture for facial image inpainting. The results comparison on Celeb HQ [25] dataset using known (NVIDIA) and unknown (QD-IMD) masks and Places2 [57] dataset with NVIDIA masks along with extensive ablation study prove the superiority of the proposed approach for image inpainting tasks. The code is available at: https://github.com/shrutiphutke/NDMA_Facial_Inpainting.
引用
收藏
页码:6067 / 6076
页数:10
相关论文
共 50 条
  • [1] Multi-Scale Generative Adversarial Network With Multi-Head External Attention for Image Inpainting
    Chen, Gang
    Feng, Qing
    He, Xiu
    Yao, Jian
    IEEE ACCESS, 2024, 12 : 133456 - 133468
  • [2] Symmetric Connected U-Net with Multi-Head Self Attention (MHSA) and WGAN for Image Inpainting
    Hou, Yanyang
    Ma, Xiaopeng
    Zhang, Junjun
    Guo, Chenxian
    SYMMETRY-BASEL, 2024, 16 (11):
  • [3] On the diversity of multi-head attention
    Li, Jian
    Wang, Xing
    Tu, Zhaopeng
    Lyu, Michael R.
    NEUROCOMPUTING, 2021, 454 : 14 - 24
  • [4] Multi-head attention with CNN and wavelet for classification of hyperspectral image
    Tulapurkar, Harshula
    Banerjee, Biplab
    Buddhiraju, Krishna Mohan
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (10): : 7595 - 7609
  • [5] Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition
    Wen, Zhengyao
    Lin, Wenzhong
    Wang, Tao
    Xu, Ge
    BIOMIMETICS, 2023, 8 (02)
  • [6] Multi-head attention with CNN and wavelet for classification of hyperspectral image
    Harshula Tulapurkar
    Biplab Banerjee
    Krishna Mohan Buddhiraju
    Neural Computing and Applications, 2023, 35 : 7595 - 7609
  • [7] Omni-Frequency Image Denoising with Multi-Head Attention
    Jiang, Jielin
    Shi, Mingyue
    Yang, Haidong
    Cui, Yan
    Computer Engineering and Applications, 60 (16): : 236 - 247
  • [8] Multi-Head Attention Affinity Diversity Sharing Network for Facial Expression Recognition
    Zheng, Caixia
    Liu, Jiayu
    Zhao, Wei
    Ge, Yingying
    Chen, Wenhe
    ELECTRONICS, 2024, 13 (22)
  • [9] Combining Multi-Head Attention and Sparse Multi-Head Attention Networks for Session-Based Recommendation
    Zhao, Zhiwei
    Wang, Xiaoye
    Xiao, Yingyuan
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [10] Local Multi-Head Channel Self-Attention for Facial Expression Recognition
    Pecoraro, Roberto
    Basile, Valerio
    Bono, Viviana
    INFORMATION, 2022, 13 (09)