Nested Deformable Multi-head Attention for Facial Image Inpainting

被引:3
|
作者
Phutke, Shruti S. [1 ]
Murala, Subrahmanyam [1 ]
机构
[1] Indian Inst Technol Ropar, CVPR Lab, Ropar, Punjab, India
关键词
NETWORK;
D O I
10.1109/WACV56688.2023.00602
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Extracting adequate contextual information is an important aspect of any image inpainting method. To achieve this, ample image inpainting methods are available that aim to focus on large receptive fields. Recent advancements in the deep learning field with the introduction of transformers for image inpainting paved the way toward plausible results. Stacking multiple transformer blocks in a single layer causes the architecture to become computationally complex. In this context, we propose a novel lightweight architecture with a nested deformable attention-based transformer layer for feature fusion. The nested attention helps the network to focus on long-term dependencies from encoder and decoder features. Also, multi-head attention consisting of a deformable convolution is proposed to delve into the diverse receptive fields. With the advantage of nested and deformable attention, we propose a lightweight architecture for facial image inpainting. The results comparison on Celeb HQ [25] dataset using known (NVIDIA) and unknown (QD-IMD) masks and Places2 [57] dataset with NVIDIA masks along with extensive ablation study prove the superiority of the proposed approach for image inpainting tasks. The code is available at: https://github.com/shrutiphutke/NDMA_Facial_Inpainting.
引用
收藏
页码:6067 / 6076
页数:10
相关论文
共 50 条
  • [21] A facial depression recognition method based on hybrid multi-head cross attention network
    Li, Yutong
    Liu, Zhenyu
    Zhou, Li
    Yuan, Xiaoyan
    Shangguan, Zixuan
    Hu, Xiping
    Hu, Bin
    FRONTIERS IN NEUROSCIENCE, 2023, 17
  • [22] Lightweight Facial Expression Recognition Based on Hybrid Multiscale and Multi-Head Collaborative Attention
    Zhang, Haitao
    Zhuang, Xufei
    Gao, Xudong
    Mao, Rui
    Ren, Qing-Dao-Er-Ji
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT II, 2025, 15032 : 304 - 316
  • [23] Improving Vision Transformers with Nested Multi-head Attentions
    Peng, Jiquan
    Li, Chaozhuo
    Zhao, Yi
    Lin, Yuting
    Fang, Xiaohan
    Gong, Jibing
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1925 - 1930
  • [24] Federated learning based multi-head attention framework for medical image classification
    Firdaus, Naima
    Raza, Zahid
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2024, 36 (27):
  • [25] Advancing spatial mapping for satellite image road segmentation with multi-head attention
    Ben Salah, Khawla
    Othmani, Mohamed
    Fourati, Jihen
    Kherallah, Monji
    VISUAL COMPUTER, 2025, 41 (04): : 2079 - 2089
  • [26] Self Multi-Head Attention for Speaker Recognition
    India, Miquel
    Safari, Pooyan
    Hernando, Javier
    INTERSPEECH 2019, 2019, : 4305 - 4309
  • [27] DOUBLE MULTI-HEAD ATTENTION FOR SPEAKER VERIFICATION
    India, Miquel
    Safari, Pooyan
    Hernando, Javier
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6144 - 6148
  • [28] Learning Sentences Similarity By Multi-Head Attention
    Wang, Ming Yang
    Li, Chen Jiang
    Sun, Jian Dong
    Xu, Wei Ran
    Gao, Sheng
    Zhang, Ya Hao
    Wang, Pu
    Li, Jun Liang
    PROCEEDINGS OF 2018 INTERNATIONAL CONFERENCE ON NETWORK INFRASTRUCTURE AND DIGITAL CONTENT (IEEE IC-NIDC), 2018, : 16 - 19
  • [29] VIDEO SUMMARIZATION WITH ANCHORS AND MULTI-HEAD ATTENTION
    Sung, Yi-Lin
    Hong, Cheng-Yao
    Hsu, Yen-Chi
    Liu, Tyng-Luh
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 2396 - 2400
  • [30] Software and Hardware Fusion Multi-Head Attention
    Hu, Wei
    Xu, Dian
    Liu, Fang
    Fan, Zimeng
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2022, PT III, 2022, 13370 : 644 - 655