Nested Deformable Multi-head Attention for Facial Image Inpainting

被引:3
|
作者
Phutke, Shruti S. [1 ]
Murala, Subrahmanyam [1 ]
机构
[1] Indian Inst Technol Ropar, CVPR Lab, Ropar, Punjab, India
来源
2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV) | 2023年
关键词
NETWORK;
D O I
10.1109/WACV56688.2023.00602
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Extracting adequate contextual information is an important aspect of any image inpainting method. To achieve this, ample image inpainting methods are available that aim to focus on large receptive fields. Recent advancements in the deep learning field with the introduction of transformers for image inpainting paved the way toward plausible results. Stacking multiple transformer blocks in a single layer causes the architecture to become computationally complex. In this context, we propose a novel lightweight architecture with a nested deformable attention-based transformer layer for feature fusion. The nested attention helps the network to focus on long-term dependencies from encoder and decoder features. Also, multi-head attention consisting of a deformable convolution is proposed to delve into the diverse receptive fields. With the advantage of nested and deformable attention, we propose a lightweight architecture for facial image inpainting. The results comparison on Celeb HQ [25] dataset using known (NVIDIA) and unknown (QD-IMD) masks and Places2 [57] dataset with NVIDIA masks along with extensive ablation study prove the superiority of the proposed approach for image inpainting tasks. The code is available at: https://github.com/shrutiphutke/NDMA_Facial_Inpainting.
引用
收藏
页码:6067 / 6076
页数:10
相关论文
共 50 条
  • [31] Classification of Heads in Multi-head Attention Mechanisms
    Huang, Feihu
    Jiang, Min
    Liu, Fang
    Xu, Dian
    Fan, Zimeng
    Wang, Yonghao
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2022, PT III, 2022, 13370 : 681 - 692
  • [32] Diversifying Multi-Head Attention in the Transformer Model
    Ampazis, Nicholas
    Sakketou, Flora
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (04): : 2618 - 2638
  • [33] A Multi-Head Convolutional Neural Network with Multi-Path Attention Improves Image Denoising
    Zhang, Jiahong
    Qu, Meijun
    Wang, Ye
    Cao, Lihong
    PRICAI 2022: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2022, 13631 : 338 - 351
  • [34] Finding the Pillars of Strength for Multi-Head Attention
    Ni, Jinjie
    Mao, Rui
    Yang, Zonglin
    Lei, Han
    Cambria, Erik
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14526 - 14540
  • [35] Improving Multi-head Attention with Capsule Networks
    Gu, Shuhao
    Feng, Yang
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 314 - 326
  • [36] Abstractive Text Summarization with Multi-Head Attention
    Li, Jinpeng
    Zhang, Chuang
    Chen, Xiaojun
    Cao, Yanan
    Liao, Pengcheng
    Zhang, Peng
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [37] Classification of Facial Expression In-the-Wild based on Ensemble of Multi-head Cross Attention Networks
    Jeong, Jae Yeop
    Hong, Yeong-Gi
    Kim, Daun
    Jeong, Jin-Woo
    Jung, Yuchul
    Kim, Sang-Ho
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 2352 - 2357
  • [38] Repulsive Attention: Rethinking Multi-head Attention as Bayesian Inference
    An, Bang
    Lyu, Jie
    Wang, Zhenyi
    Li, Chunyuan
    Hu, Changwei
    Tan, Fei
    Zhang, Ruiyi
    Hu, Yifan
    Chen, Changyou
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 236 - 255
  • [39] A multi-head adjacent attention-based pyramid layered model for nested named entity recognition
    Shengmin Cui
    Inwhee Joe
    Neural Computing and Applications, 2023, 35 : 2561 - 2574
  • [40] A multi-head adjacent attention-based pyramid layered model for nested named entity recognition
    Cui, Shengmin
    Joe, Inwhee
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (03): : 2561 - 2574