Consistent Arbitrary Style Transfer Using Consistency Training and Self-Attention Module

被引:3
|
作者
Zhou, Zheng [1 ]
Wu, Yue [2 ]
Zhou, Yicong [1 ]
机构
[1] Univ Macau, Dept Comp & Informat Sci, Taipa, Macao, Peoples R China
[2] Amazon Alexa Nat Understanding, Manhattan Beach, CA 90007 USA
关键词
Image color analysis; Adaptation models; Transformers; Learning systems; Visualization; Training; Loss measurement; Arbitrary style transfer (AST); consistent training; self-attention (SA); style inconsistency;
D O I
10.1109/TNNLS.2023.3298383
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Arbitrary style transfer (AST) has garnered considerable attention for its ability to transfer styles infinitely. Although existing methods have achieved impressive results, they may overlook style consistencies and fail to capture crucial style patterns, leading to inconsistent style transfer (ST) caused by minor disturbances. To tackle this issue, we conduct a mathematical analysis of inconsistent ST and develop a style inconsistency measure (SIM) to quantify the inconsistencies between generated images. Moreover, we propose a consistent AST (CAST) framework that effectively captures and transfers essential style features into content images. The proposed CAST framework incorporates an intersection-of-union-preserving crop (IoUPC) module to obtain style pairs with minor disturbance, a self-attention (SA) module to learn the crucial style features, and a style inconsistency loss regularization (SILR) to facilitate consistent feature learning for consistent stylization. Our proposed framework not only provides an optimal solution for consistent ST but also outperforms existing methods when embedded into the CAST framework. Extensive experiments demonstrate that the proposed CAST framework can effectively transfer style patterns while preserving consistency and achieve the state-of-the-art performance.
引用
收藏
页码:16845 / 16856
页数:12
相关论文
共 50 条
  • [41] Multilingual Constituency Parsing with Self-Attention and Pre-Training
    Kitaev, Nikita
    Cao, Steven
    Klein, Dan
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3499 - 3505
  • [42] Speaker identification for household scenarios with self-attention and adversarial training
    Li, Ruirui
    Jiang, Jyun-Yu
    Wu, Xian
    Hsieh, Chu-Cheng
    Stolcke, Andreas
    Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, 2020, 2020-October : 2272 - 2276
  • [43] Speaker Identification for Household Scenarios with Self-attention and Adversarial Training
    Li, Ruirui
    Joang, Jyun-Yu
    Wu, Xian
    Hsieh, Chu-Cheng
    Stolcke, Andreas
    INTERSPEECH 2020, 2020, : 2272 - 2276
  • [44] Improved Feature-Based Gaze Estimation Using Self-Attention Module and Synthetic Eye Images
    Oh, Jaekwang
    Lee, Youngkeun
    Yoo, Jisang
    Kwon, Soonchul
    SENSORS, 2022, 22 (11)
  • [45] Surface defect segmentation of magnetic tiles based on cross self-attention module
    Liu, Hong
    Wang, Gaihua
    Li, Qi
    Wang, Nengyuan
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (06) : 9523 - 9532
  • [46] Salient Object Detection Combining a Self-Attention Module and a Feature Pyramid Network
    Ren, Guangyu
    Dai, Tianhong
    Barmpoutis, Panagiotis
    Stathaki, Tania
    ELECTRONICS, 2020, 9 (10) : 1 - 13
  • [47] Remote Sensing Image Scene Classification Based on Global Self-Attention Module
    Li, Qingwen
    Yan, Dongmei
    Wu, Wanrong
    REMOTE SENSING, 2021, 13 (22)
  • [48] Skeleton action recognition via graph convolutional network with self-attention module
    Li, Min
    Chen, Ke
    Bai, Yunqing
    Pei, Jihong
    ELECTRONIC RESEARCH ARCHIVE, 2024, 32 (04): : 2848 - 2864
  • [49] Compression Reconstruction Network with Coordinated Self-Attention and Adaptive Gaussian Filtering Module
    Wei, Zhen
    Yan, Qiurong
    Lu, Xiaoqiang
    Zheng, Yongjian
    Sun, Shida
    Lin, Jian
    MATHEMATICS, 2023, 11 (04)
  • [50] Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention
    Springstein, Matthias
    Mueller-Budack, Eric
    Ewerth, Ralph
    MMPT '21: PROCEEDINGS OF THE 2021 WORKSHOP ON MULTI-MODAL PRE-TRAINING FOR MULTIMEDIA UNDERSTANDING, 2021, : 46 - 54