Consistent Arbitrary Style Transfer Using Consistency Training and Self-Attention Module

被引:3
|
作者
Zhou, Zheng [1 ]
Wu, Yue [2 ]
Zhou, Yicong [1 ]
机构
[1] Univ Macau, Dept Comp & Informat Sci, Taipa, Macao, Peoples R China
[2] Amazon Alexa Nat Understanding, Manhattan Beach, CA 90007 USA
关键词
Image color analysis; Adaptation models; Transformers; Learning systems; Visualization; Training; Loss measurement; Arbitrary style transfer (AST); consistent training; self-attention (SA); style inconsistency;
D O I
10.1109/TNNLS.2023.3298383
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Arbitrary style transfer (AST) has garnered considerable attention for its ability to transfer styles infinitely. Although existing methods have achieved impressive results, they may overlook style consistencies and fail to capture crucial style patterns, leading to inconsistent style transfer (ST) caused by minor disturbances. To tackle this issue, we conduct a mathematical analysis of inconsistent ST and develop a style inconsistency measure (SIM) to quantify the inconsistencies between generated images. Moreover, we propose a consistent AST (CAST) framework that effectively captures and transfers essential style features into content images. The proposed CAST framework incorporates an intersection-of-union-preserving crop (IoUPC) module to obtain style pairs with minor disturbance, a self-attention (SA) module to learn the crucial style features, and a style inconsistency loss regularization (SILR) to facilitate consistent feature learning for consistent stylization. Our proposed framework not only provides an optimal solution for consistent ST but also outperforms existing methods when embedded into the CAST framework. Extensive experiments demonstrate that the proposed CAST framework can effectively transfer style patterns while preserving consistency and achieve the state-of-the-art performance.
引用
收藏
页码:16845 / 16856
页数:12
相关论文
共 50 条
  • [31] Learning convolutional self-attention module for unmanned aerial vehicle tracking
    Jun Wang
    Chenchen Meng
    Chengzhi Deng
    Yuanyun Wang
    Signal, Image and Video Processing, 2023, 17 : 2323 - 2331
  • [32] Learning convolutional self-attention module for unmanned aerial vehicle tracking
    Wang, Jun
    Meng, Chenchen
    Deng, Chengzhi
    Wang, Yuanyun
    SIGNAL IMAGE AND VIDEO PROCESSING, 2023, 17 (05) : 2323 - 2331
  • [33] Encoder-decoder Network with Self-attention Module for Image Restoration
    Jin, Qing
    Yu, Qi
    Liu, Jiying
    Tan, Xintong
    THIRTEENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING (ICGIP 2021), 2022, 12083
  • [34] ECM: arbitrary style transfer via Enhanced-Channel Module
    Yu, Xiaoming
    Zhou, Gan
    MACHINE VISION AND APPLICATIONS, 2023, 34 (04)
  • [35] ECM: arbitrary style transfer via Enhanced-Channel Module
    Xiaoming Yu
    Gan Zhou
    Machine Vision and Applications, 2023, 34
  • [36] ST-SACLF: Style Transfer Informed Self-attention Classifier for Bias-Aware Painting Classification
    Vijendran, Mridula
    Li, Frederick W. B.
    Deng, Jingjing
    Shum, Hubert P. H.
    COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS, VISIGRAPP 2023, 2024, 2103 : 181 - 205
  • [37] Acoustic model training using self-attention for low-resource speech recognition
    Park, Hosung
    Kim, Ji-Hwan
    JOURNAL OF THE ACOUSTICAL SOCIETY OF KOREA, 2020, 39 (05): : 483 - 489
  • [38] Subgraph representation learning with self-attention and free adversarial training
    Qin, Denggao
    Tang, Xianghong
    Lu, Jianguang
    APPLIED INTELLIGENCE, 2024, : 7012 - 7029
  • [39] Semi-Supervised Formality Style Transfer with Consistency Training
    Liu, Ao
    Wang, An
    Okazaki, Naoaki
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4689 - 4701
  • [40] MASKED FACE RECOGNITION VIA SELF-ATTENTION BASED LOCAL CONSISTENCY REGULARIZATION
    Lin, Dongyun
    Li, Yiqun
    Cheng, Yi
    Prasad, Shitala
    Guo, Aiyuan
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 436 - 440