Perception-Oriented U-Shaped Transformer Network for 360-Degree No-Reference Image Quality Assessment

被引:40
|
作者
Zhou, Mingliang [1 ]
Chen, Lei [1 ]
Wei, Xuekai [1 ]
Liao, Xingran [2 ]
Mao, Qin [3 ,4 ]
Wang, Heqiang [1 ]
Pu, Huayan [5 ]
Luo, Jun [5 ]
Xiang, Tao [1 ]
Fang, Bin [1 ]
机构
[1] Chongqing Univ, Sch Comp Sci, Chongqing 400044, Peoples R China
[2] City Univ Hong Kong, Comp Sci Dept, Hong Kong, Peoples R China
[3] Qiannan Normal Coll Nationalities, Coll Comp & Informat, Duyun 558000, Peoples R China
[4] Qiannan Normal Univ Nationalities, Sch Comp & Informat, Key Lab Complex Syst & Intelligent Optimizat Guizh, Duyun 558000, Peoples R China
[5] Chongqing Univ, State Key Lab Mech Transmiss, Chongqing 400044, Peoples R China
基金
中国国家自然科学基金;
关键词
Image quality assessment; no-reference image quality assessment; 360-degree image; U-shaped transformer; OMNIDIRECTIONAL IMAGE; SALIENCY; CNN;
D O I
10.1109/TBC.2022.3231101
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Generally, 360-degree images have absolute senses of reality and three-dimensionality, providing a wide range of immersive interactions. Due to the novel rendering and display technology of 360-degree images, they have more complex perceptual characteristics than other images. It is challenging to perform comprehensive image quality assessment (IQA) learning by simply stacking multichannel neural network architectures for pre/postprocessing, compression, and rendering tasks. To thoroughly learn the global and local features in 360-degree images, reduce the complexity of multichannel neural network models and simplify the training process, this paper proposes a joint architecture with user perception and an efficient transformer dedicated to 360-degree no-reference (NR) IQA. The input of the proposed method is a 360-degree cube map projection (CMP) image. Furthermore, the proposed 360-degree NRIQA method includes a saliency map-based non-overlapping self-attention selection module and a U-shaped transformer (U-former)-based feature extraction module to account for perceptual region importance and projection distortion. The transformer-based architecture and the weighted average technique are jointly utilized for predicting local perceptual quality. Experimental results obtained on widely used databases show that the proposed model outperforms other state-of-the-art methods in NR 360-degree image quality evaluation cases. Furthermore, a cross-database evaluation and an ablation study also demonstrate the inherent robustness and generalization ability of the proposed model.
引用
收藏
页码:396 / 405
页数:10
相关论文
共 50 条
  • [31] Cubemap-Based Perception-Driven Blind Quality Assessment for 360-degree Images
    Jiang, Hao
    Jiang, Gangyi
    Yu, Mei
    Zhang, Yun
    Yang, You
    Peng, Zongju
    Chen, Fen
    Zhang, Qingbo
    IEEE Transactions on Image Processing, 2021, 30 : 2364 - 2377
  • [32] MC360IQA: A Multi-channel CNN for Blind 360-Degree Image Quality Assessment
    Sun, Wei
    Min, Xiongkuo
    Zhai, Guangtao
    Gu, Ke
    Duan, Huiyu
    Ma, Siwei
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (01) : 64 - 77
  • [33] MC360IQA: THE MULTI-CHANNEL CNN FOR BLIND 360-DEGREE IMAGE QUALITY ASSESSMENT
    Sun, Wei
    Gu, Ke
    Luo, Weike
    Min, Xiongkuo
    Zhai, Guangtao
    Ma, Siwei
    Yang, Xiaokang
    2019 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2019,
  • [34] TU-Former: A Hybrid U-Shaped Transformer Network for SAR Image Denoising
    Tian, Shikang
    Liu, Shuaiqi
    Zhao, Yuhang
    Liu, Siyuan
    Zhao, Shuhuan
    Zhao, Jie
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XI, 2024, 14435 : 377 - 389
  • [35] No-reference Image Contrast Quality Assessment based on the Degree of Uniformity in Probability Distribution
    Kim W.
    Yim C.
    IEIE Transactions on Smart Processing and Computing, 2022, 11 (02): : 85 - 91
  • [36] FreqAlign: Excavating Perception-Oriented Transferability for Blind Image Quality Assessment From a Frequency Perspective
    Li, Xin
    Lu, Yiting
    Chen, Zhibo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 4652 - 4666
  • [37] VCRNet: Visual Compensation Restoration Network for No-Reference Image Quality Assessment
    Pan, Zhaoqing
    Yuan, Feng
    Lei, Jianjun
    Fang, Yuming
    Shao, Xiao
    Kwong, Sam
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 1613 - 1627
  • [38] NESTED ERROR MAP GENERATION NETWORK FOR NO-REFERENCE IMAGE QUALITY ASSESSMENT
    Chen, Junming
    Wang, Haiqiang
    Li, Ge
    Liu, Shan
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2080 - 2084
  • [39] A Multibranch Network With Multilayer Feature Fusion for No-Reference Image Quality Assessment
    Zhao, Wenqing
    Li, Mengwei
    Xu, Lijiao
    Sun, Yue
    Zhao, Zhenbing
    Zhai, Yongjie
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73
  • [40] Convolutional Neural Network with Uncertainty Estimates for No-reference Image Quality Assessment
    Huang, Yuge
    Tian, Xiang
    Jiang, Rongxin
    Chen, Yaowu
    TENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING (ICGIP 2018), 2019, 11069