Perception-Oriented U-Shaped Transformer Network for 360-Degree No-Reference Image Quality Assessment

被引:40
|
作者
Zhou, Mingliang [1 ]
Chen, Lei [1 ]
Wei, Xuekai [1 ]
Liao, Xingran [2 ]
Mao, Qin [3 ,4 ]
Wang, Heqiang [1 ]
Pu, Huayan [5 ]
Luo, Jun [5 ]
Xiang, Tao [1 ]
Fang, Bin [1 ]
机构
[1] Chongqing Univ, Sch Comp Sci, Chongqing 400044, Peoples R China
[2] City Univ Hong Kong, Comp Sci Dept, Hong Kong, Peoples R China
[3] Qiannan Normal Coll Nationalities, Coll Comp & Informat, Duyun 558000, Peoples R China
[4] Qiannan Normal Univ Nationalities, Sch Comp & Informat, Key Lab Complex Syst & Intelligent Optimizat Guizh, Duyun 558000, Peoples R China
[5] Chongqing Univ, State Key Lab Mech Transmiss, Chongqing 400044, Peoples R China
基金
中国国家自然科学基金;
关键词
Image quality assessment; no-reference image quality assessment; 360-degree image; U-shaped transformer; OMNIDIRECTIONAL IMAGE; SALIENCY; CNN;
D O I
10.1109/TBC.2022.3231101
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Generally, 360-degree images have absolute senses of reality and three-dimensionality, providing a wide range of immersive interactions. Due to the novel rendering and display technology of 360-degree images, they have more complex perceptual characteristics than other images. It is challenging to perform comprehensive image quality assessment (IQA) learning by simply stacking multichannel neural network architectures for pre/postprocessing, compression, and rendering tasks. To thoroughly learn the global and local features in 360-degree images, reduce the complexity of multichannel neural network models and simplify the training process, this paper proposes a joint architecture with user perception and an efficient transformer dedicated to 360-degree no-reference (NR) IQA. The input of the proposed method is a 360-degree cube map projection (CMP) image. Furthermore, the proposed 360-degree NRIQA method includes a saliency map-based non-overlapping self-attention selection module and a U-shaped transformer (U-former)-based feature extraction module to account for perceptual region importance and projection distortion. The transformer-based architecture and the weighted average technique are jointly utilized for predicting local perceptual quality. Experimental results obtained on widely used databases show that the proposed model outperforms other state-of-the-art methods in NR 360-degree image quality evaluation cases. Furthermore, a cross-database evaluation and an ablation study also demonstrate the inherent robustness and generalization ability of the proposed model.
引用
收藏
页码:396 / 405
页数:10
相关论文
共 50 条
  • [41] Dual-Feature Aggregation Network for No-Reference Image Quality Assessment
    Chen, Yihua
    Chen, Zhiyuan
    Yu, Mengzhu
    Tang, Zhenjun
    MULTIMEDIA MODELING, MMM 2023, PT I, 2023, 13833 : 149 - 161
  • [42] PW-360IQA: Perceptually-Weighted Multichannel CNN for Blind 360-Degree Image Quality Assessment
    Sendjasni, Abderrezzaq
    Larabi, Mohamed-Chaker
    SENSORS, 2023, 23 (09)
  • [43] Luminance decomposition and Transformer based no-reference tone-mapped image quality assessment
    Chen, Zikang
    He, Zhouyan
    Luo, Ting
    Jin, Chongchong
    Song, Yang
    DISPLAYS, 2024, 85
  • [44] Transformer-Based No-Reference Image Quality Assessment via Supervised Contrastive Learning
    Shi, Jinsong
    Gao, Pan
    Qin, Jie
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 5, 2024, : 4829 - 4837
  • [45] No-Reference Image Quality Assessment Combining Swin-Transformer and Natural Scene Statistics
    Yang, Yuxuan
    Lei, Zhichun
    Li, Changlu
    SENSORS, 2024, 24 (16)
  • [46] GUFORMER: a gradient-aware U-shaped transformer neural network for real image denoising
    Bai, Xuefei
    Wan, Yongsong
    Wang, Weiming
    Zhou, Bin
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (01):
  • [47] No-reference synthetic image quality assessment with convolutional neural network and local image saliency
    Xiaochuan Wang
    Xiaohui Liang
    Bailin Yang
    Frederick W. B. Li
    Computational Visual Media, 2019, 5 : 193 - 208
  • [48] No-reference synthetic image quality assessment with convolutional neural network and local image saliency
    Xiaochuan Wang
    Xiaohui Liang
    Bailin Yang
    Frederick W.B.Li
    Computational Visual Media, 2019, 5 (02) : 193 - 208
  • [49] No-reference synthetic image quality assessment with convolutional neural network and local image saliency
    Wang, Xiaochuan
    Liang, Xiaohui
    Yang, Bailin
    Li, Frederick W. B.
    COMPUTATIONAL VISUAL MEDIA, 2019, 5 (02) : 193 - 208
  • [50] HVS-Based Perception-Driven No-Reference Omnidirectional Image Quality Assessment
    Liu, Yun
    Yin, Xiaohua
    Wang, Yan
    Yin, Zixuan
    Zheng, Zhi
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72