A semiautomatic saliency model and its application to video compression

被引:0
|
作者
Lyudvichenko, Vitaliy
Erofeev, Mikhail
Gitman, Yury
Vatolin, Dmitriy
机构
关键词
Eye-Tracking; Saliency; Video Compression; Visual Attention; x264; IMAGE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work aims to apply visual-attention modeling to attention-based video compression. During our comparison we found that eye-tracking data collected even from a single observer outperforms existing automatic models by a significant margin. Therefore, we offer a semiautomatic approach: using computer-vision algorithms and good initial estimation of eye-tracking data from just one observer to produce high-quality saliency maps that are similar to multi-observer eye tracking and that are appropriate for practical applications. We propose a simple algorithm that is based on temporal coherence of the visual-attention distribution and requires eye tracking of just one observer. The results are as good as an average gaze map for two observers. While preparing the saliency-model comparison, we paid special attention to the quality-measurement procedure. We observe that many modern visual-attention models can be improved by applying simple transforms such as brightness adjustment and blending with the center-prior model. The novel quality-evaluation procedure that we propose is invariant to such transforms. To show the practical use of our semiautomatic approach, we developed a saliency-aware modification of the x264 video encoder and performed subjective and objective evaluations. The modified encoder can serve with any attention model and is publicly available.
引用
收藏
页码:403 / 410
页数:8
相关论文
共 50 条
  • [41] Audiovisual focus of attention and its application to Ultra High Definition Video compression
    Rerabek, Martin
    Nemoto, Hiromi
    Lee, Jong-Seok
    Ebrahimi, Touradj
    HUMAN VISION AND ELECTRONIC IMAGING XIX, 2014, 9014
  • [42] An embedded saliency map estimator scheme: Application to video encoding
    Tsapatsoulis, Nicolas
    Rapantzikos, Konstantinos
    Pattichis, Constantinos
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2007, 17 (04) : 289 - 304
  • [43] The application of wavelet theory in video compression
    Wu, B
    Zhao, SH
    Li, SF
    IEEE 2005 International Symposium on Microwave, Antenna, Propagation and EMC Technologies for Wireless Communications Proceedings, Vols 1 and 2, 2005, : 1234 - 1236
  • [44] Research and application of compression of video image
    Electronics Department, Shengyang Institute of Aeronautical Engineering, Shenyang 110136, China
    Shu Ju Cai Ji Yu Chu Li, 2006, SUPPL. (75-77):
  • [45] Camera-Assisted Video Saliency Prediction and Its Applications
    Sun, Xiao
    Hu, Yuxing
    Zhang, Luming
    Chen, Yanxiang
    Li, Ping
    Xie, Zhao
    Liu, Zhenguang
    IEEE TRANSACTIONS ON CYBERNETICS, 2018, 48 (09) : 2520 - 2530
  • [46] Deep Saliency Features for Video Saliency Prediction
    Azaza, Aymen
    Douik, Ali
    2018 INTERNATIONAL CONFERENCE ON ADVANCED SYSTEMS AND ELECTRICAL TECHNOLOGIES (IC_ASET), 2017, : 335 - 339
  • [47] Drosophila-Vision-Inspired Motion Perception Model and Its Application in Saliency Detection
    Chen, Zhe
    Mu, Qi
    Han, Guangjie
    Lu, Huimin
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (01) : 819 - 830
  • [48] Exploring Gradient Flow Based Saliency for DNN Model Compression
    Liu, Xinyu
    Li, Baopu
    Chen, Zhen
    Yuan, Yixuan
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 3238 - 3246
  • [49] Model Compression Based on Knowledge Distillation and Its Application in HRRP
    Chen, Xiaojiao
    An, Zhenyu
    Huang, Liansheng
    He, Shiying
    Wang, Zhen
    PROCEEDINGS OF 2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2020), 2020, : 1268 - 1272
  • [50] Application of a new video image compression method to video phone
    Wang Qiu-hua
    Zhang Jian-wu
    Proceedings of 2005 Chinese Control and Decision Conference, Vols 1 and 2, 2005, : 1555 - 1558