Hybrid convolutional neural networks and optical flow for video visual attention prediction

被引:0
|
作者
Meijun Sun
Ziqi Zhou
Dong Zhang
Zheng Wang
机构
[1] Tianjin University,School of Computer Science and Technology
[2] Tianjin University of Traditional Chinese Medicine,School of Computer Software
[3] Tianjin University,undefined
来源
关键词
Convolutional neural networks; Optical flow; Spatial temporal feature; Visual attention;
D O I
暂无
中图分类号
学科分类号
摘要
In this paper, a convolutional neural networks (CNN) and optical flow based method is proposed for prediction of visual attention in the videos. First, a deep-learning framework is employed to extract spatial features in frames to replace those commonly used handcrafted features. The optical flow is calculated to obtain the temporal feature of the moving objects in video frames, which always draw audiences’ attentions. By integrating these two groups of features, a hybrid spatial temporal feature set is obtained and taken as the input of a support vector machine (SVM) to predict the degree of visual attention. Finally, two publicly available video datasets were used to test the performance of the proposed model, where the results have demonstrated the efficacy of the proposed approach.
引用
收藏
页码:29231 / 29244
页数:13
相关论文
共 50 条
  • [31] Video source traffic flow prediction using neural networks
    Bhattacharya, A
    Parlos, AG
    Atiya, AF
    PROCEEDINGS OF THE 46TH IEEE INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS & SYSTEMS, VOLS 1-3, 2003, : 33 - 38
  • [32] Neural Architecture Search for Convolutional Neural Networks with Attention
    Nakai, Kohei
    Matsubara, Takashi
    Uehara, Kuniaki
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (02) : 312 - 321
  • [33] VISUAL ATTENTION MODELING FOR 3D VIDEO USING NEURAL NETWORKS
    Iatsun, Iana
    Larabi, Mohamed-Chaker
    Fernandez-Maloigne, Christine
    2014 INTERNATIONAL CONFERENCE ON 3D IMAGING (IC3D), 2014,
  • [34] Optical Network Traffic Prediction Based on Graph Convolutional Neural Networks
    Gui, Yihan
    Wang, Danshi
    Guan, Luyao
    Zhang, Min
    2020 OPTO-ELECTRONICS AND COMMUNICATIONS CONFERENCE (OECC 2020), 2020,
  • [35] Guiding visual attention in deep convolutional neural networks based on human eye movements
    van Dyck, Leonard Elia
    Denzler, Sebastian Jochen
    Gruber, Walter Roland
    FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [36] Estimating Visual Focus of Attention in Multiparty Meetings using Deep Convolutional Neural Networks
    Otsuka, Kazuhiro
    Kasuga, Keisuke
    Koehler, Martina
    ICMI'18: PROCEEDINGS OF THE 20TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2018, : 191 - 199
  • [37] Generating self-attention activation maps for visual interpretations of convolutional neural networks
    Liang, Yu
    Li, Maozhen
    Jiang, Changjun
    NEUROCOMPUTING, 2022, 490 : 206 - 216
  • [38] DISCOVERING AFFECTIVE REGIONS IN DEEP CONVOLUTIONAL NEURAL NETWORKS FOR VISUAL SENTIMENT PREDICTION
    Sun, Ming
    Yang, Jufeng
    Wang, Kai
    Shen, Hui
    2016 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO (ICME), 2016,
  • [39] Video smoke detection base on dense optical flow and convolutional neural network
    Yuanlu Wu
    Minghao Chen
    Yan Wo
    Guoqiang Han
    Multimedia Tools and Applications, 2021, 80 : 35887 - 35901
  • [40] Video smoke detection base on dense optical flow and convolutional neural network
    Wu, Yuanlu
    Chen, Minghao
    Wo, Yan
    Han, Guoqiang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (28-29) : 35887 - 35901