Two-Stream Attention Network for Pain Recognition from Video Sequences

被引:23
|
作者
Thiam, Patrick [1 ,2 ]
Kestler, Hans A. [1 ]
Schwenker, Friedhelm [2 ]
机构
[1] Ulm Univ, Inst Med Syst Biol, Albert Einstein Allee 11, D-89081 Ulm, Germany
[2] Ulm Univ, Inst Neural Informat Proc, James Frank Ring, D-89081 Ulm, Germany
关键词
convolutional neural networks; long short-term memory recurrent neural networks; information fusion; pain recognition;
D O I
10.3390/s20030839
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Several approaches have been proposed for the analysis of pain-related facial expressions. These approaches range from common classification architectures based on a set of carefully designed handcrafted features, to deep neural networks characterised by an autonomous extraction of relevant facial descriptors and simultaneous optimisation of a classification architecture. In the current work, an end-to-end approach based on attention networks for the analysis and recognition of pain-related facial expressions is proposed. The method combines both spatial and temporal aspects of facial expressions through a weighted aggregation of attention-based neural networks' outputs, based on sequences of Motion History Images (MHIs) and Optical Flow Images (OFIs). Each input stream is fed into a specific attention network consisting of a Convolutional Neural Network (CNN) coupled to a Bidirectional Long Short-Term Memory (BiLSTM) Recurrent Neural Network (RNN). An attention mechanism generates a single weighted representation of each input stream (MHI sequence and OFI sequence), which is subsequently used to perform specific classification tasks. Simultaneously, a weighted aggregation of the classification scores specific to each input stream is performed to generate a final classification output. The assessment conducted on both the BioVid Heat Pain Database (Part A) and SenseEmotion Database points at the relevance of the proposed approach, as its classification performance is on par with state-of-the-art classification approaches proposed in the literature.
引用
收藏
页数:19
相关论文
共 50 条
  • [21] Gaze Estimation by Attention Using a Two-Stream Regression Network
    Karazor, Ahmet
    Bayar, Alperen Enes
    Topal, Cihan
    Cevikalp, Hakan
    2023 31ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU, 2023,
  • [22] Unsupervised video-based action recognition using two-stream generative adversarial network
    Lin, Wei
    Zeng, Huanqiang
    Zhu, Jianqing
    Hsia, Chih-Hsien
    Hou, Junhui
    Ma, Kai-Kuang
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (09): : 5077 - 5091
  • [24] Unsupervised video-based action recognition using two-stream generative adversarial network
    Wei Lin
    Huanqiang Zeng
    Jianqing Zhu
    Chih-Hsien Hsia
    Junhui Hou
    Kai-Kuang Ma
    Neural Computing and Applications, 2024, 36 : 5077 - 5091
  • [25] Compositional attention networks with two-stream fusion for video question answering
    Yu, Ting
    Yu, Jun
    Yu, Zhou
    Tao, Dacheng
    IEEE Transactions on Image Processing, 2020, 29 : 1204 - 1218
  • [26] Compositional Attention Networks With Two-Stream Fusion for Video Question Answering
    Yu, Ting
    Yu, Jun
    Yu, Zhou
    Tao, Dacheng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 1204 - 1218
  • [27] Spatiotemporal two-stream LSTM network for unsupervised video summarization
    Min Hu
    Ruimin Hu
    Zhongyuan Wang
    Zixiang Xiong
    Rui Zhong
    Multimedia Tools and Applications, 2022, 81 : 40489 - 40510
  • [28] Pornographic Video Detection with Convolutional Two-Stream Network Fusion
    Lee, Wonjae
    Kim, Junghak
    Lee, Nam Kyung
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1273 - 1275
  • [29] Spatiotemporal two-stream LSTM network for unsupervised video summarization
    Hu, Min
    Hu, Ruimin
    Wang, Zhongyuan
    Xiong, Zixiang
    Zhong, Rui
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (28) : 40489 - 40510
  • [30] Efficient Two-Stream Network for Online Video Action Segmentation
    Kang, Min-Seok
    Park, Rae-Hong
    Park, Hyung-Min
    IEEE ACCESS, 2022, 10 : 90635 - 90646