Im2Flow: Motion Hallucination from Static Images for Action Recognition

被引:50
|
作者
Gao, Ruohan [1 ]
Xiong, Bo [1 ]
Grauman, Kristen [1 ]
机构
[1] UT Austin, Austin, TX 78712 USA
关键词
D O I
10.1109/CVPR.2018.00622
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing methods to recognize actions in static images take the images at their face value, learning the appearances-objects, scenes, and body poses-that distinguish each action class. However, such models are deprived of the rich dynamic structure and motions that also define human activity. We propose an approach that hallucinates the unobserved future motion implied by a single snapshot to help static-image action recognition. The key idea is to learn a prior over short-term dynamics from thousands of unlabeled videos, infer the anticipated optical flow on novel static images, and then train discriminative models that exploit both streams of information. Our main contributions are twofold. First, we devise an encoder-decoder convolutional neural network and a novel optical flow encoding that can translate a static image into an accurate flow map. Second, we show the power of hallucinated flow for recognition, successfully transferring the learned motion into a standard two-stream network for activity recognition. On seven datasets, we demonstrate the power of the approach. It not only achieves state-of-the-art accuracy for dense optical flow prediction, but also consistently enhances recognition of actions and dynamic scenes.
引用
收藏
页码:5937 / 5947
页数:11
相关论文
共 50 条
  • [1] Exploiting Motion Information from Unlabeled Videos for Static Image Action Recognition
    Zhang, Yiyi
    Li Niu
    Pan, Ziqi
    Luo, Meichao
    Zhang, Jianfu
    Cheng, Dawei
    Zhang, Liqing
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12918 - 12925
  • [2] Facial action recognition for facial expression analysis from static face images
    Pantic, M
    Rothkrantz, LJM
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2004, 34 (03): : 1449 - 1461
  • [3] Recognition of human action in motion detected images with GMACA
    Peldek, Serkan
    Becerikli, Yasar
    Journal of the Faculty of Engineering and Architecture of Gazi University, 2019, 34 (02): : 1025 - 1043
  • [4] Recognition of human action in motion detected images with GMACA
    Peldek, Serkan
    Becerikli, Yasar
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2019, 34 (02): : 1026 - 1043
  • [5] Hallucinating uncertain motion and future for static image action recognition
    Niu, Li
    Huang, Shengyuan
    Zhao, Xing
    Kang, Liwei
    Zhang, Yiyi
    Zhang, Liqing
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2022, 215
  • [6] 3D human action analysis and recognition through GLAC descriptor on 2D motion and static posture images
    Bulbul, Mohammad Farhad
    Islam, Saiful
    Ali, Hazrat
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (15) : 21085 - 21111
  • [7] 3D human action analysis and recognition through GLAC descriptor on 2D motion and static posture images
    Mohammad Farhad Bulbul
    Saiful Islam
    Hazrat Ali
    Multimedia Tools and Applications, 2019, 78 : 21085 - 21111
  • [8] FACIAL EXPRESSION RECOGNITION FROM STATIC IMAGES
    Akkoca, Bilge Suheyla
    Gokmen, Muhittin
    2014 22ND SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2014, : 1291 - 1294
  • [9] Learning Motion Features from Dynamic Images of Depth Video for Human Action Recognition
    Huang, Yao
    Yang, Jianyu
    Shao, Zhanpeng
    Li, Youfu
    2021 27TH INTERNATIONAL CONFERENCE ON MECHATRONICS AND MACHINE VISION IN PRACTICE (M2VIP), 2021,
  • [10] Action Recognition based on Multiple Key Motion History Images
    Li, Dongxue
    Yu, Lejun
    He, Jun
    Sun, Bo
    Ge, Fengxiang
    PROCEEDINGS OF 2016 IEEE 13TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP 2016), 2016, : 993 - 996