Frequency-Based Motion Representation for Video Generative Adversarial Networks

被引:0
|
作者
Hyun, Sangeek [1 ]
Lew, Jaihyun [2 ]
Chung, Jiwoo [1 ]
Kim, Euiyeon [1 ]
Heo, Jae-Pil [3 ]
机构
[1] Sungkyunkwan Univ, Dept Artificial Intelligence, Suwon 16419, South Korea
[2] Seoul Natl Univ, Interdisciplinary Program Artificial Intelligence, Seoul 08826, South Korea
[3] Sungkyunkwan Univ, Dept Comp Sci & Engn, Suwon 16419, South Korea
关键词
Generative adversarial networks; video generation; sinusoidal motion representation; speed-level motion manipulation;
D O I
10.1109/TIP.2023.3293767
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Videos contain motions of various speeds. For example, the motions of one's head and mouth differ in terms of speed - the head being relatively stable and the mouth moving rapidly as one speaks. Despite its diverse nature, previous video GANs generate video based on a single unified motion representation without considering the aspect of speed. In this paper, we propose a frequency-based motion representation for video GANs to realize the concept of speed in video generation process. In detail, we represent motions as continuous sinusoidal signals of various frequencies by introducing a coordinate-based motion generator. We show, in that case, frequency is highly related to the speed of motion. Based on this observation, we present frequency-aware weight modulation that enables manipulation of motions within a specific range of speed, which could not be achieved with the previous techniques. Extensive experiments validate that the proposed method outperforms state-of-the-art video GANs in terms of generation quality by its capability to model various speed of motions. Furthermore, we also show that our temporally continuous representation enables to further synthesize intermediate and future frames of generated videos.
引用
收藏
页码:3949 / 3963
页数:15
相关论文
共 50 条
  • [41] Frequency-based Skill Analysis for Motion Pictures
    Maeda, Toshiyuki
    Yajima, Masumi
    Wakatani, Akiyoshi
    PROCEEDINGS 2018 12TH FRANCE-JAPAN AND 10TH EUROPE-ASIA CONGRESS ON MECHATRONICS, 2018, : 294 - 297
  • [42] Aftershock ground motion prediction model based on conditional convolutional generative adversarial networks
    Shen, Jiaxu
    Ni, Bo
    Ding, Yinjun
    Xiong, Jiecheng
    Zhong, Zilan
    Chen, Jun
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [43] Motion Deblurring Via Using Generative Adversarial Networks For Space-Based Imaging
    Chen, Yi
    Wu, Fengge
    Zhao, Junsuo
    2018 IEEE/ACIS 16TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING RESEARCH, MANAGEMENT AND APPLICATION (SERA), 2018, : 37 - 41
  • [44] Frequency-based methods for improving the imperceptibility and transferability of adversarial examples
    Zhu, Hegui
    Ren, Yuchen
    Liu, Chong
    Sui, Xiaoyan
    Zhang, Libo
    APPLIED SOFT COMPUTING, 2024, 150
  • [45] Pose transfer based on generative adversarial networks
    Pan, Hao
    Cao, Xincong
    2020 INTERNATIONAL CONFERENCE ON IMAGE, VIDEO PROCESSING AND ARTIFICIAL INTELLIGENCE, 2020, 11584
  • [46] Icon Generation Based on Generative Adversarial Networks
    Yang, Hongyi
    Xue, Chengqi
    Yang, Xiaoying
    Yang, Han
    APPLIED SCIENCES-BASEL, 2021, 11 (17):
  • [47] Image Inpainting Based on Generative Adversarial Networks
    Jiang, Yi
    Xu, Jiajie
    Yang, Baoqing
    Xu, Jing
    Zhu, Junwu
    IEEE ACCESS, 2020, 8 (08): : 22884 - 22892
  • [48] Data Synthesis based on Generative Adversarial Networks
    Park, Noseong
    Mohammadi, Mahmoud
    Gorde, Kshitij
    Jajodia, Sushil
    Park, Hongkyu
    Kim, Youngmin
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2018, 11 (10): : 1071 - 1083
  • [49] Generative Adversarial Networks Based on Cooperative Games
    Luo, Lie
    Cai, Jiewei
    Fan, Zouyang
    Chen, Yumin
    Jiang, Hongbo
    Journal of Network Intelligence, 2024, 9 (01): : 88 - 107
  • [50] Generative Adversarial Networks
    Goodfellow, Ian
    Pouget-Abadie, Jean
    Mirza, Mehdi
    Xu, Bing
    Warde-Farley, David
    Ozair, Sherjil
    Courville, Aaron
    Bengio, Yoshua
    COMMUNICATIONS OF THE ACM, 2020, 63 (11) : 139 - 144