Frequency-Based Motion Representation for Video Generative Adversarial Networks

被引:0
|
作者
Hyun, Sangeek [1 ]
Lew, Jaihyun [2 ]
Chung, Jiwoo [1 ]
Kim, Euiyeon [1 ]
Heo, Jae-Pil [3 ]
机构
[1] Sungkyunkwan Univ, Dept Artificial Intelligence, Suwon 16419, South Korea
[2] Seoul Natl Univ, Interdisciplinary Program Artificial Intelligence, Seoul 08826, South Korea
[3] Sungkyunkwan Univ, Dept Comp Sci & Engn, Suwon 16419, South Korea
关键词
Generative adversarial networks; video generation; sinusoidal motion representation; speed-level motion manipulation;
D O I
10.1109/TIP.2023.3293767
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Videos contain motions of various speeds. For example, the motions of one's head and mouth differ in terms of speed - the head being relatively stable and the mouth moving rapidly as one speaks. Despite its diverse nature, previous video GANs generate video based on a single unified motion representation without considering the aspect of speed. In this paper, we propose a frequency-based motion representation for video GANs to realize the concept of speed in video generation process. In detail, we represent motions as continuous sinusoidal signals of various frequencies by introducing a coordinate-based motion generator. We show, in that case, frequency is highly related to the speed of motion. Based on this observation, we present frequency-aware weight modulation that enables manipulation of motions within a specific range of speed, which could not be achieved with the previous techniques. Extensive experiments validate that the proposed method outperforms state-of-the-art video GANs in terms of generation quality by its capability to model various speed of motions. Furthermore, we also show that our temporally continuous representation enables to further synthesize intermediate and future frames of generated videos.
引用
收藏
页码:3949 / 3963
页数:15
相关论文
共 50 条
  • [1] Generative Adversarial Networks for Multimodal Representation Learning in Video Hyperlinking
    Vukotic, Vedran
    Raymond, Christian
    Gravier, Guillaume
    PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR'17), 2017, : 421 - 424
  • [2] Self-supervised time-frequency representation based on generative adversarial networks
    Liu, Naihao
    Lei, Youbo
    Yang, Yang
    Wei, Shengtao
    Gao, Jinghuai
    Jiang, Xiudi
    GEOPHYSICS, 2023, 88 (04) : IM87 - IM99
  • [3] Video Generative Adversarial Networks: A Review
    Aldausari, Nuha
    Sowmya, Arcot
    Marcus, Nadine
    Mohammadi, Gelareh
    ACM COMPUTING SURVEYS, 2023, 55 (02)
  • [4] On Evaluating Video-based Generative Adversarial Networks (GANs)
    Ronquillo, Nancy
    Harguess, Josh
    2018 IEEE APPLIED IMAGERY PATTERN RECOGNITION WORKSHOP (AIPR), 2018,
  • [5] Video anomaly detection based on ensemble generative adversarial networks
    Gu Jia-Cheng
    Long Ying-Wen
    Ji Ming-Ming
    CHINESE JOURNAL OF LIQUID CRYSTALS AND DISPLAYS, 2022, 37 (12) : 1607 - 1613
  • [6] High-resolution time-frequency representation with generative adversarial networks
    Deprem, Zeynel
    Cetin, A. Enis
    SIGNAL IMAGE AND VIDEO PROCESSING, 2023, 17 (03) : 849 - 854
  • [7] High-resolution time-frequency representation with generative adversarial networks
    Zeynel Deprem
    A. Enis Çetin
    Signal, Image and Video Processing, 2023, 17 : 849 - 854
  • [8] Orthogonal Subspace Representation for Generative Adversarial Networks
    Jiang, Hongxiang
    Luo, Xiaoyan
    Yin, Jihao
    Fu, Huazhu
    Wang, Fuxiang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 15
  • [9] Orthogonal Subspace Representation for Generative Adversarial Networks
    Jiang, Hongxiang
    Luo, Xiaoyan
    Yin, Jihao
    Fu, Huazhu
    Wang, Fuxiang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (03) : 4413 - 4427
  • [10] Application of a Deep Generative Model for Diversified Video Subtitles Based on Generative Adversarial Networks
    Shen, Lingzhi
    2023 3RD ASIA-PACIFIC CONFERENCE ON COMMUNICATIONS TECHNOLOGY AND COMPUTER SCIENCE, ACCTCS, 2023, : 176 - 181