A Markov Chain approach for video-based virtual try-on with denoising diffusion generative adversarial network

被引:1
|
作者
Hou, Jue [1 ,2 ]
Lu, Yinwen [1 ,2 ]
Wang, Mingjie [3 ]
Ouyang, Wenbing [4 ]
Yang, Yang [1 ,2 ]
Zou, Fengyuan [1 ,2 ]
Gu, Bingfei [1 ,2 ]
Liu, Zheng [2 ,5 ]
机构
[1] Zhejiang Sci Tech Univ, Sch Fash Design & Engn, CN-310018 Hangzhou, Zhejiang, Peoples R China
[2] Minist Culture & Tourism, Key Lab Silk Culture Heritage & Prod Design Digita, CN-310018 Hangzhou, Zhejiang, Peoples R China
[3] Zhejiang Sci Tech Univ, Sch Sci, Dept Math, CN-310018 Hangzhou, Zhejiang, Peoples R China
[4] Amazon Inc, 410 Terry Ave N, Seattle, WA 98109 USA
[5] Zhejiang Sci Tech Univ, Sch Int Educ, CN-310018 Hangzhou, Zhejiang, Peoples R China
关键词
Markov Chain; Diffusion model; Video synthesis; Virtual try -on;
D O I
10.1016/j.knosys.2024.112233
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video-based virtual try-ons have attracted unprecedented attention owing to the development of e-commerce. However, this problem is very challenging because of the arbitrary poses of persons and the demand for temporary consistency of frames, particularly when attempting to synthesize high-quality virtual try-on videos using single images. Specifically, there are two key challenges. 1) The existing video-based virtual try-on methods are based on generative adversarial networks (GAN), which are limited by unstable training and a lack of realism in generated details. 2) The explicit building of stronger constraints of generated frames, which aims to increase the coherence of generated videos. To address these challenges, this study proposed a novel framework, Extended Markov Chain Based Denoising Diffusion Generative Adversarial Network (EMC-DDGAN), which was derived from a denoising diffusion GAN, which is a diffusion model with efficient sampling. Moreover, we proposed an extended Markov chain that used a diffusion model to synthesize frames via sequential generation. With a carefully designed network and learning objects, the proposed approach achieved outstanding performance on public datasets. Rigorous experiments demonstrated that EMC-DDGAN could synthesize higher-quality videos compared to other state-of-the-art methods and validated the effectiveness of the proposed approach.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] An ECG Denoising Method Based on the Generative Adversarial Residual Network
    Xu, Bingxin
    Liu, Ruixia
    Shu, Minglei
    Shang, Xiaoyi
    Wang, Yinglong
    COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2021, 2021
  • [22] Toward Characteristic-Preserving Image-Based Virtual Try-On Network
    Wang, Bochao
    Zheng, Huabin
    Liang, Xiaodan
    Chen, Yimin
    Lin, Liang
    Yang, Meng
    COMPUTER VISION - ECCV 2018, PT XIII, 2018, 11217 : 607 - 623
  • [23] Instance Map Based Image Synthesis With a Denoising Generative Adversarial Network
    Zheng, Ziqiang
    Wang, Chao
    Yu, Zhibin
    Zheng, Haiyong
    Zheng, Bing
    IEEE ACCESS, 2018, 6 : 33654 - 33665
  • [24] Image Blind Denoising With Generative Adversarial Network Based Noise Modeling
    Chen, Jingwen
    Chen, Jiawei
    Chao, Hongyang
    Yang, Ming
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 3155 - 3164
  • [25] Micro-Doppler Spectrogram Denoising Based on Generative Adversarial Network
    Huang, Danyang
    Hou, Chunping
    Yang, Yang
    Lang, Yue
    Wang, Qing
    2018 48TH EUROPEAN MICROWAVE CONFERENCE (EUMC), 2018, : 909 - 912
  • [26] Conditional Denoising Diffusion Generative Adversarial Network for fast direct PET attenuation correction
    Cho, Min Jeong
    Lee, Jae Sung
    JOURNAL OF NUCLEAR MEDICINE, 2024, 65
  • [27] VTNCT: an image-based virtual try-on network by combining feature with pixel transformation
    Chang, Yuan
    Peng, Tao
    Yu, Feng
    He, Ruhan
    Hu, Xinrong
    Liu, Junping
    Zhang, Zili
    Jiang, Minghua
    VISUAL COMPUTER, 2023, 39 (07): : 2583 - 2596
  • [28] VTNCT: an image-based virtual try-on network by combining feature with pixel transformation
    Yuan Chang
    Tao Peng
    Feng Yu
    Ruhan He
    Xinrong Hu
    Junping Liu
    Zili Zhang
    Minghua Jiang
    The Visual Computer, 2023, 39 : 2583 - 2596
  • [29] Three-dimensional virtual try-on network based on attention mechanism and vision transformer
    Yuan T.
    Wang X.
    Luo W.
    Mei C.
    Wei J.
    Zhong Y.
    Fangzhi Xuebao/Journal of Textile Research, 2023, 44 (07): : 192 - 198
  • [30] KF-VTON: Keypoints-Driven Flow Based Virtual Try-On Network
    Wu, Zizhao
    Liu, Siyu
    Lu, Peioyan
    Yang, Ping
    Wong, Yongkang
    Gu, Xiaoling
    Kankanhalli, Mohan s.
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (09)