Audio-Driven Laughter Behavior Controller

被引:6
|
作者
Ding, Yu [1 ]
Huang, Jing [2 ]
Pelachaud, Catherine [3 ]
机构
[1] Univ Houston, Dept Comp Sci, Houston, TX 77204 USA
[2] Zhejiang Gongshang Univ, Sch Informat & Elect Engn, Hangzhou 310018, Zhejiang, Peoples R China
[3] Univ Paris 06, CNRS, ISIR, F-75005 Paris, France
关键词
Laughter; audio-driven; data-driven; animation synthesis; continuous-state; Kalman filter; prosody; nonverbal behaviors; virtual character; statistical framework;
D O I
10.1109/TAFFC.2017.2754365
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It has been well documented that laughter is an important communicative and expressive signal in face-to-face conversations. Our work aims at building a laughter behavior controller for a virtual character which is able to generate upper body animations from laughter audio given as input. This controller relies on the tight correlations between laughter audio and body behaviors. A unified continuous-state statistical framework, inspired by Kalman filter, is proposed to learn the correlations between laughter audio and head/torso behavior from a recorded laughter human dataset. Due to the lack of shoulder behavior data in the recorded human dataset, a rule-based method is defined to model the correlation between laughter audio and shoulder behavior. In the synthesis step, these characterized correlations are rendered in the animation of a virtual character. To validate our controller, a subjective evaluation is conducted where participants viewed the videos of a laughing virtual character. It compares the animations of a virtual character using our controller and a state of the art method. The evaluation results show that the laughter animations computed with our controller are perceived as more natural, expressing amusement more freely and appearing more authentic than with the state of the art method.
引用
收藏
页码:546 / 558
页数:13
相关论文
共 50 条
  • [31] DiffTalk: Crafting Diffusion Models for Generalized Audio-Driven Portraits Animation
    Shen, Shuai
    Zhao, Wenliang
    Meng, Zibin
    Li, Wanhua
    Zhu, Zheng
    Zhou, Jie
    Lu, Jiwen
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 1982 - 1991
  • [32] Listen, Denoise, Action! Audio-Driven Motion Synthesis with Diffusion Models
    Alexanderson, Simon
    Nagy, Rajmund
    Beskow, Jonas
    Henter, Gustav Eje
    ACM TRANSACTIONS ON GRAPHICS, 2023, 42 (04):
  • [33] Efficient Emotional Adaptation for Audio-Driven Talking-Head Generation
    Gan, Yuan
    Yang, Zongxin
    Yue, Xihang
    Sun, Lingyun
    Yang, Yi
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 22577 - 22588
  • [34] FaceTalk: Audio-Driven Motion Diffusion for Neural Parametric Head Models
    Aneja, Shivangi
    Thies, Justus
    Dail, Angela
    Niessner, Matthias
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 21263 - 21273
  • [35] Audio-Driven Lips and Expression on 3D Human Face
    Ma, Le
    Ma, Zhihao
    Meng, Weiliang
    Xu, Shibiao
    Zhang, Xiaopeng
    ADVANCES IN COMPUTER GRAPHICS, CGI 2023, PT II, 2024, 14496 : 15 - 26
  • [36] Emotional Semantic Neural Radiance Fields for Audio-Driven Talking Head
    Lin, Haodong
    Wu, Zhonghao
    Zhang, Zhenyu
    Ma, Chao
    Yang, Xiaokang
    ARTIFICIAL INTELLIGENCE, CICAI 2022, PT II, 2022, 13605 : 532 - 544
  • [37] Let's Play Music: Audio-driven Performance Video Generation
    Zhu, Hao
    Li, Yi
    Zhu, Feixia
    Zheng, Aihua
    He, Ran
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 3574 - 3581
  • [38] Audio-driven talking face generation with diverse yet realistic facial animations
    Wu, Rongliang
    Yu, Yingchen
    Zhan, Fangneng
    Zhang, Jiahui
    Zhang, Xiaoqin
    Lu, Shijian
    PATTERN RECOGNITION, 2023, 144
  • [39] Imitating Arbitrary Talking Style for Realistic Audio-Driven Talking Face Synthesis
    Wu, Haozhe
    Jia, Jia
    Wang, Haoyu
    Dou, Yishun
    Duan, Chao
    Deng, Qingshan
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 1478 - 1486
  • [40] MODA: Mapping-Once Audio-driven Portrait Animation with Dual Attentions
    Liu, Yunfei
    Lin, Lijian
    Yu, Fei
    Zhou, Changyin
    Li, Yu
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 22963 - 22972