Rigid head motion in expressive speech animation: Analysis and synthesis

被引:104
|
作者
Busso, Carlos [1 ]
Deng, Zhigang [1 ]
Grimm, Michael [1 ]
Neumann, Ulrich [1 ]
Narayanan, Shrikanth [1 ]
机构
[1] Univ So Calif, Viterbi Sch Engn, Integrated Media Syst Ctr, Los Angeles, CA 90089 USA
基金
美国国家科学基金会;
关键词
D O I
10.1109/TASL.2006.885910
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Rigid head motion is a gesture that conveys important nonverbal information in human communication, and hence it needs to be appropriately modeled and included in realistic facial animations to effectively mimic human behaviors. In this paper, head motion sequences in expressive facial animations are analyzed in terms of their naturalness and emotional salience in perception. Statistical measures are derived from an audiovisual database, comprising synchronized facial gestures and speech, which revealed characteristic patterns in emotional head motion sequences. Head motion patterns with neutral speech significantly differ from head motion patterns with emotional speech in motion activation, range, and velocity. The results show that head motion provides discriminating information about emotional categories. An approach to synthesize emotional head motion sequences driven by prosodic features is presented, expanding upon our previous framework on head motion synthesis. This method naturally models the specific temporal dynamics of emotional head motion sequences by building hidden Markov models for each emotional category (sadness, happiness, anger, and neutral state). Human raters were asked to assess the naturalness and the emotional content of the facial animations. On average, the synthesized head motion sequences were perceived even more natural than the original head motion sequences. The results also show that head motion modifies the emotional perception of the facial animation especially in the valence and activation domain. These results suggest that appropriate. head motion not only significantly improves the naturalness of the animation but can also be used to enhance the emotional content of the animation to effectively engage the users.
引用
收藏
页码:1075 / 1086
页数:12
相关论文
共 50 条
  • [21] Rigid-motion-inspired liquid character animation
    Zhang, Guijuan
    Lu, Dianjie
    Zhu, Dengming
    Lv, Lei
    Liu, Hong
    Meng, Xiangxu
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2013, 24 (3-4) : 205 - 213
  • [22] X-Portrait: Expressive Portrait Animation with Hierarchical Motion Attention
    Xie, You
    Xu, Hongyi
    Song, Guoxian
    Wang, Chao
    Shi, Yichun
    Luo, Linjie
    PROCEEDINGS OF SIGGRAPH 2024 CONFERENCE PAPERS, 2024,
  • [23] Articulatory features for speech-driven head motion synthesis
    Ben-Youssef, Atef
    Shimodaira, Hiroshi
    Braude, David A.
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2757 - 2761
  • [24] BLSTM Neural Networks for Speech Driven Head Motion Synthesis
    Ding, Chuang
    Zhu, Pengcheng
    Xie, Lei
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 3345 - 3349
  • [25] Advancements in Expressive Speech Synthesis: a Review
    Alwaisi, Shaimaa
    Nemeth, Geza
    INFOCOMMUNICATIONS JOURNAL, 2024, 16 (01): : 35 - 46
  • [26] EXPRESSIVE SPEECH SYNTHESIS FOR CRITICAL SITUATIONS
    Rusko, Milan
    Darjaa, Sakhia
    Trnka, Marian
    Sabo, Robert
    Ritomsk, Marian
    COMPUTING AND INFORMATICS, 2014, 33 (06) : 1312 - 1332
  • [27] ARTICULATORY FEATURES FOR EXPRESSIVE SPEECH SYNTHESIS
    Black, Alan W.
    Bunnell, H. Timothy
    Dou, Ying
    Muthukumar, Prasanna Kumar
    Metze, Florian
    Perry, Daniel
    Polzehl, Tim
    Prahallad, Kishore
    Steidl, Stefan
    Vaughn, Callie
    2012 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2012, : 4005 - 4008
  • [28] Expressive speech: Production, perception and application to speech synthesis
    Erickson, Donna
    ACOUSTICAL SCIENCE AND TECHNOLOGY, 2005, 26 (04) : 317 - 325
  • [29] Real-time speech-driven animation of expressive talking faces
    Liu, Jia
    You, Mingyu
    Chen, Chun
    Song, Mingli
    INTERNATIONAL JOURNAL OF GENERAL SYSTEMS, 2011, 40 (04) : 439 - 455
  • [30] Related Work in Speech Animation Synthesis
    不详
    IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2015, 35 (02) : 72 - 72