kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Emotional Speech-Driven Animation with Content-Emotion Disentanglement
Max Planck Institute for Intelligent Systems, Germany.
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Computational Science and Technology (CST).ORCID iD: 0000-0002-7414-845X
Max Planck Institute for Intelligent Systems, Germany.
Max Planck Institute for Intelligent Systems, Germany.
Show others and affiliations
2023 (English)In: Proceedings - SIGGRAPH Asia 2023 Conference Papers, SA 2023, Association for Computing Machinery (ACM) , 2023, article id 41Conference paper, Published paper (Refereed)
Abstract [en]

To be widely adopted, 3D facial avatars must be animated easily, realistically, and directly from speech signals. While the best recent methods generate 3D animations that are synchronized with the input audio, they largely ignore the impact of emotions on facial expressions. Realistic facial animation requires lip-sync together with the natural expression of emotion. To that end, we propose EMOTE (Expressive Model Optimized for Talking with Emotion), which generates 3D talking-head avatars that maintain lip-sync from speech while enabling explicit control over the expression of emotion. To achieve this, we supervise EMOTE with decoupled losses for speech (i.e., lip-sync) and emotion. These losses are based on two key observations: (1) deformations of the face due to speech are spatially localized around the mouth and have high temporal frequency, whereas (2) facial expressions may deform the whole face and occur over longer intervals. Thus we train EMOTE with a per-frame lip-reading loss to preserve the speech-dependent content, while supervising emotion at the sequence level. Furthermore, we employ a content-emotion exchange mechanism in order to supervise different emotions on the same audio, while maintaining the lip motion synchronized with the speech. To employ deep perceptual losses without getting undesirable artifacts, we devise a motion prior in the form of a temporal VAE. Due to the absence of high-quality aligned emotional 3D face datasets with speech, EMOTE is trained with 3D pseudo-ground-truth extracted from an emotional video dataset (i.e., MEAD). Extensive qualitative and perceptual evaluations demonstrate that EMOTE produces speech-driven facial animations with better lip-sync than state-of-the-art methods trained on the same data, while offering additional, high-quality emotional control.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM) , 2023. article id 41
Keywords [en]
Computer Graphics, Computer Vision, Deep learning, Facial Animation, Speech-driven Animation
National Category
Computer Vision and Robotics (Autonomous Systems)
Identifiers
URN: urn:nbn:se:kth:diva-347500DOI: 10.1145/3610548.3618183ISI: 001278296700041Scopus ID: 2-s2.0-85180390692OAI: oai:DiVA.org:kth-347500DiVA, id: diva2:1873752
Conference
2023 SIGGRAPH Asia 2023 Conference Papers, SA 2023, Sydney, Australia, Dec 12 2023 - Dec 15 2023
Note

Part of ISBN 9798400703157

QC 20240619

Available from: 2024-06-19 Created: 2024-06-19 Last updated: 2024-09-12Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Chhatre, Kiran

Search in DiVA

By author/editor
Chhatre, Kiran
By organisation
Computational Science and Technology (CST)
Computer Vision and Robotics (Autonomous Systems)

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 50 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf