Transflower: probabilistic autoregressive dance generation with multimodal attentionShow others and affiliations
2021 (English)In: ACM Transactions on Graphics, ISSN 0730-0301, E-ISSN 1557-7368, Vol. 40, no 6, article id 195Article in journal (Refereed) Published
Abstract [en]
Dance requires skillful composition of complex movements that follow rhythmic, tonal and timbral features of music. Formally, generating dance conditioned on a piece of music can be expressed as a problem of modelling a high-dimensional continuous motion signal, conditioned on an audio signal. In this work we make two contributions to tackle this problem. First, we present a novel probabilistic autoregressive architecture that models the distribution over future poses with a normalizing flow conditioned on previous poses as well as music context, using a multimodal transformer encoder. Second, we introduce the currently largest 3D dance-motion dataset, obtained with a variety of motion-capture technologies, and including both professional and casual dancers. Using this dataset, we compare our new model against two baselines, via objective metrics and a user study, and show that both the ability to model a probability distribution, as well as being able to attend over a large motion and music context are necessary to produce interesting, diverse, and realistic dance that matches the music.
Place, publisher, year, edition, pages
Association for Computing Machinery (ACM) , 2021. Vol. 40, no 6, article id 195
Keywords [en]
Generative models, machine learning, normalising flows, Glow, transformers, dance
National Category
Computer graphics and computer vision Computer Sciences Signal Processing
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:kth:diva-307028DOI: 10.1145/3478513.3480570ISI: 000729846700001Scopus ID: 2-s2.0-85125127739OAI: oai:DiVA.org:kth-307028DiVA, id: diva2:1626445
Funder
Swedish Research Council, 2018-05409Swedish Research Council, 2019-03694Knut and Alice Wallenberg Foundation, WASPMarianne and Marcus Wallenberg Foundation, 2020.0102
Note
QC 20220520
2022-01-112022-01-112025-02-01Bibliographically approved