Data-driven multimodal synthesis
2005 (English)In: Speech Communication, ISSN 0167-6393, Vol. 47, no 02-jan, 182-193 p.Article in journal (Refereed) Published
This paper is a report on current efforts at the Department of Speech, Music and Hearing, KTH, on data-driven multimodal synthesis including both visual speech synthesis and acoustic modeling. In the research we try to combine both corpus based methods with knowledge based models and to explore the best of the two approaches. In the paper an attempt to build formant-synthesis systems based on both rule-generated and database driven methods is presented. A pilot experiment is also reported showing that this approach can be a very interesting path to explore further. Two studies on visual speech synthesis are reported, one on data acquisition using a combination of motion capture techniques and one concerned with coarticulation, comparing different models.
Place, publisher, year, edition, pages
2005. Vol. 47, no 02-jan, 182-193 p.
speech synthesis, multimodal synthesis, data-driven synthesis, speech synthesis
IdentifiersURN: urn:nbn:se:kth:diva-15032DOI: 10.1016/j.specom.2005.02.015ISI: 000231788500016ScopusID: 2-s2.0-24144469759OAI: oai:DiVA.org:kth-15032DiVA: diva2:333073
QC 201005252010-08-052010-08-05Bibliographically approved