kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Analysis of and feedback on phonetic features in pronunciation training with a virtual teacher
KTH, School of Computer Science and Communication (CSC), Speech, Music and Hearing, TMH, Speech Communication and Technology.ORCID iD: 0000-0003-4532-014X
2012 (English)In: Computer Assisted Language Learning, ISSN 0958-8221, E-ISSN 1744-3210, Vol. 25, no 1, p. 37-64Article in journal (Refereed) Published
Abstract [en]

Pronunciation errors may be caused by several different deviations from the target, such as voicing, intonation, insertions or deletions of segments, or that the articulators are placed incorrectly. Computer-animated pronunciation teachers could potentially provide important assistance on correcting all these types of deviations, but they have an additional benefit for articulatory errors. By making parts of the face transparent, they can show the correct position and shape of the tongue and provide audiovisual feedback on how to change erroneous articulations. Such a scenario however requires firstly that the learner's current articulation can be estimated with precision and secondly that the learner is able to imitate the articulatory changes suggested in the audiovisual feedback. This article discusses both these aspects, with one experiment on estimating the important articulatory features from a speaker through acoustic-to-articulatory inversion and one user test with a virtual pronunciation teacher, in which the articulatory changes made by seven learners who receive audiovisual feedback are monitored using ultrasound imaging.

Place, publisher, year, edition, pages
Routledge, 2012. Vol. 25, no 1, p. 37-64
Keywords [en]
articulation analysis, acoustic-to-articulatory inversion, embodied conversational agents, articulation feedback
National Category
Computer Sciences Language Technology (Computational Linguistics)
Identifiers
URN: urn:nbn:se:kth:diva-52189DOI: 10.1080/09588221.2011.582845ISI: 000301642700003Scopus ID: 2-s2.0-84860710032OAI: oai:DiVA.org:kth-52189DiVA, id: diva2:465486
Note

QC 20120416

Available from: 2011-12-14 Created: 2011-12-14 Last updated: 2024-03-18Bibliographically approved

Open Access in DiVA

fulltext(746 kB)217 downloads
File information
File name FULLTEXT01.pdfFile size 746 kBChecksum SHA-512
149739aba63886801b14f28a79287a87f382f89d0a047799bb8699242de406575c4a9d3505400e357226992fae5dd890ea7a47153a1528c8ca6916ce9be1b6c9
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Engwall, Olov

Search in DiVA

By author/editor
Engwall, Olov
By organisation
Speech Communication and Technology
In the same journal
Computer Assisted Language Learning
Computer SciencesLanguage Technology (Computational Linguistics)

Search outside of DiVA

GoogleGoogle Scholar
Total: 217 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 272 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf