kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Is there a McGurk effect for tongue reading?
KTH, School of Computer Science and Communication (CSC), Speech, Music and Hearing, TMH, Speech Communication and Technology. KTH, School of Computer Science and Communication (CSC), Centres, Centre for Speech Technology, CTT.ORCID iD: 0000-0003-4532-014X
2010 (English)In: Proceedings of AVSP: International Conferenceon Audio-Visual Speech Processing, The International Society for Computers and Their Applications (ISCA) , 2010Conference paper, Published paper (Refereed)
Abstract [en]

Previous studies on tongue reading, i.e., speech perception ofdegraded audio supported by animations of tongue movementshave indicated that the support is weak initially and that subjectsneed training to learn to interpret the movements. Thispaper investigates if the learning is of the animation templatesas such or if subjects learn to retrieve articulatory knowledgethat they already have. Matching and conflicting animationsof tongue movements were presented randomly together withthe auditory speech signal at three different levels of noise in aconsonant identification test. The average recognition rate overthe three noise levels was significantly higher for the matchedaudiovisual condition than for the conflicting and the auditoryonly. Audiovisual integration effects were also found for conflictingstimuli. However, the visual modality is given much lessweight in the perception than for a normal face view, and intersubjectdifferences in the use of visual information are large.

Place, publisher, year, edition, pages
The International Society for Computers and Their Applications (ISCA) , 2010.
Keywords [en]
audiovisual speech perception, augmented reality, McGurk, Acoustic noise, Speech processing, Audio-visual speech, Consonant identification, Learn+, Matchings, McGurk effect, Noise levels, Speech perception, Speech signals
National Category
Computer Sciences Language Technology (Computational Linguistics)
Identifiers
URN: urn:nbn:se:kth:diva-52167Scopus ID: 2-s2.0-85133376544OAI: oai:DiVA.org:kth-52167DiVA, id: diva2:465462
Conference
Auditory-Visual Speech Processing (AVSP) 2010. Hakone, Kanagawa, Japan. September 30-October 3, 2010
Note

QC 20230404

Available from: 2011-12-14 Created: 2011-12-14 Last updated: 2024-03-18Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Scopushttp://www.isca-speech.org/archive/avsp10/papers/av10_S2-2.pdf

Authority records

Engwall, Olov

Search in DiVA

By author/editor
Engwall, Olov
By organisation
Speech Communication and TechnologyCentre for Speech Technology, CTT
Computer SciencesLanguage Technology (Computational Linguistics)

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 316 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf