Endre søk
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Transfer learning of articulatory information through phone information
KTH, Skolan för elektroteknik och datavetenskap (EECS), Intelligenta system, Tal, musik och hörsel, TMH.ORCID-id: 0000-0002-3323-5311
Vise andre og tillknytning
2020 (engelsk)Inngår i: Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, International Speech Communication Association , 2020, s. 2877-2881Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

Articulatory information has been argued to be useful for several speech tasks. However, in most practical scenarios this information is not readily available. We propose a novel transfer learning framework to obtain reliable articulatory information in such cases. We demonstrate its reliability both in terms of estimating parameters of speech production and its ability to enhance the accuracy of an end-to-end phone recognizer. Articulatory information is estimated from speaker independent phonemic features, using a small speech corpus, with electromagnetic articulography (EMA) measurements. Next, we employ a teacher-student model to learn estimation of articulatory features from acoustic features for the targeted phone recognition task. Phone recognition experiments, demonstrate that the proposed transfer learning approach outperforms the baseline transfer learning system acquired directly from an acoustic-to-articulatory (AAI) model. The articulatory features estimated by the proposed method, in conjunction with acoustic features, improved the phone error rate (PER) by 6.7% and 6% on the TIMIT core test and development sets, respectively, compared to standalone static acoustic features. Interestingly, this improvement is slightly higher than what is obtained by static+dynamic acoustic features, but with a significantly less. Adding articulatory features on top of static+dynamic acoustic features yields a small but positive PER improvement.

sted, utgiver, år, opplag, sider
International Speech Communication Association , 2020. s. 2877-2881
Emneord [en]
Articulatory inversion, Deep learning, Speech recognition, Transfer learning, Learning systems, Speech communication, Telephone sets, Acoustic features, Articulatory features, Articulatory informations, Electromagnetic articulography, Estimating parameters, Learning frameworks, Speaker independents, Speech production
HSV kategori
Identifikatorer
URN: urn:nbn:se:kth:diva-302931DOI: 10.21437/Interspeech.2020-1139ISI: 000833594103003Scopus ID: 2-s2.0-85098223486OAI: oai:DiVA.org:kth-302931DiVA, id: diva2:1599907
Konferanse
21st Annual Conference of the International Speech Communication Association, INTERSPEECH 2020, 25 October 2020 through 29 October 2020
Merknad

QC 20211003

Tilgjengelig fra: 2021-10-03 Laget: 2021-10-03 Sist oppdatert: 2025-02-07bibliografisk kontrollert

Open Access i DiVA

Fulltekst mangler i DiVA

Andre lenker

Forlagets fulltekstScopus

Person

Salvi, Giampiero

Søk i DiVA

Av forfatter/redaktør
Salvi, Giampiero
Av organisasjonen

Søk utenfor DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric

doi
urn-nbn
Totalt: 41 treff
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf