kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Cross-Domain Transfer of Generative Explanations Using Text-to-Text Models
KTH, School of Electrical Engineering and Computer Science (EECS), Electrical Engineering. Peltarion, Stockholm, Sweden.ORCID iD: 0000-0003-2436-8566
Peltarion, Stockholm, Sweden.
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0000-0002-4722-0823
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0000-0002-2748-8929
2021 (English)In: Lecture Notes in Computer Science, Springer Nature , 2021, p. 76-89Conference paper, Published paper (Refereed)
Abstract [en]

Deep learning models based on the Transformers architecture have achieved impressive state-of-the-art results and even surpassed human-level performance across various natural language processing tasks. However, these models remain opaque and hard to explain due to their vast complexity and size. This limits adoption in highly-regulated domains like medicine and finance, and often there is a lack of trust from non-expert end-users. In this paper, we show that by teaching a model to generate explanations alongside its predictions on a large annotated dataset, we can transfer this capability to a low-resource task in another domain. Our proposed three-step training procedure improves explanation quality by up to 7% and avoids sacrificing classification performance on the downstream task, while at the same time reducing the need for human annotations.

Place, publisher, year, edition, pages
Springer Nature , 2021. p. 76-89
Keywords [en]
Explainable AI, Generative explanations, Transfer learning, Deep learning, Information systems, Information use, Large dataset, Classification performance, Human annotations, Human-level performance, Learning models, NAtural language processing, State of the art, Time-reducing, Training procedures, Natural language processing systems
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-310724DOI: 10.1007/978-3-030-80599-9_8ISI: 000884368500008Scopus ID: 2-s2.0-85111441259OAI: oai:DiVA.org:kth-310724DiVA, id: diva2:1651800
Conference
26th International Conference on Applications of Natural Language to Information Systems, NLDB 2021, Virtual, Online, 23-25 June 2021
Note

Part of proceedings ISBN: 978-3-030-80598-2

QC 20220413

Available from: 2022-04-13 Created: 2022-04-13 Last updated: 2023-12-05Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Erliksson, Karl FredrikMatskin, MihhailPayberah, Amir H.

Search in DiVA

By author/editor
Erliksson, Karl FredrikMatskin, MihhailPayberah, Amir H.
By organisation
Electrical EngineeringSoftware and Computer systems, SCS
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 52 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf