Cross-Domain Transfer of Generative Explanations Using Text-to-Text Models
2021 (English)In: Lecture Notes in Computer Science, Springer Nature , 2021, p. 76-89Conference paper, Published paper (Refereed)
Abstract [en]
Deep learning models based on the Transformers architecture have achieved impressive state-of-the-art results and even surpassed human-level performance across various natural language processing tasks. However, these models remain opaque and hard to explain due to their vast complexity and size. This limits adoption in highly-regulated domains like medicine and finance, and often there is a lack of trust from non-expert end-users. In this paper, we show that by teaching a model to generate explanations alongside its predictions on a large annotated dataset, we can transfer this capability to a low-resource task in another domain. Our proposed three-step training procedure improves explanation quality by up to 7% and avoids sacrificing classification performance on the downstream task, while at the same time reducing the need for human annotations.
Place, publisher, year, edition, pages
Springer Nature , 2021. p. 76-89
Keywords [en]
Explainable AI, Generative explanations, Transfer learning, Deep learning, Information systems, Information use, Large dataset, Classification performance, Human annotations, Human-level performance, Learning models, NAtural language processing, State of the art, Time-reducing, Training procedures, Natural language processing systems
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-310724DOI: 10.1007/978-3-030-80599-9_8ISI: 000884368500008Scopus ID: 2-s2.0-85111441259OAI: oai:DiVA.org:kth-310724DiVA, id: diva2:1651800
Conference
26th International Conference on Applications of Natural Language to Information Systems, NLDB 2021, Virtual, Online, 23-25 June 2021
Note
Part of proceedings ISBN: 978-3-030-80598-2
QC 20220413
2022-04-132022-04-132023-12-05Bibliographically approved