kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A Toxic Style Transfer Method Based on the Delete–Retrieve–Generate Framework Exploiting Toxic Lexicon Semantic Similarity
KTH, School of Electrical Engineering and Computer Science (EECS).
Intelligent Systems Group, ETSI Telecomunicación, Universidad Politécnica de Madrid, Avda. Complutense 30, Madrid, 28040, Spain..
Intelligent Systems Group, ETSI Telecomunicación, Universidad Politécnica de Madrid, Avda. Complutense 30, Madrid, 28040, Spain..
2023 (English)In: Applied Sciences, E-ISSN 2076-3417, Vol. 13, no 15, article id 8590Article in journal (Refereed) Published
Abstract [en]

Whether consciously or inadvertently, our messages can include toxic language which contributes to the polarization of social networks. Intelligent techniques can help us detect these expressions and even change them into kinder expressions by applying style transfer techniques. This work aims to advance detoxification style transfer techniques using deep learning and semantic similarity technologies. The article explores the advantages of a toxicity-deletion method that uses linguistic resources in a detoxification system. For this purpose, we propose a method that removes toxic words from the source sentence using a similarity function with a toxic vocabulary. We present two models that leverage it, namely, LexiconGST and MultiLexiconGST, which are based on the Delete –Retrieve–Generate framework. Experimental results show that our models perform well in the detoxification task compared to other state-of-the-art methods. Finally, this research confirms that linguistic resources can guide deep learning techniques and improve their performance.

Place, publisher, year, edition, pages
MDPI AG , 2023. Vol. 13, no 15, article id 8590
Keywords [en]
deep learning, detoxifcation, linguistics, NLP, text style transfer, transformers
National Category
Natural Language Processing
Identifiers
URN: urn:nbn:se:kth:diva-334959DOI: 10.3390/app13158590ISI: 001046084800001Scopus ID: 2-s2.0-85167907477OAI: oai:DiVA.org:kth-334959DiVA, id: diva2:1792799
Note

QC 20230830

Available from: 2023-08-30 Created: 2023-08-30 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Iglesias, Martín

Search in DiVA

By author/editor
Iglesias, Martín
By organisation
School of Electrical Engineering and Computer Science (EECS)
In the same journal
Applied Sciences
Natural Language Processing

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 29 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf