The Large Language Model GreekLegalRoBERTa
2024 (English)In: PROCEEDINGS OF THE 13TH HELLENIC CONFERENCE ON ARTIFICIAL INTELLIGENCE, SETN 2024, Association for Computing Machinery (ACM) , 2024, article id 18Conference paper, Published paper (Refereed)
Abstract [en]
We develop four versions of GreekLegalRoBERTa, which are four large language models trained on Greek legal and nonlegal text. We show that our models surpass the performance of GreekLegalBERT, Greek- LegalBERT-v2, and GreekBERT in two tasks involving Greek legal documents: named entity recognition and multi-class legal topic classification. We view our work as a contribution to the study of domain-specific NLP tasks in low-resource languages, like Greek, using modern NLP techniques and methodologies.
Place, publisher, year, edition, pages
Association for Computing Machinery (ACM) , 2024. article id 18
Keywords [en]
Natural Language Processing, Pre-trained Language Models, Greek NLP Resources, Greek Legislation, Classification, Named Entity Recognition
National Category
Natural Language Processing
Identifiers
URN: urn:nbn:se:kth:diva-361612DOI: 10.1145/3688671.3688770ISI: 001426422600018Scopus ID: 2-s2.0-85216010719OAI: oai:DiVA.org:kth-361612DiVA, id: diva2:1947665
Conference
13th Hellenic Conference on Artificial Intelligence, SEP 11-13, 2024, Piraeus, GREECE
Note
Part of ISBN 979-8-4007-0982-1
2025-03-262025-03-262025-03-26Bibliographically approved