kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Explainable Artificial Intelligence for Telecommunications
KTH, School of Industrial Engineering and Management (ITM), Engineering Design, Mechatronics and Embedded Control Systems.ORCID iD: 0000-0002-6650-2789
2026 (English)Doctoral thesis, comprehensive summary (Other academic)
Sustainable development
SDG 9: Industry, innovation and infrastructure, SDG 12: Responsible consumption and production, SDG 5: Gender equality, SDG 10: Reduced inequalities, SDG 16: Peace, justice and strong institutions
Abstract [en]

Artificial Intelligence (AI) is a key driver of technological development in many industrial sectors. It is being embedded into many components of telecommunications networks to optimize their functionality in various ways. AI technologies are advancing rapidly, with increasingly sophisticated techniques being introduced. Therefore, understanding how an AI model operates and arrives at its output is crucial to ensure the integrity of the overall system. One way to achieve this is by applying Explainable  Artificial Intelligence (XAI) techniques to generate information about the operation of an AI model. This thesis develops and evaluates XAI techniques to improve the transparency of AI models.

In supervised learning, several XAI methods that compute feature importance were applied to identify the root cause of network operation issues. Their characteristics were compared and analyzed for local, cohort, and global scopes. However, the generated attributive explanations do not provide actionable insight to resolve the underlying issue. Therefore, another type of explanation, namely counterfactual, was explored during the study. This type of explanation indicates the changes necessary to obtain a different result. Counterfactual explanations were utilized to prevent potential issues such as Service Level Agreement (SLA) violations from occurring. This method was shown to significantly reduce SLA violations in an emulated network, but requires explanation-to-action conversion.

Unlike the previous method, a Reinforcement Learning (RL) agent can perform an action in its environment to achieve its goal, eliminating the need for explanation-to-action conversion. Therefore, understanding its behavior becomes important, especially when it controls a critical infrastructure. In this thesis, two state-of-the-art Explainable Reinforcement Learning (XRL) methods, namely reward decomposition and Autonomous Policy Explanation (APE), were investigated and implemented to generate explanations for different users, technical and non-technical, respectively. While the reward decomposition explains the output of a model and the feature attribution explains the input, the connection between them was missing in the literature. In this thesis, the combination of feature importance and reward decomposition methods was proposed to generate detailed explanations as well as to identify and mitigate bias in the AI models. In addition, a detailed contrastive explanation can be generated to explain why an action is preferred over another. For non-technical users, APE was integrated with the attribution method to generate explanations for a certain condition. APE was also integrated with a counterfactual method to generate a meaningful explanation. However, APE has a limitation in scaling up with the number of predicates. Therefore, an alternative textual explainer, namely Clustering-Based Summarizer (CBS), was proposed to address this limitation. The evaluation of textual explanations is limited in the literature. Therefore, a rule extraction technique was proposed to evaluate textual explanations based on their characteristics, fidelity, and performance. In addition, two refinement techniques were proposed to improve the F1 score and reduce the number of duplicate conditions. 

In summary, this thesis has developed the following contributions: a) implementation and analysis of different XAI methods; b) methods to utilize explanations and explainers; c) evaluation methods for AI explanations; and d) methods to improve explanation quality. This thesis revolves around network automation in the telecommunications field. The explainability methods for supervised learning were applied to a network slice assurance use case, and for reinforcement learning, it was applied to a network optimization use case (namely, Remote Electrical Tilt (RET)). In addition, applications in other open-source environments were also presented, showing broader applications in different use cases.

Abstract [sv]

Artificiell Intelligens (AI) är en viktig drivkraft för teknologisk utveckling inom många industriella sektorer. Den implementeras i många delar av telekommunikationsnätverk för att optimera deras funktionalitet på olika sätt. AI-teknologier utvecklats snabbt, med alltmer sofistikerade tekniker som introduceras. Därför är det avgörande att förstå hur en AI-modell fungerar och kommer fram till sitt resultat för att säkerställa systemets integritet. Ett sätt att uppnå detta är att tillämpa förklarbara-AI-tekniker för att generera information om en AI-modells funktion. Denna avhandling utvecklar och utvärderar förklarbara-AI-tekniker för att förbättra transparensen hos AI-modeller.

Inom övervakad inlärning tillämpades flera förklarbara-AI-metoder som beräknar variabelsviktighet för att identifiera den bakomliggande orsaken till nätverksdriftsproblem. Deras egenskaper jämfördes och analyserades på lokal, grupp- och global nivå. Dock ger de genererade attributiva förklaringarna ingen handlingsbar insikt för att lösa det underliggande problemet. Därför utforskades en annan typ av förklaring, nämligen kontrafaktisk, under studien. Den här typen av förklaring indikerar de förändringar som krävs för att erhålla ett annat resultat. Kontrafaktiska förklaringar användes för att förhindra potentiella problem, såsom brott mot servicenivåavtal (Service Level Agreements, SLA). Den här metoden visade sig minska SLA-överträdelser i ett emulerat nätverk avsevärt, men kräver en konvertering från förklaring till handling.

Till skillnad från den tidigare metoden kan en förstärkningsinlärningsagent utföra en handling i sin miljö för att uppnå sitt mål, vilket eliminerar behovet av konvertering från förklaring till handling. Därför blir det viktigt att förstå dess beteende, särskilt när den styr en kritisk infrastruktur. I denna avhandling undersöktes och implementerades två nya metoder för förklarbar förstärkningsinlärning, nämligen belöningsdekomposition och Autonomous Policy Explanation (APE), för att generera förklaringar för olika användare, tekniska och icke-tekniska. Medan belöningsdekomposition förklarar en modells utdata och variabelsattribution förklarar indata, saknades kopplingen mellan dessa i litteraturen. I denna avhandling föreslogs en kombination av metoder för variabelsviktighet och belöningsdekomposition för att generera detaljerade förklaringar samt för att identifiera och mildra bias i AI-modellerna. Dessutom kan en detaljerad kontrastiv förklaring genereras för att förklara varför en åtgärd föredras framför en annan. För icke-tekniska användare integrerades APE med attributmetoden för att generera förklaringar för ett visst tillstånd. APE integrerades också med en kontrafaktisk metod för att skapa en meningsfull förklaring. Dock har APE en begränsning när det gäller skalbarhet med antalet predikat. Därför föreslogs en alternativ textförklarare, nämligen Clustering-Based Summarizer (CBS), för att hantera denna begränsning. Utvärderingen av textuella förklaringar är begränsad i litteraturen. Därför föreslogs en regelutvinningsmetod för att utvärdera textuella förklaringar baserat på deras egenskaper, tillförlitlighet och prestanda.Dessutom föreslogs två förfiningstekniker för att förbättra F1-poängen och minska antalet duplicerade villkor.

Sammanfattningsvis, avhandlingen har utvecklat följande bidrag: a) implementering och analys av olika förklarbara-AI-metoder; b) metoder för att använda förklaringar och förklarare; c) utvärderingsmetoder för AI-förklaringar; och d) metoder för att förbättra förklaringars kvalitet. Avhandlingen kretsar kring nätverksautomation inom telekommunikationsområdet. Förklaringsbarhetsmetoderna för övervakad inlärning tillämpades på ett fall för nätverksskiveförsäkran, och för förstärkningsinlärning tillämpades de på ett nätverksoptimeringsfall (nämligen fjärrstyrd elektrisk tilt). Dessutom presenterades tillämpningar i andra öppna miljöer, vilket visar bredare användningsområden i olika tillämpningsfall.

Place, publisher, year, edition, pages
Stockholm, Sweden: KTH Royal Institute of Technology, 2026. , p. 68
Series
TRITA-ITM-AVL ; 2026:3
Keywords [en]
Artificial Intelligence, Explainability, Supervised Learning, Reinforcement Learning, Network Slice Assurance, Network Optimization, Telecommunications
Keywords [sv]
Artificiell intelligens, Förklarbarhet, Övervakad inlärning, Förstärkande inlärning, Nätverk Slice Assurance, Nätverksoptimering, Telekommunikation
National Category
Computer Sciences Telecommunications Artificial Intelligence
Research subject
Machine Design
Identifiers
URN: urn:nbn:se:kth:diva-375341ISBN: 978-91-8106-506-0 (print)OAI: oai:DiVA.org:kth-375341DiVA, id: diva2:2027356
Public defence
2026-02-06, F3 / https://kth-se.zoom.us/j/63010540491, Lindstedtsvägen 26 & 28, Stockholm, 09:30 (English)
Opponent
Supervisors
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Available from: 2026-01-13 Created: 2026-01-12 Last updated: 2026-02-02Bibliographically approved
List of papers
1. Explainability Methods for Identifying Root-Cause of SLA Violation Prediction in 5G Network
Open this publication in new window or tab >>Explainability Methods for Identifying Root-Cause of SLA Violation Prediction in 5G Network
Show others...
2020 (English)In: GLOBECOM 2020 - 2020 IEEE Global Communications Conference, Institute of Electrical and Electronics Engineers (IEEE) , 2020Conference paper, Published paper (Refereed)
Abstract [en]

Artificial Intelligence (AI) is implemented in various applications of telecommunication domain, ranging from managing the network, controlling a specific hardware function, preventing a failure, or troubleshooting a problem till automating the network slice management in 5G. The greater levels of autonomy increase the need for explainability of the decisions made by AI so that humans can understand them (e.g. the underlying data evidence and causal reasoning) consequently enabling trust. This paper presents first, the application of multiple global and local explainability methods with the main purpose to analyze the root-cause of Service Level Agreement violation prediction in a 5G network slicing setup by identifying important features contributing to the decision. Second, it performs a comparative analysis of the applied methods to analyze explainability of the predicted violation. Further, the global explainability results are validated using statistical Causal Dataframe method in order to improve the identified cause of the problem and thus validating the explainability.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2020
Keywords
Forecasting; Queueing networks; 5g slicing; Causality; Explainability; Global and local explanation; Hardware functions; Level of autonomies; Root cause; Service level agreement violation prediction; Servicelevel agreement (SLA); Specific hardware; 5G mobile communication systems
National Category
Artificial Intelligence
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-373648 (URN)10.1109/GLOBECOM42002.2020.9322496 (DOI)000668970502120 ()2-s2.0-85100415521 (Scopus ID)
Conference
IEEE Global Communications Conference, GLOBECOM, Taipei, Taiwan, December 7-11, 2020
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

Part of ISBN 9781728182988

QC 20251204

Available from: 2025-12-04 Created: 2025-12-04 Last updated: 2026-01-12Bibliographically approved
2. Using Counterfactuals to Proactively Solve Service Level Agreement Violations in 5G Networks
Open this publication in new window or tab >>Using Counterfactuals to Proactively Solve Service Level Agreement Violations in 5G Networks
2022 (English)In: 2022 IEEE 20TH INTERNATIONAL CONFERENCE ON INDUSTRIAL INFORMATICS (INDIN), Institute of Electrical and Electronics Engineers (IEEE) , 2022, p. 552-559Conference paper, Published paper (Refereed)
Abstract [en]

A main challenge of using 5G network slices is to meet all the quality of service requirements of the slices (which are agreed with the customer in a service level agreement (SLA)), throughout the network slices' lifecycle. To avoid the penalty for violation, a proactive solution is presented, including predicting the SLA violation, explaining the violation cause, and then providing an adaptation to traffic. This work uses counterfactual (CF) explanations to 1) explain the main factors affecting the identified model's SLA violation prediction and 2) present modifications in the input values, which are required to configure the network traffic to avoid such a violation. We evaluate the CF explanation at two different levels where the generated CF instance is fed to the predictive model, and then actuation data are generated to evaluate the result in the real network. Our solution minimizes the violation up to 98%. This information can be utilized to reconfigure the system, either by humans or by the system automatically, to make it fully autonomous on the one hand and comply with the 'right to explanation' introduced by the European Union's General Data Protection Regulation on the other hand.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2022
Series
IEEE International Conference on Industrial Informatics INDIN, ISSN 1935-4576
Keywords
5G network slicing, SLA violation predictions, explainable artificial intelligence, counterfactual
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-324791 (URN)10.1109/INDIN51773.2022.9976075 (DOI)000907121600087 ()2-s2.0-85145777745 (Scopus ID)
Conference
20th IEEE International Conference on Industrial Informatics (INDIN), JUL 25-28, 2022, ELECTR NETWORK
Note

QC 20230316

Available from: 2023-03-16 Created: 2023-03-16 Last updated: 2026-01-12Bibliographically approved
3. Explainable Reinforcement Learning for Human-Robot Collaboration
Open this publication in new window or tab >>Explainable Reinforcement Learning for Human-Robot Collaboration
Show others...
2021 (English)In: 2021 20Th International Conference On Advanced Robotics (ICAR), Institute of Electrical and Electronics Engineers (IEEE) , 2021, p. 927-934Conference paper, Published paper (Refereed)
Abstract [en]

Reinforcement learning (RL) is getting popular in the robotics field due to its nature to learn from dynamic environments. However, it is unable to provide explanations of why an output was generated. Explainability becomes therefore important in situations where humans interact with robots, such as in human-robot collaboration (HRC) scenarios. Attempts to address explainability in robotics usually are restricted to explain a specific decision taken by the RL model, but not to understand the complete behavior of the robot. In addition, the explainability methods are restricted to be used by domain experts as queries and responses are not translated to natural language. This work overcomes these limitations by proposing an explainability solution for RL models applied to HRC. It is mainly formed by the adaptation of two methods: (i) Reward decomposition gives an insight into the factors that impacted the robot's choice by decomposing the reward function. It further provides sets of relevant reasons for each decision taken during the robot's operation; (ii) Autonomous policy explanation provides a global explanation of the robot's behavior by answering queries in the form of natural language, thus making understandable to any human user. Experiments in simulated HRC scenarios revealed an increased understanding of the optimal choices made by the robots. Additionally, our solution demonstrated as a powerful debugging tool to find weaknesses in the robot's policy and assist in its improvement.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2021
Keywords
Explainable reinforcement learning, explainable artificial intelligence, human-robot collaboration, risk mitigation, reward decomposition, autonomous policy explanation, collaborative robots, safety
National Category
Robotics and automation Computer Sciences
Identifiers
urn:nbn:se:kth:diva-310885 (URN)10.1109/ICAR53236.2021.9659472 (DOI)000766318900140 ()2-s2.0-85124696608 (Scopus ID)
Conference
20th International Conference on Advanced Robotics (ICAR), DEC 07-10, 2021, ELECTR NETWORK
Note

Part of proceedings: ISBN 978-1-6654-3684-7

QC 20220413

Available from: 2022-04-13 Created: 2022-04-13 Last updated: 2026-01-12Bibliographically approved
4. BEERL: Both Ends Explanations for Reinforcement Learning
Open this publication in new window or tab >>BEERL: Both Ends Explanations for Reinforcement Learning
2022 (English)In: Applied Sciences, E-ISSN 2076-3417, Vol. 12, no 21, p. 10947-, article id 10947Article in journal (Refereed) Published
Abstract [en]

Deep Reinforcement Learning (RL) is a black-box method and is hard to understand because the agent employs a neural network (NN). To explain the behavior and decisions made by the agent, different eXplainable RL (XRL) methods are developed; for example, feature importance methods are applied to analyze the contribution of the input side of the model, and reward decomposition methods are applied to explain the components of the output end of the RL model. In this study, we present a novel method to connect explanations from both input and output ends of a black-box model, which results in fine-grained explanations. Our method exposes the reward prioritization to the user, which in turn generates two different levels of explanation and allows RL agent reconfigurations when unwanted behaviors are observed. The method further summarizes the detailed explanations into a focus value that takes into account all reward components and quantifies the fulfillment of the explanation of desired properties. We evaluated our method by applying it to a remote electrical telecom-antenna-tilt use case and two openAI gym environments: lunar lander and cartpole. The results demonstrated fine-grained explanations by detailing input features’ contributions to certain rewards and revealed biases of the reward components, which are then addressed by adjusting the reward’s weights.

Place, publisher, year, edition, pages
MDPI AG, 2022
Keywords
bias, deep reinforcement learning, explainability, explainable reinforcement learning, reward decomposition, reward prioritization
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-329022 (URN)10.3390/app122110947 (DOI)000880918200001 ()2-s2.0-85141824419 (Scopus ID)
Note

QC 20230614

Available from: 2023-06-14 Created: 2023-06-14 Last updated: 2026-01-12Bibliographically approved
5. Comprehensive Reinforcement Learning Explanations Using Queries
Open this publication in new window or tab >>Comprehensive Reinforcement Learning Explanations Using Queries
2026 (English)In: Explainable Artificial Intelligence: Third World Conference, xAI 2025, Istanbul, Turkey, July 9–11, 2025, Proceedings, Part V / [ed] Guidotti R. and Schmid U. and Longo L., Springer Nature , 2026, p. 27-40Conference paper, Published paper (Refereed)
Abstract [en]

Generating detailed explanations that are easy to comprehend and interact with is a challenging problem for complex Reinforcement Learning (RL) agents. While various methods explain different aspects of the agents, it is difficult to aggregate and generate tailored insights for different users. Thus, we propose a comprehensive explainability approach that utilizes interactive natural language queries and generates different types of explanations. First, we introduce a new approach to generate meaningful counterfactual explanations using natural language queries. Further, we complement the natural language explanations with customized feature attributions for detailed insights. This helps in facilitating the interaction with explanations as well as tailoring the explanations for different purposes and levels of expertise. We demonstrate our proposal using an industrial telecommunication use case which shows its applicability and utility in a complex real-world scenario.

Place, publisher, year, edition, pages
Springer Nature, 2026
Series
Communications in Computer and Information Science, ISSN 1865-0929, E-ISSN 1865-0937 ; 2580
Keywords
Natural language processing systems; Counterfactuals; Explainability; Explainable reinforcement learning; Natural language explanations; Natural language queries; New approaches; Real-world scenario; Reinforcement learning agent; Reinforcement learnings; Reinforcement learning
National Category
Artificial Intelligence
Identifiers
urn:nbn:se:kth:diva-373645 (URN)10.1007/978-3-032-08333-3_2 (DOI)2-s2.0-105020691300 (Scopus ID)
Conference
3rd World Conference on Explainable Artificial Intelligence, xAI 2025, Istanbul, Turkey, July 9–11, 2025
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

Part of ISBN 9783032083326, 9783032083333

QC 20251204

Available from: 2025-12-04 Created: 2025-12-04 Last updated: 2026-01-12Bibliographically approved
6. Textual Explanations and Their Evaluations for Reinforcement Learning Policy
Open this publication in new window or tab >>Textual Explanations and Their Evaluations for Reinforcement Learning Policy
Show others...
2026 (English)Manuscript (preprint) (Other academic)
Abstract [en]

Understanding a Reinforcement Learning (RL) policy is crucial for ensuring that autonomous agents behave according to human expectations. This goal can be achieved using Explainable Reinforcement Learning (XRL) techniques. Although textual explanations are easily understood by humans, ensuring their correctness remains a challenge, and evaluations in state-of-the-art remain limited. We present a novel XRL framework for generating textual explanations, converting them into a set of transparent rules, improving their quality, and evaluating them. Expert's knowledge can be incorporated into this framework, and an automatic predicate generator is also proposed to determine the semantic information of a state. Textual explanations are generated using a Large Language Model (LLM) and a clustering technique to identify frequent conditions. These conditions are then converted into rules to evaluate their properties, fidelity, and performance in the deployed environment. Two refinement techniques are proposed to improve the quality of explanations and reduce conflicting information. Experiments were conducted in three open-source environments to enable reproducibility, and in a telecom use case to evaluate the industrial applicability of the proposed XRL framework. This framework addresses the limitations of an existing method, Autonomous Policy Explanation, and the generated transparent rules can achieve satisfactory performance on certain tasks. This framework also enables a systematic and quantitative evaluation of textual explanations, providing valuable insights for the XRL field.

Keywords
Explainable reinforcement learning, textual explanations, rule extraction, explanation evaluation
National Category
Artificial Intelligence Computer Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-375036 (URN)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

QC 20260109

Available from: 2026-01-08 Created: 2026-01-08 Last updated: 2026-01-12Bibliographically approved

Open Access in DiVA

kappa(1975 kB)151 downloads
File information
File name FULLTEXT01.pdfFile size 1975 kBChecksum SHA-512
4834d61ab4ca91d52f5a06cc7467fd1f331dac655f94575f432e6b80e769ef7f7265893b47469a03f8fa8c3d6ef06900b0b8973cc87aae89a9c1f4eda1838b15
Type fulltextMimetype application/pdf

Authority records

Terra, Ahmad

Search in DiVA

By author/editor
Terra, Ahmad
By organisation
Mechatronics and Embedded Control Systems
Computer SciencesTelecommunicationsArtificial Intelligence

Search outside of DiVA

GoogleGoogle Scholar
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 10304 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf