kth.sePublications
Change search
Link to record
Permanent link

Direct link
Girdzijauskas, SarunasORCID iD iconorcid.org/0000-0003-4516-7317
Alternative names
Publications (10 of 95) Show all publications
Komini, V. & Girdzijauskas, S. (2025). Integrating Logit Space Embeddings for Reliable Out-of-Distribution Detection. In: Machine Learning, Optimization, and Data Science - 10th International Conference, LOD 2024, Revised Selected Papers: . Paper presented at 10th International Conference on Machine Learning, Optimization, and Data Science, LOD 2024, Castiglione della Pescaia, Italy, September 22-25, 2024 (pp. 255-269). Springer Nature
Open this publication in new window or tab >>Integrating Logit Space Embeddings for Reliable Out-of-Distribution Detection
2025 (English)In: Machine Learning, Optimization, and Data Science - 10th International Conference, LOD 2024, Revised Selected Papers, Springer Nature , 2025, p. 255-269Conference paper, Published paper (Refereed)
Abstract [en]

Deep learning (DL) models have significantly transformed machine learning (ML), particularly with their prowess in classification tasks. However, these models struggle to differentiate between in-distribution (ID) and out-of-distribution (OOD) data at the testing phase. This challenge has curtailed their deployment in sensitive fields like biotechnology, where misidentifying OOD data, such as unclear or unknown bacterial genomic sequences, as known ID classes could lead to dire consequences. To address this, we propose an approach to make DL models OOD-sensitive by exploiting the configuration of the logit space embeddings, into the model’s decision-making process. Leveraging the effect observed in recent studies that there is minimal overlap between the embeddings of ID and OOD data, we use a density estimator to model the ID logit distribution based on the training data. This allows us to reliably flag data that do not match the ID distribution as OOD. Our methodology is designed to be independent of the specific data or model architecture and can seamlessly augment existing trained models without the need to expose them to OOD data. Testing our method on widely recognized image datasets, we achieve leading-edge results, including a substantial 10% enhancement in the area under the receiver operating characteristic curve (AUCROC) on the Google genome dataset.

Place, publisher, year, edition, pages
Springer Nature, 2025
National Category
Computer Sciences Bioinformatics (Computational Biology) Software Engineering
Identifiers
urn:nbn:se:kth:diva-361973 (URN)10.1007/978-3-031-82484-5_19 (DOI)2-s2.0-105000982628 (Scopus ID)
Conference
10th International Conference on Machine Learning, Optimization, and Data Science, LOD 2024, Castiglione della Pescaia, Italy, September 22-25, 2024
Note

Part of ISBN 9783031824838

QC 20250404

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-04Bibliographically approved
Samy, A. E., Kefato, Z. T. & Girdzijauskas, Š. (2025). Leap: Inductive Link Prediction via Learnable Topology Augmentation. In: Machine Learning, Optimization, and Data Science - 10th International Conference, LOD 2024, Revised Selected Papers: . Paper presented at 10th International Conference on Machine Learning, Optimization, and Data Science, LOD 2024, Castiglione della Pescaia, Italy, Sep 22 2024 - Sep 25 2024 (pp. 448-463). Springer Nature
Open this publication in new window or tab >>Leap: Inductive Link Prediction via Learnable Topology Augmentation
2025 (English)In: Machine Learning, Optimization, and Data Science - 10th International Conference, LOD 2024, Revised Selected Papers, Springer Nature , 2025, p. 448-463Conference paper, Published paper (Refereed)
Abstract [en]

Link prediction is a crucial task in many downstream applications of graph machine learning. To this end, Graph Neural Network (GNN) is a widely used technique for link prediction, mainly in transductive settings, where the goal is to predict missing links between existing nodes. However, many real-life applications require an inductive setting that accommodates for new nodes, coming into an existing graph. Thus, recently inductive link prediction has attracted considerable attention, and a multi-layer perceptron (MLP) is the popular choice of most studies to learn node representations. However, these approaches have limited expressivity and do not fully capture the graph’s structural signal. Therefore, in this work we propose LEAP, an inductive link prediction method based on LEArnable toPology augmentation. Unlike previous methods, LEAP models the inductive bias from both the structure and node features, and hence is more expressive. To the best of our knowledge, this is the first attempt to provide structural contexts for new nodes via learnable augmentation in inductive settings. Extensive experiments on seven real-world homogeneous and heterogeneous graphs demonstrates that LEAP significantly surpasses SOTA methods. The improvements are up to 22% and 17% in terms of AUC and average precision, respectively. The code and datasets are available on GitHub (1https://github.com/AhmedESamy/LEAP/).

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Graph Neural Networks, Heterogeneous graphs, Inductive link prediction, Learnable augmentation
National Category
Computer Sciences Applied Mechanics
Identifiers
urn:nbn:se:kth:diva-361992 (URN)10.1007/978-3-031-82481-4_31 (DOI)2-s2.0-105000770530 (Scopus ID)
Conference
10th International Conference on Machine Learning, Optimization, and Data Science, LOD 2024, Castiglione della Pescaia, Italy, Sep 22 2024 - Sep 25 2024
Note

Part of ISBN 9783031824807

QC 20250403

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-03Bibliographically approved
Listo Zec, E., Hagander, T., Ihre-Thomason, E. & Girdzijauskas, S. (2025). On the effects of similarity metrics in decentralized deep learning under distributional shift. Transactions on Machine Learning Research, 2025-January, 1-23
Open this publication in new window or tab >>On the effects of similarity metrics in decentralized deep learning under distributional shift
2025 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856, Vol. 2025-January, p. 1-23Article in journal (Refereed) Published
Abstract [en]

Decentralized Learning (DL) enables privacy-preserving collaboration among organizations or users to enhance the performance of local deep learning models. However, model aggregation becomes challenging when client data is heterogeneous, and identifying compatible collaborators without direct data exchange remains a pressing issue. In this paper, we investigate the effectiveness of various similarity metrics in DL for identifying peers for model merging, conducting an empirical analysis across multiple datasets with distribution shifts. Our research provides insights into the performance of these metrics, examining their role in facilitating effective collaboration. By exploring the strengths and limitations of these metrics, we contribute to the development of robust DL methods.

Place, publisher, year, edition, pages
Transactions on Machine Learning Research, 2025
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-361192 (URN)2-s2.0-85219582623 (Scopus ID)
Note

QC 20250313

Available from: 2025-03-12 Created: 2025-03-12 Last updated: 2025-03-13Bibliographically approved
Cornell, F., Jin, Y., Karlgren, J. & Girdzijauskas, S. (2025). Unsupervised Ontology- and Taxonomy Construction Through Hyperbolic Relational Domains and Ranges. In: Machine Learning and Principles and Practice of Knowledge Discovery in Databases - International Workshops of ECML PKDD 2023, Revised Selected Papers: . Paper presented at Joint European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML PKDD 2023, Turin, Italy, September 18-22, 2023 (pp. 339-348). Springer Nature
Open this publication in new window or tab >>Unsupervised Ontology- and Taxonomy Construction Through Hyperbolic Relational Domains and Ranges
2025 (English)In: Machine Learning and Principles and Practice of Knowledge Discovery in Databases - International Workshops of ECML PKDD 2023, Revised Selected Papers, Springer Nature , 2025, p. 339-348Conference paper, Published paper (Refereed)
Abstract [en]

The sets of possible heads and tails for relations in a Knowledge Graph and a type taxonomy of participating entities are important aspects of a Knowledge Graph and constitute a large portion of a Knowledge Graph’s ontology. Making the ontology explicit helps to ensure consistency of the knowledge represented in the graph and allows for less costly maintenance and update of graph content. However, Knowledge Graphs are often conveniently formed without an explicitly described ontology, using only (head, relation, tail)-tuples. For such graphs without predefined structure, we propose learning an ontology and a hierarchical type taxonomy from the graph itself, taking advantage of the reciprocity of entity types and sets of possible heads (relational domains) and tails (relational ranges). Experiments on real-world datasets validate our approach and demonstrate the promise for leveraging machine learning methodologies to efficiently generate taxonomies and ontologies jointly for Knowledge Graphs (Our implementation can be found at https://tinyurl.com/5n6sw6yj).

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Hyperbolic learning, Knowledge Graphs, Ontology Learning, Taxonomy Learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-359658 (URN)10.1007/978-3-031-74633-8_23 (DOI)2-s2.0-85216103400 (Scopus ID)
Conference
Joint European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML PKDD 2023, Turin, Italy, September 18-22, 2023
Note

Part of ISBN 9783031746321

QC 20250207

Available from: 2025-02-06 Created: 2025-02-06 Last updated: 2025-02-07Bibliographically approved
Isaksson, M., Listo Zec, E., Coster, R., Gillblad, D. & Girdzijauskas, S. (2023). Adaptive Expert Models for Federated Learning. In: Goebel, R Yu, H Faltings, B Fan, L Xiong, Z (Ed.), Trustworthy Federated Learning: First International Workshop, FL 2022. Paper presented at Trustworthy Federated Learning - First International Workshop, FL 2022, Held in Conjunction with IJCAI 2022, Vienna, Austria, July 23, 2022 (pp. 1-16). Springer Nature, 13448
Open this publication in new window or tab >>Adaptive Expert Models for Federated Learning
Show others...
2023 (English)In: Trustworthy Federated Learning: First International Workshop, FL 2022 / [ed] Goebel, R Yu, H Faltings, B Fan, L Xiong, Z, Springer Nature , 2023, Vol. 13448, p. 1-16Conference paper, Published paper (Refereed)
Abstract [en]

Federated Learning (FL) is a promising framework for distributed learning when data is private and sensitive. However, the state-of-the-art solutions in this framework are not optimal when data is heterogeneous and non-IID. We propose a practical and robust approach to personalization in FL that adjusts to heterogeneous and non-IID data by balancing exploration and exploitation of several global models. To achieve our aim of personalization, we use a Mixture of Experts (MoE) that learns to group clients that are similar to each other, while using the global models more efficiently. We show that our approach achieves an accuracy up to 29.78% better than the state-of-the-art and up to 4.38% better compared to a local model in a pathological non-IID setting, even though we tune our approach in the IID setting.

Place, publisher, year, edition, pages
Springer Nature, 2023
Series
Lecture Notes in Artificial Intelligence, ISSN 2945-9133
Keywords
Federated learning, Personalization, Privacy preserving
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-330493 (URN)10.1007/978-3-031-28996-5_1 (DOI)000999818400001 ()2-s2.0-85152565856 (Scopus ID)
Conference
Trustworthy Federated Learning - First International Workshop, FL 2022, Held in Conjunction with IJCAI 2022, Vienna, Austria, July 23, 2022
Note

Part of proceedings ISBN 978-3-031-28995-8  978-3-031-28996-5

QC 20230630

Available from: 2023-06-30 Created: 2023-06-30 Last updated: 2024-05-27Bibliographically approved
Roy, D., Komini, V. & Girdzijauskas, S. (2023). Classifying falls using out-of-distribution detection in human activity recognition. AI Communications, 36(4), 251-267
Open this publication in new window or tab >>Classifying falls using out-of-distribution detection in human activity recognition
2023 (English)In: AI Communications, ISSN 0921-7126, E-ISSN 1875-8452, Vol. 36, no 4, p. 251-267Article in journal (Refereed) Published
Abstract [en]

As the research community focuses on improving the reliability of deep learning, identifying out-of-distribution (OOD) data has become crucial. Detecting OOD inputs during test/prediction allows the model to account for discriminative features unknown to the model. This capability increases the model's reliability since this model provides a class prediction solely at incoming data similar to the training one. Although OOD detection is well-established in computer vision, it is relatively unexplored in other areas, like time series-based human activity recognition (HAR). Since uncertainty has been a critical driver for OOD in vision-based models, the same component has proven effective in time-series applications. In this work, we propose an ensemble-based temporal learning framework to address the OOD detection problem in HAR with time-series data. First, we define different types of OOD for HAR that arise from realistic scenarios. Then we apply our ensemble-based temporal learning framework incorporating uncertainty to detect OODs for the defined HAR workloads. This particular formulation also allows a novel approach to fall detection. We train our model on non-fall activities and detect falls as OOD. Our method shows state-of-The-Art performance in a fall detection task using much lesser data. Furthermore, the ensemble framework outperformed the traditional deep-learning method (our baseline) on the OOD detection task across all the other chosen datasets.

Place, publisher, year, edition, pages
IOS Press, 2023
Keywords
deep learning, human activity recognition, Out-of-distribution detection, time-series classification, uncertainty estimation
National Category
Computer Sciences Bioinformatics (Computational Biology)
Identifiers
urn:nbn:se:kth:diva-339522 (URN)10.3233/AIC-220205 (DOI)001087274200001 ()2-s2.0-85175210057 (Scopus ID)
Note

QC 20231114

Available from: 2023-11-14 Created: 2023-11-14 Last updated: 2024-04-11Bibliographically approved
Samy, A. E., Kefato, Z. T. & Girdzijauskas, S. (2023). Data-Driven Self-Supervised Graph Representation Learning. In: ECAI 2023: 26th European Conference on Artificial Intelligence, including 12th Conference on Prestigious Applications of Intelligent Systems, PAIS 2023 - Proceedings. Paper presented at 26th European Conference on Artificial Intelligence, ECAI 2023, Krakow, Poland, Sep 30 2023 - Oct 4 2023 (pp. 629-636). IOS Press
Open this publication in new window or tab >>Data-Driven Self-Supervised Graph Representation Learning
2023 (English)In: ECAI 2023: 26th European Conference on Artificial Intelligence, including 12th Conference on Prestigious Applications of Intelligent Systems, PAIS 2023 - Proceedings, IOS Press , 2023, p. 629-636Conference paper, Published paper (Refereed)
Abstract [en]

Self-supervised graph representation learning (SSGRL) is a representation learning paradigm used to reduce or avoid manual labeling. An essential part of SSGRL is graph data augmentation. Existing methods usually rely on heuristics commonly identified through trial and error and are effective only within some application domains. Also, it is not clear why one heuristic is better than another. Moreover, recent studies have argued against some techniques (e.g., dropout: that can change the properties of molecular graphs or destroy relevant signals for graph-based document classification tasks). In this study, we propose a novel data-driven SSGRL approach that automatically learns a suitable graph augmentation from the signal encoded in the graph (i.e., the nodes' predictive feature and topological information). We propose two complementary approaches that produce learnable feature and topological augmentations. The former learns multi-view augmentation of node features, and the latter learns a high-order view of the topology. Moreover, the augmentations are jointly learned with the representation. Our approach is general that it can be applied to homogeneous and heterogeneous graphs. We perform extensive experiments on node classification (using nine homogeneous and heterogeneous datasets) and graph property prediction (using another eight datasets). The results show that the proposed method matches or outperforms the SOTA SSGRL baselines and performs similarly to semi-supervised methods. The anonymised source code is available at https://github.com/AhmedESamy/dsgrl/

Place, publisher, year, edition, pages
IOS Press, 2023
National Category
Computer Sciences Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-339683 (URN)10.3233/FAIA230325 (DOI)2-s2.0-85175858097 (Scopus ID)
Conference
26th European Conference on Artificial Intelligence, ECAI 2023, Krakow, Poland, Sep 30 2023 - Oct 4 2023
Note

Part of ISBN 9781643684369

QC 20231116

Available from: 2023-11-16 Created: 2023-11-16 Last updated: 2025-04-03Bibliographically approved
Listo Zec, E., Ekblom, E., Willbo, M., Mogren, O. & Girdzijauskas, S. (2023). Decentralized Adaptive Clustering of Deep Nets is Beneficial for Client Collaboration. In: Goebel, R Yu, H Faltings, B Fan, L Xiong, Z (Ed.), FL 2022: Trustworthy Federated Learning. Paper presented at 1st International Workshop on Trustworthy Federated Learning (FL), JUL 23, 2022, Vienna, AUSTRIA (pp. 59-71). Springer Nature, 13448
Open this publication in new window or tab >>Decentralized Adaptive Clustering of Deep Nets is Beneficial for Client Collaboration
Show others...
2023 (English)In: FL 2022: Trustworthy Federated Learning / [ed] Goebel, R Yu, H Faltings, B Fan, L Xiong, Z, Springer Nature , 2023, Vol. 13448, p. 59-71Conference paper, Published paper (Refereed)
Abstract [en]

We study the problem of training personalized deep learning models in a decentralized peer-to-peer setting, focusing on the setting where data distributions differ between the clients and where different clients have different local learning tasks. We study both covariate and label shift, and our contribution is an algorithm which for each client finds beneficial collaborations based on a similarity estimate for the local task. Our method does not rely on hyperparameters which are hard to estimate, such as the number of client clusters, but rather continuously adapts to the network topology using soft cluster assignment based on a novel adaptive gossip algorithm. We test the proposed method in various settings where data is not independent and identically distributed among the clients. The experimental evaluation shows that the proposed method performs better than previous state-of-the-art algorithms for this problem setting, and handles situations well where previous methods fail.

Place, publisher, year, edition, pages
Springer Nature, 2023
Series
Lecture Notes in Artificial Intelligence, ISSN 2945-9133
Keywords
decentralized learning, federated learning, deep learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-330522 (URN)10.1007/978-3-031-28996-5_5 (DOI)000999818400005 ()2-s2.0-85152516432 (Scopus ID)
Conference
1st International Workshop on Trustworthy Federated Learning (FL), JUL 23, 2022, Vienna, AUSTRIA
Note

QC 20230630

Available from: 2023-06-30 Created: 2023-06-30 Last updated: 2024-12-13Bibliographically approved
Bonvalet, M., Kefato, Z. T. & Girdzijauskas, S. (2023). Graph2Feat: Inductive Link Prediction via Knowledge Distillation. In: ACM Web Conference 2023: Companion of the World Wide Web Conference, WWW 2023. Paper presented at 2023 World Wide Web Conference, WWW 2023, Austin, United States of America, Apr 30 2023 - May 4 2023 (pp. 805-812). Association for Computing Machinery (ACM)
Open this publication in new window or tab >>Graph2Feat: Inductive Link Prediction via Knowledge Distillation
2023 (English)In: ACM Web Conference 2023: Companion of the World Wide Web Conference, WWW 2023, Association for Computing Machinery (ACM) , 2023, p. 805-812Conference paper, Published paper (Refereed)
Abstract [en]

Link prediction between two nodes is a critical task in graph machine learning. Most approaches are based on variants of graph neural networks (GNNs) that focus on transductive link prediction and have high inference latency. However, many real-world applications require fast inference over new nodes in inductive settings where no information on connectivity is available for these nodes. Thereby, node features provide an inevitable alternative in the latter scenario. To that end, we propose Graph2Feat, which enables inductive link prediction by exploiting knowledge distillation (KD) through the Student-Teacher learning framework. In particular, Graph2Feat learns to match the representations of a lightweight student multi-layer perceptron (MLP) with a more expressive teacher GNN while learning to predict missing links based on the node features, thus attaining both GNN's expressiveness and MLP's fast inference. Furthermore, our approach is general; it is suitable for transductive and inductive link predictions on different types of graphs regardless of them being homogeneous or heterogeneous, directed or undirected. We carry out extensive experiments on seven real-world datasets including homogeneous and heterogeneous graphs. Our experiments demonstrate that Graph2Feat significantly outperforms SOTA methods in terms of AUC and average precision in homogeneous and heterogeneous graphs. Finally, Graph2Feat has the minimum inference time compared to the SOTA methods, and 100x acceleration compared to GNNs. The code and datasets are available on GitHub1.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2023
Keywords
graph representation learning, heterogeneous networks, inductive link prediction, knowledge distillation
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-333310 (URN)10.1145/3543873.3587596 (DOI)001124276300163 ()2-s2.0-85159575698 (Scopus ID)
Conference
2023 World Wide Web Conference, WWW 2023, Austin, United States of America, Apr 30 2023 - May 4 2023
Note

Part of ISBN 9781450394161

QC 20230801

Available from: 2023-08-01 Created: 2023-08-01 Last updated: 2024-03-05Bibliographically approved
Jin, Y., Daoutis, M., Girdzijauskas, Š. & Gionis, A. (2023). Learning Cellular Coverage from Real Network Configurations using GNNs. In: 2023 IEEE 97th Vehicular Technology Conference, VTC 2023-Spring - Proceedings: . Paper presented at 97th IEEE Vehicular Technology Conference, VTC 2023-Spring, Florence, Italy, Jun 20 2023 - Jun 23 2023. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Learning Cellular Coverage from Real Network Configurations using GNNs
2023 (English)In: 2023 IEEE 97th Vehicular Technology Conference, VTC 2023-Spring - Proceedings, Institute of Electrical and Electronics Engineers (IEEE) , 2023Conference paper, Published paper (Refereed)
Abstract [en]

Cellular coverage quality estimation has been a critical task for self-organized networks. In real-world scenarios, deep-learning-powered coverage quality estimation methods cannot scale up to large areas due to little ground truth can be provided during network design & optimization. In addition, they fall short in producing expressive embeddings to adequately capture the variations of the cells' configurations. To deal with this challenge, we formulate the task in a graph representation and so that we can apply state-of-the-art graph neural networks, that show exemplary performance. We propose a novel training framework that can both produce quality cell configuration embeddings for estimating multiple KPIs, while we show it is capable of generalising to large (area-wide) scenarios given very few labeled cells. We show that our framework yields comparable accuracy with models that have been trained using massively labeled samples.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2023
Keywords
Cellular Coverage Estimation, Few-shot Learning, Graph Neural Network, Self-supervised Learning
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-336725 (URN)10.1109/VTC2023-Spring57618.2023.10199469 (DOI)2-s2.0-85169786270 (Scopus ID)
Conference
97th IEEE Vehicular Technology Conference, VTC 2023-Spring, Florence, Italy, Jun 20 2023 - Jun 23 2023
Note

Part of ISBN 9798350311143

QC 20230919

Available from: 2023-09-19 Created: 2023-09-19 Last updated: 2025-01-27Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-4516-7317

Search in DiVA

Show all publications