kth.sePublications
Change search
Link to record
Permanent link

Direct link
Kefato, Zekarias TilahunORCID iD iconorcid.org/0000-0001-7898-0879
Publications (10 of 12) Show all publications
Samy, A. E., Kefato, Z. T. & Girdzijauskas, Š. (2025). Leap: Inductive Link Prediction via Learnable Topology Augmentation. In: Machine Learning, Optimization, and Data Science - 10th International Conference, LOD 2024, Revised Selected Papers: . Paper presented at 10th International Conference on Machine Learning, Optimization, and Data Science, LOD 2024, Castiglione della Pescaia, Italy, Sep 22 2024 - Sep 25 2024 (pp. 448-463). Springer Nature
Open this publication in new window or tab >>Leap: Inductive Link Prediction via Learnable Topology Augmentation
2025 (English)In: Machine Learning, Optimization, and Data Science - 10th International Conference, LOD 2024, Revised Selected Papers, Springer Nature , 2025, p. 448-463Conference paper, Published paper (Refereed)
Abstract [en]

Link prediction is a crucial task in many downstream applications of graph machine learning. To this end, Graph Neural Network (GNN) is a widely used technique for link prediction, mainly in transductive settings, where the goal is to predict missing links between existing nodes. However, many real-life applications require an inductive setting that accommodates for new nodes, coming into an existing graph. Thus, recently inductive link prediction has attracted considerable attention, and a multi-layer perceptron (MLP) is the popular choice of most studies to learn node representations. However, these approaches have limited expressivity and do not fully capture the graph’s structural signal. Therefore, in this work we propose LEAP, an inductive link prediction method based on LEArnable toPology augmentation. Unlike previous methods, LEAP models the inductive bias from both the structure and node features, and hence is more expressive. To the best of our knowledge, this is the first attempt to provide structural contexts for new nodes via learnable augmentation in inductive settings. Extensive experiments on seven real-world homogeneous and heterogeneous graphs demonstrates that LEAP significantly surpasses SOTA methods. The improvements are up to 22% and 17% in terms of AUC and average precision, respectively. The code and datasets are available on GitHub (1https://github.com/AhmedESamy/LEAP/).

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Graph Neural Networks, Heterogeneous graphs, Inductive link prediction, Learnable augmentation
National Category
Computer Sciences Applied Mechanics
Identifiers
urn:nbn:se:kth:diva-361992 (URN)10.1007/978-3-031-82481-4_31 (DOI)2-s2.0-105000770530 (Scopus ID)
Conference
10th International Conference on Machine Learning, Optimization, and Data Science, LOD 2024, Castiglione della Pescaia, Italy, Sep 22 2024 - Sep 25 2024
Note

Part of ISBN 9783031824807

QC 20250403

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-03Bibliographically approved
Samy, A. E., Kefato, Z. T. & Girdzijauskas, S. (2023). Data-Driven Self-Supervised Graph Representation Learning. In: ECAI 2023: 26th European Conference on Artificial Intelligence, including 12th Conference on Prestigious Applications of Intelligent Systems, PAIS 2023 - Proceedings. Paper presented at 26th European Conference on Artificial Intelligence, ECAI 2023, Krakow, Poland, Sep 30 2023 - Oct 4 2023 (pp. 629-636). IOS Press
Open this publication in new window or tab >>Data-Driven Self-Supervised Graph Representation Learning
2023 (English)In: ECAI 2023: 26th European Conference on Artificial Intelligence, including 12th Conference on Prestigious Applications of Intelligent Systems, PAIS 2023 - Proceedings, IOS Press , 2023, p. 629-636Conference paper, Published paper (Refereed)
Abstract [en]

Self-supervised graph representation learning (SSGRL) is a representation learning paradigm used to reduce or avoid manual labeling. An essential part of SSGRL is graph data augmentation. Existing methods usually rely on heuristics commonly identified through trial and error and are effective only within some application domains. Also, it is not clear why one heuristic is better than another. Moreover, recent studies have argued against some techniques (e.g., dropout: that can change the properties of molecular graphs or destroy relevant signals for graph-based document classification tasks). In this study, we propose a novel data-driven SSGRL approach that automatically learns a suitable graph augmentation from the signal encoded in the graph (i.e., the nodes' predictive feature and topological information). We propose two complementary approaches that produce learnable feature and topological augmentations. The former learns multi-view augmentation of node features, and the latter learns a high-order view of the topology. Moreover, the augmentations are jointly learned with the representation. Our approach is general that it can be applied to homogeneous and heterogeneous graphs. We perform extensive experiments on node classification (using nine homogeneous and heterogeneous datasets) and graph property prediction (using another eight datasets). The results show that the proposed method matches or outperforms the SOTA SSGRL baselines and performs similarly to semi-supervised methods. The anonymised source code is available at https://github.com/AhmedESamy/dsgrl/

Place, publisher, year, edition, pages
IOS Press, 2023
National Category
Computer Sciences Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-339683 (URN)10.3233/FAIA230325 (DOI)2-s2.0-85175858097 (Scopus ID)
Conference
26th European Conference on Artificial Intelligence, ECAI 2023, Krakow, Poland, Sep 30 2023 - Oct 4 2023
Note

Part of ISBN 9781643684369

QC 20231116

Available from: 2023-11-16 Created: 2023-11-16 Last updated: 2025-04-03Bibliographically approved
Bonvalet, M., Kefato, Z. T. & Girdzijauskas, S. (2023). Graph2Feat: Inductive Link Prediction via Knowledge Distillation. In: ACM Web Conference 2023: Companion of the World Wide Web Conference, WWW 2023. Paper presented at 2023 World Wide Web Conference, WWW 2023, Austin, United States of America, Apr 30 2023 - May 4 2023 (pp. 805-812). Association for Computing Machinery (ACM)
Open this publication in new window or tab >>Graph2Feat: Inductive Link Prediction via Knowledge Distillation
2023 (English)In: ACM Web Conference 2023: Companion of the World Wide Web Conference, WWW 2023, Association for Computing Machinery (ACM) , 2023, p. 805-812Conference paper, Published paper (Refereed)
Abstract [en]

Link prediction between two nodes is a critical task in graph machine learning. Most approaches are based on variants of graph neural networks (GNNs) that focus on transductive link prediction and have high inference latency. However, many real-world applications require fast inference over new nodes in inductive settings where no information on connectivity is available for these nodes. Thereby, node features provide an inevitable alternative in the latter scenario. To that end, we propose Graph2Feat, which enables inductive link prediction by exploiting knowledge distillation (KD) through the Student-Teacher learning framework. In particular, Graph2Feat learns to match the representations of a lightweight student multi-layer perceptron (MLP) with a more expressive teacher GNN while learning to predict missing links based on the node features, thus attaining both GNN's expressiveness and MLP's fast inference. Furthermore, our approach is general; it is suitable for transductive and inductive link predictions on different types of graphs regardless of them being homogeneous or heterogeneous, directed or undirected. We carry out extensive experiments on seven real-world datasets including homogeneous and heterogeneous graphs. Our experiments demonstrate that Graph2Feat significantly outperforms SOTA methods in terms of AUC and average precision in homogeneous and heterogeneous graphs. Finally, Graph2Feat has the minimum inference time compared to the SOTA methods, and 100x acceleration compared to GNNs. The code and datasets are available on GitHub1.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2023
Keywords
graph representation learning, heterogeneous networks, inductive link prediction, knowledge distillation
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-333310 (URN)10.1145/3543873.3587596 (DOI)001124276300163 ()2-s2.0-85159575698 (Scopus ID)
Conference
2023 World Wide Web Conference, WWW 2023, Austin, United States of America, Apr 30 2023 - May 4 2023
Note

Part of ISBN 9781450394161

QC 20230801

Available from: 2023-08-01 Created: 2023-08-01 Last updated: 2024-03-05Bibliographically approved
Stefanoni, A., Girdzijauskas, S., Jenkins, C., Kefato, Z. T., Sbattella, L., Scotti, V. & Wåreus, E. (2022). Detecting Security Patches in Java Projects Using NLP Technology. In: ICNLSP 2022: Proceedings of the 5th International Conference on Natural Language and Speech Processing. Paper presented at 5th International Conference on Natural Language and Speech Processing, ICNLSP 2022, Virtual, Online, Dec 16 2022 - Dec 17 2022 (pp. 50-56). Association for Computational Linguistics (ACL)
Open this publication in new window or tab >>Detecting Security Patches in Java Projects Using NLP Technology
Show others...
2022 (English)In: ICNLSP 2022: Proceedings of the 5th International Conference on Natural Language and Speech Processing, Association for Computational Linguistics (ACL) , 2022, p. 50-56Conference paper, Published paper (Refereed)
Abstract [en]

Known vulnerabilities in software are solved through security patches; thus, applying such patches as soon as they are released is crucial to protect from cyber-attacks. The diffusion of open source software allowed to inspect the patches to understand whether they are security related or not. In this paper, we propose some solutions based on state-of-the-art deep learning technologies for Natural Language Processing for security patches detection. In the experiments, we benchmarked our solutions on two data sets for Java security patches detection. Our models showed promising results, outperforming all the others we used for comparison. Interestingly, we achieved better results training the classifiers from scratch than fine tuning existing models.

Place, publisher, year, edition, pages
Association for Computational Linguistics (ACL), 2022
National Category
Computer Sciences Reliability and Maintenance
Identifiers
urn:nbn:se:kth:diva-333357 (URN)2-s2.0-85152141641 (Scopus ID)
Conference
5th International Conference on Natural Language and Speech Processing, ICNLSP 2022, Virtual, Online, Dec 16 2022 - Dec 17 2022
Note

Part of ISBN 9781959429364

QC 20230801

Available from: 2023-08-01 Created: 2023-08-01 Last updated: 2023-08-01Bibliographically approved
Zamboni, S., Kefato, Z. T., Girdzijauskas, S., Norén, C. & Dal Col, L. (2022). Pedestrian trajectory prediction with convolutional neural networks. Pattern Recognition, 121, Article ID 108252.
Open this publication in new window or tab >>Pedestrian trajectory prediction with convolutional neural networks
Show others...
2022 (English)In: Pattern Recognition, ISSN 0031-3203, E-ISSN 1873-5142, Vol. 121, article id 108252Article in journal (Refereed) Published
Abstract [en]

Predicting the future trajectories of pedestrians is a challenging problem that has a range of application, from crowd surveillance to autonomous driving. In literature, methods to approach pedestrian trajectory prediction have evolved, transitioning from physics-based models to data-driven models based on recurrent neural networks. In this work, we propose a new approach to pedestrian trajectory prediction, with the introduction of a novel 2D convolutional model. This new model outperforms recurrent models, and it achieves state-of-the-art results on the ETH and TrajNet datasets. We also present an effective system to represent pedestrian positions and powerful data augmentation techniques, such as the addition of Gaussian noise and the use of random rotations, which can be applied to any model. As an additional exploratory analysis, we present experimental results on the inclusion of occupancy methods to model social information, which empirically show that these methods are ineffective in capturing social interaction.

Place, publisher, year, edition, pages
Elsevier BV, 2022
Keywords
Convolutional neural networks, Pedestrian prediction, Trajectory prediction, Convolution, Gaussian noise (electronic), Recurrent neural networks, Trajectories, Autonomous driving, Convolutional neural network, Crowd surveillance, Data driven modelling, Model-based OPC, Neural-networks, Pedestrian trajectories, Physics-based modeling, Forecasting
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-311167 (URN)10.1016/j.patcog.2021.108252 (DOI)000697551500006 ()2-s2.0-85113188585 (Scopus ID)
Note

QC 20220517

Available from: 2022-05-17 Created: 2022-05-17 Last updated: 2025-02-07Bibliographically approved
Samy, A. E., Giaretta, L., Kefato, Z. T. & Girdzijauskas, S. (2022). SchemaWalk: Schema Aware Random Walks for Heterogeneous Graph Embedding. In: WWW 2022 - Companion Proceedings of the Web Conference 2022: . Paper presented at 31st ACM Web Conference, WWW 2022, 25 April 2022 (pp. 1157-1166). Association for Computing Machinery (ACM)
Open this publication in new window or tab >>SchemaWalk: Schema Aware Random Walks for Heterogeneous Graph Embedding
2022 (English)In: WWW 2022 - Companion Proceedings of the Web Conference 2022, Association for Computing Machinery (ACM) , 2022, p. 1157-1166Conference paper, Published paper (Refereed)
Abstract [en]

Heterogeneous Information Network (HIN) embedding has been a prevalent approach to learn representations off semantically-rich heterogeneous networks. Most HIN embedding methods exploit meta-paths to retain high-order structures, yet, their performance is conditioned on the quality of the (generated/manually-defined) meta-paths and their suitability for the specific label set. Whereas other methods adjust random walks to harness or skip certain heterogeneous structures (e.g. node type(s)), in doing so, the adjusted random walker may casually omit other node/edge types. Our key insight is with no domain knowledge, the random walker should hold no assumptions about heterogeneous structure (i.e. edge types). Thus, aiming for a flexible and general method, we utilize network schema as a unique blueprint of HIN, and propose SchemaWalk, a random walk to uniformly sample all edge types within the network schema. Moreover, we identify the starvation phenomenon which induces random walkers on HINs to under- or over-sample certain edge types. Accordingly, we design SchemaWalkHO to skip local deficient connectivity to preserve uniform sampling distribution. Finally, we carry out node classification experiments on four real-world HINs, and provide in-depth qualitative analysis. The results highlight the robustness of our method regardless to the graph structure in contrast with the state-of-the-art baselines. 

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2022
Keywords
Heterogeneous Information Network, Network Embeddings, Random Walk, Representation Learning, Domain Knowledge, Graphic methods, Information services, Random processes, Graph embeddings, Heterogeneous graph, Heterogeneous information, Heterogeneous structures, Information networks, Network embedding, Random walkers, Heterogeneous networks
National Category
Computer Systems
Identifiers
urn:nbn:se:kth:diva-327049 (URN)10.1145/3487553.3524728 (DOI)001147592700198 ()2-s2.0-85137448476 (Scopus ID)
Conference
31st ACM Web Conference, WWW 2022, 25 April 2022
Note

QC 20230523

Available from: 2023-05-23 Created: 2023-05-23 Last updated: 2025-04-03Bibliographically approved
Kefato, Z. T., Girdzijauskas, S., Sheikh, N. & Montresor, A. (2021). Dynamic embeddings for interaction prediction. In: The Web Conference 2021 - Proceedings of the World Wide Web Conference, WWW 2021: . Paper presented at 2021 World Wide Web Conference, WWW 2021, 19 April 2021 through 23 April 2021, Ljubljana Slovenia (pp. 1609-1618). Association for Computing Machinery, Inc
Open this publication in new window or tab >>Dynamic embeddings for interaction prediction
2021 (English)In: The Web Conference 2021 - Proceedings of the World Wide Web Conference, WWW 2021, Association for Computing Machinery, Inc , 2021, p. 1609-1618Conference paper, Published paper (Refereed)
Abstract [en]

In recommender systems (RSs), predicting the next item that a user interacts with is critical for user retention. While the last decade has seen an explosion of RSs aimed at identifying relevant items that match user preferences, there is still a range of aspects that could be considered to further improve their performance. For example, often RSs are centered around the user, who is modeled using her recent sequence of activities. Recent studies, however, have shown the effectiveness of modeling the mutual interactions between users and items using separate user and item embeddings. Building on the success of these studies, we propose a novel method called DeePRed that addresses some of their limitations. In particular, we avoid recursive and costly interactions between consecutive short-term embeddings by using long-term (stationary) embeddings as a proxy. This enable us to train DeePRed using simple mini-batches without the overhead of specialized mini-batches proposed in previous studies. Moreover, DeePRed's effectiveness comes from the aforementioned design and a multi-way attention mechanism that inspects user-item compatibility. Experiments show that DeePRed outperforms the best state-of-the-art approach by at least 14% of Mean Reciprocal Rank (MRR) on next item prediction task, while gaining more than an order of magnitude speedup over the best performing baselines. Although this study is mainly concerned with temporal interaction networks, we also show the power and flexibility of DeePRed by adapting it to the case of static interaction networks, substituting the short- and long-term aspects with local and global ones. 

Place, publisher, year, edition, pages
Association for Computing Machinery, Inc, 2021
Keywords
Dynamic embeddings, Interaction prediction, Multi-way attention, Mutual RNN, Recommender systems, Forecasting, World Wide Web, Attention mechanisms, Best state, Interaction networks, Mean reciprocal ranks, Mutual interaction, Prediction tasks, Static interaction, Embeddings
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-309946 (URN)10.1145/3442381.3450020 (DOI)000733621801054 ()2-s2.0-85107936313 (Scopus ID)
Conference
2021 World Wide Web Conference, WWW 2021, 19 April 2021 through 23 April 2021, Ljubljana Slovenia
Note

Part of proceedings: ISBN 978-1-4503-8312-7

QC 20220517

Available from: 2022-03-21 Created: 2022-03-21 Last updated: 2023-01-18Bibliographically approved
Kefato, Z. T. & Girdzijauskas, S. (2021). Self-supervised Graph Neural Networks without explicit negative sampling. In: : . Paper presented at The International Workshop on Self-Supervised Learning for the Web (SSL'21), at WWW'21.
Open this publication in new window or tab >>Self-supervised Graph Neural Networks without explicit negative sampling
2021 (English)Conference paper, Oral presentation only (Refereed)
Abstract [en]

Real world data is mostly unlabeled or only few instances are labeled. Manually labeling data is a very expensive and daunting task. This calls for unsupervised learning techniques that are powerful enough to achieve comparable results as semi-supervised/supervised techniques. Contrastive self-supervised learning has emerged as a powerful direction, in some cases outperforming supervised techniques. In this study, we propose, SelfGNN, a novel contrastive self-supervised graph neural network (GNN) without relying on explicit contrastive terms. We leverage Batch Normalization, which introduces implicit contrastive terms, without sacrificing performance. Furthermore, as data augmentation is key in contrastive learning, we introduce four feature augmentation (FA) techniques for graphs. Though graph topological augmentation (TA) is commonly used, our empirical findings show that FA perform as good as TA. Moreover, FA incurs no computational overhead, unlike TA, which often has O(N^3) time complexity, N-number of nodes. Our empirical evaluation on seven publicly available real-world data shows that, SelfGNN is powerful and leads to a performance comparable with SOTA supervised GNNs and always better than SOTA semi-supervised and unsupervised GNNs.

National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-296172 (URN)
Conference
The International Workshop on Self-Supervised Learning for the Web (SSL'21), at WWW'21
Note

QCR 20210802

Available from: 2021-05-31 Created: 2021-05-31 Last updated: 2022-06-25Bibliographically approved
Sheikh, N., Kefato, Z. T. & Montresor, A. (2020). A Simple Approach to Attributed Graph Embedding via Enhanced Autoencoder. In: Complex Networks and Their Applications VIII: Volume 1 Proceedings of the Eighth International Conference on Complex Networks and Their Applications COMPLEX NETWORKS 2019. Paper presented at The Eighth International Conference on Complex Networks and Their Applications COMPLEX NETWORKS 2019, Lisbon, Portugal, December 10-12, 2019. (pp. 797-809). Springer, 881
Open this publication in new window or tab >>A Simple Approach to Attributed Graph Embedding via Enhanced Autoencoder
2020 (English)In: Complex Networks and Their Applications VIII: Volume 1 Proceedings of the Eighth International Conference on Complex Networks and Their Applications COMPLEX NETWORKS 2019, Springer , 2020, Vol. 881, p. 797-809Conference paper, Published paper (Refereed)
Abstract [en]

Network Representation Learning (NRL) aims at learning a low-dimensional representation of nodes in a graph such that its properties are preserved in the learned embedding. NRL methods may exploit different sources of information such as the structural or attribute information of the graph. Recent efforts have shown that jointly using both structure and attributes helps in learning a better representation. Most of these methods rely on highly complex procedures, such as sampling, which makes them non-scalable to large graphs. In this paper, we propose a simple and scalable deep neural network model that learns an embedding by jointly incorporating the network structure and the attribute information. Specifically, the model employs an enhanced decoder that preserves global network structure and also handles the non-linearities of both the network structure and network attributes. We discuss node classification, link prediction, and network reconstruction experiments on four real-world datasets, demonstrating that our approach achieves better performance against the state-of-the-art baselines.

Place, publisher, year, edition, pages
Springer, 2020
Keywords
Attributed graphs, Network embedding, Unsupervised Learning
National Category
Astronomy, Astrophysics and Cosmology
Identifiers
urn:nbn:se:kth:diva-272307 (URN)10.1007/978-3-030-36687-2_66 (DOI)000843927300066 ()2-s2.0-85076693572 (Scopus ID)
Conference
The Eighth International Conference on Complex Networks and Their Applications COMPLEX NETWORKS 2019, Lisbon, Portugal, December 10-12, 2019.
Note

QC 20200513

Available from: 2020-05-13 Created: 2020-05-13 Last updated: 2023-03-27Bibliographically approved
Kefato, Z. T. & Girdzijauskas, S. (2020). Gossip and Attend: Context-Sensitive Graph Representation Learning. In: In Proc. of the 14-th International AAAI Conference on Web and Social Media, ICWSM'20: . Paper presented at 14-th International Conference on Web and Social Media, ICWSM'20.
Open this publication in new window or tab >>Gossip and Attend: Context-Sensitive Graph Representation Learning
2020 (English)In: In Proc. of the 14-th International AAAI Conference on Web and Social Media, ICWSM'20, 2020Conference paper, Published paper (Refereed)
Abstract [en]

Graph representation learning (GRL) is a powerful techniquefor learning low-dimensional vector representation of high-dimensional and often sparse graphs. Most studies explore thestructure and metadata associated with the graph using ran-dom walks and employ an unsupervised or semi-supervisedlearning schemes. Learning in these methods is context-free,resulting in only a single representation per node. Recentlystudies have argued on the adequacy of a single representationand proposed context-sensitive approaches, which are capa-ble of extracting multiple node representations for differentcontexts. This proved to be highly effective in applicationssuch as link prediction and ranking.However, most of these methods rely on additional textualfeatures that require complex and expensive RNNs or CNNsto capture high-level features or rely on a community detec-tion algorithm to identify multiple contexts of a node.In this study we show that in-order to extract high-qualitycontext-sensitive node representations it is not needed to relyon supplementary node features, nor to employ computa-tionally heavy and complex models. We propose GOAT,acontext-sensitive algorithm inspired by gossip communica-tion and a mutual attention mechanism simply over the struc-ture of the graph. We show the efficacy of GOATusing 6 real-world datasets on link prediction and node clustering tasksand compare it against 12 popular and state-of-the-art (SOTA)baselines. GOATconsistently outperforms them and achievesup to 12% and 19% gain over the best performing methodson link prediction and clustering tasks, respectively.

National Category
Other Computer and Information Science Computer Sciences
Identifiers
urn:nbn:se:kth:diva-282661 (URN)2-s2.0-85098821723 (Scopus ID)
Conference
14-th International Conference on Web and Social Media, ICWSM'20
Note

QC 20200930

Available from: 2020-09-30 Created: 2020-09-30 Last updated: 2022-06-25Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0001-7898-0879

Search in DiVA

Show all publications