kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Leap: Inductive Link Prediction via Learnable Topology Augmentation
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0000-0002-5392-6531
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0000-0001-7898-0879
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0000-0003-4516-7317
2025 (English)In: Machine Learning, Optimization, and Data Science - 10th International Conference, LOD 2024, Revised Selected Papers, Springer Nature , 2025, p. 448-463Conference paper, Published paper (Refereed)
Abstract [en]

Link prediction is a crucial task in many downstream applications of graph machine learning. To this end, Graph Neural Network (GNN) is a widely used technique for link prediction, mainly in transductive settings, where the goal is to predict missing links between existing nodes. However, many real-life applications require an inductive setting that accommodates for new nodes, coming into an existing graph. Thus, recently inductive link prediction has attracted considerable attention, and a multi-layer perceptron (MLP) is the popular choice of most studies to learn node representations. However, these approaches have limited expressivity and do not fully capture the graph’s structural signal. Therefore, in this work we propose LEAP, an inductive link prediction method based on LEArnable toPology augmentation. Unlike previous methods, LEAP models the inductive bias from both the structure and node features, and hence is more expressive. To the best of our knowledge, this is the first attempt to provide structural contexts for new nodes via learnable augmentation in inductive settings. Extensive experiments on seven real-world homogeneous and heterogeneous graphs demonstrates that LEAP significantly surpasses SOTA methods. The improvements are up to 22% and 17% in terms of AUC and average precision, respectively. The code and datasets are available on GitHub (1https://github.com/AhmedESamy/LEAP/).

Place, publisher, year, edition, pages
Springer Nature , 2025. p. 448-463
Keywords [en]
Graph Neural Networks, Heterogeneous graphs, Inductive link prediction, Learnable augmentation
National Category
Computer Sciences Applied Mechanics
Identifiers
URN: urn:nbn:se:kth:diva-361992DOI: 10.1007/978-3-031-82481-4_31ISI: 001530948600031Scopus ID: 2-s2.0-105000770530OAI: oai:DiVA.org:kth-361992DiVA, id: diva2:1949665
Conference
10th International Conference on Machine Learning, Optimization, and Data Science, LOD 2024, Castiglione della Pescaia, Italy, Sep 22 2024 - Sep 25 2024
Note

Part of ISBN 9783031824807

QC 20250403

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-12-05Bibliographically approved
In thesis
1. Representation Learning on Graphs: Investigating and Overcoming Common Challenges
Open this publication in new window or tab >>Representation Learning on Graphs: Investigating and Overcoming Common Challenges
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Graph Representation Learning (GRL) has emerged as a crucial area for modeling and understanding the structure of graph-structured data across diverse applications. This thesis advances GRL by addressing key challenges in both homogeneous and heterogeneous graphs, including modeling complex heterogeneous relational structures, designing generalizable augmentations for self-supervised learning, improving inductive link prediction in cold-start scenarios, and mitigating over-squashing in message-passing architectures.

Heterogeneous graphs present modeling difficulties due to the presence of multiple node and edge types. To address this, we propose a flexible random walk framework that removes the need for predefined domain knowledge such as meta-paths, enabling more effective and scalable modeling of complex relational structures.

In the self-supervised learning setting, current GRL methods often rely on manually designed graph augmentations that limit generalizability. This thesis introduces augmentation techniques that are task- and domain-agnostic, improving performance across varied graph types and structures.

Inductive link prediction remains challenging for GNNs, particularly in cold-start scenarios where target nodes lack topological context. We propose methods that support efficient and accurate inference without requiring access to neighborhood information of unseen nodes, addressing both scalability and generalization.

While GNNs are effective at capturing local structure, they often suffer from over-squashing, which restricts information propagation across long-range dependencies. To overcome this, we present strategies that improve the aggregation process, enabling GNNs to better preserve and prioritize critical signals from distant parts of the graph.

Through extensive experiments on benchmark datasets, the proposed methods demonstrate consistent improvements in node classification, link prediction, and graph property prediction tasks. Our approaches outperform strong baselines in settings involving heterogeneity, inductive generalization, and large-diameter graphs. Some methods significantly reduce inference cost, while others enhance model expressiveness and robustness by improving structural generalization. Collectively, these contributions show that principled and general-purpose solutions can effectively address long-standing challenges in graph representation learning.

Place, publisher, year, edition, pages
Stockholm: KTH Royal Institute of Technology, 2025. p. ix, 70
Series
TRITA-EECS-AVL ; 2025:92
Keywords
Graph Machine Learning, Representation Learning
National Category
Artificial Intelligence
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-370617 (URN)978-91-8106-426-1 (ISBN)
Public defence
2025-11-07, F3, Lindstedtsvägen 26 & 28, KTH Campus, Stocholm, 09:00 (English)
Opponent
Supervisors
Note

QC 20250929

Available from: 2025-09-29 Created: 2025-09-29 Last updated: 2025-09-29Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Samy, Ahmed E.Kefato, Zekarias TilahunGirdzijauskas, Šarūnas

Search in DiVA

By author/editor
Samy, Ahmed E.Kefato, Zekarias TilahunGirdzijauskas, Šarūnas
By organisation
Software and Computer systems, SCS
Computer SciencesApplied Mechanics

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 45 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf