kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
HopNet: Addressing Over-Squashing with Learnable Rewiring in GNNs
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS. (GALE)ORCID iD: 0000-0002-5392-6531
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS. Research Institute of Sweden.ORCID iD: 0000-0002-0223-8907
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0000-0003-4516-7317
2025 (English)Manuscript (preprint) (Other academic)
Abstract [en]

Graph Neural Networks (GNNs) have emerged as powerful tools for extracting insights from graph-structured data. However, GNNs often encounter challenges when aggregating information across distant and constrained connections, leading to performance degradation due to a phenomenon known as over-squashing. Previous solutions, such as static rewiring, address this issue by increasing graph density but compromise the inherent inductive bias of GNNs with respect to node distances. In this study, we introduce HopNet, a novel model designed to facilitate effective long-range information propagation through learnable rewiring. HopNet employs attention mechanisms to dynamically create targeted shortcuts, enabling efficient communication between distant nodes while maintaining a balance between local and global interactions. Extensive experiments across diverse tasks, including node classification, link prediction, graph regression, and graph classification, on established real-world benchmark datasets demonstrate HopNet’s ability to overcome GNN limitations, consistently achieving superior performance over state-of-the-art methods.

Place, publisher, year, edition, pages
2025.
Keywords [en]
Graph Neural Networks, Bottlenecks, Over-Squashing, Learnable Rewiring.
National Category
Artificial Intelligence
Identifiers
URN: urn:nbn:se:kth:diva-370615OAI: oai:DiVA.org:kth-370615DiVA, id: diva2:2001843
Note

Submitted to ACM WebConference 2026

QC 20250929

Available from: 2025-09-29 Created: 2025-09-29 Last updated: 2025-10-21Bibliographically approved
In thesis
1. Representation Learning on Graphs: Investigating and Overcoming Common Challenges
Open this publication in new window or tab >>Representation Learning on Graphs: Investigating and Overcoming Common Challenges
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Graph Representation Learning (GRL) has emerged as a crucial area for modeling and understanding the structure of graph-structured data across diverse applications. This thesis advances GRL by addressing key challenges in both homogeneous and heterogeneous graphs, including modeling complex heterogeneous relational structures, designing generalizable augmentations for self-supervised learning, improving inductive link prediction in cold-start scenarios, and mitigating over-squashing in message-passing architectures.

Heterogeneous graphs present modeling difficulties due to the presence of multiple node and edge types. To address this, we propose a flexible random walk framework that removes the need for predefined domain knowledge such as meta-paths, enabling more effective and scalable modeling of complex relational structures.

In the self-supervised learning setting, current GRL methods often rely on manually designed graph augmentations that limit generalizability. This thesis introduces augmentation techniques that are task- and domain-agnostic, improving performance across varied graph types and structures.

Inductive link prediction remains challenging for GNNs, particularly in cold-start scenarios where target nodes lack topological context. We propose methods that support efficient and accurate inference without requiring access to neighborhood information of unseen nodes, addressing both scalability and generalization.

While GNNs are effective at capturing local structure, they often suffer from over-squashing, which restricts information propagation across long-range dependencies. To overcome this, we present strategies that improve the aggregation process, enabling GNNs to better preserve and prioritize critical signals from distant parts of the graph.

Through extensive experiments on benchmark datasets, the proposed methods demonstrate consistent improvements in node classification, link prediction, and graph property prediction tasks. Our approaches outperform strong baselines in settings involving heterogeneity, inductive generalization, and large-diameter graphs. Some methods significantly reduce inference cost, while others enhance model expressiveness and robustness by improving structural generalization. Collectively, these contributions show that principled and general-purpose solutions can effectively address long-standing challenges in graph representation learning.

Place, publisher, year, edition, pages
Stockholm: KTH Royal Institute of Technology, 2025. p. ix, 70
Series
TRITA-EECS-AVL ; 2025:92
Keywords
Graph Machine Learning, Representation Learning
National Category
Artificial Intelligence
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-370617 (URN)978-91-8106-426-1 (ISBN)
Public defence
2025-11-07, F3, Lindstedtsvägen 26 & 28, KTH Campus, Stocholm, 09:00 (English)
Opponent
Supervisors
Note

QC 20250929

Available from: 2025-09-29 Created: 2025-09-29 Last updated: 2025-09-29Bibliographically approved

Open Access in DiVA

Paper(767 kB)90 downloads
File information
File name FULLTEXT02.pdfFile size 767 kBChecksum SHA-512
9e8c430593ce6c2e34879b17a005192c365aad0e462b599b223b6548630c7df7685fba1755752f39a319e6df26f0170890d0b7870aeb37922ec01bd4f2209f6c
Type fulltextMimetype application/pdf

Authority records

Samy, Ahmed E.Giaretta, LodovicoGirdzijauskas, Sarunas

Search in DiVA

By author/editor
Samy, Ahmed E.Giaretta, LodovicoGirdzijauskas, Sarunas
By organisation
Software and Computer systems, SCS
Artificial Intelligence

Search outside of DiVA

GoogleGoogle Scholar
Total: 114 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 534 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf