kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Publications (5 of 5) Show all publications
Cornell, F., Jin, Y., Karlgren, J. & Girdzijauskas, S. (2025). Are We Wasting Time? A Fast, Accurate Performance Evaluation Framework for Knowledge Graph Link Predictors. In: Proceedings - 2025 IEEE 41st International Conference on Data Engineering, ICDE 2025: . Paper presented at 41st IEEE International Conference on Data Engineering, ICDE 2025, Hong Kong, China, May 19-23, 2025 (pp. 1650-1663). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Are We Wasting Time? A Fast, Accurate Performance Evaluation Framework for Knowledge Graph Link Predictors
2025 (English)In: Proceedings - 2025 IEEE 41st International Conference on Data Engineering, ICDE 2025, Institute of Electrical and Electronics Engineers (IEEE) , 2025, p. 1650-1663Conference paper, Published paper (Refereed)
Abstract [en]

The standard evaluation protocol for measuring the quality of Knowledge Graph Completion methods - the task of inferring new links to be added to a graph - typically involves a step which ranks every entity of a Knowledge Graph to assess their fit as a head or tail of a candidate link to be added. In Knowledge Graphs on a larger scale, this task rapidly becomes prohibitively heavy. Previous approaches mitigate this problem by using random sampling of entities to assess the quality of links predicted or suggested by a method. However, we show that this approach has serious limitations since the ranking metrics produced do not properly reflect true outcomes. In this paper, we present a thorough analysis of these effects along with the following findings. First, we empirically find and theoretically motivate why sampling uniformly at random vastly overestimates the ranking performance of a method. We show that this can be attributed to the effect of easy versus hard negatives. Second, we propose a framework that uses relational recommenders to guide the selection of candidates for evaluation. We provide both theoretical and empirical justification of our methodology, and find that simple and fast methods work extremely well, matching advanced neural approaches. Even when a large portion of the true candidates for a property are missed, the estimation of the ranking metrics on a downstream model barely deteriorates. With our proposed framework, we can reduce the time and computation needed similar to random sampling strategies while vastly improving the estimation; on ogbl-wikikg2, we show that accurate estimations of the full ranking can be obtained in 20 seconds instead of 30 minutes. We conclude that considerable computational effort can be saved by effective preprocessing and sampling methods and still reliably predict performance accurately of the true performance for the entire ranking procedure. We make our code available to the community1. 1Accessible at https://github.com/Filco306/are-we-wasting-time.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
evaluation, Knowledge Graph, link prediction, sampling
National Category
Probability Theory and Statistics Computer Sciences
Identifiers
urn:nbn:se:kth:diva-370773 (URN)10.1109/ICDE65448.2025.00127 (DOI)2-s2.0-105015357159 (Scopus ID)
Conference
41st IEEE International Conference on Data Engineering, ICDE 2025, Hong Kong, China, May 19-23, 2025
Note

Part of ISBN 9798331536039

QC 20251001

Available from: 2025-10-01 Created: 2025-10-01 Last updated: 2025-10-01Bibliographically approved
Ennadir, S., Gandler, G. Z., Cornell, F., Cao, L., Smirnov, O., Wang, T., . . . Asadi, S. (2025). Expressivity of Representation Learning on Continuous-Time Dynamic Graphs: An Information-Flow Centric Review. Transactions on Machine Learning Research, 2025-April
Open this publication in new window or tab >>Expressivity of Representation Learning on Continuous-Time Dynamic Graphs: An Information-Flow Centric Review
Show others...
2025 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856, Vol. 2025-AprilArticle in journal (Refereed) Published
Abstract [en]

Graphs are ubiquitous in real-world applications, ranging from social networks to biological systems, and have inspired the development of Graph Neural Networks (GNNs) for learning expressive representations. While most research has centered on static graphs, many real-world scenarios involve dynamic, temporally evolving graphs, motivating the need for Continuous-Time Dynamic Graph (CTDG) models. This paper provides a comprehensive review of Graph Representation Learning (GRL) on CTDGs with a focus on Self-Supervised Representation Learning (SSRL). We introduce a novel theoretical framework that analyzes the expressivity of CTDG models through an Information-Flow (IF) lens, quantifying their ability to propagate and encode temporal and structural information. Leveraging this frame-work, we categorize existing CTDG methods based on their suitability for different graph types and application scenarios. Within the same scope, we examine the design of SSRL methods tailored to CTDGs, such as predictive and contrastive approaches, highlighting their potential to mitigate the reliance on labeled data. Empirical evaluations on synthetic and real-world datasets validate our theoretical insights, demonstrating the strengths and limitations of various methods across long-range, bi-partite and community-based graphs. This work offers both a theoretical foundation and practical guidance for selecting and developing CTDG models, advancing the understanding of GRL in dynamic settings.

Place, publisher, year, edition, pages
Transactions on Machine Learning Research, 2025
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-363405 (URN)2-s2.0-105004389835 (Scopus ID)
Note

QC 20250519

Available from: 2025-05-15 Created: 2025-05-15 Last updated: 2025-05-19Bibliographically approved
Cornell, F., Jin, Y., Karlgren, J. & Girdzijauskas, S. (2025). Unsupervised Ontology- and Taxonomy Construction Through Hyperbolic Relational Domains and Ranges. In: Machine Learning and Principles and Practice of Knowledge Discovery in Databases - International Workshops of ECML PKDD 2023, Revised Selected Papers: . Paper presented at Joint European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML PKDD 2023, Turin, Italy, September 18-22, 2023 (pp. 339-348). Springer Nature
Open this publication in new window or tab >>Unsupervised Ontology- and Taxonomy Construction Through Hyperbolic Relational Domains and Ranges
2025 (English)In: Machine Learning and Principles and Practice of Knowledge Discovery in Databases - International Workshops of ECML PKDD 2023, Revised Selected Papers, Springer Nature , 2025, p. 339-348Conference paper, Published paper (Refereed)
Abstract [en]

The sets of possible heads and tails for relations in a Knowledge Graph and a type taxonomy of participating entities are important aspects of a Knowledge Graph and constitute a large portion of a Knowledge Graph’s ontology. Making the ontology explicit helps to ensure consistency of the knowledge represented in the graph and allows for less costly maintenance and update of graph content. However, Knowledge Graphs are often conveniently formed without an explicitly described ontology, using only (head, relation, tail)-tuples. For such graphs without predefined structure, we propose learning an ontology and a hierarchical type taxonomy from the graph itself, taking advantage of the reciprocity of entity types and sets of possible heads (relational domains) and tails (relational ranges). Experiments on real-world datasets validate our approach and demonstrate the promise for leveraging machine learning methodologies to efficiently generate taxonomies and ontologies jointly for Knowledge Graphs (Our implementation can be found at https://tinyurl.com/5n6sw6yj).

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Hyperbolic learning, Knowledge Graphs, Ontology Learning, Taxonomy Learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-359658 (URN)10.1007/978-3-031-74633-8_23 (DOI)001437448200023 ()2-s2.0-85216103400 (Scopus ID)
Conference
Joint European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML PKDD 2023, Turin, Italy, September 18-22, 2023
Note

Part of ISBN 9783031746321

QC 20250207

Available from: 2025-02-06 Created: 2025-02-06 Last updated: 2025-12-05Bibliographically approved
Cornell, F., Zhang, C., Girdzijauskas, S. & Karlgren, J. (2022). Challenging the Assumption of Structure-based embeddings in Few- and Zero-shot Knowledge Graph Completion. In: Calzolari, N Bechet, F Blache, P Choukri, K Cieri, C Declerck, T Goggi, S Isahara, H Maegaard, B Mazo, H Odijk, H Piperidis, S (Ed.), LREC 2022: Thirteen International Conference On Language Resources And Evaluation. Paper presented at 13th International Conference on Language Resources and Evaluation (LREC), JUN 20-25, 2022, Marseille, FRANCE (pp. 6300-6309). European Language Resources Association (ELRA)
Open this publication in new window or tab >>Challenging the Assumption of Structure-based embeddings in Few- and Zero-shot Knowledge Graph Completion
2022 (English)In: LREC 2022: Thirteen International Conference On Language Resources And Evaluation / [ed] Calzolari, N Bechet, F Blache, P Choukri, K Cieri, C Declerck, T Goggi, S Isahara, H Maegaard, B Mazo, H Odijk, H Piperidis, S, European Language Resources Association (ELRA) , 2022, p. 6300-6309Conference paper, Published paper (Refereed)
Abstract [en]

In this paper, we report experiments on Few- and Zero-shot Knowledge Graph completion, where the objective is to add missing relational links between entities into an existing Knowledge Graph with few or no previous examples of the relation in question. While previous work has used pre-trained embeddings based on the structure of the graph as input for a neural network, nobody has, to the best of our knowledge, addressed the task by only using textual descriptive data associated with the entities and relations, much since current standard benchmark data sets lack such information. We therefore enrich the benchmark data sets for these tasks by collecting textual description data to provide a new resource for future research to bridge the gap between structural and textual Knowledge Graph completion. Our results show that we can improve the results for Knowledge Graph completion for both Few- and Zero-shot scenarios with up to a two-fold increase of all metrics in the Zero-shot setting. From a more general perspective, our experiments demonstrate the value of using textual resources to enrich more formal representations of human knowledge and in the utility of transfer learning from textual data and text collections to enrich and maintain knowledge resources.

Place, publisher, year, edition, pages
European Language Resources Association (ELRA), 2022
Keywords
Knowledge Graph completion, Meta-learning, Zero-shot learning, textual enrichment
National Category
Natural Language Processing
Identifiers
urn:nbn:se:kth:diva-324341 (URN)000889371706045 ()2-s2.0-85144370104 (Scopus ID)
Conference
13th International Conference on Language Resources and Evaluation (LREC), JUN 20-25, 2022, Marseille, FRANCE
Note

QC 20230228

Available from: 2023-02-28 Created: 2023-02-28 Last updated: 2025-02-07Bibliographically approved
Cornell, F., Karlgren, J., Sachan, A. & Girdzijauskas, S. (2022). Symbolic Hyperdimensional Vectors with Sparse Graph Convolutional Neural Networks. In: 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN): . Paper presented at IEEE International Conference on Fuzzy Systems (FUZZ-IEEE) / IEEE World Congress on Computational Intelligence (IEEE WCCI) / International Joint Conference on Neural Networks (IJCNN) / IEEE Congress on Evolutionary Computation (IEEE CEC), JUL 18-23, 2022, Padua, ITALY. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Symbolic Hyperdimensional Vectors with Sparse Graph Convolutional Neural Networks
2022 (English)In: 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), Institute of Electrical and Electronics Engineers (IEEE) , 2022Conference paper, Published paper (Refereed)
Abstract [en]

In this paper, we propose a novel way of representing graphs for processing in Graph Neural Networks. We reduce the dimensionality of the input data by using Random Indexing, a Vector Symbolic Architectural framework; we implement a new trainable neural layer, also inspired by Vector Symbolic Architectures; we leverage the sparseness of the incoming data in a Sparse Neural Network framework. Our experiments on a number of publicly available datasets and standard benchmarks demonstrate that we can reduce the number of parameters by up to two orders of magnitude. We show how this parsimonious approach not only delivers competitive results but even improves performance for node classification and link prediction. We find that this holds in particular for cases where the graph lacks node features.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2022
Series
IEEE International Joint Conference on Neural Networks (IJCNN), ISSN 2161-4393
Keywords
vector symbolic architectures, graph neural networks, random indexing
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-323028 (URN)10.1109/IJCNN55064.2022.9892300 (DOI)000867070903060 ()2-s2.0-85140763914 (Scopus ID)
Conference
IEEE International Conference on Fuzzy Systems (FUZZ-IEEE) / IEEE World Congress on Computational Intelligence (IEEE WCCI) / International Joint Conference on Neural Networks (IJCNN) / IEEE Congress on Evolutionary Computation (IEEE CEC), JUL 18-23, 2022, Padua, ITALY
Note

Part of proceedings: ISBN 978-1-7281-8671-9

QC 20230112

Available from: 2023-01-12 Created: 2023-01-12 Last updated: 2023-12-11Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0002-8346-610X

Search in DiVA

Show all publications