kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Learning Node Representations Using Stationary Flow Prediction on Large Payment and Cash Transaction Networks
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.
SEB Grp, Stockholm, Sweden..
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0003-1114-6040
2021 (English)In: International Conference On Machine Learning, Vol 139 / [ed] Meila, M Zhang, T, JMLR-JOURNAL MACHINE LEARNING RESEARCH , 2021, Vol. 139Conference paper, Published paper (Refereed)
Abstract [en]

Banks are required to analyse large transaction datasets as a part of the fight against financial crime. Today, this analysis is either performed manually by domain experts or using expensive feature engineering. Gradient flow analysis allows for basic representation learning as node potentials can be inferred directly from network transaction data. However, the gradient model has a fundamental limitation: it cannot represent all types of of network flows. Furthermore, standard methods for learning the gradient flow are not appropriate for flow signals that span multiple orders of magnitude and contain outliers, i.e. transaction data. In this work, the gradient model is extended to a gated version and we prove that it, unlike the gradient model, is a universal approximator for flows on graphs. To tackle the mentioned challenges of transaction data, we propose a multi-scale and outlier robust loss function based on the Student-t log-likelihood. Ethereum transaction data is used for evaluation and the gradient models outperform MLP models using hand-engineered and node2vec features in terms of relative error. These results extend to 60 synthetic datasets, with experiments also showing that the gated gradient model learns qualitative information about the underlying synthetic generative flow distributions.

Place, publisher, year, edition, pages
JMLR-JOURNAL MACHINE LEARNING RESEARCH , 2021. Vol. 139
Series
Proceedings of Machine Learning Research, ISSN 2640-3498
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-303379ISI: 000683104601036OAI: oai:DiVA.org:kth-303379DiVA, id: diva2:1603381
Conference
International Conference on Machine Learning (ICML), JUL 18-24, 2021, ELECTR NETWORK
Note

QC 20211015

Available from: 2021-10-15 Created: 2021-10-15 Last updated: 2022-06-25Bibliographically approved

Open Access in DiVA

No full text in DiVA

Authority records

Ceylan, CiwanPokorny, Florian T.

Search in DiVA

By author/editor
Ceylan, CiwanPokorny, Florian T.
By organisation
Robotics, Perception and Learning, RPL
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 103 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf