kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Workload-Aware Materialization of Junction Trees
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Theoretical Computer Science, TCS.ORCID iD: 0000-0002-5211-112X
2022 (English)In: Advances in Database Technology - EDBT, OpenProceedings.org , 2022, p. 65-77Conference paper, Published paper (Refereed)
Abstract [en]

Bayesian networks are popular probabilistic models that capture the conditional dependencies among a set of variables. Inference in Bayesian networks is a fundamental task for answering probabilistic queries over a subset of variables in the data. However, exact inference in Bayesian networks is NP-hard, which has prompted the development of many practical inference methods. In this paper, we focus on improving the performance of the junction-tree algorithm, a well-known method for exact inference in Bayesian networks. In particular, we seek to leverage information in the workload of probabilistic queries to obtain an optimal workload-aware materialization of junction trees, with the aim to accelerate the processing of inference queries. We devise an optimal pseudo-polynomial algorithm to tackle this problem and discuss approximation schemes. Compared to state-of-the-art approaches for efficient processing of inference queries via junction trees, our methods are the first to exploit the information provided in query workloads. Our experimentation on several real-world Bayesian networks confirms the effectiveness of our techniques in speeding-up query processing. 

Place, publisher, year, edition, pages
OpenProceedings.org , 2022. p. 65-77
Keywords [en]
Approximation algorithms, Bayesian networks, Inference engines, Optimization, Polynomial approximation, Query processing, Trees (mathematics), Bayesia n networks, Exact inference, Inference methods, Junction tree algorithms, Junction trees, NP-hard, Performance, Probabilistic models, Probabilistics, Pseudopolynomial algorithms, Forestry
National Category
Computer Sciences Probability Theory and Statistics
Identifiers
URN: urn:nbn:se:kth:diva-322404DOI: 10.5441/002/edbt.2022.06Scopus ID: 2-s2.0-85127247652OAI: oai:DiVA.org:kth-322404DiVA, id: diva2:1718906
Conference
25th International Conference on Extending Database Technology, EDBT 2022, 29 March 2022 through 1 April 2022
Note

QC 20221214

Available from: 2022-12-14 Created: 2022-12-14 Last updated: 2025-01-27Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Gionis, Aristides

Search in DiVA

By author/editor
Gionis, Aristides
By organisation
Theoretical Computer Science, TCS
Computer SciencesProbability Theory and Statistics

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 59 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf