kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A global structural em algorithm for a model of cancer progression
School of Computer Science, McGill Centre for Bioinformatics, McGill University, Canada.
Stockholm Bioinformatics Center, Stockholm University, Sweden.
Department of Oncology, Lund University, Sweden.
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Computational Science and Technology (CST). KTH, Centres, Science for Life Laboratory, SciLifeLab. KTH, Centres, SeRC - Swedish e-Science Research Centre.ORCID iD: 0000-0002-4552-0240
2011 (English)In: Adv. Neural Inf. Process. Syst.: Annu. Conf. Neural Inf. Process. Syst., NIPS, Neural Information Processing Systems , 2011Conference paper, Published paper (Refereed)
Abstract [en]

Cancer has complex patterns of progression that include converging as well as diverging progressional pathways. Vogelstein's path model of colon cancer was a pioneering contribution to cancer research. Since then, several attempts have been made at obtaining mathematical models of cancer progression, devising learning algorithms, and applying these to cross-sectional data. Beerenwinkel et al. provided, what they coined, EM-like algorithms for Oncogenetic Trees (OTs) and mixtures of such. Given the small size of current and future data sets, it is important to minimize the number of parameters of a model. For this reason, we too focus on tree-based models and introduce Hidden-variable Oncogenetic Trees (HOTs). In contrast to OTs, HOTs allow for errors in the data and thereby provide more realistic modeling. We also design global structural EM algorithms for learning HOTs and mixtures of HOTs (HOT-mixtures). The algorithms are global in the sense that, during the M-step, they find a structure that yields a global maximum of the expected complete log-likelihood rather than merely one that improves it. The algorithm for single HOTs performs very well on reasonable-sized data sets, while that for HOT-mixtures requires data sets of sizes obtainable only with tomorrow's more cost-efficient technologies.

Place, publisher, year, edition, pages
Neural Information Processing Systems , 2011.
Keywords [en]
Algorithms, Diseases, Forestry, Mathematical Models, Mixtures, Learning algorithms, 'current, Cancer progression, Cancer research, Colon cancer, Complex pattern, Data set, EM algorithms, Hidden variable, Path models, Tree-based model
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-350908Scopus ID: 2-s2.0-85162483791OAI: oai:DiVA.org:kth-350908DiVA, id: diva2:1885476
Conference
Advances in Neural Information Processing Systems 24: 25th Annual Conference on Neural Information Processing Systems 2011, 12-14 December 2011, Granada
Note

QC 20240723

Available from: 2024-07-23 Created: 2024-07-23 Last updated: 2024-07-23Bibliographically approved

Open Access in DiVA

No full text in DiVA

Scopus

Authority records

Lagergren, Jens

Search in DiVA

By author/editor
Lagergren, Jens
By organisation
Computational Science and Technology (CST)Science for Life Laboratory, SciLifeLabSeRC - Swedish e-Science Research Centre
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 13 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf