Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Have I seen you before?: Principles of Bayesian predictive classification revisited
University of Helsinki. (Department of mathematics and statistics)
University of Helsinki. (Department of mathematics and statistics)
KTH, School of Engineering Sciences (SCI), Mathematics (Dept.), Mathematical Statistics. (computational biostatistics)ORCID iD: 0000-0003-1489-8512
University of Helsinki. (Department of mathematics and statistics)
2013 (English)In: Statistics and computing, ISSN 0960-3174, E-ISSN 1573-1375, Vol. 23, no 1, 59-73 p.Article in journal (Refereed) Published
Abstract [en]

A general inductive Bayesian classification framework is considered using a simultaneous predictive distribution for test items. We introduce a principle of generative supervised and semi-supervised classification based on marginalizing the joint posterior distribution of labels for all test items. The simultaneous and marginalized classifiers arise under different loss functions, while both acknowledge jointly all uncertainty about the labels of test items and the generating probability measures of the classes. We illustrate for data from multiple finite alphabets that such classifiers achieve higher correct classification rates than a standard marginal predictive classifier which labels all test items independently, when training data are sparse. In the supervised case for multiple finite alphabets the simultaneous and the marginal classifiers are proven to become equal under generalized exchangeability when the amount of training data increases. Hence, the marginal classifier can be interpreted as an asymptotic approximation to the simultaneous classifier for finite sets of training data. It is also shown that such convergence is not guaranteed in the semi-supervised setting, where the marginal classifier does not provide a consistent approximation.

Place, publisher, year, edition, pages
Springer Berlin/Heidelberg, 2013. Vol. 23, no 1, 59-73 p.
Keyword [en]
Classification, Exchangeability, Inductive learning, Predictive inference
National Category
Computer and Information Science
Identifiers
URN: urn:nbn:se:kth:diva-88088DOI: 10.1007/s11222-011-9291-7ISI: 000313731400005Scopus ID: 2-s2.0-84872607314OAI: oai:DiVA.org:kth-88088DiVA: diva2:502193
Funder
EU, European Research Council, 239784
Note

QC 20130204

Available from: 2012-02-14 Created: 2012-02-14 Last updated: 2017-12-07Bibliographically approved

Open Access in DiVA

No full text

Other links

Publisher's full textScopus

Authority records BETA

Koski, Timo

Search in DiVA

By author/editor
Koski, Timo
By organisation
Mathematical Statistics
In the same journal
Statistics and computing
Computer and Information Science

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 111 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf